diff options
Diffstat (limited to 'recipes-containers/k3s/README.md')
-rw-r--r-- | recipes-containers/k3s/README.md | 208 |
1 files changed, 208 insertions, 0 deletions
diff --git a/recipes-containers/k3s/README.md b/recipes-containers/k3s/README.md index 3fe5ccd1..e4cb3e33 100644 --- a/recipes-containers/k3s/README.md +++ b/recipes-containers/k3s/README.md @@ -28,3 +28,211 @@ k3s-agent -t <token> -s https://<master>:6443 (Here `<token>` is found in `/var/lib/rancher/k3s/server/node-token` at the k3s master.) + +Example: +```shell +k3s-agent -t /var/lib/rancher/k3s/server/node-token -s https://localhost:6443 +``` + +## Notes: + +if running under qemu, the default of 256M of memory is not enough, k3s will +OOM and exit. + +Boot with qemuparams="-m 2048" to boot with 2G of memory (or choose the +appropriate amount for your configuration) + +Disk: if using qemu and core-image* you'll need to add extra space in your disks +to ensure containers can start. The following in your image recipe, or local.conf +would add 2G of extra space to the rootfs: + +```shell +IMAGE_ROOTFS_EXTRA_SPACE = "2097152" +``` + +## Example output from qemux86-64 running k3s server: + +```shell +root@qemux86-64:~# kubectl get nodes +NAME STATUS ROLES AGE VERSION +qemux86-64 Ready master 46s v1.18.9-k3s1 + + +root@qemux86-64:~# kubectl get pods -n kube-system +NAME READY STATUS RESTARTS AGE +local-path-provisioner-6d59f47c7-h7lxk 1/1 Running 0 2m32s +metrics-server-7566d596c8-mwntr 1/1 Running 0 2m32s +helm-install-traefik-229v7 0/1 Completed 0 2m32s +coredns-7944c66d8d-9rfj7 1/1 Running 0 2m32s +svclb-traefik-pb5j4 2/2 Running 0 89s +traefik-758cd5fc85-lxpr8 1/1 Running 0 89s + +root@qemux86-64:~# kubectl describe pods -n kube-system + +root@qemux86-64:~# ip a s +1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 + link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 + inet 127.0.0.1/8 scope host lo + valid_lft forever preferred_lft forever + inet6 ::1/128 scope host + valid_lft forever preferred_lft forever +2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP group default qlen 1000 + link/ether 52:54:00:12:35:02 brd ff:ff:ff:ff:ff:ff + inet 10.0.2.15/24 brd 10.0.2.255 scope global eth0 + valid_lft forever preferred_lft forever + inet6 fec0::5054:ff:fe12:3502/64 scope site dynamic mngtmpaddr + valid_lft 86239sec preferred_lft 14239sec + inet6 fe80::5054:ff:fe12:3502/64 scope link + valid_lft forever preferred_lft forever +3: sit0@NONE: <NOARP> mtu 1480 qdisc noop state DOWN group default qlen 1000 + link/sit 0.0.0.0 brd 0.0.0.0 +4: flannel.1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue state UNKNOWN group default + link/ether e2:aa:04:89:e6:0a brd ff:ff:ff:ff:ff:ff + inet 10.42.0.0/32 brd 10.42.0.0 scope global flannel.1 + valid_lft forever preferred_lft forever + inet6 fe80::e0aa:4ff:fe89:e60a/64 scope link + valid_lft forever preferred_lft forever +5: docker0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc noqueue state DOWN group default + link/ether 02:42:be:3e:25:e7 brd ff:ff:ff:ff:ff:ff + inet 172.17.0.1/16 brd 172.17.255.255 scope global docker0 + valid_lft forever preferred_lft forever +6: cni0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue state UP group default qlen 1000 + link/ether 82:8e:b4:f8:06:e7 brd ff:ff:ff:ff:ff:ff + inet 10.42.0.1/24 brd 10.42.0.255 scope global cni0 + valid_lft forever preferred_lft forever + inet6 fe80::808e:b4ff:fef8:6e7/64 scope link + valid_lft forever preferred_lft forever +7: veth82ac482e@if4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master cni0 state UP group default + link/ether ea:9d:14:c1:00:70 brd ff:ff:ff:ff:ff:ff link-netns cni-c52e6e09-f6e0-a47b-aea3-d6c47d3e2d01 + inet6 fe80::e89d:14ff:fec1:70/64 scope link + valid_lft forever preferred_lft forever +8: vethb94745ed@if4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master cni0 state UP group default + link/ether 1e:7f:7e:d3:ca:e8 brd ff:ff:ff:ff:ff:ff link-netns cni-86958efe-2462-016f-292d-81dbccc16a83 + inet6 fe80::8046:3cff:fe23:ced1/64 scope link + valid_lft forever preferred_lft forever +9: veth81ffb276@if4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master cni0 state UP group default + link/ether 2a:1d:48:54:76:50 brd ff:ff:ff:ff:ff:ff link-netns cni-5d77238e-6452-4fa3-40d2-91d48386080b + inet6 fe80::acf4:7fff:fe11:b6f2/64 scope link + valid_lft forever preferred_lft forever +10: vethce261f6a@if4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master cni0 state UP group default + link/ether 72:a3:90:4a:c5:12 brd ff:ff:ff:ff:ff:ff link-netns cni-55675948-77f2-a952-31ce-615f2bdb0093 + inet6 fe80::4d5:1bff:fe5d:db3a/64 scope link + valid_lft forever preferred_lft forever +11: vethee199cf4@if4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master cni0 state UP group default + link/ether e6:90:a4:a3:bc:a1 brd ff:ff:ff:ff:ff:ff link-netns cni-4aeccd16-2976-8a78-b2c4-e028da3bb1ea + inet6 fe80::c85a:8bff:fe0b:aea0/64 scope link + valid_lft forever preferred_lft forever + + +root@qemux86-64:~# kubectl describe nodes + +Name: qemux86-64 +Roles: master +Labels: beta.kubernetes.io/arch=amd64 + beta.kubernetes.io/instance-type=k3s + beta.kubernetes.io/os=linux + k3s.io/hostname=qemux86-64 + k3s.io/internal-ip=10.0.2.15 + kubernetes.io/arch=amd64 + kubernetes.io/hostname=qemux86-64 + kubernetes.io/os=linux + node-role.kubernetes.io/master=true + node.kubernetes.io/instance-type=k3s +Annotations: flannel.alpha.coreos.com/backend-data: {"VtepMAC":"2e:52:6a:1b:76:d4"} + flannel.alpha.coreos.com/backend-type: vxlan + flannel.alpha.coreos.com/kube-subnet-manager: true + flannel.alpha.coreos.com/public-ip: 10.0.2.15 + k3s.io/node-args: ["server"] + k3s.io/node-config-hash: MLFMUCBMRVINLJJKSG32TOUFWB4CN55GMSNY25AZPESQXZCYRN2A==== + k3s.io/node-env: {} + node.alpha.kubernetes.io/ttl: 0 + volumes.kubernetes.io/controller-managed-attach-detach: true +CreationTimestamp: Tue, 10 Nov 2020 14:01:28 +0000 +Taints: <none> +Unschedulable: false +Lease: + HolderIdentity: qemux86-64 + AcquireTime: <unset> + RenewTime: Tue, 10 Nov 2020 14:56:27 +0000 +Conditions: + Type Status LastHeartbeatTime LastTransitionTime Reason Message + ---- ------ ----------------- ------------------ ------ ------- + NetworkUnavailable False Tue, 10 Nov 2020 14:43:46 +0000 Tue, 10 Nov 2020 14:43:46 +0000 FlannelIsUp Flannel is running on this node + MemoryPressure False Tue, 10 Nov 2020 14:51:48 +0000 Tue, 10 Nov 2020 14:45:46 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available + DiskPressure False Tue, 10 Nov 2020 14:51:48 +0000 Tue, 10 Nov 2020 14:45:46 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure + PIDPressure False Tue, 10 Nov 2020 14:51:48 +0000 Tue, 10 Nov 2020 14:45:46 +0000 KubeletHasSufficientPID kubelet has sufficient PID available + Ready True Tue, 10 Nov 2020 14:51:48 +0000 Tue, 10 Nov 2020 14:45:46 +0000 KubeletReady kubelet is posting ready status +Addresses: + InternalIP: 10.0.2.15 + Hostname: qemux86-64 +Capacity: + cpu: 1 + ephemeral-storage: 39748144Ki + memory: 2040164Ki + pods: 110 +Allocatable: + cpu: 1 + ephemeral-storage: 38666994453 + memory: 2040164Ki + pods: 110 +System Info: + Machine ID: 6a4abfacbf83457e9a0cbb5777457c5d + System UUID: 6a4abfacbf83457e9a0cbb5777457c5d + Boot ID: f5ddf6c8-1abf-4aef-9e29-106488e3c337 + Kernel Version: 5.8.13-yocto-standard + OS Image: Poky (Yocto Project Reference Distro) 3.2+snapshot-20201105 (master) + Operating System: linux + Architecture: amd64 + Container Runtime Version: containerd://1.4.1-4-ge44e8ebea.m + Kubelet Version: v1.18.9-k3s1 + Kube-Proxy Version: v1.18.9-k3s1 +PodCIDR: 10.42.0.0/24 +PodCIDRs: 10.42.0.0/24 +ProviderID: k3s://qemux86-64 +Non-terminated Pods: (5 in total) + Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits AGE + --------- ---- ------------ ---------- --------------- ------------- --- + kube-system svclb-traefik-jpmnd 0 (0%) 0 (0%) 0 (0%) 0 (0%) 54m + kube-system metrics-server-7566d596c8-wh29d 0 (0%) 0 (0%) 0 (0%) 0 (0%) 56m + kube-system local-path-provisioner-6d59f47c7-npn4d 0 (0%) 0 (0%) 0 (0%) 0 (0%) 56m + kube-system coredns-7944c66d8d-md8hr 100m (10%) 0 (0%) 70Mi (3%) 170Mi (8%) 56m + kube-system traefik-758cd5fc85-phjr2 0 (0%) 0 (0%) 0 (0%) 0 (0%) 54m +Allocated resources: + (Total limits may be over 100 percent, i.e., overcommitted.) + Resource Requests Limits + -------- -------- ------ + cpu 100m (10%) 0 (0%) + memory 70Mi (3%) 170Mi (8%) + ephemeral-storage 0 (0%) 0 (0%) +Events: + Type Reason Age From Message + ---- ------ ---- ---- ------- + Normal Starting 56m kube-proxy Starting kube-proxy. + Normal Starting 55m kubelet Starting kubelet. + Warning InvalidDiskCapacity 55m kubelet invalid capacity 0 on image filesystem + Normal NodeHasSufficientPID 55m (x2 over 55m) kubelet Node qemux86-64 status is now: NodeHasSufficientPID + Normal NodeHasSufficientMemory 55m (x2 over 55m) kubelet Node qemux86-64 status is now: NodeHasSufficientMemory + Normal NodeHasNoDiskPressure 55m (x2 over 55m) kubelet Node qemux86-64 status is now: NodeHasNoDiskPressure + Normal NodeAllocatableEnforced 55m kubelet Updated Node Allocatable limit across pods + Normal NodeReady 54m kubelet Node qemux86-64 status is now: NodeReady + Normal Starting 52m kube-proxy Starting kube-proxy. + Normal NodeReady 50m kubelet Node qemux86-64 status is now: NodeReady + Normal NodeAllocatableEnforced 50m kubelet Updated Node Allocatable limit across pods + Warning Rebooted 50m kubelet Node qemux86-64 has been rebooted, boot id: a4e4d2d8-ddb4-49b8-b0a9-e81d12707113 + Normal NodeHasSufficientMemory 50m (x2 over 50m) kubelet Node qemux86-64 status is now: NodeHasSufficientMemory + Normal Starting 50m kubelet Starting kubelet. + Normal NodeHasSufficientPID 50m (x2 over 50m) kubelet Node qemux86-64 status is now: NodeHasSufficientPID + Normal NodeHasNoDiskPressure 50m (x2 over 50m) kubelet Node qemux86-64 status is now: NodeHasNoDiskPressure + Normal NodeNotReady 17m kubelet Node qemux86-64 status is now: NodeNotReady + Warning InvalidDiskCapacity 15m (x2 over 50m) kubelet invalid capacity 0 on image filesystem + Normal Starting 12m kube-proxy Starting kube-proxy. + Normal Starting 10m kubelet Starting kubelet. + Warning InvalidDiskCapacity 10m kubelet invalid capacity 0 on image filesystem + Normal NodeAllocatableEnforced 10m kubelet Updated Node Allocatable limit across pods + Warning Rebooted 10m kubelet Node qemux86-64 has been rebooted, boot id: f5ddf6c8-1abf-4aef-9e29-106488e3c337 + Normal NodeHasSufficientMemory 10m (x2 over 10m) kubelet Node qemux86-64 status is now: NodeHasSufficientMemory + Normal NodeHasNoDiskPressure 10m (x2 over 10m) kubelet Node qemux86-64 status is now: NodeHasNoDiskPressure + Normal NodeHasSufficientPID 10m (x2 over 10m) kubelet Node qemux86-64 status is now: NodeHasSufficientPID + Normal NodeReady 10m kubelet Node qemux86-64 status is now: NodeReady + +```shell |