On Tue, Nov 10, 2020 at 8:17 AM Bruce Ashfield via
lists.yoctoproject.org
<[email protected]> wrote:
>
> On Tue, Nov 10, 2020 at 7:46 AM Bruce Ashfield via
> lists.yoctoproject.org
> <[email protected]> wrote:
> >
> > On Tue, Nov 10, 2020 at 1:43 AM Lance Yang <[email protected]> wrote:
> > >
> > > Hi Bruce and Joakim,
> > >
> > > Thanks for sharing this branch: k3s-wip. I have tested against my yocto 
> > > build.
> >
> > The branch will be more functional shortly, I have quite a few changes
> > to factor things for
> > k8s and generally more usable :D
> >
> >         modified:   classes/cni_networking.bbclass
> >         modified:   conf/layer.conf
> >         modified:   recipes-containers/containerd/containerd-docker_git.bb
> >         modified:
> > recipes-containers/containerd/containerd-opencontainers_git.bb
> >         modified:   recipes-containers/k3s/README.md
> >         modified:   recipes-containers/k3s/k3s_git.bb
> >         modified:   recipes-kernel/linux/linux-yocto/kubernetes.cfg
> >         modified:   recipes-networking/cni/cni_git.bb
> >         container-deploy.txt
> >         recipes-core/packagegroups/
> >
> > >
> > > My Image: Linux qemuarm64 by yocto.
> > >
> > > The master node can be ready after I started the k3s server. However, the 
> > > pods in kube-system (which are essential components for k3s) cannot turn 
> > > to ready state on qemuarm64.
> > >
> >
> > That's interesting, since in my configuration, the master never comes ready:
> >
> > root@qemux86-64:~# kubectl get nodes
> > NAME         STATUS     ROLES    AGE   VERSION
> > qemux86-64   NotReady   master   15h   v1.18.9-k3s1
> >
>
> Hah.
>
> I finally got the node to show up as ready:
>
> root@qemux86-64:~# kubectl get nodes
> NAME         STATUS   ROLES    AGE    VERSION
> qemux86-64   Ready    master   112s   v1.18.9-k3s1
>

Lance,

What image type were you building ? I'm pulling in dependencies to
packagegroups and the recipes themselves.

I'm not seeing the mount issue on my master/server node:

root@qemux86-64:~# kubectl get pods -n kube-system
NAME                                     READY   STATUS      RESTARTS   AGE
local-path-provisioner-6d59f47c7-h7lxk   1/1     Running     0          3m32s
metrics-server-7566d596c8-mwntr          1/1     Running     0          3m32s
helm-install-traefik-229v7               0/1     Completed   0          3m32s
coredns-7944c66d8d-9rfj7                 1/1     Running     0          3m32s
svclb-traefik-pb5j4                      2/2     Running     0          2m29s
traefik-758cd5fc85-lxpr8                 1/1     Running     0          2m29s

I'm going back to all-in-one node debugging, but can look into the
mount issue more later.

Bruce

> I'm attempting to build an all-in-one node, and that is likely causing
> me some issues.
>
> I'm revisiting those potential conflicts now.
>
> But if anyone else does have an all in one working and has some tips,
> feel free to share :D
>
> Bruce
>
> > I've sorted out more of the dependencies, and have packagegroups to
> > make them easier
> > now.
> >
> > Hopefully, I can figure out what is now missing and keeping my master
> > from moving into
> > ready today.
> >
> > Bruce
> >
> > > After the master node itself turned to ready state, I check the pods with 
> > > kubectl:
> > >
> > > kubectl get nodes
> > > NAME        STATUS   ROLES    AGE   VERSION
> > > qemuarm64   Ready    master   11m   v1.18.9-k3s1
> > > root@qemuarm64:~# ls
> > > root@qemuarm64:~# kubectl get pods -n kube-system
> > > NAME                                     READY   STATUS              
> > > RESTARTS   AGE
> > > local-path-provisioner-6d59f47c7-xxvbl   0/1     ContainerCreating   0    
> > >       12m
> > > coredns-7944c66d8d-tlrm9                 0/1     ContainerCreating   0    
> > >       12m
> > > metrics-server-7566d596c8-svkff          0/1     ContainerCreating   0    
> > >       12m
> > > helm-install-traefik-s8p5g               0/1     ContainerCreating   0    
> > >       12m
> > >
> > > Then I describe the pods with:
> > >
> > > Events:
> > >   Type     Reason       Age                  From               Message
> > >   ----     ------       ----                 ----               -------
> > >   Normal   Scheduled    16m                  default-scheduler  
> > > Successfully assigned kube-system/coredns-7944c66d8d-tlrm9 to qemuarm64
> > >   Warning  FailedMount  5m23s (x3 over 14m)  kubelet            Unable to 
> > > attach or mount volumes: unmounted volumes=[coredns-token-b7nlh], 
> > > unattached volumes=[coredns-token-b7nlh config-volume]: timed out waiting 
> > > for the condition
> > >   Warning  FailedMount  50s (x4 over 12m)    kubelet            Unable to 
> > > attach or mount volumes: unmounted volumes=[coredns-token-b7nlh], 
> > > unattached volumes=[config-volume coredns-token-b7nlh]: timed out waiting 
> > > for the condition
> > >   Warning  FailedMount  11s (x16 over 16m)   kubelet            
> > > MountVolume.SetUp failed for volume "coredns-token-b7nlh" : mount failed: 
> > > exec: "mount": executable file not found in $PATH
> > >
> > > I found the "mount" binary is not found in $PATH. However, I confirmed 
> > > the $PATH and mount binary on my qemuarm64 image:
> > >
> > > root@qemuarm64:~# echo $PATH
> > > /usr/local/bin:/usr/bin:/bin:/usr/local/sbin:/usr/sbin:/sbin
> > > root@qemuarm64:~# which mount
> > > /bin/mount
> > >
> > > When I type mount command, it worked fine:
> > >
> > > /dev/root on / type ext4 (rw,relatime)
> > > devtmpfs on /dev type devtmpfs 
> > > (rw,relatime,size=2016212k,nr_inodes=504053,mode=755)
> > > proc on /proc type proc (rw,relatime)
> > > sysfs on /sys type sysfs (rw,relatime)
> > > debugfs on /sys/kernel/debug type debugfs (rw,relatime)
> > > tmpfs on /run type tmpfs (rw,nosuid,nodev,mode=755)
> > > ...
> > > ... (skipped the verbose output)
> > >
> > > I would like to know whether you have met this "mount" issue ever?
> > >
> > > Best Regards,
> > > Lance
> > >
> > > > -----Original Message-----
> > > > From: [email protected] 
> > > > <[email protected]>
> > > > On Behalf Of Bruce Ashfield via lists.yoctoproject.org
> > > > Sent: Monday, October 26, 2020 11:46 PM
> > > > To: Joakim Roubert <[email protected]>
> > > > Cc: [email protected]
> > > > Subject: Re: [meta-virtualization][PATCH v5] Adding k3s recipe
> > > >
> > > > On Wed, Oct 21, 2020 at 2:00 AM Joakim Roubert 
> > > > <[email protected]> wrote:
> > > > >
> > > > > On 2020-10-21 05:10, Bruce Ashfield wrote:
> > > > > > Ha!!!!
> > > > > >
> > > > > > This applies.
> > > > >
> > > > > Wonderful, thank you! I guess this is what is called "five times 
> > > > > lucky"...
> > > > >
> > > > > > I'm now testing and completing some of my networking factoring, as
> > > > > > well as importing / forking some recipes to avoid extra layer
> > > > > > depends.
> > > > >
> > > > > Excellent!
> > > >
> > > > I've pushed some of my WIP to:
> > > > https://git.yoctoproject.org/cgit/cgit.cgi/meta-virtualization/log/?h=k3s-wip
> > > >
> > > > That includes the split of the networking, the import of some of the 
> > > > dependencies and some
> > > > small tweaks I'm working on.
> > > >
> > > > I did have a couple of questions on the k3s packaging itself, I was 
> > > > getting the following
> > > > error:
> > > >
> > > > ERROR: k3s-v1.18.9+k3s1-dirty-r0 do_package: QA Issue: k3s:
> > > > Files/directories were installed but not shipped in any package:
> > > >   /usr/local/bin/k3s-clean
> > > >   /usr/local/bin/crictl
> > > >   /usr/local/bin/kubectl
> > > >   /usr/local/bin/k3s
> > > >
> > > > So I added them to the FILES of the k3s package itself (so both 
> > > > k3s-server and k3s-agent will get
> > > > them), is that the split you were looking for ?
> > > >
> > > > Bruce
> > > >
> > > > >
> > > > > BR,
> > > > >
> > > > > /Joakim
> > > > > --
> > > > > Joakim Roubert
> > > > > Senior Engineer
> > > > >
> > > > > Axis Communications AB
> > > > > Emdalavägen 14, SE-223 69 Lund, Sweden
> > > > > Tel: +46 46 272 18 00, Tel (direct): +46 46 272 27 48
> > > > > Fax: +46 46 13 61 30, www.axis.com
> > > > >
> > > > --
> > > > - Thou shalt not follow the NULL pointer, for chaos and madness await 
> > > > thee at its end
> > > > - "Use the force Harry" - Gandalf, Star Trek II
> > > IMPORTANT NOTICE: The contents of this email and any attachments are 
> > > confidential and may also be privileged. If you are not the intended 
> > > recipient, please notify the sender immediately and do not disclose the 
> > > contents to any other person, use it for any purpose, or store or copy 
> > > the information in any medium. Thank you.
> >
> >
> >
> > --
> > - Thou shalt not follow the NULL pointer, for chaos and madness await
> > thee at its end
> > - "Use the force Harry" - Gandalf, Star Trek II
> >
> >
> >
>
>
> --
> - Thou shalt not follow the NULL pointer, for chaos and madness await
> thee at its end
> - "Use the force Harry" - Gandalf, Star Trek II
>
> 
>


-- 
- Thou shalt not follow the NULL pointer, for chaos and madness await
thee at its end
- "Use the force Harry" - Gandalf, Star Trek II
-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.
View/Reply Online (#6053): 
https://lists.yoctoproject.org/g/meta-virtualization/message/6053
Mute This Topic: https://lists.yoctoproject.org/mt/77679236/21656
Group Owner: [email protected]
Unsubscribe: https://lists.yoctoproject.org/g/meta-virtualization/unsub 
[[email protected]]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to