Re: [yocto] Understanding kernel patching in linux-yocto
Le mer. 12 mai 2021 à 20:33, Diego Santa Cruz a écrit : > > > -Original Message- > > From: yocto@lists.yoctoproject.org On > > Behalf Of Bruce Ashfield via lists.yoctoproject.org > > Sent: 12 May 2021 16:25 > > To: Yann Dirson > > Cc: Yocto discussion list > > Subject: Re: [yocto] Understanding kernel patching in linux-yocto > > > > On Wed, May 12, 2021 at 10:07 AM Yann Dirson > > wrote: > > > > > > Thanks for those clarifications! > > > > > > Some additional questions below > > > > > > Le mer. 12 mai 2021 à 15:19, Bruce Ashfield a > > écrit : > > > > > > > > On Wed, May 12, 2021 at 7:14 AM Yann Dirson > group.com> wrote: > > > > > > > > > > I am currently working on a kmeta BSP for the rockchip-based NanoPI > > M4 > > > > > [1], and I'm wondering how I should be providing kernel patches, as > > > > > just add ing "patch" directives in the .scc does not get them applied > > > > > unless the particular .scc gets included in KERNEL_FEATURES (see [2]). > > > > > > > > > > From an old thread [3] I understand that the patches from the standard > > > > > kmeta snippets are already applied to the tree, and that to get the > > > > > patches from my BSP I'd need to reference it explicitly in SRC_URI > > > > > (along with using "nopatch" in the right places to avoid the > > > > > already-applied patches to get applied twice). > > > > > > > > > > I have the feeling that I'm lacking the rationale behind this, and > > > > > would need to understand this better to make things right in this BSP. > > > > > Especially: > > > > > - at first sight, having the patches both applied to linux-yocto and > > > > > referenced in yocto-kernel-cache just to be skipped on parsing looks > > > > > like both information duplication and parsing of unused lines > > > > > > > > At least some of this is mentioned in the advanced section of the > > > > kernel-dev manual, but I can summarize/reword things here, and > > > > I'm also doing a presentation related to this in the Yocto summit at > > > > the end of this month. > > > > > > > > The big thing to remember, is that the configuration and changes > > > > you see in that repository, are not only for yocto purposes. The > > > > concepts and structure pre-date when they were first brought in > > > > to generate reference kernels over 10 years ago (the implementation > > > > has changed, but the concepts are still the same). To this day, > > > > there still are cases that they are used with just a kernel tree and > > > > cross toolchain. > > > > > > > > With that in mind, the meta-data is used for many different things > > > > > > > > - It organizes patches / features and their configuration into > > > >reusable blocks. At the same time documenting the changes > > > >that we have applied to a tree > > > > - It makes those patches and configuration blocks available to > > > >other kernel trees (for whatever reason). > > > > - It configures the tree during the build process, reusing both > > > >configuration only and patch + configuration blocks > > > > > > > - It is used to generate a history clean tree from scratch for > > > >each new supported kernel. Which is what I do when creating > > > >new linux-yocto-dev references, and the new /standard/* > > > >branches in linux-yocto. > > > > > > I'd think (and I take your further remarks about workflow as confirming > > > this), that when upgrading the kernel the best tool would be git-rebase. > > > Then, regenerating the linux-yocto branches would only be a akin to a > > > check that the metadata is in sync with the new tree you rebased ? > > > > The best of anything is a matter of opinion. I heavily use git-rebase and > > sure, you could use it to do something similar here. But the result is > > the same. There's still heavy use of quilt in kernel circles. Workflows > > don't change easily, and as long as they work for the maintainer, they > > tend to stay put. Asking someone to change their workflow, rarely goes > > over well. > > > > > > > > If that conclusion is correct, wouldn't it
Re: [yocto] Understanding kernel patching in linux-yocto
> -Original Message- > From: yocto@lists.yoctoproject.org On > Behalf Of Bruce Ashfield via lists.yoctoproject.org > Sent: 12 May 2021 16:25 > To: Yann Dirson > Cc: Yocto discussion list > Subject: Re: [yocto] Understanding kernel patching in linux-yocto > > On Wed, May 12, 2021 at 10:07 AM Yann Dirson > wrote: > > > > Thanks for those clarifications! > > > > Some additional questions below > > > > Le mer. 12 mai 2021 à 15:19, Bruce Ashfield a > écrit : > > > > > > On Wed, May 12, 2021 at 7:14 AM Yann Dirson group.com> wrote: > > > > > > > > I am currently working on a kmeta BSP for the rockchip-based NanoPI > M4 > > > > [1], and I'm wondering how I should be providing kernel patches, as > > > > just add ing "patch" directives in the .scc does not get them applied > > > > unless the particular .scc gets included in KERNEL_FEATURES (see [2]). > > > > > > > > From an old thread [3] I understand that the patches from the standard > > > > kmeta snippets are already applied to the tree, and that to get the > > > > patches from my BSP I'd need to reference it explicitly in SRC_URI > > > > (along with using "nopatch" in the right places to avoid the > > > > already-applied patches to get applied twice). > > > > > > > > I have the feeling that I'm lacking the rationale behind this, and > > > > would need to understand this better to make things right in this BSP. > > > > Especially: > > > > - at first sight, having the patches both applied to linux-yocto and > > > > referenced in yocto-kernel-cache just to be skipped on parsing looks > > > > like both information duplication and parsing of unused lines > > > > > > At least some of this is mentioned in the advanced section of the > > > kernel-dev manual, but I can summarize/reword things here, and > > > I'm also doing a presentation related to this in the Yocto summit at > > > the end of this month. > > > > > > The big thing to remember, is that the configuration and changes > > > you see in that repository, are not only for yocto purposes. The > > > concepts and structure pre-date when they were first brought in > > > to generate reference kernels over 10 years ago (the implementation > > > has changed, but the concepts are still the same). To this day, > > > there still are cases that they are used with just a kernel tree and > > > cross toolchain. > > > > > > With that in mind, the meta-data is used for many different things > > > > > > - It organizes patches / features and their configuration into > > >reusable blocks. At the same time documenting the changes > > >that we have applied to a tree > > > - It makes those patches and configuration blocks available to > > >other kernel trees (for whatever reason). > > > - It configures the tree during the build process, reusing both > > >configuration only and patch + configuration blocks > > > > > - It is used to generate a history clean tree from scratch for > > >each new supported kernel. Which is what I do when creating > > >new linux-yocto-dev references, and the new /standard/* > > >branches in linux-yocto. > > > > I'd think (and I take your further remarks about workflow as confirming > > this), that when upgrading the kernel the best tool would be git-rebase. > > Then, regenerating the linux-yocto branches would only be a akin to a > > check that the metadata is in sync with the new tree you rebased ? > > The best of anything is a matter of opinion. I heavily use git-rebase and > sure, you could use it to do something similar here. But the result is > the same. There's still heavy use of quilt in kernel circles. Workflows > don't change easily, and as long as they work for the maintainer, they > tend to stay put. Asking someone to change their workflow, rarely goes > over well. > > > > > If that conclusion is correct, wouldn't it be possible to avoid using the > > linux-yocto branches directly, and let all the patches be applied at > > do_patch time ? That would be much more similar to the standard > > package workflow (and thus lower the barrier for approaching the > > kernel packages). > > That's something we did in the past, and sure, you can do anything. > But patching hundreds of changes at build time means constant > failures .. again, I
Re: [yocto] Understanding kernel patching in linux-yocto
Le mer. 12 mai 2021 à 16:25, Bruce Ashfield a écrit : > > On Wed, May 12, 2021 at 10:07 AM Yann Dirson > wrote: > > > > Thanks for those clarifications! > > > > Some additional questions below > > > > Le mer. 12 mai 2021 à 15:19, Bruce Ashfield a > > écrit : > > > > > > On Wed, May 12, 2021 at 7:14 AM Yann Dirson > > > wrote: > > > > > > > > I am currently working on a kmeta BSP for the rockchip-based NanoPI M4 > > > > [1], and I'm wondering how I should be providing kernel patches, as > > > > just add ing "patch" directives in the .scc does not get them applied > > > > unless the particular .scc gets included in KERNEL_FEATURES (see [2]). > > > > > > > > From an old thread [3] I understand that the patches from the standard > > > > kmeta snippets are already applied to the tree, and that to get the > > > > patches from my BSP I'd need to reference it explicitly in SRC_URI > > > > (along with using "nopatch" in the right places to avoid the > > > > already-applied patches to get applied twice). > > > > > > > > I have the feeling that I'm lacking the rationale behind this, and > > > > would need to understand this better to make things right in this BSP. > > > > Especially: > > > > - at first sight, having the patches both applied to linux-yocto and > > > > referenced in yocto-kernel-cache just to be skipped on parsing looks > > > > like both information duplication and parsing of unused lines > > > > > > At least some of this is mentioned in the advanced section of the > > > kernel-dev manual, but I can summarize/reword things here, and > > > I'm also doing a presentation related to this in the Yocto summit at > > > the end of this month. > > > > > > The big thing to remember, is that the configuration and changes > > > you see in that repository, are not only for yocto purposes. The > > > concepts and structure pre-date when they were first brought in > > > to generate reference kernels over 10 years ago (the implementation > > > has changed, but the concepts are still the same). To this day, > > > there still are cases that they are used with just a kernel tree and > > > cross toolchain. > > > > > > With that in mind, the meta-data is used for many different things > > > > > > - It organizes patches / features and their configuration into > > >reusable blocks. At the same time documenting the changes > > >that we have applied to a tree > > > - It makes those patches and configuration blocks available to > > >other kernel trees (for whatever reason). > > > - It configures the tree during the build process, reusing both > > >configuration only and patch + configuration blocks > > > > > - It is used to generate a history clean tree from scratch for > > >each new supported kernel. Which is what I do when creating > > >new linux-yocto-dev references, and the new /standard/* > > >branches in linux-yocto. > > > > I'd think (and I take your further remarks about workflow as confirming > > this), that when upgrading the kernel the best tool would be git-rebase. > > Then, regenerating the linux-yocto branches would only be a akin to a > > check that the metadata is in sync with the new tree you rebased ? > > The best of anything is a matter of opinion. I heavily use git-rebase and > sure, you could use it to do something similar here. But the result is > the same. There's still heavy use of quilt in kernel circles. Workflows > don't change easily, and as long as they work for the maintainer, they > tend to stay put. Asking someone to change their workflow, rarely goes > over well. > > > > > If that conclusion is correct, wouldn't it be possible to avoid using the > > linux-yocto branches directly, and let all the patches be applied at > > do_patch time ? That would be much more similar to the standard > > package workflow (and thus lower the barrier for approaching the > > kernel packages). > > That's something we did in the past, and sure, you can do anything. > But patching hundreds of changes at build time means constant > failures .. again, I've been there and done that. We use similar patches > in many different contexts and optional stackings. You simply cannot > maintain them and stay sane by whacking patches onto the SRC_URI. > The last impression you want when someone builds your kernel is that > they can't even get past the patch phase. So that's a hard no, to how > the reference kernels are maintained (and that hard no has been around > for 11 years now). > > Also, we maintain contributed reference BSPs in that same tree, that > are yanking in SDKs from vendors, etc, they are in the thousands of > patches. So you need the tree and the BSP branches to support that. That pretty much clarifies the whole thing, thanks for taking the time for this! > > > > > > > > So why not just drop all the patches in the SRC_URI ? Been there, > > > done that. It fails spectacularly when you are managing queues of > > > hundreds of potentially conflicting patches (rt, yaffs, aufs,
Re: [yocto] Understanding kernel patching in linux-yocto
On Wed, May 12, 2021 at 10:07 AM Yann Dirson wrote: > > Thanks for those clarifications! > > Some additional questions below > > Le mer. 12 mai 2021 à 15:19, Bruce Ashfield a > écrit : > > > > On Wed, May 12, 2021 at 7:14 AM Yann Dirson > > wrote: > > > > > > I am currently working on a kmeta BSP for the rockchip-based NanoPI M4 > > > [1], and I'm wondering how I should be providing kernel patches, as > > > just add ing "patch" directives in the .scc does not get them applied > > > unless the particular .scc gets included in KERNEL_FEATURES (see [2]). > > > > > > From an old thread [3] I understand that the patches from the standard > > > kmeta snippets are already applied to the tree, and that to get the > > > patches from my BSP I'd need to reference it explicitly in SRC_URI > > > (along with using "nopatch" in the right places to avoid the > > > already-applied patches to get applied twice). > > > > > > I have the feeling that I'm lacking the rationale behind this, and > > > would need to understand this better to make things right in this BSP. > > > Especially: > > > - at first sight, having the patches both applied to linux-yocto and > > > referenced in yocto-kernel-cache just to be skipped on parsing looks > > > like both information duplication and parsing of unused lines > > > > At least some of this is mentioned in the advanced section of the > > kernel-dev manual, but I can summarize/reword things here, and > > I'm also doing a presentation related to this in the Yocto summit at > > the end of this month. > > > > The big thing to remember, is that the configuration and changes > > you see in that repository, are not only for yocto purposes. The > > concepts and structure pre-date when they were first brought in > > to generate reference kernels over 10 years ago (the implementation > > has changed, but the concepts are still the same). To this day, > > there still are cases that they are used with just a kernel tree and > > cross toolchain. > > > > With that in mind, the meta-data is used for many different things > > > > - It organizes patches / features and their configuration into > >reusable blocks. At the same time documenting the changes > >that we have applied to a tree > > - It makes those patches and configuration blocks available to > >other kernel trees (for whatever reason). > > - It configures the tree during the build process, reusing both > >configuration only and patch + configuration blocks > > > - It is used to generate a history clean tree from scratch for > >each new supported kernel. Which is what I do when creating > >new linux-yocto-dev references, and the new /standard/* > >branches in linux-yocto. > > I'd think (and I take your further remarks about workflow as confirming > this), that when upgrading the kernel the best tool would be git-rebase. > Then, regenerating the linux-yocto branches would only be a akin to a > check that the metadata is in sync with the new tree you rebased ? The best of anything is a matter of opinion. I heavily use git-rebase and sure, you could use it to do something similar here. But the result is the same. There's still heavy use of quilt in kernel circles. Workflows don't change easily, and as long as they work for the maintainer, they tend to stay put. Asking someone to change their workflow, rarely goes over well. > > If that conclusion is correct, wouldn't it be possible to avoid using the > linux-yocto branches directly, and let all the patches be applied at > do_patch time ? That would be much more similar to the standard > package workflow (and thus lower the barrier for approaching the > kernel packages). That's something we did in the past, and sure, you can do anything. But patching hundreds of changes at build time means constant failures .. again, I've been there and done that. We use similar patches in many different contexts and optional stackings. You simply cannot maintain them and stay sane by whacking patches onto the SRC_URI. The last impression you want when someone builds your kernel is that they can't even get past the patch phase. So that's a hard no, to how the reference kernels are maintained (and that hard no has been around for 11 years now). Also, we maintain contributed reference BSPs in that same tree, that are yanking in SDKs from vendors, etc, they are in the thousands of patches. So you need the tree and the BSP branches to support that. > > > > So why not just drop all the patches in the SRC_URI ? Been there, > > done that. It fails spectacularly when you are managing queues of > > hundreds of potentially conflicting patches (rt, yaffs, aufs, ... etc, etc) > > and then attempting to constantly merge -stable and other kernel > > trees into the repository. git is the tool for managing that, not stacks > > of patches. You spend your entire life fixing patch errors and refreshing > > fuzz (again, been there, done that). > > > > So why not just keep a history and constan
Re: [yocto] Understanding kernel patching in linux-yocto
Thanks for those clarifications! Some additional questions below Le mer. 12 mai 2021 à 15:19, Bruce Ashfield a écrit : > > On Wed, May 12, 2021 at 7:14 AM Yann Dirson > wrote: > > > > I am currently working on a kmeta BSP for the rockchip-based NanoPI M4 > > [1], and I'm wondering how I should be providing kernel patches, as > > just add ing "patch" directives in the .scc does not get them applied > > unless the particular .scc gets included in KERNEL_FEATURES (see [2]). > > > > From an old thread [3] I understand that the patches from the standard > > kmeta snippets are already applied to the tree, and that to get the > > patches from my BSP I'd need to reference it explicitly in SRC_URI > > (along with using "nopatch" in the right places to avoid the > > already-applied patches to get applied twice). > > > > I have the feeling that I'm lacking the rationale behind this, and > > would need to understand this better to make things right in this BSP. > > Especially: > > - at first sight, having the patches both applied to linux-yocto and > > referenced in yocto-kernel-cache just to be skipped on parsing looks > > like both information duplication and parsing of unused lines > > At least some of this is mentioned in the advanced section of the > kernel-dev manual, but I can summarize/reword things here, and > I'm also doing a presentation related to this in the Yocto summit at > the end of this month. > > The big thing to remember, is that the configuration and changes > you see in that repository, are not only for yocto purposes. The > concepts and structure pre-date when they were first brought in > to generate reference kernels over 10 years ago (the implementation > has changed, but the concepts are still the same). To this day, > there still are cases that they are used with just a kernel tree and > cross toolchain. > > With that in mind, the meta-data is used for many different things > > - It organizes patches / features and their configuration into >reusable blocks. At the same time documenting the changes >that we have applied to a tree > - It makes those patches and configuration blocks available to >other kernel trees (for whatever reason). > - It configures the tree during the build process, reusing both >configuration only and patch + configuration blocks > - It is used to generate a history clean tree from scratch for >each new supported kernel. Which is what I do when creating >new linux-yocto-dev references, and the new /standard/* >branches in linux-yocto. I'd think (and I take your further remarks about workflow as confirming this), that when upgrading the kernel the best tool would be git-rebase. Then, regenerating the linux-yocto branches would only be a akin to a check that the metadata is in sync with the new tree you rebased ? If that conclusion is correct, wouldn't it be possible to avoid using the linux-yocto branches directly, and let all the patches be applied at do_patch time ? That would be much more similar to the standard package workflow (and thus lower the barrier for approaching the kernel packages). > So why not just drop all the patches in the SRC_URI ? Been there, > done that. It fails spectacularly when you are managing queues of > hundreds of potentially conflicting patches (rt, yaffs, aufs, ... etc, etc) > and then attempting to constantly merge -stable and other kernel > trees into the repository. git is the tool for managing that, not stacks > of patches. You spend your entire life fixing patch errors and refreshing > fuzz (again, been there, done that). > > So why not just keep a history and constantly merge new versions > into it ? Been there, done that. You end up with an absolute garbage > history of octopus merges and changes that are completely hidden, > non-obvious and useless for collaborating with other kernel projects. > Try merging a new kernel version into those same big features, it's > nearly impossible and you have a franken-kernel that you end up trying > to support and fix yourself. All the bugs are yours and yours alone. > > So that's why there's a repository that tracks the patches and the > configuration and is used for multiple purposes. Keeping the patches > and config blocks separate would just lead to even more errors as > I update one and forget the other, etc, etc. There have been various > incarnations of the tools that also did different things with the patches, > and they weren't skipped, but detected as applied or not on-the-fly, > so there are other historical reasons for the structure as well. > > > - kernel-yocto.bbclass does its own generic job of locating a proper > > BSP using the KMACHINE/KTYPE/KARCH tags in BSP, it looks like > > specifying a specific BSP file would just defeat of this: how should I > > deal with this case where I'm providing both "standard" and "tiny" > > KTYPE's ? > > I'm not quite following the question here, so I can try to answer badly > and you can clarify based on
Re: [yocto] Understanding kernel patching in linux-yocto
On Wed, May 12, 2021 at 7:14 AM Yann Dirson wrote: > > I am currently working on a kmeta BSP for the rockchip-based NanoPI M4 > [1], and I'm wondering how I should be providing kernel patches, as > just add ing "patch" directives in the .scc does not get them applied > unless the particular .scc gets included in KERNEL_FEATURES (see [2]). > > From an old thread [3] I understand that the patches from the standard > kmeta snippets are already applied to the tree, and that to get the > patches from my BSP I'd need to reference it explicitly in SRC_URI > (along with using "nopatch" in the right places to avoid the > already-applied patches to get applied twice). > > I have the feeling that I'm lacking the rationale behind this, and > would need to understand this better to make things right in this BSP. > Especially: > - at first sight, having the patches both applied to linux-yocto and > referenced in yocto-kernel-cache just to be skipped on parsing looks > like both information duplication and parsing of unused lines At least some of this is mentioned in the advanced section of the kernel-dev manual, but I can summarize/reword things here, and I'm also doing a presentation related to this in the Yocto summit at the end of this month. The big thing to remember, is that the configuration and changes you see in that repository, are not only for yocto purposes. The concepts and structure pre-date when they were first brought in to generate reference kernels over 10 years ago (the implementation has changed, but the concepts are still the same). To this day, there still are cases that they are used with just a kernel tree and cross toolchain. With that in mind, the meta-data is used for many different things - It organizes patches / features and their configuration into reusable blocks. At the same time documenting the changes that we have applied to a tree - It makes those patches and configuration blocks available to other kernel trees (for whatever reason). - It configures the tree during the build process, reusing both configuration only and patch + configuration blocks - It is used to generate a history clean tree from scratch for each new supported kernel. Which is what I do when creating new linux-yocto-dev references, and the new /standard/* branches in linux-yocto. So why not just drop all the patches in the SRC_URI ? Been there, done that. It fails spectacularly when you are managing queues of hundreds of potentially conflicting patches (rt, yaffs, aufs, ... etc, etc) and then attempting to constantly merge -stable and other kernel trees into the repository. git is the tool for managing that, not stacks of patches. You spend your entire life fixing patch errors and refreshing fuzz (again, been there, done that). So why not just keep a history and constantly merge new versions into it ? Been there, done that. You end up with an absolute garbage history of octopus merges and changes that are completely hidden, non-obvious and useless for collaborating with other kernel projects. Try merging a new kernel version into those same big features, it's nearly impossible and you have a franken-kernel that you end up trying to support and fix yourself. All the bugs are yours and yours alone. So that's why there's a repository that tracks the patches and the configuration and is used for multiple purposes. Keeping the patches and config blocks separate would just lead to even more errors as I update one and forget the other, etc, etc. There have been various incarnations of the tools that also did different things with the patches, and they weren't skipped, but detected as applied or not on-the-fly, so there are other historical reasons for the structure as well. > - kernel-yocto.bbclass does its own generic job of locating a proper > BSP using the KMACHINE/KTYPE/KARCH tags in BSP, it looks like > specifying a specific BSP file would just defeat of this: how should I > deal with this case where I'm providing both "standard" and "tiny" > KTYPE's ? I'm not quite following the question here, so I can try to answer badly and you can clarify based on my terrible answer. The tools can locate your "bsp entry point" / "bsp definition" in your layer. Either provided by something on the SRC_URI or something in a kmeta repository (also specified on the SRC_URI). Since both of those are added to the search paths they check. Those are just .scc files with a specified KMACHINE/KTYPE that match, and as you could guess from my first term I used, they are the entry point into building the configuration queue. That's where you start inheriting the base configuration(s) and including feature blocks, etc. Those definitions are exactly the same as the internal ones in the kernel-cache repository. By default, that located BSP definition is excluded from inheriting patches .. because as you noted, it would start trying to re-apply changes to the tree. It is there to get the configuration blocks, patc