Hi Andrii,

On Thu Jan 15, 2026 at 6:58 PM CET, Andrii Nakryiko wrote:
> On Wed, Jan 14, 2026 at 12:59 AM Alexis Lothoré (eBPF Foundation)
> <[email protected]> wrote:
>>
>> Hello,
>> this series is part of the larger effort aiming to convert all
>> standalone tests to the CI runners so that they are properly executed on
>> patches submission.
>>
>> Some of those tests are validating bpftool behavior(test_bpftool_map.sh,
>> test_bpftool_metadata.sh, test_bpftool_synctypes.py, test_bpftool.py...)
>> and so they do not integrate well in test_progs. This series proposes to
>
> Can you elaborate why they do not integrate well? In my mind,
> test_progs should be the only runner into which we invest effort
> (parallel tests, all the different filtering, etc; why would we have
> to reimplement subsets of this). The fact that we have test_maps and
> test_verifier is historical and if we had enough time we'd merge all
> of them into test_progs.
>
> What exactly in test_progs would prevent us from implementing bpftool
> test runner?

I don't think there is any strong technical blocker preventing from
integrating those tests directly into test_progs. That's rather about
the fact that test_progs tests depends (almost) exclusively on
libbpf/skeletons. Those bpftool tests rather need to directly execute
bpftool and parse its stdout output, so I thought that it made sense to
have a dedicated runner for this. If I'm wrong and so if those tests
should rather be moved in the test_progs runner (eg to avoid duplicating
the runner features), I'm fine with it. Any additional opinion on this
is welcome.

Thanks,

Alexis
-- 
Alexis Lothoré, Bootlin
Embedded Linux and Kernel engineering
https://bootlin.com


Reply via email to