Krause, Karl-Heinz wrote: > Hi Jan > > Sorry for attaching the wrong patch. Here is the right one.
Thanks. Am I right, it also contains the full A&D RT-kernel? > From all the ipipe patches/hooks we use only a few one. > The ipipe changes we did fall into two categories > - one additional hook in entry.s which checks at system call exit whether > the thread has to migrate to the realtime domain > - optimizations for a dual domain system concerning the dispatching of > events and interrupt handling. The latter makes me wonder if you already analysed the "wired" top-most domain optimisation in recent I-pipe and compared it to your approach. > So if just that kind of dual domain system is of interest, the patches may be > of general interest and we are willing to support it. > To allow for a possible integration we made it configurable as follows > [ ] interrupt pipeline > [ ] rtx domain system > [ ] fast rtx dual domain system A wired ipipe domain is known to stay the top-most one in the pipeline, but this optimisation still allows multiple lower domains, not just Linux. Maybe it is worth looking at the I-pipe abstractions again with the hard-wired scenario "dual domain system" in mind, checking for further optimisation possibilities that do not kill the common model and its clean realisation. Optimisations are not bad, but they also have to be balanced against potentially increased porting effort for other archs. > > Common for both is the check for preemption which we have in the optimized > handle_irq routine as well as in sync_stage. > > More generic are the patches we made for the POSIX message queues and for > saving/restoring the fpu context IIRC, sanitising xxx_fpu services in order to make them domain-agnostic without hacking preempt_enable/disable is something of generic interest for ipipe as well (Or are you patching fpu services for another reason?). I recall to stumble over this on my own when trying to detect invalid calls of Linux services from non-root domains automatically. > > We are not specifically locked to 126.96.36.199 and we don't expect any surprises > if we would upgrade. Hmm, I wouldn't be surprised if things like genirq or upcoming clockevent/hrtimers _will_ cause a few surprises. The latter is actually also a to-do for vanilla ipipe in order to move to 2.6.21 and beyond. > (Of course we know that we have to do some additions to hook into PI > mutexes, despite the fact that for priority changes itself the domain > boundary is already transparent). It is only caused by our general policy to > keep the number of different versions minimal, since we cannot simply forget > older versions, we have to maintain it.(It is not only a kernel for x86, it > is for ARM with and without MMU and for MIPS as well and it is the complete > tool chain). So we always will do some skipping and we haven't decided yet > what our next version will be. Does your RT-kernel run on ARM and MIPS as well already? Did you port ipipe (or a subset) to MIPS? > > Concerning testing we took mainly the existing POSIX test suites which we > had to make suitable for static priorities > 0. We added a few tests for > checking the transparency of the domain boundary for POSIX message queues, > futexes and priority changes and we combined them with performance > measurement that is all. Those changes may be of some interest for POSIX testing with Xenomai as well, that's why I was asking for a complete, instrumented testsuite. If you could share it with the community, that would be great. > > Just let me know what else you need. > > > Karl-Heinz > Thanks for making your code available. I think both sides can benefit from discussing new approaches based on public code. Jan
Description: OpenPGP digital signature
_______________________________________________ Xenomai-core mailing list Xenomaifirstname.lastname@example.org https://mail.gna.org/listinfo/xenomai-core