Re: [Adeos-main] Re: [Xenomai-core] I-pipe latency tracer for ppc
Jan Kiszka wrote: Wolfgang Grandegger wrote: Jan Kiszka wrote: Wolfgang Grandegger wrote: ... This trace covers some long stall of the root domain, not the maximum stall of the Xenomai domain that caused the 140 us above. See the explanation for the stall bits above. Ah, some feed for the real experts. Indeed :), comments below. Maybe the maximum irq latency tracing is broken, I haven't tested it with latest ipipe patch revision yet. Well, that max path should not trigger on root-domain stalls as long as there is some higher priority domain running - and it should have been reset on xenomai mounting. Hmm, maybe only that reset does not work, please check by running "echo > /proc/ipipe/trace/max" before the latency test. Yes, that changed something. I have attached the new output listing. You could also test if the -f option of latency works for you. You will find the result of the back-trace freezing on max latency under /proc/ipipe/trace/frozen. The number of back-trace points can be tuned even after the event (only post-tracing required ahead-of-time tuning). Seem to work. The output has been added to the attached file. Wolfgang. bash-3.00# cat /proc/ipipe/trace/max I-pipe worst-case tracing service on 2.6.14/ipipe-1.3-00 Begin: 90358006525 cycles, Trace Points: 25 (-10/+4), Length: 69 us +- Hard IRQs ('|': locked) |+ ||+--- |||+-- Xenomai +- Linux ('*': domain stalled) | +-- Delay flag ('+': > 1 us, '!': > 10 us) | |+- NMI noise ('N') | || Type User Val. TimeDelay Function (Parent) fn -251.935 schedule+0x14 (syscall_exit_work+0x110) fn -232.075 profile_hit+0x14 (schedule+0x68) fn -212.390 sched_clock+0x14 (schedule+0xa0) fn -194.400 __ipipe_stall_root+0x14 (schedule+0x118) *fn -142.040 __ipipe_dispatch_event+0x14 (schedule+0x300) | *begin 0x8001 -123.190 __ipipe_dispatch_event+0x240 (schedule+0x300) | *end0x8001-91.990 __ipipe_dispatch_event+0x228 (schedule+0x300) *fn -74.115 schedule_event+0x14 (__ipipe_dispatch_event+0x160) | *begin 0x8001-31.835 __ipipe_dispatch_event+0x198 (schedule+0x300) | *end0x8001-11.815 __ipipe_dispatch_event+0x1b8 (schedule+0x300) | *begin 0x8000 0+ 4.715 schedule+0x6ac (syscall_exit_work+0x110) :| *fn4+ 1.225 __switch_to+0x14 (schedule+0x358) :| *fn5+ 5.025 __ipipe_test_and_stall_root+0x14 (__switch_to+0x30) :*fn 10+ 1.445 schedule_tail+0x14 (ret_from_fork+0x60) :*fn 12+ 3.450 __ipipe_unstall_root+0x14 (schedule_tail+0x48) Instrumentation of the maximum IRQs-off path seems to be broken. The two functions above were entered with hard-IRQs on, so they should have interrupted the trace path. You mean that something in the instrumentation of local_irq*_hw() might be broken (asm/hw_irq.h on this arch), or would some tracepoint be missing somewhere? :| *fn 15+ 1.145 __ipipe_grab_timer+0x14 (__ipipe_ret_from_except+0x0) :| *begin 0xc00466f416+ 1.615 __ipipe_grab_timer+0x34 (__ipipe_ret_from_except+0x0) Hmm, the User Value on x86 would contain the IRQ number here. Some forgotten dereferencing? Looks like a pointer. That was intended, but merely useless actually. It's the value of the link register when the interrupt handler is branched to, and this value is always __ipipe_ret_from_exception. Will fix. :| *fn 18+ 3.665 __ipipe_handle_irq+0x14 (__ipipe_grab_timer+0x4c) :| *fn 22+ 1.980 __ipipe_walk_pipeline+0x14 (__ipipe_handle_irq+0x174) :| *fn 24+ 2.540 ipipe_suspend_domain+0x14 (__ipipe_walk_pipeline+0xe4) :| *fn 26+ 3.600 __ipipe_sync_stage+0x14 (ipipe_suspend_domain+0xa8) :| **fn 30+ 1.650 xnintr_clock_handler+0x14 (__ipipe_sync_stage+0x1ac) :| **fn 31+ 2.150 xnintr_irq_handler+0x14 (xnintr_clock_handler+0x38) :| **fn 34+ 1.980 xnpod_announce_tick+0x14 (xnintr_irq_handler+0x58) :| **fn 36+ 3.395 xntimer_do_tick_aperiodic+0x14 (xnpod_announce_tick+0x3c) :| **fn 39+ 1.645 xnthread_periodic_handler+0x14 (xntimer_do_tick_aperiodic+0x2b4) :| **fn 41+ 1.420 xnpod_resume_thread+0x14 (xnthread_periodic_handler+0x3c) :| **fn 42+ 4.010 ipipe_test_and_stall_pipeline_head+0x14 (xnpod_resume_thread+0x34) :| **fn 46+ 4.910 ipip
Re: [Adeos-main] Re: [Xenomai-core] I-pipe latency tracer for ppc
Wolfgang Grandegger wrote: > Jan Kiszka wrote: >> Wolfgang Grandegger wrote: >> ... >> This trace covers some long stall of the root domain, not the maximum >> stall of the Xenomai domain that caused the 140 us above. See the >> explanation for the stall bits above. > > Ah, some feed for the real experts. Indeed :), comments below. > >> >> Maybe the maximum irq latency tracing is broken, I haven't tested it >> with latest ipipe patch revision yet. Well, that max path should not >> trigger on root-domain stalls as long as there is some higher priority >> domain running - and it should have been reset on xenomai mounting. Hmm, >> maybe only that reset does not work, please check by running "echo > >> /proc/ipipe/trace/max" before the latency test. > > Yes, that changed something. I have attached the new output listing. > >> You could also test if the -f option of latency works for you. You will >> find the result of the back-trace freezing on max latency under >> /proc/ipipe/trace/frozen. The number of back-trace points can be tuned >> even after the event (only post-tracing required ahead-of-time tuning). > > Seem to work. The output has been added to the attached file. > > Wolfgang. > > > > > bash-3.00# cat /proc/ipipe/trace/max > I-pipe worst-case tracing service on 2.6.14/ipipe-1.3-00 > > Begin: 90358006525 cycles, Trace Points: 25 (-10/+4), Length: 69 us > > +- Hard IRQs ('|': locked) > |+ > ||+--- > |||+-- Xenomai > +- Linux ('*': domain stalled) > | +-- Delay flag ('+': > 1 us, '!': > 10 > us) > | |+- NMI noise ('N') > | || > Type User Val. TimeDelay Function (Parent) > fn -251.935 schedule+0x14 > (syscall_exit_work+0x110) > fn -232.075 profile_hit+0x14 (schedule+0x68) > fn -212.390 sched_clock+0x14 (schedule+0xa0) > fn -194.400 __ipipe_stall_root+0x14 > (schedule+0x118) > *fn -142.040 __ipipe_dispatch_event+0x14 > (schedule+0x300) > | *begin 0x8001 -123.190 __ipipe_dispatch_event+0x240 > (schedule+0x300) > | *end0x8001-91.990 __ipipe_dispatch_event+0x228 > (schedule+0x300) > *fn -74.115 schedule_event+0x14 > (__ipipe_dispatch_event+0x160) > | *begin 0x8001-31.835 __ipipe_dispatch_event+0x198 > (schedule+0x300) > | *end0x8001-11.815 __ipipe_dispatch_event+0x1b8 > (schedule+0x300) >> | *begin 0x8000 0+ 4.715 schedule+0x6ac >> (syscall_exit_work+0x110) > :| *fn4+ 1.225 __switch_to+0x14 (schedule+0x358) > :| *fn5+ 5.025 __ipipe_test_and_stall_root+0x14 > (__switch_to+0x30) > :*fn 10+ 1.445 schedule_tail+0x14 > (ret_from_fork+0x60) > :*fn 12+ 3.450 __ipipe_unstall_root+0x14 > (schedule_tail+0x48) Instrumentation of the maximum IRQs-off path seems to be broken. The two functions above were entered with hard-IRQs on, so they should have interrupted the trace path. > :| *fn 15+ 1.145 __ipipe_grab_timer+0x14 > (__ipipe_ret_from_except+0x0) > :| *begin 0xc00466f416+ 1.615 __ipipe_grab_timer+0x34 > (__ipipe_ret_from_except+0x0) Hmm, the User Value on x86 would contain the IRQ number here. Some forgotten dereferencing? Looks like a pointer. > :| *fn 18+ 3.665 __ipipe_handle_irq+0x14 > (__ipipe_grab_timer+0x4c) > :| *fn 22+ 1.980 __ipipe_walk_pipeline+0x14 > (__ipipe_handle_irq+0x174) > :| *fn 24+ 2.540 ipipe_suspend_domain+0x14 > (__ipipe_walk_pipeline+0xe4) > :| *fn 26+ 3.600 __ipipe_sync_stage+0x14 > (ipipe_suspend_domain+0xa8) > :| **fn 30+ 1.650 xnintr_clock_handler+0x14 > (__ipipe_sync_stage+0x1ac) > :| **fn 31+ 2.150 xnintr_irq_handler+0x14 > (xnintr_clock_handler+0x38) > :| **fn 34+ 1.980 xnpod_announce_tick+0x14 > (xnintr_irq_handler+0x58) > :| **fn 36+ 3.395 xntimer_do_tick_aperiodic+0x14 > (xnpod_announce_tick+0x3c) > :| **fn 39+ 1.645 xnthread_periodic_handler+0x14 > (xntimer_do_tick_aperiodic+0x2b4) > :| **fn 41+ 1.420 xnpod_resume_thread+0x14 > (xnthread_periodic_handler+0x3c) > :| **fn 42+ 4.010 > ipipe_test_and_stall_pipeline_head+0x14 (xnpod_resume_thread+0x34) > :| **fn 46+ 4.910 ipipe_restore_pipeline_head+0x14 > (xnpod_resume_thread+0x120) > :| **fn 51+ 1.870 xnpod_schedule+0x14 > (xnintr_irq_handler+0xf0) > :| **fn
Re: [Adeos-main] Re: [Xenomai-core] I-pipe latency tracer for ppc
Jan Kiszka wrote: Wolfgang Grandegger wrote: Philippe Gerum wrote: As promised, the I-pipe tracer has been ported to ppc. People working on this architecture are invited to give it a try, it's a great tool to find out where the cycles are actually going. Just apply the tracer patch on top of the Adeos patch bearing the same revision number, and select the feature from the kernel configuration. If you happen to have some issue while booting the instrumented kernel on embedded setups (e.g. freeze after kernel decompression), try reducing the size of the trace log (CONFIG_IPIPE_TRACE_SHIFT). http://download.gna.org/adeos/patches/v2.6/ppc/tracer/ I gave it a try on a rather low-end PowerPC Walnut board (AMCC PowerPC 405GP, Rev. E at 200 MHz, 16 kB I-Cache 8 kB D-Cache). A nice tool, indeed, and it works fine. I have attached the results from running the Xenomai latency "-t0" test with some load showing latencies up to 140 us. The trace seems not to show any obvious problems, I think. Thanks. Wolfgang. bash-3.00# cat /proc/ipipe/version 1.3-00 bash-3.00# ./latency -p500 == Sampling period: 500 us == Test mode: periodic user-mode task == All results in microseconds warming up... ... RTH|-lat min|-lat avg|-lat max|-overrun|lat best|---lat worst RTS| 59.020| 67.260| 138.910| 0|00:54:31/00:54:31 bash-3.00# cat /proc/ipipe/trace/max I-pipe worst-case tracing service on 2.6.14/ipipe-1.3-00 Begin: 182095500 cycles, Trace Points: 192 (-10/+3), Length: 265 us +- Hard IRQs ('|': locked) |+ ||+--- |||+-- Xenomai +- Linux ('*': domain stalled) | +-- Delay flag ('+': > 1 us, '!': > 10 us) | |+- NMI noise ('N') | || Type User Val. TimeDelay Function (Parent) *fn -303.465 timer_interrupt+0x14 (__ipipe_do_timer+0x30) *fn -262.430 profile_tick+0x14 (timer_interrupt+0x130) *fn -241.865 profile_hit+0x14 (profile_tick+0x78) *fn -221.700 update_process_times+0x14 (timer_interrupt+0x13c) *fn -203.505 account_system_time+0x14 (update_process_times+0xac) *fn -172.195 update_mem_hiwater+0x14 (account_system_time+0x78) *fn -151.200 run_local_timers+0x14 (update_process_times+0xb0) *fn -14 11.905 raise_softirq+0x14 (run_local_timers+0x30) *fn -21.200 __ipipe_restore_root+0x14 (raise_softirq+0x84) *fn -11.095 __ipipe_stall_root+0x14 (__ipipe_restore_root+0x34) | *begin 0x8000 0+ 3.950 __ipipe_stall_root+0x98 (__ipipe_restore_root+0x34) :| *fn3+ 1.500 scheduler_tick+0x14 (update_process_times+0x7c) :| *fn5+ 4.040 sched_clock+0x14 (scheduler_tick+0x34) This trace covers some long stall of the root domain, not the maximum stall of the Xenomai domain that caused the 140 us above. See the explanation for the stall bits above. Ah, some feed for the real experts. Maybe the maximum irq latency tracing is broken, I haven't tested it with latest ipipe patch revision yet. Well, that max path should not trigger on root-domain stalls as long as there is some higher priority domain running - and it should have been reset on xenomai mounting. Hmm, maybe only that reset does not work, please check by running "echo > /proc/ipipe/trace/max" before the latency test. Yes, that changed something. I have attached the new output listing. You could also test if the -f option of latency works for you. You will find the result of the back-trace freezing on max latency under /proc/ipipe/trace/frozen. The number of back-trace points can be tuned even after the event (only post-tracing required ahead-of-time tuning). Seem to work. The output has been added to the attached file. Wolfgang. bash-3.00# cat /proc/ipipe/trace/max I-pipe worst-case tracing service on 2.6.14/ipipe-1.3-00 Begin: 90358006525 cycles, Trace Points: 25 (-10/+4), Length: 69 us +- Hard IRQs ('|': locked) |+ ||+--- |||+-- Xenomai +- Linux ('*': domain stalled) | +-- Delay flag ('+': > 1 us, '!': > 10 us) | |+- NMI noise ('N') | || Type User Val. TimeDelay Function (Parent) fn -251.935 schedule+0x14 (syscall_exit_work+0x110) fn -232.075 profile_hit+0x14 (schedule+0x68) fn -212.390 sched_clock+0x14 (schedule+0xa0) fn -
Re: [Adeos-main] Re: [Xenomai-core] I-pipe latency tracer for ppc
Wolfgang Grandegger wrote: > Philippe Gerum wrote: >> >> As promised, the I-pipe tracer has been ported to ppc. People working >> on this architecture are invited to give it a try, it's a great tool >> to find out where the cycles are actually going. >> >> Just apply the tracer patch on top of the Adeos patch bearing the same >> revision number, and select the feature from the kernel configuration. >> If you happen to have some issue while booting the instrumented kernel >> on embedded setups (e.g. freeze after kernel decompression), try >> reducing the size of the trace log (CONFIG_IPIPE_TRACE_SHIFT). >> >> http://download.gna.org/adeos/patches/v2.6/ppc/tracer/ > > I gave it a try on a rather low-end PowerPC Walnut board (AMCC PowerPC > 405GP, Rev. E at 200 MHz, 16 kB I-Cache 8 kB D-Cache). A nice tool, > indeed, and it works fine. I have attached the results from running the > Xenomai latency "-t0" test with some load showing latencies up to 140 > us. The trace seems not to show any obvious problems, I think. > > Thanks. > > Wolfgang. > > > > > > bash-3.00# cat /proc/ipipe/version > 1.3-00 > > bash-3.00# ./latency -p500 > == Sampling period: 500 us > == Test mode: periodic user-mode task > == All results in microseconds > warming up... > ... > RTH|-lat min|-lat avg|-lat max|-overrun|lat best|---lat worst > RTS| 59.020| 67.260| 138.910| 0|00:54:31/00:54:31 > > bash-3.00# cat /proc/ipipe/trace/max > I-pipe worst-case tracing service on 2.6.14/ipipe-1.3-00 > > Begin: 182095500 cycles, Trace Points: 192 (-10/+3), Length: 265 us > > +- Hard IRQs ('|': locked) > |+ > ||+--- > |||+-- Xenomai > +- Linux ('*': domain stalled) > | +-- Delay flag ('+': > 1 us, '!': > 10 > us) > | |+- NMI noise ('N') > | || > Type User Val. TimeDelay Function (Parent) > *fn -303.465 timer_interrupt+0x14 > (__ipipe_do_timer+0x30) > *fn -262.430 profile_tick+0x14 > (timer_interrupt+0x130) > *fn -241.865 profile_hit+0x14 (profile_tick+0x78) > *fn -221.700 update_process_times+0x14 > (timer_interrupt+0x13c) > *fn -203.505 account_system_time+0x14 > (update_process_times+0xac) > *fn -172.195 update_mem_hiwater+0x14 > (account_system_time+0x78) > *fn -151.200 run_local_timers+0x14 > (update_process_times+0xb0) > *fn -14 11.905 raise_softirq+0x14 > (run_local_timers+0x30) > *fn -21.200 __ipipe_restore_root+0x14 > (raise_softirq+0x84) > *fn -11.095 __ipipe_stall_root+0x14 > (__ipipe_restore_root+0x34) >> | *begin 0x8000 0+ 3.950 __ipipe_stall_root+0x98 >> (__ipipe_restore_root+0x34) > :| *fn3+ 1.500 scheduler_tick+0x14 > (update_process_times+0x7c) > :| *fn5+ 4.040 sched_clock+0x14 (scheduler_tick+0x34) This trace covers some long stall of the root domain, not the maximum stall of the Xenomai domain that caused the 140 us above. See the explanation for the stall bits above. Maybe the maximum irq latency tracing is broken, I haven't tested it with latest ipipe patch revision yet. Well, that max path should not trigger on root-domain stalls as long as there is some higher priority domain running - and it should have been reset on xenomai mounting. Hmm, maybe only that reset does not work, please check by running "echo > /proc/ipipe/trace/max" before the latency test. You could also test if the -f option of latency works for you. You will find the result of the back-trace freezing on max latency under /proc/ipipe/trace/frozen. The number of back-trace points can be tuned even after the event (only post-tracing required ahead-of-time tuning). Jan signature.asc Description: OpenPGP digital signature ___ Xenomai-core mailing list Xenomai-core@gna.org https://mail.gna.org/listinfo/xenomai-core