Am Mittwoch, den 06.11.2013, 20:59 +0100 schrieb Richard Weinberger: > Am 06.11.2013 20:52, schrieb Thomas Meyer: > > Am Mittwoch, den 06.11.2013, 13:40 +0100 schrieb Richard Weinberger: > >> On Tue, Nov 5, 2013 at 9:21 PM, Thomas Meyer <tho...@m3y3r.de> wrote: > >>> Hi, > >>> > >>> I'm running Fedora 20 inside a 3.12 UML kernel and the "yum upgrade -y" > >>> command seems to get stuck after a while/few minutes. > >>> > >>> Any ideas what's going one here? How to debug this? > >>> > >>> It looks like the process running yum is in state ptrace stopped, but > >>> doesn't continue. > >> > >> Got only yum stuck or the whole UML kernel? > > > > How to tell? It feels like the whole kernel got stuck. > > Login on another shell... :)
oops, yes, of course! I tried that and the getty was able to receive my username, but after pressing enter nothing happened any more, some timeout message appeared. I logged in via mconsole and did an emergency sync and a halt command and restarted uml. now I'm receiving funny warnings/bad pages: systemd-udevd[599]: starting version 208 EXT4-fs (ubda): re-mounted. Opts: (null) systemd-journald[338]: Received request to flush runtime journal from PID 1 systemd-journald[338]: File /var/log/journal/8e4cbfea404512ae70096c6202c9a3bf/system.journal corrupted or uncleanly shut down, renaming and replacing. ------------[ cut here ]------------ WARNING: CPU: 0 PID: 0 at lib/timerqueue.c:74 timerqueue_del+0x57/0x60() Modules linked in: CPU: 0 PID: 0 Comm: swapper Not tainted 3.12.0-00048-gbe408cd #22 6036d968 60370947 00000000 00000000 00000009 60324f52 0000004a 6036d9b0 602972e5 6036da00 60030bc1 6036d9d0 90b153c0 6036da10 603816c0 6037e4e8 00000000 00000001 8dccbcf0 6036da10 60030d15 6036da30 601cf007 603816c0 Call Trace: 6036d9a8: [<602972e5>] dump_stack+0x17/0x19 6036d9b8: [<60030bc1>] warn_slowpath_common+0x71/0x90 6036da08: [<60030d15>] warn_slowpath_null+0x15/0x20 6036da18: [<601cf007>] timerqueue_del+0x57/0x60 6036da38: [<6004dfe6>] __remove_hrtimer+0x46/0xa0 6036da78: [<6004e548>] __hrtimer_start_range_ns+0xd8/0x1e0 6036dad8: [<6004e683>] hrtimer_start+0x13/0x20 6036dae8: [<60067e12>] __tick_nohz_idle_enter.constprop.28+0x2d2/0x360 6036db58: [<60068329>] tick_nohz_irq_exit+0x19/0x20 6036db68: [<60034f45>] irq_exit+0x85/0xa0 6036db78: [<6005dac8>] generic_handle_irq+0x28/0x30 6036db88: [<60014fa8>] do_IRQ+0x28/0x40 6036dba8: [<60016af0>] timer_handler+0x20/0x30 6036dbc8: [<6002834e>] real_alarm_handler+0x3e/0x50 6036dd38: [<60028b53>] os_nsecs+0x13/0x30 6036dd58: [<60028b53>] os_nsecs+0x13/0x30 6036dd78: [<60028a98>] timer_one_shot+0x68/0x80 6036dda8: [<60016acc>] itimer_next_event+0xc/0x10 6036ddb8: [<6006704a>] clockevents_program_event+0x6a/0xf0 6036dde8: [<6006780a>] tick_program_event+0x1a/0x20 6036ddf8: [<6004df9b>] hrtimer_force_reprogram+0x6b/0x70 6036de08: [<6004e03c>] __remove_hrtimer+0x9c/0xa0 6036de28: [<601cef40>] timerqueue_add+0x60/0xb0 6036de48: [<6004e514>] __hrtimer_start_range_ns+0xa4/0x1e0 6036dea8: [<6004e683>] hrtimer_start+0x13/0x20 6036deb8: [<6007c298>] rcu_sched_qs+0x78/0x90 6036dee8: [<60028224>] unblock_signals+0x64/0x80 6036df08: [<60015b0a>] arch_cpu_idle+0x3a/0x50 6036df18: [<6007c3cf>] rcu_idle_enter+0x6f/0xb0 6036df38: [<6005da4e>] cpu_startup_entry+0x8e/0xe0 6036df48: [<602991b9>] schedule_preempt_disabled+0x9/0x10 6036df58: [<60294228>] rest_init+0x68/0x70 6036df68: [<600027f5>] check_bugs+0xe/0x19 6036df78: [<60001635>] start_kernel+0x27f/0x286 6036df80: [<600011c5>] unknown_bootoption+0x0/0x185 6036dfb8: [<6000289a>] start_kernel_proc+0x31/0x35 6036dfd8: [<6001561a>] new_thread_handler+0x7a/0xa0 ---[ end trace 41ecadffe5cf650c ]--- BUG: Bad page state in process systemd-journal pfn:2d25c page:0000000062c97420 count:0 mapcount:1 mapping: (null) index:0x2 page flags: 0x80008(uptodate|swapbacked) Modules linked in: CPU: 0 PID: 338 Comm: systemd-journal Tainted: G W 3.12.0-00048-gbe408cd #22 8dc759b8 60370947 60309985 62c97420 00000000 62c97458 60383ac0 8dc75a00 602972e5 8dc75a30 60082ce0 ffffffffff0a0210 8dc75adc 8dc75a50 62c97420 8dc75a60 60082df0 60309982 62c97420 00080008 00000000 8dc75aa0 60084561 00000000 Call Trace: 8dc759f8: [<602972e5>] dump_stack+0x17/0x19 8dc75a08: [<60082ce0>] bad_page+0xb0/0x100 8dc75a38: [<60082df0>] free_pages_prepare+0xc0/0xd0 8dc75a68: [<60084561>] free_hot_cold_page+0x21/0x130 8dc75aa8: [<60084ace>] free_hot_cold_page_list+0x3e/0x60 8dc75ad8: [<60087958>] release_pages+0x158/0x1a0 8dc75b38: [<60087a62>] pagevec_lru_move_fn+0xc2/0xe0 8dc75b50: [<60087120>] __pagevec_lru_add_fn+0x0/0xc0 8dc75b98: [<60087e91>] lru_add_drain_cpu+0x71/0x80 8dc75bb8: [<60087f6b>] lru_add_drain+0xb/0x10 8dc75bc8: [<600a0136>] exit_mmap+0x46/0x170 8dc75bf0: [<60018a50>] copy_chunk_to_user+0x0/0x30 8dc75c28: [<6002e2a7>] mmput.part.62+0x27/0xc0 8dc75c48: [<6002e359>] mmput+0x19/0x20 8dc75c58: [<6003117a>] exit_mm+0x10a/0x150 8dc75ca8: [<60031de1>] do_exit+0x331/0x4a0 8dc75ce8: [<600330de>] do_group_exit+0x3e/0xd0 8dc75d18: [<6003de7b>] get_signal_to_deliver+0x1bb/0x4d0 8dc75d48: [<600535cb>] wake_up_state+0xb/0x10 8dc75d88: [<60016757>] kern_do_signal+0x57/0x150 8dc75e08: [<60028460>] set_signals+0x30/0x40 8dc75e28: [<6003d175>] force_sig_info+0xb5/0xd0 8dc75e68: [<60016871>] do_signal+0x21/0x30 8dc75e88: [<60017d14>] fatal_sigsegv+0x24/0x30 8dc75ea8: [<6002b2d3>] userspace+0x2c3/0x4d0 8dc75f78: [<600273d7>] save_registers+0x17/0x30 8dc75f88: [<6002df20>] arch_prctl+0x150/0x180 8dc75fd8: [<600156a9>] fork_handler+0x69/0x70 Disabling lock debugging due to kernel taint BUG: Bad page state in process systemd-journal pfn:2cc8c page:0000000062c82ea0 count:0 mapcount:1 mapping: (null) index:0x3 page flags: 0x80008(uptodate|swapbacked) Modules linked in: CPU: 0 PID: 338 Comm: systemd-journal Tainted: G B W 3.12.0-00048-gbe408cd #22 8dc759b8 60370947 60309985 62c82ea0 00000000 62c82ed8 60383ac0 8dc75a00 602972e5 8dc75a30 60082ce0 ffffffffff0a0210 8dc75adc 8dc75a50 62c82ea0 8dc75a60 60082df0 60309982 62c82ea0 00080008 00000000 8dc75aa0 60084561 00000000 Call Trace: 8dc759f8: [<602972e5>] dump_stack+0x17/0x19 8dc75a08: [<60082ce0>] bad_page+0xb0/0x100 8dc75a38: [<60082df0>] free_pages_prepare+0xc0/0xd0 8dc75a68: [<60084561>] free_hot_cold_page+0x21/0x130 8dc75aa8: [<60084ace>] free_hot_cold_page_list+0x3e/0x60 8dc75ad8: [<60087958>] release_pages+0x158/0x1a0 8dc75b38: [<60087a62>] pagevec_lru_move_fn+0xc2/0xe0 8dc75b50: [<60087120>] __pagevec_lru_add_fn+0x0/0xc0 8dc75b98: [<60087e91>] lru_add_drain_cpu+0x71/0x80 8dc75bb8: [<60087f6b>] lru_add_drain+0xb/0x10 8dc75bc8: [<600a0136>] exit_mmap+0x46/0x170 8dc75bf0: [<60018a50>] copy_chunk_to_user+0x0/0x30 8dc75c28: [<6002e2a7>] mmput.part.62+0x27/0xc0 8dc75c48: [<6002e359>] mmput+0x19/0x20 8dc75c58: [<6003117a>] exit_mm+0x10a/0x150 8dc75ca8: [<60031de1>] do_exit+0x331/0x4a0 8dc75ce8: [<600330de>] do_group_exit+0x3e/0xd0 8dc75d18: [<6003de7b>] get_signal_to_deliver+0x1bb/0x4d0 8dc75d48: [<600535cb>] wake_up_state+0xb/0x10 8dc75d88: [<60016757>] kern_do_signal+0x57/0x150 8dc75e08: [<60028460>] set_signals+0x30/0x40 8dc75e28: [<6003d175>] force_sig_info+0xb5/0xd0 8dc75e68: [<60016871>] do_signal+0x21/0x30 8dc75e88: [<60017d14>] fatal_sigsegv+0x24/0x30 8dc75ea8: [<6002b2d3>] userspace+0x2c3/0x4d0 8dc75f78: [<600273d7>] save_registers+0x17/0x30 8dc75f88: [<6002df20>] arch_prctl+0x150/0x180 8dc75fd8: [<600156a9>] fork_handler+0x69/0x70 BUG: failure at mm/slab.c:1813/kmem_freepages()! Kernel panic - not syncing: BUG! CPU: 0 PID: 338 Comm: systemd-journal Tainted: G B W 3.12.0-00048-gbe408cd #22 8dc75988 60370947 00000000 602ffc6e 8de9a000 9080cdf0 00000000 8dc759d0 602972e5 8dc75ad0 60294a4a 00000000 3000000008 8dc75ae0 8dc75a00 8dc75a30 60423e00 00000007 00000006 8dc75720 000000f7 00000715 602a4944 ffffffffffffffff Call Trace: 8dc759c8: [<602972e5>] dump_stack+0x17/0x19 8dc759d8: [<60294a4a>] panic+0xf4/0x1e2 8dc75a68: [<60084619>] free_hot_cold_page+0xd9/0x130 8dc75aa8: [<60087362>] __put_single_page+0x22/0x30 8dc75ad8: [<600ae4a5>] kmem_freepages.isra.69+0x135/0x140 8dc75af8: [<600aedb9>] slab_destroy+0x29/0x60 8dc75b18: [<600aef2c>] free_block+0x13c/0x150 8dc75b48: [<60296210>] cache_flusharray+0x60/0x84 8dc75b78: [<600aed87>] kmem_cache_free+0xa7/0xb0 8dc75ba8: [<6009d1f5>] remove_vma+0x45/0x50 8dc75bc8: [<600a01b4>] exit_mmap+0xc4/0x170 8dc75c28: [<6002e2a7>] mmput.part.62+0x27/0xc0 8dc75c48: [<6002e359>] mmput+0x19/0x20 8dc75c58: [<6003117a>] exit_mm+0x10a/0x150 8dc75ca8: [<60031de1>] do_exit+0x331/0x4a0 8dc75ce8: [<600330de>] do_group_exit+0x3e/0xd0 8dc75d18: [<6003de7b>] get_signal_to_deliver+0x1bb/0x4d0 8dc75d48: [<600535cb>] wake_up_state+0xb/0x10 8dc75d88: [<60016757>] kern_do_signal+0x57/0x150 8dc75e08: [<60028460>] set_signals+0x30/0x40 8dc75e28: [<6003d175>] force_sig_info+0xb5/0xd0 8dc75e68: [<60016871>] do_signal+0x21/0x30 8dc75e88: [<60017d14>] fatal_sigsegv+0x24/0x30 8dc75ea8: [<6002b2d3>] userspace+0x2c3/0x4d0 8dc75f78: [<600273d7>] save_registers+0x17/0x30 8dc75f88: [<6002df20>] arch_prctl+0x150/0x180 8dc75fd8: [<600156a9>] fork_handler+0x69/0x70 Modules linked in: Pid: 338, comm: systemd-journal Tainted: G B W 3.12.0-00048-gbe408cd RIP: 0033:[<0000000041621463>] RSP: 0000007fbfe99b28 EFLAGS: 00000246 RAX: 0000000000000001 RBX: 0000007fbfe99b40 RCX: ffffffffffffffff RDX: 0000000000000001 RSI: 0000007fbfe99b30 RDI: 0000000000000007 RBP: 00000000ffffffff R08: 0000000000000001 R09: 000000552ace5943 R10: 00000000ffffffff R11: 0000000000000246 R12: 0000007fbfe99b30 R13: 00000000000003e8 R14: 0004ea87f4e0a28c R15: 0000000000000000 Call Trace: 8dc75968: [<6001837b>] panic_exit+0x2b/0x50 8dc75978: [<60016990>] show_stack+0x40/0xe0 8dc75988: [<6004fb2c>] notifier_call_chain+0x4c/0x70 8dc759c8: [<6004fb71>] atomic_notifier_call_chain+0x11/0x20 8dc759d8: [<60294a5b>] panic+0x105/0x1e2 8dc75a68: [<60084619>] free_hot_cold_page+0xd9/0x130 8dc75aa8: [<60087362>] __put_single_page+0x22/0x30 8dc75ad8: [<600ae4a5>] kmem_freepages.isra.69+0x135/0x140 8dc75af8: [<600aedb9>] slab_destroy+0x29/0x60 8dc75b18: [<600aef2c>] free_block+0x13c/0x150 8dc75b48: [<60296210>] cache_flusharray+0x60/0x84 8dc75b78: [<600aed87>] kmem_cache_free+0xa7/0xb0 8dc75ba8: [<6009d1f5>] remove_vma+0x45/0x50 8dc75bc8: [<600a01b4>] exit_mmap+0xc4/0x170 8dc75c28: [<6002e2a7>] mmput.part.62+0x27/0xc0 8dc75c48: [<6002e359>] mmput+0x19/0x20 8dc75c58: [<6003117a>] exit_mm+0x10a/0x150 8dc75ca8: [<60031de1>] do_exit+0x331/0x4a0 8dc75ce8: [<600330de>] do_group_exit+0x3e/0xd0 8dc75d18: [<6003de7b>] get_signal_to_deliver+0x1bb/0x4d0 8dc75d48: [<600535cb>] wake_up_state+0xb/0x10 8dc75d88: [<60016757>] kern_do_signal+0x57/0x150 8dc75e08: [<60028460>] set_signals+0x30/0x40 8dc75e28: [<6003d175>] force_sig_info+0xb5/0xd0 8dc75e68: [<60016871>] do_signal+0x21/0x30 8dc75e88: [<60017d14>] fatal_sigsegv+0x24/0x30 8dc75ea8: [<6002b2d3>] userspace+0x2c3/0x4d0 8dc75f78: [<600273d7>] save_registers+0x17/0x30 8dc75f88: [<6002df20>] arch_prctl+0x150/0x180 8dc75fd8: [<600156a9>] fork_handler+0x69/0x70 ------------[ cut here ]------------ WARNING: CPU: 0 PID: 618 at lib/timerqueue.c:74 timerqueue_del +0x57/0x60() Modules linked in: CPU: 0 PID: 618 Comm: plymouthd Not tainted 3.12.0-00048-gbe408cd #22 6036f5b8 60370947 6004fc41 00000000 00000009 60324f52 0000004a 6036f600 602972e5 6036f650 60030bc1 6036f830 6036f630 00000000 603816c0 6037e4e8 00000002 00000000 603816c0 6036f660 60030d15 6036f680 601cf007 603816c0 Call Trace: 6036f5c8: [<6004fc41>] raw_notifier_call_chain+0x11/0x20 6036f5f8: [<602972e5>] dump_stack+0x17/0x19 6036f608: [<60030bc1>] warn_slowpath_common+0x71/0x90 6036f658: [<60030d15>] warn_slowpath_null+0x15/0x20 6036f668: [<601cf007>] timerqueue_del+0x57/0x60 6036f688: [<6004dfe6>] __remove_hrtimer+0x46/0xa0 6036f6c8: [<6004e2fd>] __run_hrtimer.isra.35+0x2d/0xf0 6036f6e8: [<6004e7c7>] hrtimer_interrupt+0xb7/0x210 6036f738: [<60016a3f>] um_timer+0xf/0x20 6036f748: [<6005e0f1>] handle_irq_event_percpu+0x31/0x130 6036f798: [<6005e213>] handle_irq_event+0x23/0x40 6036f7b8: [<60060287>] handle_edge_irq+0x67/0x120 6036f7d8: [<6005dac8>] generic_handle_irq+0x28/0x30 6036f7e8: [<60014fa3>] do_IRQ+0x23/0x40 6036f808: [<60016af0>] timer_handler+0x20/0x30 6036f828: [<6002834e>] real_alarm_handler+0x3e/0x50 6036f838: [<6001a400>] winch_thread+0x0/0x1b0 6036f8b0: [<6003f612>] sigsuspend+0x22/0x90 6036fb48: [<6002840a>] alarm_handler+0x3a/0x50 6036fb68: [<60027f7d>] hard_handler+0x7d/0xc0 6036fba0: [<6001a400>] winch_thread+0x0/0x1b0 6036fc18: [<6001a400>] winch_thread+0x0/0x1b0 6036fc68: [<6003f612>] sigsuspend+0x22/0x90 ---[ end trace 41ecadffe5cf650c ]--- > > >> Does yum a ptrace() within UML or did you observe that from the outside? > > > > I saw it from outside in below process listing. > > Okay. All UML childs do ptrace() as UML uses ptrace() for system call > emulation. > > >> > >>> The process tree looks also strange: > >>> > >>> 20330 pts/3 S+ 1:18 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20337 pts/3 S+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20338 pts/3 S+ 0:03 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20339 pts/3 S+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20347 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20405 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20469 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20615 pts/3 S+ 0:00 | \_ xterm -T Virtual Console #1 > >>> (fedora20) -e port-helper -uml-socket /tmp/xterm-pipeiW6d5k > >>> 20625 ? Ss 0:00 | | \_ port-helper -uml-socket > >>> /tmp/xterm-pipeiW6d5k > >>> 20626 ? Zs 0:00 | \_ [linux] <defunct> > >>> 20630 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20642 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20650 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20651 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20663 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20681 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20684 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20690 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20691 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20699 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20709 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20722 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20754 pts/3 S+ 0:00 | \_ xterm -T Virtual Console #2 > >>> (fedora20) -e port-helper -uml-socket /tmp/xterm-pipetxRIbS > >>> 20757 ? Ss 0:00 | | \_ port-helper -uml-socket > >>> /tmp/xterm-pipetxRIbS > >>> 20755 pts/3 S+ 0:00 | \_ xterm -T Virtual Console #6 > >>> (fedora20) -e port-helper -uml-socket /tmp/xterm-pipedhXmGp > >>> 20762 ? Ss 0:00 | | \_ port-helper -uml-socket > >>> /tmp/xterm-pipedhXmGp > >>> 20758 ? Zs 0:00 | \_ [linux] <defunct> > >>> 20760 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20763 ? Zs 0:00 | \_ [linux] <defunct> > >>> 20797 pts/3 S+ 0:00 | \_ xterm -T Virtual Console #3 > >>> (fedora20) -e port-helper -uml-socket /tmp/xterm-pipeULItXd > >>> 20812 ? Ss 0:00 | | \_ port-helper -uml-socket > >>> /tmp/xterm-pipeULItXd > >>> 20813 ? Zs 0:00 | \_ [linux] <defunct> > >>> 20815 pts/3 S+ 0:00 | \_ xterm -T Virtual Console #5 > >>> (fedora20) -e port-helper -uml-socket /tmp/xterm-pipeaKUbD3 > >>> 20876 ? Ss 0:00 | | \_ port-helper -uml-socket > >>> /tmp/xterm-pipeaKUbD3 > >>> 20877 ? Zs 0:00 | \_ [linux] <defunct> > >>> 20896 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 20909 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 21005 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 21007 pts/3 Z+ 0:00 | \_ [uml_net] <defunct> > >>> 21019 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 21112 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 21125 pts/3 Z+ 0:00 | \_ [uml_net] <defunct> > >>> 22164 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 22211 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 22224 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 22380 pts/3 t+ 0:51 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 21965 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 21968 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 21983 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 22053 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 22058 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >>> 22887 pts/3 t+ 0:00 | \_ ./linux ubd0=ext3fs.img mem=768M > >>> systemd.unit=multi-user.target umid=fedora20 > >> > >> Remain the tasks in state Z or are they fipping around? > > > > All process remain in thier state. nothing seems to happen any more. > > Ok. The it would be nice to find out what the UML main thread does. > gdb can tell you. > > >> Maybe the UML userspace creates many threads and on the host side UML > >> didn't call wait() jet... > > > > I don't think so. yum is probably not a big thread user, I guess. > > Isn't it a huge chunk of python? ;-) > > Thanks, > //richard ------------------------------------------------------------------------------ November Webinars for C, C++, Fortran Developers Accelerate application performance with scalable programming models. Explore techniques for threading, error checking, porting, and tuning. Get the most from the latest Intel processors and coprocessors. See abstracts and register http://pubads.g.doubleclick.net/gampad/clk?id=60136231&iu=/4140/ostg.clktrk _______________________________________________ User-mode-linux-user mailing list User-mode-linux-user@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/user-mode-linux-user