We have 3 test servers running SmartOS, host B and C have zpools zp01 and zp02 respectively, host A can see disks of zp01 and zp02 (disks of zp01 are connected to HBA ports of A and B; disks of zp02 are connected to HBA ports of A and C).
And I have a loop to test zpool import/export: for i in {1..100} do on B: zpool export -f zp01 on A: zpool import -o cachefile=none zp01 on A: zpool export -f zp01 on B: zpool import -o cachefile=none zp01 on C: zpool export -f zp02 on A: zpool import -o cachefile=none zp02 on A: zpool export -f zp02 on C: zpool import -o cachefile=none zp02 done After a few iterations, zpool import hangs. This can be reproduced consistently. Is this a known issue? Should I file a bug report? I have captured the stack of the hanging zpool import process. Thanks, Youzhong ---------------------------------------------------------------------------------------- # pstack 29207 29207: zpool import -o cachefile=none zp01 # mdb -ke '0t29207::pid2proc | ::walk thread | ::findstack -v' stack pointer for thread ffffff3289f39040: ffffff01ed3041e0 [ ffffff01ed3041e0 _resume_from_idle+0xf4() ] ffffff01ed304210 swtch+0x141() ffffff01ed304250 cv_wait+0x70(ffffff32330f901e, ffffff32330f9020) ffffff01ed304380 vmem_xalloc+0x630(ffffff32330f9000, 6000, 1000, 0, 0, 0, 0, ffffffff00000100) ffffff01ed3043f0 vmem_alloc+0x137(ffffff32330f9000, 6000, 100) ffffff01ed304510 segkp_get_internal+0x11b(fffffffffbc33760, 5000, e, ffffff01ed304528, 0) ffffff01ed304570 segkp_cache_get+0x103(1) ffffff01ed304610 thread_create+0x544(0, 0, fffffffffbaf35b0, ffffff32c9520b40 , 0, fffffffffbc30540, ffffff0100000002, ffffffff0000003c) ffffff01ed304660 taskq_thread_create+0x108(ffffff32c9520b40) ffffff01ed304710 taskq_create_common+0x1a7(fffffffff7e7e238, 0, 32, 3c, a, 7fffffff, fffffffffbc30540, ffffff3200000000, ffffff0100040008) ffffff01ed304770 taskq_create+0x50(fffffffff7e7e238, 32, 3c, a, 7fffffff, 8) ffffff01ed3047b0 metaslab_group_create+0x96(ffffff36fbca70a8, ffffff328784d000 ) ffffff01ed304860 vdev_alloc+0x54a(ffffff32a78fe000, ffffff01ed304928, ffffff3285910a80, ffffff3284e74540, a, 0) ffffff01ed304900 spa_config_parse+0x48(ffffff32a78fe000, ffffff01ed304928, ffffff3285910a80, ffffff3284e74540, a, 0) ffffff01ed3049a0 spa_config_parse+0xda(ffffff32a78fe000, ffffff01ed304a18, ffffff36fbca7f88, 0, 0, 0) ffffff01ed304a90 spa_load_impl+0xf4(ffffff32a78fe000, d5c8b305012c90c8, ffffff32d3417d30, 3, 0, 1, ffffff01ed304ad8) ffffff01ed304b30 spa_load+0x14e(ffffff32a78fe000, 3, 0, 1) ffffff01ed304b80 spa_tryimport+0xaa(ffffff3286740180) ffffff01ed304bd0 zfs_ioc_pool_tryimport+0x51(ffffff335c22a000) ffffff01ed304c80 zfsdev_ioctl+0x4a7(5a00000000, 5a06, 804258c, 100003, ffffff32578b3458, ffffff01ed304e68) ffffff01ed304cc0 cdev_ioctl+0x39(5a00000000, 5a06, 804258c, 100003, ffffff32578b3458, ffffff01ed304e68) ffffff01ed304d10 spec_ioctl+0x60(ffffff3284335d80, 5a06, 804258c, 100003, ffffff32578b3458, ffffff01ed304e68, 0) ffffff01ed304da0 fop_ioctl+0x55(ffffff3284335d80, 5a06, 804258c, 100003, ffffff32578b3458, ffffff01ed304e68, 0) ffffff01ed304ec0 ioctl+0x9b(3, 5a06, 804258c) ffffff01ed304f10 _sys_sysenter_post_swapgs+0x149() ------------------------------------------- smartos-discuss Archives: https://www.listbox.com/member/archive/184463/=now RSS Feed: https://www.listbox.com/member/archive/rss/184463/25769125-55cfbc00 Modify Your Subscription: https://www.listbox.com/member/?member_id=25769125&id_secret=25769125-7688e9fb Powered by Listbox: http://www.listbox.com