> A couple more questions here.
...
> What do you have zfs compresison set to? The gzip level is
> tunable, according to zfs set, anyway:
>
> PROPERTY EDIT INHERIT VALUES
> compression YES YES on | off | lzjb | gzip | gzip-[1-9]
I've used the "default" gzip compression level, that is I used
zfs set compression=gzip gzip_pool
> You still have idle time in this lockstat (and mpstat).
>
> What do you get for a lockstat -A -D 20 sleep 30?
# lockstat -A -D 20 /usr/tmp/fill /gzip_pool/junk
lockstat: warning: 723388 aggregation drops on CPU 0
lockstat: warning: 239335 aggregation drops on CPU 1
lockstat: warning: 62366 aggregation drops on CPU 0
lockstat: warning: 51856 aggregation drops on CPU 1
lockstat: warning: 45187 aggregation drops on CPU 0
lockstat: warning: 46536 aggregation drops on CPU 1
lockstat: warning: 687832 aggregation drops on CPU 0
lockstat: warning: 575675 aggregation drops on CPU 1
lockstat: warning: 46504 aggregation drops on CPU 0
lockstat: warning: 40874 aggregation drops on CPU 1
lockstat: warning: 45571 aggregation drops on CPU 0
lockstat: warning: 33422 aggregation drops on CPU 1
lockstat: warning: 501063 aggregation drops on CPU 0
lockstat: warning: 361041 aggregation drops on CPU 1
lockstat: warning: 651 aggregation drops on CPU 0
lockstat: warning: 7011 aggregation drops on CPU 1
lockstat: warning: 61600 aggregation drops on CPU 0
lockstat: warning: 19386 aggregation drops on CPU 1
lockstat: warning: 566156 aggregation drops on CPU 0
lockstat: warning: 105502 aggregation drops on CPU 1
lockstat: warning: 25362 aggregation drops on CPU 0
lockstat: warning: 8700 aggregation drops on CPU 1
lockstat: warning: 585002 aggregation drops on CPU 0
lockstat: warning: 645299 aggregation drops on CPU 1
lockstat: warning: 237841 aggregation drops on CPU 0
lockstat: warning: 20931 aggregation drops on CPU 1
lockstat: warning: 320102 aggregation drops on CPU 0
lockstat: warning: 435898 aggregation drops on CPU 1
lockstat: warning: 115 dynamic variable drops with non-empty dirty list
lockstat: warning: 385192 aggregation drops on CPU 0
lockstat: warning: 81833 aggregation drops on CPU 1
lockstat: warning: 259105 aggregation drops on CPU 0
lockstat: warning: 255812 aggregation drops on CPU 1
lockstat: warning: 486712 aggregation drops on CPU 0
lockstat: warning: 61607 aggregation drops on CPU 1
lockstat: warning: 1865 dynamic variable drops with non-empty dirty list
lockstat: warning: 250425 aggregation drops on CPU 0
lockstat: warning: 171415 aggregation drops on CPU 1
lockstat: warning: 166277 aggregation drops on CPU 0
lockstat: warning: 74819 aggregation drops on CPU 1
lockstat: warning: 39342 aggregation drops on CPU 0
lockstat: warning: 3556 aggregation drops on CPU 1
lockstat: warning: ran out of data records (use -n for more)
Adaptive mutex spin: 4701 events in 64.812 seconds (73 events/sec)
Count indv cuml rcnt spin Lock Caller
-------------------------------------------------------------------------------
1726 37% 37% 0.00 2 vph_mutex+0x17e8 pvn_write_done+0x10c
1518 32% 69% 0.00 1 vph_mutex+0x17e8 hat_page_setattr+0x70
264 6% 75% 0.00 2 vph_mutex+0x2000 page_hashin+0xad
194 4% 79% 0.00 4 0xfffffffed2ee0a88 cv_wait+0x69
106 2% 81% 0.00 2 vph_mutex+0x2000 page_hashout+0xdd
91 2% 83% 0.00 4 0xfffffffed2ee0a88 taskq_dispatch+0x2c9
83 2% 85% 0.00 4 0xfffffffed2ee0a88 taskq_thread+0x1cb
83 2% 86% 0.00 1 0xfffffffec17a56b0 ufs_iodone+0x3d
47 1% 87% 0.00 4 0xfffffffec1e4ce98 vdev_queue_io+0x85
43 1% 88% 0.00 6 0xfffffffec139a2c0 trap+0xf66
38 1% 89% 0.00 6 0xfffffffecb5f8cd0 cv_wait+0x69
37 1% 90% 0.00 4 0xfffffffec143ee90 dmult_deque+0x36
26 1% 91% 0.00 2 htable_mutex+0x108 htable_release+0x79
26 1% 91% 0.00 1 0xfffffffec17a56b0 ufs_putpage+0xa4
18 0% 91% 0.00 4 0xfffffffec00dca48 ghd_intr+0xa8
17 0% 92% 0.00 2 0xfffffffec00dca48 ghd_waitq_delete+0x35
12 0% 92% 0.00 2 htable_mutex+0x248 htable_release+0x79
11 0% 92% 0.00 8 0xfffffffec1e4ce98 vdev_queue_io_done+0x3b
10 0% 93% 0.00 3 0xfffffffec00dca48 ghd_transport+0x71
10 0% 93% 0.00 2 0xffffff00077dc138
page_get_mnode_freelist+0xdb
-------------------------------------------------------------------------------
Adaptive mutex block: 167 events in 64.812 seconds (3 events/sec)
Count indv cuml rcnt nsec Lock Caller
-------------------------------------------------------------------------------
78 47% 47% 0.00 31623 vph_mutex+0x17e8 pvn_write_done+0x10c
25 15% 62% 0.00 97897 0xfffffffed2ee0a88 cv_wait+0x69
13 8% 69% 0.00 70426 0xfffffffed2ee0a88 taskq_dispatch+0x2c9
10 6% 75% 0.00 30559 0xfffffffec17a56b0 ufs_iodone+0x3d
9 5% 81% 0.00 34569 0xfffffffec143ee90 dmult_deque+0x36
4 2% 83% 0.00 30579 0xfffffffec00dca48 ghd_waitq_delete+0x35
3 2% 85% 0.00 117726 0xfffffffec14365f0
ohci_hcdi_pipe_bulk_xfer+0x50
2 1% 86% 0.00 32324 pse_mutex+0x1a80 page_unlock+0x3b
1 1% 87% 0.00 381871279 ph_mutex+0xac0 page_create_va+0x334
1 1% 87% 0.00 34696 pse_mutex+0x1cc0 page_unlock+0x3b
1 1% 88% 0.00 29577 pse_mutex+0x1e80 page_unlock+0x3b
1 1% 89% 0.00 29324 pse_mutex+0x1580 page_unlock+0x3b
1 1% 89% 0.00 29029 pse_mutex+0x1400 page_unlock+0x3b
1 1% 90% 0.00 29806 pse_mutex+0x1700 page_unlock+0x3b
1 1% 90% 0.00 30873 pse_mutex+0x8c0 page_unlock+0x3b
1 1% 91% 0.00 34591 pse_mutex+0x700 page_unlock+0x3b
1 1% 92% 0.00 32196 pse_mutex+0x140 page_unlock+0x3b
1 1% 92% 0.00 31986 pse_mutex+0x440 page_unlock+0x3b
1 1% 93% 0.00 27500 pse_mutex+0xd00 page_unlock+0x3b
1 1% 93% 0.00 36400 pse_mutex+0xb00 page_unlock+0x3b
-------------------------------------------------------------------------------
Spin lock spin: 252 events in 64.812 seconds (4 events/sec)
Count indv cuml rcnt spin Lock Caller
-------------------------------------------------------------------------------
97 38% 38% 0.00 4 cpu0_disp disp_lock_enter+0x31
44 17% 56% 0.00 3 cpu0_disp disp_lock_enter_high+0x11
44 17% 73% 0.00 3 0xfffffffec1a82708 disp_lock_enter+0x31
36 14% 88% 0.00 3 xc_mbox_lock+0x11 mutex_vector_enter+0x4ee
16 6% 94% 0.00 4 0xfffffffec1a82708 disp_lock_enter_high+0x11
14 6% 100% 0.00 5 turnstile_table+0xd38 disp_lock_enter+0x31
1 0% 100% 0.00 4 turnstile_table+0xad8 disp_lock_enter+0x31
-------------------------------------------------------------------------------
R/W writer blocked by writer: 8281 events in 64.812 seconds (128 events/sec)
Count indv cuml rcnt nsec Lock Caller
-------------------------------------------------------------------------------
8281 100% 100% 0.00 20609451 0xfffffffec17a56a0 ufs_rwlock+0xfd
-------------------------------------------------------------------------------
Adaptive mutex hold: 14476537 events in 64.812 seconds (223362 events/sec)
Count indv cuml rcnt nsec Lock Caller
-------------------------------------------------------------------------------
289666 2% 2% 0.00 1242 vph_mutex+0x2000 page_hashout+0xed
274484 2% 4% 0.00 1295 vph_mutex+0x2000 page_hashin+0xc6
256452 2% 6% 0.00 1359 vph_mutex+0x17e8 hat_page_setattr+0xbf
255344 2% 7% 0.00 1245 vph_mutex+0x17e8 pvn_write_done+0x14b
213470 1% 9% 0.00 1181 pcf+0x40 page_free+0x1ba
146344 1% 10% 0.00 4043 0xfffffffec14742a0 getblk_common+0x1e1
146344 1% 11% 0.00 1238 0xfffffffec14742a0 brelse+0x15d
131168 1% 12% 0.00 1274 0xfffffffec142fdf0 deltamap_add+0x1f5
106877 1% 13% 0.00 1277 0xfffffffec17a56b0 ufs_putpage+0xf8
64537 0% 13% 0.00 1283 0xfffffffec0208cc8 free_vpmap+0x85
64508 0% 13% 0.00 1293 0xfffffffec0208d60 free_vpmap+0x85
61387 0% 14% 0.00 4756 0xfffffffec0208d08 get_free_vpmap+0x357
61261 0% 14% 0.00 4640 0xfffffffec0208da0 get_free_vpmap+0x357
59765 0% 15% 0.00 1294 0xfffffffec0208da0 free_vpmap+0x85
59731 0% 15% 0.00 1251 0xfffffffec0208d08 free_vpmap+0x85
57656 0% 16% 0.00 4686 0xfffffffec0208cc8 get_free_vpmap+0x357
57655 0% 16% 0.00 4555 0xfffffffec0208d60 get_free_vpmap+0x357
49703 0% 16% 0.00 1270 0xffffff00077dd718 page_ctr_sub+0x80
48208 0% 17% 0.00 1273 0xffffff00077dd720 page_ctr_sub+0x80
47333 0% 17% 0.00 1276 0xffffff00077dd710 page_ctr_sub+0x80
-------------------------------------------------------------------------------
Spin lock hold: 1092671 events in 64.812 seconds (16859 events/sec)
Count indv cuml rcnt nsec Lock Caller
-------------------------------------------------------------------------------
508900 47% 47% 0.00 2965 xc_mbox_lock+0x11 mutex_vector_exit+0xf1
292686 27% 73% 0.00 1353 sleepq_head+0x12f8 disp_lock_exit+0x56
32656 3% 76% 0.00 1537 pcicfg_mutex+0x1 mutex_vector_exit+0xf1
24638 2% 79% 0.00 1555 cpu0_disp disp_lock_exit_high+0x34
21397 2% 81% 0.00 1580 0xfffffffec1a82708 disp_lock_exit_high+0x34
13948 1% 82% 0.00 1787 cpu[0]+0xe8 disp_lock_exit_high+0x34
12960 1% 83% 0.00 1742 hres_lock dtrace_hres_tick+0x55
12763 1% 84% 0.00 1943 cpu[1]+0xe8 disp_lock_exit_high+0x34
8280 1% 85% 0.00 6327 turnstile_table+0xd38 disp_lock_exit+0x56
8268 1% 86% 0.00 4912 turnstile_table+0xd38 disp_lock_exit_high+0x34
5670 1% 86% 0.00 5425 sleepq_head+0x1ab8 disp_lock_exit+0x56
5237 0% 87% 0.00 2626 cpu0_disp
disp_lock_exit_nopreempt+0x44
4854 0% 87% 0.00 1204 sleepq_head+0x1eb8 disp_lock_exit+0x56
4843 0% 88% 0.00 1223 sleepq_head+0x1a08 disp_lock_exit+0x56
4671 0% 88% 0.00 1382 sleepq_head+0xf58 disp_lock_exit+0x56
4599 0% 88% 0.00 1272 sleepq_head+0x1f98 disp_lock_exit+0x56
4308 0% 89% 0.00 3844 sleepq_head+0x1ab8
disp_lock_exit_nopreempt+0x44
4138 0% 89% 0.00 1405 cpu0_disp disp_lock_exit+0x56
4130 0% 90% 0.00 1353 sleepq_head+0x17c8 disp_lock_exit+0x56
4111 0% 90% 0.00 1526 0xfffffffec1a82708 disp_lock_exit+0x56
-------------------------------------------------------------------------------
R/W writer hold: 164764 events in 64.812 seconds (2542 events/sec)
Count indv cuml rcnt nsec Lock Caller
-------------------------------------------------------------------------------
125018 76% 76% 0.00 24819 0xfffffffec17a56a8 wrip+0x489
12533 8% 83% 0.00 7098 0xfffffffec17a56a8 ufs_write+0x54b
12532 8% 91% 0.00 1414263 0xfffffffec17a56a0 ufs_rwunlock+0x24
4729 3% 94% 0.00 1444 0xffffff01efb81d08 dnode_new_blkid+0x19d
3630 2% 96% 0.00 2009 0xfffffffec17a56a8 wrip+0x3cd
1648 1% 97% 0.00 31342 0xfffffffed09c3490 as_map_locked+0x305
1647 1% 98% 0.00 2829 0xfffffffec2b2bc88 segvn_extend_prev+0x149
438 0% 98% 0.00 40571 0xfffffffed09c33b0 as_unmap+0x25c
196 0% 99% 0.00 1468 0xfffffffed1ebc7d8 dnode_new_blkid+0x19d
89 0% 99% 0.00 1551 0xfffffffecba72700 spa_sync+0x4cc
89 0% 99% 0.00 1449 0xfffffffecb5f8cd8 txg_sync_thread+0x1c6
86 0% 99% 0.00 8430 0xfffffffecb5f8cd8 txg_sync_thread+0x186
61 0% 99% 0.00 910403 0xfffffffec142fef8 ldl_waito+0x6d
61 0% 99% 0.00 4446 0xfffffffec142fef8 get_write_bp+0xac
61 0% 99% 0.00 7958 0xfffffffec142fef8 inval_range+0xbd
53 0% 99% 0.00 1454 0xfffffffed37a2560 dnode_new_blkid+0x19d
40 0% 99% 0.00 1469 0xfffffffed3e58ad8 dnode_new_blkid+0x19d
34 0% 99% 0.00 1549 0xfffffffed379ea78 dnode_new_blkid+0x19d
32 0% 99% 0.00 16405 0xfffffffec7bec560 zfs_getpage+0x2ff
24 0% 99% 0.00 1497 0xfffffffed37a32d0 dnode_new_blkid+0x19d
-------------------------------------------------------------------------------
R/W reader hold: 290702 events in 64.812 seconds (4485 events/sec)
Count indv cuml rcnt nsec Lock Caller
-------------------------------------------------------------------------------
125018 43% 43% 0.00 28827 0xfffffffec1435648 wrip+0x49e
23000 8% 51% 0.00 188307 0xfffffffec17a56a8 ufs_putpages+0x340
22368 8% 59% 0.00 8206227 0xfffffffed2ee0a90 taskq_thread+0x1c3
14403 5% 64% 0.00 1508 0xfffffffed37a3518 dmu_zfetch_find+0x3c1
14403 5% 69% 0.00 6107 0xfffffffed37a32d0 dnode_hold_impl+0xbc
14403 5% 73% 0.00 6606 0xfffffffed37a32d0 dbuf_read+0x250
12533 4% 78% 0.00 10450 0xfffffffec1435648 ufs_write+0x554
6722 2% 80% 0.00 75280 0xfffffffed09c3490 as_fault+0x658
6704 2% 82% 0.00 70784 0xfffffffed1e95958 segvn_fault+0xbfc
6704 2% 85% 0.00 66973 0xfffffffec2b2bc88 segvn_fault+0xbf4
4766 2% 86% 0.00 6106 0xffffff01efb81d08 dbuf_dirty+0x39c
4734 2% 88% 0.00 3757 0xffffff01efb81d08 dbuf_read+0x250
4729 2% 90% 0.00 180997 0xffffffff76afd928 zfs_write+0x3f5
4728 2% 91% 0.00 21145 0xffffff01efb81d08
dmu_buf_hold_array_by_dnode+0x1eb
3630 1% 92% 0.00 5639 0xfffffffec1435648 wrip+0x3e2
3576 1% 94% 0.00 25200843 0xfffffffed2ee0b68 taskq_thread+0x1c3
2764 1% 95% 0.00 5248 0xfffffffed37a3a68 dbuf_read+0x250
2519 1% 96% 0.00 4934 0xfffffffec17a56a8 ufs_getpage+0x8d5
1379 0% 96% 0.00 68439 0xfffffffed09c33b0 as_fault+0x658
1229 0% 96% 0.00 56389 0xffffffff76b28d88 segvn_fault+0xbf4
-------------------------------------------------------------------------------
> Do you see anyone with long lock hold times, long
> sleeps, or excessive spinning?
Not sure if the above lockstat shows anything excessive.
> The largest numbers from mpstat are for interrupts
> and cross calls.
> What does intrstat(1M) show?
The first two are for the idle machine; starting with
#3 filling of the gzip compressed zpool starts:
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 0 0.0 0 0.0
ata#1 | 0 0.0 0 0.0
audiohd#0 | 1 0.0 0 0.0
ehci#0 | 1 0.0 0 0.0
nge#0 | 0 0.0 0 0.0
nvidia#0 | 60 0.1 0 0.0
ohci#0 | 0 0.0 0 0.0
pci-ide#1 | 0 0.0 0 0.0
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 0 0.0 0 0.0
ata#1 | 0 0.0 0 0.0
audiohd#0 | 2 0.0 0 0.0
ehci#0 | 2 0.0 0 0.0
nge#0 | 0 0.0 158 0.0
nvidia#0 | 61 0.2 0 0.0
ohci#0 | 0 0.0 0 0.0
pci-ide#1 | 0 0.0 158 0.1
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 0 0.0 0 0.0
ata#1 | 2 0.0 0 0.0
audiohd#0 | 4 0.0 0 0.0
ehci#0 | 4 0.0 0 0.0
nge#0 | 0 0.0 3547 0.3
nvidia#0 | 60 0.2 0 0.0
ohci#0 | 0 0.0 35 0.1
pci-ide#1 | 0 0.0 3547 2.8
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 0 0.0 0 0.0
ata#1 | 0 0.0 0 0.0
audiohd#0 | 1 0.0 0 0.0
ehci#0 | 1 0.0 0 0.0
nge#0 | 0 0.0 2861 0.2
nvidia#0 | 60 0.1 0 0.0
ohci#0 | 0 0.0 0 0.0
pci-ide#1 | 0 0.0 2861 2.2
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 0 0.0 0 0.0
ata#1 | 1 0.0 0 0.0
audiohd#0 | 2 0.0 0 0.0
ehci#0 | 2 0.0 0 0.0
nge#0 | 0 0.0 380 0.0
nvidia#0 | 61 0.3 0 0.0
ohci#0 | 0 0.0 0 0.0
pci-ide#1 | 0 0.0 380 0.3
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 0 0.0 0 0.0
ata#1 | 0 0.0 0 0.0
audiohd#0 | 0 0.0 0 0.0
ehci#0 | 0 0.0 0 0.0
nge#0 | 0 0.0 3302 0.2
nvidia#0 | 60 0.2 0 0.0
ohci#0 | 0 0.0 30 0.1
pci-ide#1 | 0 0.0 3302 2.6
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 0 0.0 0 0.0
ata#1 | 0 0.0 0 0.0
audiohd#0 | 2 0.0 0 0.0
ehci#0 | 2 0.0 0 0.0
nge#0 | 0 0.0 3165 0.3
nvidia#0 | 61 0.2 0 0.0
ohci#0 | 0 0.0 0 0.0
pci-ide#1 | 0 0.0 3165 2.5
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 0 0.0 0 0.0
ata#1 | 1 0.0 0 0.0
audiohd#0 | 3 0.0 0 0.0
ehci#0 | 3 0.0 0 0.0
nge#0 | 0 0.0 304 0.0
nvidia#0 | 61 0.3 0 0.0
ohci#0 | 0 0.0 0 0.0
pci-ide#1 | 0 0.0 304 0.2
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 0 0.0 0 0.0
ata#1 | 2 0.0 0 0.0
audiohd#0 | 0 0.0 0 0.0
ehci#0 | 0 0.0 0 0.0
nge#0 | 0 0.0 3878 0.3
nvidia#0 | 60 0.2 0 0.0
ohci#0 | 0 0.0 49 0.1
pci-ide#1 | 0 0.0 3878 3.0
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 0 0.0 0 0.0
ata#1 | 0 0.0 0 0.0
audiohd#0 | 2 0.0 0 0.0
ehci#0 | 2 0.0 0 0.0
nge#0 | 0 0.0 2679 0.2
nvidia#0 | 61 0.1 0 0.0
ohci#0 | 0 0.0 0 0.0
pci-ide#1 | 0 0.0 2679 2.1
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 1 0.0 0 0.0
ata#1 | 1 0.0 0 0.0
audiohd#0 | 3 0.0 0 0.0
ehci#0 | 3 0.0 0 0.0
nge#0 | 0 0.0 312 0.0
nvidia#0 | 60 0.2 0 0.0
ohci#0 | 0 0.0 0 0.0
pci-ide#1 | 0 0.0 312 0.2
device | cpu0 %tim cpu1 %tim
-------------+------------------------------
ata#0 | 0 0.0 0 0.0
ata#1 | 0 0.0 0 0.0
audiohd#0 | 0 0.0 0 0.0
ehci#0 | 0 0.0 0 0.0
nge#0 | 0 0.0 6403 0.5
nvidia#0 | 103 0.5 0 0.0
ohci#0 | 0 0.0 413 1.2
pci-ide#1 | 0 0.0 6403 5.1
> Have you run dtrace to determine the most frequent
> cross-callers?
>
> #!/usr/sbin/dtrace -s
>
> sysinfo:::xcalls
> {
> @a[stack(30)] = count();
>
> trunc(@a, 30);
> }
>
> is an easy way to do this.
Seems to taskq_threads, running zfs' gzip compression:
dtrace: script '/home/jk/src/dtrace/xcalls.d' matched 3 probes
^C
CPU ID FUNCTION:NAME
1 2 :END
unix`xc_do_call+0xfd
unix`xc_wait_sync+0x2b
unix`x86pte_inval+0x139
unix`hat_pte_unmap+0xfc
unix`hat_unload_callback+0x1df
unix`hat_unload+0x41
genunix`segkp_release_internal+0xb2
genunix`segkp_release+0xb6
genunix`schedctl_freepage+0x30
genunix`schedctl_proc_cleanup+0x5c
genunix`exec_args+0x1c2
elfexec`elf32exec+0x410
genunix`gexec+0x36d
genunix`exec_common+0x417
genunix`exece+0x1b
unix`sys_syscall32+0x101
1
unix`xc_do_call+0xfd
unix`xc_sync+0x2b
unix`dtrace_xcall+0x6f
dtrace`dtrace_state_go+0x3ec
dtrace`dtrace_ioctl+0x79c
genunix`cdev_ioctl+0x48
specfs`spec_ioctl+0x86
genunix`fop_ioctl+0x37
genunix`ioctl+0x16b
unix`sys_syscall+0x17b
1
unix`xc_do_call+0xfd
unix`xc_wait_sync+0x2b
unix`x86pte_inval+0x139
unix`hat_pte_unmap+0xfc
unix`hat_unload_callback+0x1df
unix`hat_unload+0x41
swrand`physmem_ent_gen+0x190
swrand`rnd_handler+0x29
genunix`callout_execute+0xb1
genunix`taskq_thread+0x1a7
unix`thread_start+0x8
24
unix`xc_do_call+0xfd
unix`xc_wait_sync+0x2b
unix`x86pte_inval+0x139
unix`hat_pte_unmap+0xfc
unix`hat_unload_callback+0x148
unix`hat_unload+0x41
genunix`segkp_release_internal+0xb2
genunix`segkp_release+0xb6
genunix`thread_free+0x238
genunix`thread_reap_list+0x1f
genunix`thread_reaper+0x130
unix`thread_start+0x8
45
unix`xc_do_call+0xfd
unix`xc_wait_sync+0x2b
unix`x86pte_inval+0x139
unix`hat_pte_unmap+0xfc
unix`hat_unload_callback+0x148
unix`hat_unload+0x41
unix`segkmem_free_vn+0x6a
unix`segkmem_free+0x20
genunix`vmem_xfree+0x10c
genunix`vmem_free+0x25
genunix`kmem_free+0x47
kstat`read_kstat_data+0x484
kstat`kstat_ioctl+0x4a
genunix`cdev_ioctl+0x48
specfs`spec_ioctl+0x86
genunix`fop_ioctl+0x37
genunix`ioctl+0x16b
unix`sys_syscall32+0x101
76
unix`xc_do_call+0xfd
unix`xc_wait_sync+0x2b
unix`x86pte_inval+0x139
unix`hat_pte_unmap+0xfc
unix`hat_unload_callback+0x148
unix`hat_unload+0x41
unix`segkmem_free_vn+0x6a
unix`segkmem_free+0x20
genunix`vmem_xfree+0x10c
genunix`vmem_free+0x25
unix`kfreea+0x5a
unix`i_ddi_mem_free+0x5d
genunix`ddi_dma_mem_free+0x26
ehci`ehci_free_tw+0x45
ehci`ehci_deallocate_tw+0xb6
ehci`ehci_traverse_active_qtd_list+0xdf
ehci`ehci_intr+0x162
unix`av_dispatch_autovect+0x78
unix`dispatch_hardint+0x2f
unix`switch_sp_and_call+0x13
120
unix`xc_do_call+0xfd
unix`xc_sync+0x2b
unix`dtrace_xcall+0x6f
unix`dtrace_sync+0x18
dtrace`dtrace_state_deadman+0x18
genunix`cyclic_softint+0xc9
unix`cbe_low_level+0x17
unix`av_dispatch_softvect+0x5f
unix`dispatch_softint+0x34
unix`switch_sp_and_call+0x13
149
unix`xc_do_call+0xfd
unix`xc_wait_sync+0x2b
unix`x86pte_inval+0x139
unix`hat_pte_unmap+0xfc
unix`hat_unload_callback+0x148
unix`hat_unload+0x41
unix`segkmem_free_vn+0x6a
unix`segkmem_free+0x20
genunix`vmem_xfree+0x10c
genunix`vmem_free+0x25
unix`kfreea+0x5a
unix`i_ddi_mem_free+0x5d
rootnex`rootnex_teardown_copybuf+0x20
rootnex`rootnex_dma_unbindhdl+0xbe
genunix`ddi_dma_unbind_handle+0x22
ata`ghd_dmafree_attr+0x29
ata`ata_disk_memfree+0x20
gda`gda_free+0x39
dadk`dadk_iodone+0xbf
dadk`dadk_pktcb+0xa5
ata`ata_disk_complete+0x115
ata`ata_hba_complete+0x34
ata`ghd_doneq_process+0x85
unix`av_dispatch_softvect+0x5f
unix`dispatch_softint+0x34
unix`switch_sp_and_call+0x13
290
unix`xc_do_call+0xfd
unix`xc_sync+0x2b
unix`dtrace_xcall+0x6f
dtrace`dtrace_ioctl+0x47f
genunix`cdev_ioctl+0x48
specfs`spec_ioctl+0x86
genunix`fop_ioctl+0x37
genunix`ioctl+0x16b
unix`sys_syscall+0x17b
300
unix`xc_do_call+0xfd
unix`xc_wait_sync+0x2b
unix`x86pte_inval+0x139
unix`hat_pte_unmap+0xfc
unix`hat_unload_callback+0x148
unix`hat_unload+0x41
unix`segkmem_free_vn+0x6a
unix`segkmem_free+0x20
genunix`vmem_xfree+0x10c
genunix`vmem_free+0x25
unix`kfreea+0x5a
unix`i_ddi_mem_free+0x5d
genunix`ddi_dma_mem_free+0x26
ohci`ohci_free_tw+0x59
ohci`ohci_deallocate_tw_resources+0xa4
ohci`ohci_traverse_done_list+0xfc
ohci`ohci_intr+0x1d0
unix`av_dispatch_autovect+0x78
unix`dispatch_hardint+0x2f
unix`switch_sp_and_call+0x13
1020
unix`xc_do_call+0xfd
unix`xc_wait_sync+0x2b
unix`x86pte_inval+0x139
unix`hat_pte_unmap+0xfc
unix`hat_unload_callback+0x148
unix`hat_unload+0x41
unix`segkmem_free_vn+0x6a
unix`segkmem_free+0x20
genunix`vmem_xfree+0x10c
genunix`vmem_free+0x25
genunix`kmem_free+0x47
unix`kobj_free+0x23
unix`zcfree+0x36
unix`z_deflateEnd+0x74
unix`z_compress_level+0x9d
zfs`gzip_compress+0x4b
zfs`zio_compress_data+0xbc
zfs`zio_write_compress+0x8d
genunix`taskq_thread+0x1a7
unix`thread_start+0x8
133773
unix`xc_do_call+0xfd
unix`xc_wait_sync+0x2b
unix`x86pte_inval+0x139
unix`hat_pte_unmap+0xfc
unix`hat_unload_callback+0x148
unix`hat_unload+0x41
unix`segkmem_free_vn+0x6a
unix`segkmem_free+0x20
genunix`vmem_xfree+0x10c
genunix`vmem_free+0x25
genunix`kmem_free+0x47
unix`kobj_free+0x23
unix`zcfree+0x36
unix`z_deflateEnd+0x91
unix`z_compress_level+0x9d
zfs`gzip_compress+0x4b
zfs`zio_compress_data+0xbc
zfs`zio_write_compress+0x8d
genunix`taskq_thread+0x1a7
unix`thread_start+0x8
133773
unix`xc_do_call+0xfd
unix`xc_wait_sync+0x2b
unix`x86pte_inval+0x139
unix`hat_pte_unmap+0xfc
unix`hat_unload_callback+0x148
unix`hat_unload+0x41
unix`segkmem_free_vn+0x6a
unix`segkmem_free+0x20
genunix`vmem_xfree+0x10c
genunix`vmem_free+0x25
genunix`kmem_free+0x47
unix`kobj_free+0x23
unix`zcfree+0x36
unix`z_deflateEnd+0xae
unix`z_compress_level+0x9d
zfs`gzip_compress+0x4b
zfs`zio_compress_data+0xbc
zfs`zio_write_compress+0x8d
genunix`taskq_thread+0x1a7
unix`thread_start+0x8
133773
unix`xc_do_call+0xfd
unix`xc_wait_sync+0x2b
unix`x86pte_inval+0x139
unix`hat_pte_unmap+0xfc
unix`hat_unload_callback+0x148
unix`hat_unload+0x41
unix`segkmem_free_vn+0x6a
unix`segkmem_free+0x20
genunix`vmem_xfree+0x10c
genunix`vmem_free+0x25
genunix`kmem_free+0x47
unix`kobj_free+0x23
unix`zcfree+0x36
unix`z_deflateEnd+0xcb
unix`z_compress_level+0x9d
zfs`gzip_compress+0x4b
zfs`zio_compress_data+0xbc
zfs`zio_write_compress+0x8d
genunix`taskq_thread+0x1a7
unix`thread_start+0x8
133773
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
[email protected]
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss