Hello,

I have a recent FreeBSD 11-STABLE which is mainly used as an iSCSI target. The system has 64G of RAM but is swapping intensively. Yup, about of half of the memory is used as ZFS ARC (isn't capped in loader.conf), and another half is eaten by the kernel, but it oly uses only about half of it (thus 25% of the total amount).

Could this be tweaked by some sysctl oids (I suppose not, but worth asking).

top, vmstat 1 snapshots and zfs-stats -a are listed below.


Thanks.


[root@san01:nginx/vhost.d]# vmstat 1
procs  memory       page                    disks faults         cpu
r b w  avm   fre   flt  re  pi  po    fr   sr da0 da1   in sy    cs us sy id
0 0 38 23G 609M 1544 68 118 64 895 839 0 0 3644 2678 649 0 13 87 0 0 53 23G 601M 1507 185 742 315 1780 33523 651 664 56438 785 476583 0 28 72 0 0 53 23G 548M 1727 330 809 380 2377 33256 758 763 55555 1273 468545 0 26 73 0 0 53 23G 528M 1702 239 660 305 1347 32335 611 631 59962 1025 490365 0 22 78 0 0 52 23G 854M 2409 309 693 203 97943 16944 525 515 64309 1570 540533 0 29 71 3 0 54 23G 1.1G 2756 639 641 149 124049 19531 542 538 64777 1576 553946 0 35 65 0 0 53 23G 982M 1694 236 680 282 2754 35602 597 603 66540 1385 583687 0 28 72 0 0 41 23G 867M 1882 223 767 307 1162 34936 682 638 67284 780 568818 0 33 67 0 0 39 23G 769M 1542 167 673 336 1187 35123 646 610 65925 1176 551623 0 23 77 2 0 41 23G 700M 3602 535 688 327 2192 37109 622 594 65862 4256 518934 0 33 67 0 0 54 23G 650M 2957 219 726 464 4838 36464 852 868 65384 4110 558132 1 37 62 0 0 54 23G 641M 1576 245 730 344 1139 33681 740 679 67216 970 560379 0 31 69


[root@san01:nginx/vhost.d]# top
last pid: 55190; load averages: 11.32, 12.15, 10.76 up 10+16:05:14 14:38:58
101 processes: 1 running, 100 sleeping
CPU:  0.2% user,  0.0% nice, 28.9% system,  1.6% interrupt, 69.3% idle
Mem: 85M Active, 1528K Inact, 12K Laundry, 62G Wired, 540M Free
ARC: 31G Total, 19G MFU, 6935M MRU, 2979M Anon, 556M Header, 1046M Other
     25G Compressed, 34G Uncompressed, 1.39:1 Ratio
Swap: 32G Total, 1186M Used, 31G Free, 3% Inuse, 7920K In, 3752K Out
PID USERNAME THR PRI NICE SIZE RES STATE C TIME WCPU COMMAND
40132 root         131  52    0  3152M 75876K uwait  14  36:59   6.10% java
55142 root           1  20    0  7904K  2728K CPU20  20   0:00   0.72% top
20026 root           1  20    0   106M  5676K nanslp 28   1:23   0.60% gstat
53642 root           1  20    0  7904K  2896K select 14   0:03   0.58% top
977 zfsreplica 1 20 0 30300K 3568K kqread 21 4:00 0.42% uwsgi 968 zfsreplica 1 20 0 30300K 2224K swread 11 2:03 0.21% uwsgi 973 zfsreplica 1 20 0 30300K 2264K swread 13 12:26 0.13% uwsgi
53000 www            1  20    0 23376K  1372K kqread 24   0:00   0.05% nginx
1292 root 1 20 0 6584K 2040K select 29 0:23 0.04% blacklistd 776 zabbix 1 20 0 12408K 4236K nanslp 26 4:42 0.03% zabbix_agentd 1289 root 1 20 0 67760K 5148K select 13 9:50 0.03% bsnmpd 777 zabbix 1 20 0 12408K 1408K select 25 5:06 0.03% zabbix_agentd 785 zfsreplica 1 20 0 27688K 3960K kqread 28 2:04 0.02% uwsgi 975 zfsreplica 1 20 0 30300K 464K kqread 18 2:33 0.02% uwsgi 974 zfsreplica 1 20 0 30300K 480K kqread 30 3:39 0.02% uwsgi 965 zfsreplica 1 20 0 30300K 464K kqread 4 3:23 0.02% uwsgi 976 zfsreplica 1 20 0 30300K 464K kqread 14 2:59 0.01% uwsgi 972 zfsreplica 1 20 0 30300K 464K kqread 10 2:57 0.01% uwsgi 963 zfsreplica 1 20 0 30300K 460K kqread 3 2:45 0.01% uwsgi 971 zfsreplica 1 20 0 30300K 464K kqread 13 3:16 0.01% uwsgi
69644 emz            1  20    0 13148K  4596K select 24   0:05   0.01% sshd
18203 vryabov        1  20    0 13148K  4624K select  9   0:02   0.01% sshd
636 root 1 20 0 6412K 1884K select 17 4:10 0.01% syslogd
51266 emz            1  20    0 13148K  4576K select  5   0:00   0.01% sshd
964 zfsreplica 1 20 0 30300K 460K kqread 18 11:02 0.01% uwsgi 962 zfsreplica 1 20 0 30300K 460K kqread 28 6:56 0.01% uwsgi 969 zfsreplica 1 20 0 30300K 464K kqread 12 2:07 0.01% uwsgi 967 zfsreplica 1 20 0 30300K 464K kqread 27 5:18 0.01% uwsgi 970 zfsreplica 1 20 0 30300K 464K kqread 0 4:25 0.01% uwsgi 966 zfsreplica 1 22 0 30300K 468K kqread 14 4:29 0.01% uwsgi
53001 www            1  20    0 23376K  1256K kqread 10   0:00   0.01% nginx
791 zfsreplica 1 20 0 27664K 4244K kqread 17 1:34 0.01% uwsgi
52431 root           1  20    0 17132K  4492K select 21   0:00   0.01% mc
70013 root           1  20    0 17132K  4492K select  4   0:03   0.01% mc
  870 root           1  20    0 12448K 12544K select 19   0:51   0.01% ntpd
[root@san01:nginx/vhost.d]# zfs-stats -a
------------------------------------------------------------------------
ZFS Subsystem Report                            Tue Nov 20 14:39:05 2018
------------------------------------------------------------------------
System Information:
        Kernel Version: 1102503 (osreldate)
        Hardware Platform:  amd64
        Processor Architecture: amd64
        ZFS Storage pool Version: 5000
        ZFS Filesystem Version:                 5
FreeBSD 11.2-STABLE #0 r340287M: Fri Nov 9 22:23:22 +05 2018 emz
14:39  up 10 days, 16:05, 5 users, load averages: 10,96 12,05 10,74
------------------------------------------------------------------------
System Memory:
        0.14%   90.20   MiB Active,     0.01% 8.62    MiB Inact
        98.97%  61.57   GiB Wired,      0.00%   0 Cache
        0.88%   560.02  MiB Free,       -0.00%  -184320 Bytes Gap
        Real Installed: 64.00   GiB
        Real Available:                 99.77%  63.85   GiB
        Real Managed:                   97.43%  62.21   GiB
        Logical Total:  64.00   GiB
        Logical Used:                   99.13%  63.44   GiB
        Logical Free:                   0.87% 568.64  MiB
Kernel Memory:  22.77   GiB
        Data:                           99.84%  22.73   GiB
        Text:                           0.16% 36.52   MiB
Kernel Memory Map:  62.21   GiB
        Size:                           46.85%  29.15   GiB
        Free:                           53.15%  33.06   GiB
------------------------------------------------------------------------
ARC Summary: (HEALTHY)
        Memory Throttle Count:                  0
ARC Misc:
        Deleted:  5.61b
        Recycle Misses:                         0
        Mutex Misses: 64.25m
        Evict Skips:  98.33m
ARC Size:                               50.02%  30.62   GiB
        Target Size: (Adaptive)         50.02%  30.62   GiB
        Min Size (Hard Limit):          12.50%  7.65    GiB
        Max Size (High Water):          8:1 61.21   GiB
ARC Size Breakdown:
        Recently Used Cache Size:       50.69%  15.52   GiB
        Frequently Used Cache Size:     49.31%  15.10   GiB
ARC Hash Breakdown:
        Elements Max: 8.35m
        Elements Current:               30.08%  2.51m
        Collisions: 2.18b
        Chain Max:                              10
        Chains: 308.03k
------------------------------------------------------------------------
ARC Efficiency: 48.52b
        Cache Hit Ratio:                84.49%  40.99b
        Cache Miss Ratio:               15.51%  7.53b
        Actual Hit Ratio:               84.19%  40.85b
        Data Demand Efficiency:         83.84%  13.06b
        Data Prefetch Efficiency:       40.66%  1.42b
        CACHE HITS BY CACHE LIST:
          Most Recently Used:           15.11%  6.19b
          Most Frequently Used:         84.54%  34.66b
          Most Recently Used Ghost:     0.97% 396.23m
          Most Frequently Used Ghost:   0.18% 75.49m
        CACHE HITS BY DATA TYPE:
          Demand Data:                  26.70%  10.95b
          Prefetch Data:                1.41% 576.89m
          Demand Metadata:              70.83%  29.04b
          Prefetch Metadata:            1.06% 434.19m
        CACHE MISSES BY DATA TYPE:
          Demand Data:                  28.04%  2.11b
          Prefetch Data:                11.18%  841.81m
          Demand Metadata:              60.37%  4.54b
          Prefetch Metadata:            0.40% 30.27m
------------------------------------------------------------------------
L2ARC is disabled
------------------------------------------------------------------------
File-Level Prefetch: (HEALTHY)
DMU Efficiency: 9.20b
        Hit Ratio:                      5.80% 533.41m
        Miss Ratio:                     94.20%  8.66b
        Colinear:                               0
          Hit Ratio:                    100.00% 0
          Miss Ratio:                   100.00% 0
        Stride:                                 0
          Hit Ratio:                    100.00% 0
          Miss Ratio:                   100.00% 0
DMU Misc:
        Reclaim:                                0
          Successes:                    100.00% 0
          Failures:                     100.00% 0
        Streams:                                0
          +Resets:                      100.00% 0
          -Resets:                      100.00% 0
          Bogus:                                0
------------------------------------------------------------------------
VDEV cache is disabled
------------------------------------------------------------------------
ZFS Tunables (sysctl):
        kern.maxusers 4422
        vm.kmem_size  66799345664
        vm.kmem_size_scale                      1
        vm.kmem_size_min                        0
        vm.kmem_size_max  1319413950874
        vfs.zfs.trim.max_interval               1
        vfs.zfs.trim.timeout                    30
        vfs.zfs.trim.txg_delay                  32
        vfs.zfs.trim.enabled                    0
        vfs.zfs.vol.immediate_write_sz  131072
        vfs.zfs.vol.unmap_sync_enabled          0
        vfs.zfs.vol.unmap_enabled               1
        vfs.zfs.vol.recursive                   0
        vfs.zfs.vol.mode                        1
        vfs.zfs.version.zpl                     5
        vfs.zfs.version.spa 5000
        vfs.zfs.version.acl                     1
        vfs.zfs.version.ioctl                   7
        vfs.zfs.debug                           0
        vfs.zfs.super_owner                     0
        vfs.zfs.immediate_write_sz  32768
        vfs.zfs.sync_pass_rewrite               2
        vfs.zfs.sync_pass_dont_compress         5
        vfs.zfs.sync_pass_deferred_free         2
        vfs.zfs.zio.dva_throttle_enabled        1
        vfs.zfs.zio.exclude_metadata            0
        vfs.zfs.zio.use_uma                     1
        vfs.zfs.zil_slog_bulk 786432
        vfs.zfs.cache_flush_disable             0
        vfs.zfs.zil_replay_disable              0
        vfs.zfs.standard_sm_blksz 131072
        vfs.zfs.dtl_sm_blksz  4096
        vfs.zfs.min_auto_ashift                 9
        vfs.zfs.max_auto_ashift                 13
        vfs.zfs.vdev.trim_max_pending 10000
        vfs.zfs.vdev.bio_delete_disable         0
        vfs.zfs.vdev.bio_flush_disable          0
        vfs.zfs.vdev.def_queue_depth            32
        vfs.zfs.vdev.queue_depth_pct  1000
        vfs.zfs.vdev.write_gap_limit  4096
        vfs.zfs.vdev.read_gap_limit 32768
        vfs.zfs.vdev.aggregation_limit  1048576
        vfs.zfs.vdev.trim_max_active            64
        vfs.zfs.vdev.trim_min_active            1
        vfs.zfs.vdev.scrub_max_active           2
        vfs.zfs.vdev.scrub_min_active           1
        vfs.zfs.vdev.async_write_max_active     10
        vfs.zfs.vdev.async_write_min_active     1
        vfs.zfs.vdev.async_read_max_active      3
        vfs.zfs.vdev.async_read_min_active      1
        vfs.zfs.vdev.sync_write_max_active      10
        vfs.zfs.vdev.sync_write_min_active      10
        vfs.zfs.vdev.sync_read_max_active       10
        vfs.zfs.vdev.sync_read_min_active       10
        vfs.zfs.vdev.max_active 1000
vfs.zfs.vdev.async_write_active_max_dirty_percent60
vfs.zfs.vdev.async_write_active_min_dirty_percent30
        vfs.zfs.vdev.mirror.non_rotating_seek_inc1
        vfs.zfs.vdev.mirror.non_rotating_inc    0
vfs.zfs.vdev.mirror.rotating_seek_offset1048576
        vfs.zfs.vdev.mirror.rotating_seek_inc   5
        vfs.zfs.vdev.mirror.rotating_inc        0
        vfs.zfs.vdev.trim_on_init               1
        vfs.zfs.vdev.cache.bshift               16
        vfs.zfs.vdev.cache.size                 0
        vfs.zfs.vdev.cache.max  16384
        vfs.zfs.vdev.default_ms_shift           29
        vfs.zfs.vdev.min_ms_count               16
        vfs.zfs.vdev.max_ms_count               200
        vfs.zfs.txg.timeout                     5
        vfs.zfs.space_map_ibs                   14
        vfs.zfs.spa_allocators                  4
        vfs.zfs.spa_min_slop  134217728
        vfs.zfs.spa_slop_shift                  5
        vfs.zfs.spa_asize_inflation             24
        vfs.zfs.deadman_enabled                 1
        vfs.zfs.deadman_checktime_ms  5000
        vfs.zfs.deadman_synctime_ms 1000000
        vfs.zfs.debug_flags                     0
        vfs.zfs.debugflags                      0
        vfs.zfs.recover                         0
        vfs.zfs.spa_load_verify_data            1
        vfs.zfs.spa_load_verify_metadata        1
        vfs.zfs.spa_load_verify_maxinflight 10000
        vfs.zfs.max_missing_tvds_scan           0
        vfs.zfs.max_missing_tvds_cachefile      2
        vfs.zfs.max_missing_tvds                0
        vfs.zfs.spa_load_print_vdev_tree        0
        vfs.zfs.ccw_retry_interval              300
        vfs.zfs.check_hostid                    1
        vfs.zfs.mg_fragmentation_threshold      85
        vfs.zfs.mg_noalloc_threshold            0
        vfs.zfs.condense_pct                    200
        vfs.zfs.metaslab_sm_blksz 4096
        vfs.zfs.metaslab.bias_enabled           1
        vfs.zfs.metaslab.lba_weighting_enabled  1
vfs.zfs.metaslab.fragmentation_factor_enabled1
        vfs.zfs.metaslab.preload_enabled        1
        vfs.zfs.metaslab.preload_limit          3
        vfs.zfs.metaslab.unload_delay           8
        vfs.zfs.metaslab.load_pct               50
        vfs.zfs.metaslab.min_alloc_size 33554432
        vfs.zfs.metaslab.df_free_pct            4
        vfs.zfs.metaslab.df_alloc_threshold 131072
        vfs.zfs.metaslab.debug_unload           0
        vfs.zfs.metaslab.debug_load             0
        vfs.zfs.metaslab.fragmentation_threshold70
        vfs.zfs.metaslab.force_ganging  16777217
        vfs.zfs.free_bpobj_enabled              1
        vfs.zfs.free_max_blocks                 -1
        vfs.zfs.zfs_scan_checkpoint_interval  7200
        vfs.zfs.zfs_scan_legacy                 0
        vfs.zfs.no_scrub_prefetch               0
        vfs.zfs.no_scrub_io                     0
        vfs.zfs.resilver_min_time_ms  3000
        vfs.zfs.free_min_time_ms  1000
        vfs.zfs.scan_min_time_ms  1000
        vfs.zfs.scan_idle                       50
        vfs.zfs.scrub_delay                     4
        vfs.zfs.resilver_delay                  2
        vfs.zfs.top_maxinflight                 32
        vfs.zfs.zfetch.array_rd_sz  1048576
        vfs.zfs.zfetch.max_idistance  67108864
        vfs.zfs.zfetch.max_distance 8388608
        vfs.zfs.zfetch.min_sec_reap             2
        vfs.zfs.zfetch.max_streams              8
        vfs.zfs.prefetch_disable                0
        vfs.zfs.delay_scale 500000
        vfs.zfs.delay_min_dirty_percent         60
        vfs.zfs.dirty_data_sync 67108864
        vfs.zfs.dirty_data_max_percent          10
        vfs.zfs.dirty_data_max_max  4294967296
        vfs.zfs.dirty_data_max  4294967296
        vfs.zfs.max_recordsize  1048576
        vfs.zfs.default_ibs                     17
        vfs.zfs.default_bs                      9
        vfs.zfs.send_holes_without_birth_time   1
        vfs.zfs.mdcomp_disable                  0
        vfs.zfs.per_txg_dirty_frees_percent     30
        vfs.zfs.nopwrite_enabled                1
        vfs.zfs.dedup.prefetch                  1
        vfs.zfs.dbuf_cache_lowater_pct          10
        vfs.zfs.dbuf_cache_hiwater_pct          10
        vfs.zfs.dbuf_metadata_cache_overflow    0
        vfs.zfs.dbuf_metadata_cache_shift       6
        vfs.zfs.dbuf_cache_shift                5
        vfs.zfs.dbuf_metadata_cache_max_bytes 1026962560
        vfs.zfs.dbuf_cache_max_bytes  2053925120
        vfs.zfs.arc_min_prescient_prefetch_ms   6
        vfs.zfs.arc_min_prefetch_ms             1
        vfs.zfs.l2c_only_size                   0
        vfs.zfs.mfu_ghost_data_esize  1910587392
        vfs.zfs.mfu_ghost_metadata_esize  5158840832
        vfs.zfs.mfu_ghost_size  7069428224
        vfs.zfs.mfu_data_esize  17620227072
        vfs.zfs.mfu_metadata_esize  950300160
        vfs.zfs.mfu_size  20773338624
        vfs.zfs.mru_ghost_data_esize  6989578240
        vfs.zfs.mru_ghost_metadata_esize  18479132160
        vfs.zfs.mru_ghost_size  25468710400
        vfs.zfs.mru_data_esize  4455460352
        vfs.zfs.mru_metadata_esize  70236672
        vfs.zfs.mru_size  7413314560
        vfs.zfs.anon_data_esize                 0
        vfs.zfs.anon_metadata_esize             0
        vfs.zfs.anon_size 3040037888
        vfs.zfs.l2arc_norw                      1
        vfs.zfs.l2arc_feed_again                1
        vfs.zfs.l2arc_noprefetch                1
        vfs.zfs.l2arc_feed_min_ms               200
        vfs.zfs.l2arc_feed_secs                 1
        vfs.zfs.l2arc_headroom                  2
        vfs.zfs.l2arc_write_boost 8388608
        vfs.zfs.l2arc_write_max 8388608
        vfs.zfs.arc_meta_limit  16431400960
        vfs.zfs.arc_free_target 113124
        vfs.zfs.arc_kmem_cache_reap_retry_ms    0
        vfs.zfs.compressed_arc_enabled          1
        vfs.zfs.arc_grow_retry                  60
        vfs.zfs.arc_shrink_shift                7
        vfs.zfs.arc_average_blocksize 8192
        vfs.zfs.arc_no_grow_shift               5
        vfs.zfs.arc_min 8215700480
        vfs.zfs.arc_max 65725603840
        vfs.zfs.abd_chunk_size  4096
        vfs.zfs.abd_scatter_enabled             1
------------------------------------------------------------------------
_______________________________________________
[email protected] mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "[email protected]"

Reply via email to