This is what we call an "/etc/system virus"
Unless you understand each of the tunables and what they impact, do yourself a 
favor
and avoid putting them on your system. In some cases, they are documented in the
Solaris kernel tunables guide, in other cases they are documented only in 
source code.
 — richard


> On Feb 23, 2017, at 10:47 AM, Humberto Ramirez <[email protected]> wrote:
> 
> Did you re-run the dd tests after tweaking those parameters? 
> 
> 
> 
> On Feb 22, 2017 11:52 AM, "Will Beazley" <[email protected] 
> <mailto:[email protected]>> wrote:
> Mille Grazie!
> 
> 
> On 2/21/17 23:12, Artem Penner wrote:
>> Read about this kernel parameters
>> zfs:zfs_dirty_data_max
>> zfs:zfs_txg_timeout
>> zfs:zfs_dirty_data_sync
>> 
>> It's limut your i/o.
>> 
>> Example of /etc/config
>> set ibft_noprobe=1
>> set noexec_user_stack=1
>> set noexec_user_stack_log=1
>> set idle_cpu_no_deep_c=1
>> set idle_cpu_prefer_mwait = 0
>> set hires_tick=1
>> set ip:ip_squeue_fanout=1
>> set pcplusmp:apic_panic_on_nmi=1
>> set apix:apic_panic_on_nmi=1
>> set dump_plat_mincpu=0
>> set dump_bzip2_level=1
>> set dump_metrics_on=1
>> set sata:sata_auto_online=1
>> set sd:sd_max_throttle = 128
>> set sd:sd_io_time=10
>> set rlim_fd_max = 131072
>> set rlim_fd_cur = 65536
>> set ndd:tcp_wscale_always = 1
>> set ndd:tstamp_if_wscale = 1
>> set ndd:tcp_max_buf = 166777216
>> set nfs:nfs_allow_preepoch_time = 1
>> set nfs:nfs3_max_threads = 256
>> set nfs:nfs4_max_threads = 256
>> set nfs:nfs3_nra = 32
>> set nfs:nfs4_nra = 32
>> set nfs:nfs3_bsize = 1048576
>> set nfs:nfs4_bsize = 1048576
>> set nfs3:max_transfer_size = 1048576
>> set nfs4:max_transfer_size = 1048576
>> set nfs:nfs4_async_clusters = 16
>> set rpcmod:svc_default_stksize=0x6000
>> set rpcmod:cotsmaxdupreqs = 4096
>> set rpcmod:maxdupreqs = 4096
>> set rpcmod:clnt_max_conns = 8
>> set maxphys=1048576
>> set zfs:zfs_dirty_data_max = 0x600000000
>> set zfs:zfs_txg_timeout = 0xc
>> set zfs:zfs_dirty_data_sync = 0x400000000
>> set zfs:zfs_arc_max = 0x6400000000
>> set zfs:zfs_arc_shrink_shift=12
>> set zfs:l2arc_write_max = 0x6400000
>> set zfs:l2arc_write_boost = 0xC800000
>> set zfs:l2arc_headroom = 12
>> set zfs:l2arc_norw=0
>> 
>> 22 февр. 2017 г. 7:28 пользователь "Will Beazley" 
>> <[email protected] <mailto:[email protected]>> написал:
>> Christopher, et al.,
>> 
>> I am trying to get my head around why the performance of ramdisk is so much 
>> poorer than that of HDD-pool+SSD-slog.
>> 
>> /usbkey/test_dir]#  time dd if=/dev/zero of=/tmpfs/testfile bs=64k 
>> count=32768;time dd if=/dev/zero of=/usbkey/test_dir/testfile bs=64k 
>> count=32768
>> 32768+0 records in
>> 32768+0 records out
>> 2147483648 <tel:(214)%20748-3648> bytes transferred in 2.279053 secs 
>> (942270169 bytes/sec)
>> 
>> real    0m2.312s
>> user    0m0.021s
>> sys     0m1.062s
>> 32768+0 records in
>> 32768+0 records out
>> 2147483648 <tel:(214)%20748-3648> bytes transferred in 0.743729 secs 
>> (2887453957 bytes/sec)
>> 
>> real    0m0.760s
>> user    0m0.016s
>> sys     0m0.652s
>> 
>> I created the ramdisk thus:
>> ramdiskadm -a rd1 3072m
>> ...
>> zfs create -o  mountpoint=/tmpfs -o sync=disabled  ramdsk1/rd1
>> 
>> I've run it many times and although the results vary yet the tale is always 
>> the same.
>> 
>> Thank You,
>> Will
>> 
> 
> 
> 
> smartos-discuss | Archives 
> <https://www.listbox.com/member/archive/184463/=now>  
> <https://www.listbox.com/member/archive/rss/184463/21953302-fd56db47> | 
> Modify <https://www.listbox.com/member/?&;> Your Subscription   
> <http://www.listbox.com/>



-------------------------------------------
smartos-discuss
Archives: https://www.listbox.com/member/archive/184463/=now
RSS Feed: https://www.listbox.com/member/archive/rss/184463/25769125-55cfbc00
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=25769125&id_secret=25769125-7688e9fb
Powered by Listbox: http://www.listbox.com

Reply via email to