Re: [zfs-discuss] ZFS performance question over NFS
Hi Bob > I don't know what the request pattern from filebench looks like but it seems > like your ZEUS RAM devices are not keeping up or > else many requests are bypassing the ZEUS RAM devices. > > Note that very large synchronous writes will bypass your ZEUS RAM device and > go directly to a log in the main store. Small (<= > 128K) writes should directly benefit from the dedicated zil device. > > Find a copy of zilstat.ksh and run it while filebench is running in order to > understand more about what is going on. > > Bob The pattern looks like: N-Bytes N-Bytes/s N-Max-RateB-Bytes B-Bytes/s B-Max-Rateops <=4kB 4-32kB >=32kB 958865695886569588656 88399872 88399872 88399872 90 0 0 90 666228066622806662280 87031808 87031808 87031808 83 0 0 83 636672863667286366728 72790016 72790016 72790016 79 0 0 79 631635263163526316352 83886080 83886080 83886080 80 0 0 80 668761666876166687616 84594688 84594688 84594688 92 0 0 92 490904849090484909048 69238784 69238784 69238784 73 0 0 73 660528066052806605280 81924096 81924096 81924096 79 0 0 79 689533668953366895336 81625088 81625088 81625088 85 0 0 85 653212865321286532128 87486464 87486464 87486464 90 0 0 90 692513669251366925136 86118400 86118400 86118400 83 0 0 83 So does it look good, bad or ugly ;) Thomas ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS performance question over NFS
On Thu, 18 Aug 2011, Thomas Nau wrote: Tim the client is identical as the server but no SAS drives attached. Also right now only one 1gbit Intel NIC Is available I don't know what the request pattern from filebench looks like but it seems like your ZEUS RAM devices are not keeping up or else many requests are bypassing the ZEUS RAM devices. Note that very large synchronous writes will bypass your ZEUS RAM device and go directly to a log in the main store. Small (<= 128K) writes should directly benefit from the dedicated zil device. Find a copy of zilstat.ksh and run it while filebench is running in order to understand more about what is going on. Bob -- Bob Friesenhahn bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/ GraphicsMagick Maintainer,http://www.GraphicsMagick.org/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS performance question over NFS
Tim the client is identical as the server but no SAS drives attached. Also right now only one 1gbit Intel NIC Is available Thomas Am 18.08.2011 um 17:49 schrieb Tim Cook : > What are the specs on the client? > > On Aug 18, 2011 10:28 AM, "Thomas Nau" wrote: > > Dear all. > > We finally got all the parts for our new fileserver following several > > recommendations we got over this list. We use > > > > Dell R715, 96GB RAM, dual 8-core Opterons > > 1 10GE Intel dual-port NIC > > 2 LSI 9205-8e SAS controllers > > 2 DataON DNS-1600 JBOD chassis > > 46 Seagate constellation SAS drives > > 2 STEC ZEUS RAM > > > > > > The base zpool config utilizes 42 drives plus the STECs as mirrored > > log devices. The Seagates are setup as a stripe of 7 times 6-drive-RAIDZ2 > > junks plus as said a dedicated ZIL made of the mirrored STECs. > > > > As a quick'n dirty check we ran "filebench" with the "fileserver" > > workload. Running locally we get > > > > statfile1 5476ops/s 0.0mb/s 0.6ms/op 179us/op-cpu > > deletefile1 5476ops/s 0.0mb/s 1.0ms/op 454us/op-cpu > > closefile3 5476ops/s 0.0mb/s 0.0ms/op 5us/op-cpu > > readfile1 5476ops/s 729.5mb/s 0.2ms/op 128us/op-cpu > > openfile2 5477ops/s 0.0mb/s 0.8ms/op 204us/op-cpu > > closefile2 5477ops/s 0.0mb/s 0.0ms/op 5us/op-cpu > > appendfilerand1 5477ops/s 42.8mb/s 0.3ms/op 184us/op-cpu > > openfile1 5477ops/s 0.0mb/s 0.9ms/op 209us/op-cpu > > closefile1 5477ops/s 0.0mb/s 0.0ms/op 6us/op-cpu > > wrtfile1 5477ops/s 688.4mb/s 0.4ms/op 220us/op-cpu > > createfile1 5477ops/s 0.0mb/s 2.7ms/op 1068us/op-cpu > > > > > > > > with a single remote client (similar Dell System) using NFS > > > > statfile1 90ops/s 0.0mb/s 27.6ms/op 145us/op-cpu > > deletefile1 90ops/s 0.0mb/s 64.5ms/op 401us/op-cpu > > closefile3 90ops/s 0.0mb/s 25.8ms/op 40us/op-cpu > > readfile1 90ops/s 11.4mb/s 3.1ms/op 363us/op-cpu > > openfile2 90ops/s 0.0mb/s 66.0ms/op 263us/op-cpu > > closefile2 90ops/s 0.0mb/s 22.6ms/op 124us/op-cpu > > appendfilerand1 90ops/s 0.7mb/s 0.5ms/op 101us/op-cpu > > openfile1 90ops/s 0.0mb/s 72.6ms/op 269us/op-cpu > > closefile1 90ops/s 0.0mb/s 43.6ms/op 189us/op-cpu > > wrtfile1 90ops/s 11.2mb/s 0.2ms/op 211us/op-cpu > > createfile1 90ops/s 0.0mb/s 226.5ms/op 709us/op-cpu > > > > > > > > the same remote client with zpool sync disabled on the server > > > > statfile1 479ops/s 0.0mb/s 6.2ms/op 130us/op-cpu > > deletefile1 479ops/s 0.0mb/s 13.0ms/op 351us/op-cpu > > closefile3 480ops/s 0.0mb/s 3.0ms/op 37us/op-cpu > > readfile1 480ops/s 62.7mb/s 0.8ms/op 174us/op-cpu > > openfile2 480ops/s 0.0mb/s 14.1ms/op 235us/op-cpu > > closefile2 480ops/s 0.0mb/s 6.0ms/op 123us/op-cpu > > appendfilerand1 480ops/s 3.7mb/s 0.2ms/op 53us/op-cpu > > openfile1 480ops/s 0.0mb/s 13.7ms/op 235us/op-cpu > > closefile1 480ops/s 0.0mb/s 11.1ms/op 190us/op-cpu > > wrtfile1 480ops/s 60.3mb/s 0.2ms/op 233us/op-cpu > > createfile1 480ops/s 0.0mb/s 35.6ms/op 683us/op-cpu > > > > > > Disabling ZIL is no option but I expected a much better performance > > especially the ZEUS RAM only gets us a speed-up of about 1.8x > > > > Is this test realistic for a typical fileserver scenario or does it require > > many > > more clients to push the limits? > > > > Thanks > > Thomas > > ___ > > zfs-discuss mailing list > > zfs-discuss@opensolaris.org > > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS performance question over NFS
What are the specs on the client? On Aug 18, 2011 10:28 AM, "Thomas Nau" wrote: > Dear all. > We finally got all the parts for our new fileserver following several > recommendations we got over this list. We use > > Dell R715, 96GB RAM, dual 8-core Opterons > 1 10GE Intel dual-port NIC > 2 LSI 9205-8e SAS controllers > 2 DataON DNS-1600 JBOD chassis > 46 Seagate constellation SAS drives > 2 STEC ZEUS RAM > > > The base zpool config utilizes 42 drives plus the STECs as mirrored > log devices. The Seagates are setup as a stripe of 7 times 6-drive-RAIDZ2 > junks plus as said a dedicated ZIL made of the mirrored STECs. > > As a quick'n dirty check we ran "filebench" with the "fileserver" > workload. Running locally we get > > statfile1 5476ops/s 0.0mb/s 0.6ms/op 179us/op-cpu > deletefile1 5476ops/s 0.0mb/s 1.0ms/op 454us/op-cpu > closefile3 5476ops/s 0.0mb/s 0.0ms/op 5us/op-cpu > readfile1 5476ops/s 729.5mb/s 0.2ms/op 128us/op-cpu > openfile2 5477ops/s 0.0mb/s 0.8ms/op 204us/op-cpu > closefile2 5477ops/s 0.0mb/s 0.0ms/op 5us/op-cpu > appendfilerand1 5477ops/s 42.8mb/s 0.3ms/op 184us/op-cpu > openfile1 5477ops/s 0.0mb/s 0.9ms/op 209us/op-cpu > closefile1 5477ops/s 0.0mb/s 0.0ms/op 6us/op-cpu > wrtfile1 5477ops/s 688.4mb/s 0.4ms/op 220us/op-cpu > createfile1 5477ops/s 0.0mb/s 2.7ms/op 1068us/op-cpu > > > > with a single remote client (similar Dell System) using NFS > > statfile1 90ops/s 0.0mb/s 27.6ms/op 145us/op-cpu > deletefile1 90ops/s 0.0mb/s 64.5ms/op 401us/op-cpu > closefile3 90ops/s 0.0mb/s 25.8ms/op 40us/op-cpu > readfile1 90ops/s 11.4mb/s 3.1ms/op 363us/op-cpu > openfile2 90ops/s 0.0mb/s 66.0ms/op 263us/op-cpu > closefile2 90ops/s 0.0mb/s 22.6ms/op 124us/op-cpu > appendfilerand1 90ops/s 0.7mb/s 0.5ms/op 101us/op-cpu > openfile1 90ops/s 0.0mb/s 72.6ms/op 269us/op-cpu > closefile1 90ops/s 0.0mb/s 43.6ms/op 189us/op-cpu > wrtfile1 90ops/s 11.2mb/s 0.2ms/op 211us/op-cpu > createfile1 90ops/s 0.0mb/s 226.5ms/op 709us/op-cpu > > > > the same remote client with zpool sync disabled on the server > > statfile1 479ops/s 0.0mb/s 6.2ms/op 130us/op-cpu > deletefile1 479ops/s 0.0mb/s 13.0ms/op 351us/op-cpu > closefile3 480ops/s 0.0mb/s 3.0ms/op 37us/op-cpu > readfile1 480ops/s 62.7mb/s 0.8ms/op 174us/op-cpu > openfile2 480ops/s 0.0mb/s 14.1ms/op 235us/op-cpu > closefile2 480ops/s 0.0mb/s 6.0ms/op 123us/op-cpu > appendfilerand1 480ops/s 3.7mb/s 0.2ms/op 53us/op-cpu > openfile1 480ops/s 0.0mb/s 13.7ms/op 235us/op-cpu > closefile1 480ops/s 0.0mb/s 11.1ms/op 190us/op-cpu > wrtfile1 480ops/s 60.3mb/s 0.2ms/op 233us/op-cpu > createfile1 480ops/s 0.0mb/s 35.6ms/op 683us/op-cpu > > > Disabling ZIL is no option but I expected a much better performance > especially the ZEUS RAM only gets us a speed-up of about 1.8x > > Is this test realistic for a typical fileserver scenario or does it require many > more clients to push the limits? > > Thanks > Thomas > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss