Under FreeBSD I've seen zpool scrub sustain nearly 500 MB/s in pools with large
files (a pool with eight MIRROR vdevs on two Silicon Image 3124 controllers).
You need to carefully look for bottlenecks in the hardware. You don't indicate
how the disks are attached. I would measure the total
Something like this, maybe
http://blogs.sun.com/constantin/entry/x4500_solaris_zfs_iscsi_perfect
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
The database is MySQL, it runs on a Linux box that connects to the Nexenta
server through 10GbE using iSCSI. Just a short question - wouldn't it be
easier, and perhaps faster, to just have the MySQL DB on an NFS share? iSCSI
adds complexity, both on the target and the initiator.
Also,
Just a short question - wouldn't it be easier, and perhaps faster, to just
have the MySQL DB on an NFS share? iSCSI adds
complexity, both on the target and the initiator.
Yes, we did tried both and we didn't notice any difference in term of
performances. I've read conflicting opinions on
Just a short question - wouldn't it be easier, and perhaps faster, to just
have the MySQL DB on an NFS share? iSCSI adds
complexity, both on the target and the initiator.
Yes, we did tried both and we didn't notice any difference in term of
performances. I've read conflicting
On Jul 5, 2010, at 4:19 AM, Ian D wrote:
Also, are you using jumbo frames? That can usually help a bit with either
access protocol
Yes. It was off early on and we did notice a significant difference once we
switched it on. Turning naggle off as suggested by Richard also seem to
have
To summarise, putting 28 disks in a single vdev is nothing you would do if you
want performance. You'll end up with as many IOPS a single drive can do.
Split it up into smaller (10 disk) vdevs and try again. If you need high
performance, put them in a striped mirror (aka RAID1+0)
A
Ok... so we've rebuilt the pool as 14 pairs of mirrors, each pair having one
disk in each of the two JBODs. Now we're getting about 500-1000 IOPS
(according to zpool iostat) and 20-30MB/sec in random read on a big database.
Does that sounds right?Seems right, as Erik said. Btw, do you use
Ok... so we've rebuilt the pool as 14 pairs of mirrors, each pair
having one disk in each of the two JBODs. Now we're getting about
500-1000 IOPS (according to zpool iostat) and 20-30MB/sec in random
read on a big database. Does that sounds right?
I am not sure who wrote the above text
On Sun, Jul 4, 2010 at 11:28 AM, Bob Friesenhahn
bfrie...@simple.dallas.tx.us wrote:
Ok... so we've rebuilt the pool as 14 pairs of mirrors, each pair having
one disk in each of the two JBODs. Now we're getting about 500-1000 IOPS
(according to zpool iostat) and 20-30MB/sec in random read on
On Sun, Jul 4, 2010 at 10:08 AM, Ian D rewar...@hotmail.com wrote:
What I don't understand is why, when I run a single query I get 100 IOPS
and 3MB/sec. The setup can obviously do better, so where is the
bottleneck? I don't see any CPU core on any side being maxed out so it
can't be it...
In what way is CPU contention being monitored? prstat without
options is nearly useless for a multithreaded app on a multi-CPU (or
multi-core/multi-thread) system. mpstat is only useful if threads
never migrate between CPU's. prstat -mL gives a nice picture of how
busy each LWP (thread)
On Jul 4, 2010, at 8:08 AM, Ian D wrote:
Ok... so we've rebuilt the pool as 14 pairs of mirrors, each pair having one
disk in each of the two JBODs. Now we're getting about 500-1000 IOPS
(according to zpool iostat) and 20-30MB/sec in random read on a big
database. Does that sounds right?
On Sun, Jul 4, 2010 at 2:08 PM, Ian D rewar...@hotmail.com wrote:
Mem: 74098512k total, 73910728k used, 187784k free, 96948k buffers
Swap: 2104488k total, 208k used, 2104280k free, 63210472k cached
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
17652 mysql
Is that 38% of one CPU or 38% of all CPU's? How many CPU's does the
Linux box have? I don't mean the number of sockets, I mean number of
sockets * number of cores * number of threads per core. My
The server has two Intel X5570s, they are quad core and have hyperthreading.
It would say
To summarise, putting 28 disks in a single vdev is nothing you would do if you
want performance. You'll end up with as many IOPS a single drive can do.
Split it up into smaller (10 disk) vdevs and try again. If you need high
performance, put them in a striped mirror (aka RAID1+0)A little
To summarise, putting 28 disks in a single vdev is nothing you would do if you
want performance. You'll end up with as many IOPS a single drive can do.
Split it up into smaller (10 disk) vdevs and try again. If you need high
performance, put them in a striped mirror (aka RAID1+0)
A
On 7/3/2010 2:22 PM, Roy Sigurd Karlsbakk wrote:
To summarise, putting 28 disks in a single vdev is nothing you
would do if you want performance. You'll end up with as many
IOPS a single drive
Hi! We've put 28x 750GB SATA drives in a RAIDZ2 pool (a single vdev) and we
get about 80MB/s in sequential read or write. We're running local tests on the
server itself (no network involved). Is that what we should be expecting? It
seems slow to me.
Thanks
Hi! We've put 28x 750GB SATA drives in a RAIDZ2 pool (a single vdev) and we get
about 80MB/s in sequential read or write. We're running local tests on the
server itself (no network involved). Is that what we should be expecting? It
seems slow to me.
Please read the ZFS best practices
Hi! We've put 28x 750GB SATA drives in a RAIDZ2 pool (a single vdev) and we get
about 80MB/s in sequential read or write. We're running local tests on the
server itself (no network involved). Is that what we should be expecting? It
seems slow to me.
Please read the ZFS best
21 matches
Mail list logo