Re: [zfs-discuss] Faster than 1G Ether... ESX to ZFS

2010-11-20 Thread Edward Ned Harvey
 From: Saxon, Will [mailto:will.sa...@sage.com]
 
 What I am wondering is whether this is really worth it. Are you planning
to
 share the storage out to other VM hosts, or are all the VMs running on the
 host using the 'local' storage? I know we like ZFS vs. traditional RAID
and
 volume management, and I get that being able to boot any ZFS-capable OS is
 good for disaster recovery, but what I don't get is how this ends up
working
 better than a larger dedicated ZFS system and a storage network. Is it
 cheaper over several hosts? Are you getting better performance through
 e.g. the vmxnet3 adapter and NFS than you would just using the disks
 directly?

I also don't know enough details of how this works out.  In particular:

If your goal is high performance storage, snapshots, backups, and data
integrity for Linux or some other OS (AKA, ZFS on Linux or Windows) then you
should be able to win with this method of Linux  ZFS server both in VM's of
a single physical server, utilizing a vmnet switch and either NFS or iSCSI
or CIFS.  But until some benchmarks are done, to show that vmware isn't
adding undue overhead, I must consider it still unproven.  As compared to
one big ZFS server being used as the backend SAN for a bunch of vmware
hosts...  If your goal is high performance for distributed computing, then
you always need to use local disk attached independently to each of the
compute nodes.  There's simply no way you can scale any central server large
enough to handle a bunch of hosts without any performance loss.  Assuming
the ZFS server is able to max out its local disks... If there exists a bus
which is fast enough for a remote server to max out those disks...  Then the
remote server should have the storage attached locally, because the physical
disks are the performance bottleneck.

If your goal is just use ZFS datastore for all your vmware hosts, ... AKA
you're mostly interested in checksumming and snapshots, you're not terribly
concerned with performance as long as it's fast enough, then most likely
you'll be fine with using 1Gb ether because it's so cheap.  Maybe you
upgrade to a faster or different type of bus (fc or ib).

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Faster than 1G Ether... ESX to ZFS

2010-11-20 Thread Edward Ned Harvey
Suppose if you wanted to boot from an iscsi target, just to get vmware  a
ZFS server up.  And then you could pass-thru the entire local storage
bus(es) to the ZFS server, and you could create other VM's whose storage is
backed by the ZFS server on local disk.

One way you could do this is to buy FC or IB adapter, and I presume these
have some BIOS support to configure bootable iscsi targets.  But is there
any such thing for 1Gb ether?  For this purpose, I think it would be fine
for vmware  ZFS server OSes to be physically remote via iscsi 1Gb.  OTOH
... maybe PXE is the way to go.  I've never done anything with PXE, but I
certainly know it's pretty darn universal.  I guess I'll have to look into
it.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS - Sudden decrease in write performance

2010-11-20 Thread Richard L. Hamilton
arc-discuss doesn't have anything specifically to do with ZFS;
in particular, it has nothing to do with the ZFS ARC.  Just an
unfortunate overlap of acronyms.

Cross-posted to zfs-discuss, where this probably belongs.


 Hey all1
 
 Recently I've decided to implement OpenSolaris as a
 target for BackupExec.
 
 The server I've converted into a Storage Appliance
 is an IBM x3650 M2 w/ ~4TB of on board storage via
 ~10 local SATA drives and I'm using OpenSolaris
 svn_134. I'm using a QLogic 4Gb FC HBA w/ the QLT
 driver and presented an 8TB sparse volume to the host
 due to dedup and compression being turned on for the
 zpool.
 
 When writes begin, I see anywhere from 4.5GB/Min to
 5.5GB/Min and then it drops of quickly (I mean down
 to 1GB/Min or less). I've already swapped out the
 card, cable, and port with no results. I have since
 ensured that every piece of equipment on the box had
 it's firmware updated. While doing so, I installed
 Windows Server 2008 to flash all the firmware (IBM
 doesn't have a Solaris installer). 
 
 While in Server 2008, I decided to just attempt a
 backup via share on the 1Gbs copper connection. I saw
 speeds of up to 5.5GB/Min consistently and they were
 sustained throughout 3 days of testing. Today I
 decided to move back to OpenSolaris with confidence.
 All writes began at 5.5GB/Min and quickly dropped
 off.
 
 In my troubleshooting efforts, I have also dropped
 the fiber connection and made it an iSCSI target with
 no performance gains. I have let the on board RAID
 controller do the RAID portion instead of creating a
 zpool of multiple disks with no performance gains.
 And, I have created the target LUN using both rdsk
 and dsk paths.
 
 I did notice today though, that there is a direct
 correlation between the ARC memory usage and speed.
 Using arcstat.pl, as soon as arcsz hits 1G (half of c
 column [commit?]), my throughput hits the floor (i.e.
 600MB/Min or less). I can't figure it out. I tried
 every configuration possible.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Running on Dell hardware?

2010-11-20 Thread Edward Ned Harvey
 From: Edward Ned Harvey [mailto:sh...@nedharvey.com]
 
 I have a Dell R710 which has been flaky for some time.  It crashes about
once
 per week.  I have literally replaced every piece of hardware in it, and
 reinstalled Sol 10u9 fresh and clean.

It has been over 3 weeks now, with no crashes, and  me doing everything I
can to get it to crash again.  So I'm going to call this one resolved...
Tentatively acknowledging the remote possibility that the problem could
still come back.

All I did was disable the built-in Broadcom network cards, and buy an add-on
Intel network card (EXPI9400PT).  It is worth noting, that the built-in bcom
card cannot be completely disabled if you want to use ipmi...  It's disabled
for OS only, but the iDRAC ipmi traffic still goes across the bcom
interface.  So now I have two network cables running to the machine, one of
which is only used for ipmi.  No big deal.  I had ports to spare on my
switch, and the system is stable.

It's the fault of the Broadcom card.  Rumor has it (from dell support
technician) that the bcom cards have been problematic in other OSes too...
It's not isolated to solaris.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zpool lockup and dedupratio meanings

2010-11-20 Thread Don
I've previously posted about some lockups I've experienced with ZFS.

There were two suspected causes at the time: one was deduplication, and one was 
the 2009.06 code we were running.

I upgraded to OpenIndiana 147 (initially without upgrading the zpool and zfs 
disk versions). The lockups reduced in frequency but still occurred. I've since 
upgraded the zpool and zfs versions and we'll see what happens.

Dedup was the more likely causes and so we turned it off and recreated all the 
iscsi LUNS that were being exported so as to eliminate the deduplicated data. 
That almost entirely eliminated the lockups.

Having said all that- I have two questions:
When I query for the dedupratio, I still a value of 2.37x
--
r...@nas:~# zpool get dedupratio pool0
NAME   PROPERTYVALUE  SOURCE
pool0  dedupratio  2.37x  -
--
Considering the fact that all of the iscsi LUNS that were created when dedup 
was on were subsequently deleted and recreated with dedup disabled- I don't 
understand why the value is still 2.37x. It should be near zero (There are 
probably a couple of small luns that were not removed but they are rarely 
used). Am I misinterpreting the meaning of this number? 

Second question:
The most recent pool lockup was caused when a zpool scrub was kicked off. 
Initially we see 0 values for the write bandwidth in a zpool iostat and average 
numbers for the read. After a few minutes we see the read numbers jump to 
several hundred megs/second and the write performance fluctuate between 0 and a 
few kilobytes/second. Anyone else see this behavior and can provide some 
insight?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss