On Mon, 7 Jun 2010, Jefferson Cowart wrote:
> I've thought about Oracle on NFS, but not Oracle. (Our Oracle install is
> reasonably small and we don't really manage it ourselves. We have two
> applications that run on top of Oracle. We know enough to keep it
> running/backed up/etc, but we rely on the application vendors for the most
> things Oracle related. As a result we'd probably leave Oracle on block since
> that's the way they deal with it.)
Ok (I assume you meant you're thinking about VMWare on NFS but not
Oracle). Keep an open mind, even (and perhaps especially) for small
Oracle instances. The snapshot/flexclone functionality on the filer
works extremely well over NFS for testing database changes, running
warehouse-type queries against the database, and giving copies to
developers for a "playground". You can do it with LUNs, but not as easily
or elegantly.
> Right now our server access layer is via 1G (X6748-GE-TX blades in a Cisco
> 6500), but that will likely get upgraded in the 12-18 month time frame. At
> that point we'll likely move to the Nexus 5000 line. As a result I'm making
> sure we have 10G (iSCSI + NFS) and FCoE upgrade options available. (With
> NetApp this unfortunately pushes us out of the FAS2000 line which otherwise
> does everything we'd like but has no upgrade options.)
Sounds about right. I wouldn't worry too much about FCoE on the storage
end -- it seems more useful at the host / client end. The storage can
still use regular ethernet and FC (with the Cisco gear, you're going to
need MDS for your FC SAN for the foreseeable future anyway).
> Our plan is to start out at ~10TB usable. That would be about 6-7TB CIFS/NFS
> to end-users, ~2TB of VMware (including some MS SQL and Exchange that's on top
> of VMware), and some small other misc bits . The rest would be available for
> future expansion.
At those numbers, the NetApp should definitely be near the top of your
list, especially if you move the 2TB of VMWare to NFS instead of block.
> We are in the midst of an Exchange 2010 upgrade/domain consolidation. (I'm
> amused by MS telling everyone to use DAS for Exchange 2010 since that doesn't
> work very well in a virtaul environment.)
Heh, yeah, we're in the same boat, although in our case MS said that
because of the new I/O profile, you *can* use DAS. This somehow
translated to our Exchange team as "you *should* use DAS". Oy! :)
> Sounds like a nice idea, but I don't think we'd be able to do that. (Plus the
> NetApp V series stuff doesn't appear to support Compellent on the backend.)
Really? That's disappointing. I'll have to look into that further -- we
have a pretty good relationship with both Compellent and NetApp.
> I think the biggest question I've got is real-world utilization rates. (i.e.
> how much raw storage does 10TB of data actually take after accounting for all
> overhead?) I currently have proposals from both NetApp and Compellent (and
> expect to have an EMC one shortly) and I'm trying to get as much of an
> apples-to-apples comparison as I can. I've seen various things online (mostly
> vendor blogs) and heard statements from resellers, but I'd appreciate hearing
> from customers what they are seeing in actual installs.
I've found that first of all you should require the vendors to give you
quotes for usable storage. Give them as much information as you can about
what you want, and tell them you want X TB of usable for that data. That
way, you have recourse if they screw up (we had a case at a previous job
where we asked for 7TB of usable from NetApp, and they sold us 7TB raw.
We didn't realize it, but once we got the shelf and didn't have what we
needed, they ended up giving us a shelf because we documented that we
asked for 7TB usable).
So, the things to know for raw vs. usable:
1) The size of a disk is not the size of a disk. Disks are "rightsized"
so that various vendors can provide disks to the storage companies and
still come up looking like the same disk. In NetApp's case, for example,
a 144GB FC disk starts out as 136GB to the system. All vendors do this.
SATA drives lose more space than FC drives because of the block size (520
vs. 512).
2) Once the disks are rightsized, there's RAID. You usually have options
for how to lay out RAID spindles (for NetApp / EMC) or RAID protection at
the block level (Compellent). This affects performance and capacity.
NetApp lets you select how many disks per raid group. EMC has fixed
choices. I believe the default for NetApp is still 14-2 -- 14 data disks
with 2 parity, so 2 out of every 16 drives will be non-data.
3) After that, there's the filesystem. WAFL takes up space, and there's
overhead (about 10%) also. Any filesystem will take up space, of course,
and all unix filesystems attempt to prevent you from using 100% of the
space anyway, so there's overhead there. On an EMC LUN, though,
it won't take up space until the host sees it and formats it. This is
where the apples-to-apples comparison becomes crucial.
4) After the filesystem, there's the snapshot space. NetApp snapshots are
golden -- that is, they take precedence over anything else *if you're
using them*. So, you need to know your rate of change to calculate how
much snapshot space you need. The default used to be 20%, but I'm not
sure if it still is.
5) For a NetApp LUN, after all that is calculated, you'll then present the
LUN to a host and the host will format it, which means you'll lose more
space. So, for comparison purposes, a NetApp LUN is less efficient and
uses more space than an EMC or Compellent LUN. That's why it's important
to classify what will be NAS vs. SAN. If you minimize your block
requirements, NetApp makes sense, but if it's mostly block, it'll cost a
lot (er, even more :) ) than it otherwise would have.
EMC and NetApp have had an ongoing battle about raw vs. usable. EMC used
to claim that their LUNs were 80% usable while NetApp's were 55% (or
sometimes even less) usable. They were sort of right, but they were not
comparing equivalent configurations.
Here's an example from our FAS3020 with one shelf of 14 144GB FC drives:
Aggregate 'aggr1'
Total space WAFL reserve Snap reserve Usable space BSR
NVLOG A-SIS
974701056KB 97470104KB 0KB 877230952KB
0KB 3305148KB
Aggregate Allocated Used Avail
Total space 747156988KB 317262708KB 126724900KB
Snap reserve 0KB 0KB 0KB
WAFL reserve 97470104KB 8614364KB 88855740KB
Aggregate 'aggr0'
Total space WAFL reserve Snap reserve Usable space BSR
NVLOG A-SIS
139243008KB 13924300KB 0KB 125318708KB
0KB 0KB
Aggregate Allocated Used Avail
Total space 119595480KB 3448220KB 5714504KB
Snap reserve 0KB 0KB 0KB
WAFL reserve 13924300KB 1407224KB 12517076KB
The raw space would be 14*144= 2,113,929,216 KB.
Rightsized space: 14*136 = 1,996,488,704 KB.
We use RAID DP, and there are two aggregates, one 3 disk and one 9 disk
(our root vol is on its own aggregate, which is a bit of wasted space but
in a larger filer is better for management).
So, for the 9 disk aggregate:
9*136 = 1,283,457,024 KB (or 9*144=1,358,954,496 KB)
7*136 = 998,244,352 KB.
The filer itself is reporting 974,701,056 KB. That's probably a
difference in 1000 vs. 1024 or something similar.
The filer then takes 10% for WAFL reserve and there's an aggregate
snapshot reserve that's not required but defaults to 5% -- you should make
that 0% unless you're using a very specific data protection scheme (you'll
see it in the documentation somewhere -- almost nobody uses it :) ).
So we're told the usable space (before volume snap reserve) is 877,230,952
KB out of 1,358,954,496 KB raw (worst case calculation) for an overhead of
35%.
If you add a standard 20% snap reserve to that, you can see where the 55%
overhead comes from.
Now, we have an EMC CX4-120 with 1TB SATA drives configured as a RAID6,
which is roughly equivalent to NetApp's RAID-DP (I don't have any 1TB
SATAs on the NetApp, unfortunately).
The 1TB drives are rightsized to 917.18GB on the CX4. There are 16 of the
drives configured as RAID6, and the system tells me I have 12840.088GB
total space available there.
16*1TB = 17,179,869,184 KB
12840GB = 13,463,715,840 KB
So after rightsizing and RAID we're at 22% before formatting a LUN with a
filesystem. I would expect that after formatting we'd lose at least 10%
to filesystem and overhead, just like WAFL, which puts us about on par
with the NetApp.
So, the only real difference is when you take the NetApp, create a LUN on
WAFL, and then format it with another filesystem like UFS. You'll lose
another 10%.
Anyone care to check my math? :)
-Adam
_______________________________________________
Tech mailing list
[email protected]
http://lopsa.org/cgi-bin/mailman/listinfo/tech
This list provided by the League of Professional System Administrators
http://lopsa.org/