[zfs-discuss] Migrate from iscsitgt to comstar?

2009-09-21 Thread Markus Kovero
Is it possible to migrate data from iscsitgt for comstar iscsi target? I guess 
comstar wants metadata at beginning of volume and this makes things difficult?

Yours
Markus Kovero
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Real help

2009-09-21 Thread Chris Ridd


On 20 Sep 2009, at 19:46, dick hoogendijk wrote:



On Sun, 2009-09-20 at 11:41 -0700, vattini giacomo wrote:
Hi there,i'm in a bad situation,under Ubuntu i was tring to import  
a solaris zpool that is in /dev/sda1,while the Ubuntu is in /dev/ 
sda5;not being able to mount the solaris pool i decide to destroy  
the pool created like that

sudo zfs-fuse
sudo zpool  create hazz0 /dev/sda1
sudo zpool destroy hazz0
sudo reboot
Now opensolaris is not booting everything is vanished
Is there anyhow to restore everything?


Any idea about the meaning of the verb DESTROY ?


Does zpool destroy prompt are you sure in any way? Some admin tools  
do (beadm destroy for example) but there's not a lot of consistency.


Cheers,

Chris
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re-import RAID-Z2 with faulted disk

2009-09-21 Thread Kyle J. Aleshire
Hi all, I have a RAID-Z2 setup with 6x 500Gb SATA disks. I exported the
array to use under a different system but during or after the export one of
the disks failed:

k...@localhost:~$ pfexec zpool import
  pool: chronicle
id: 11592382930413748377
 state: DEGRADED
status: One or more devices are missing from the system.
action: The pool can be imported despite missing or damaged devices.  The
fault tolerance of the pool may be compromised if imported.
   see: http://www.sun.com/msg/ZFS-8000-2Q
config:

chronicle   DEGRADED
  raidz2DEGRADED
c9t2d0  UNAVAIL  cannot open
c9t1d0  ONLINE
c9t0d0  ONLINE
c9t4d0  ONLINE
c9t5d0  ONLINE
c9t3d0  ONLINE

I have no success trying to reimport the pool:
k...@localhost:~$ pfexec zpool import -f chronicle
cannot import 'chronicle': one or more devices is currently unavailable

The disk has since been replaced, so now:
k...@localhost:~$ pfexec zpool import
  pool: chronicle
id: 11592382930413748377
 state: DEGRADED
status: One or more devices contains corrupted data.
action: The pool can be imported despite missing or damaged devices.  The
fault tolerance of the pool may be compromised if imported.
   see: http://www.sun.com/msg/ZFS-8000-4J
config:

chronicle   DEGRADED
  raidz2DEGRADED
c9t2d0  FAULTED  corrupted data
c9t1d0  ONLINE
c9t0d0  ONLINE
c9t4d0  ONLINE
c9t5d0  ONLINE
c9t3d0  ONLINE

but pfexec zpool import -f chronicle still fails with the same message.

I've Google'd this several times for a fix but to no avail. Any assistance
is appreciated.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Real help

2009-09-21 Thread David Magda

On Sep 21, 2009, at 06:52, Chris Ridd wrote:

Does zpool destroy prompt are you sure in any way? Some admin  
tools do (beadm destroy for example) but there's not a lot of  
consistency.


No it doesn't, which I always found strange.

Personally I always thought you should be queried for a zfs destroy,  
but add a -f option for things like scripts. Not sure if things can  
be changed now.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re-import RAID-Z2 with faulted disk

2009-09-21 Thread Casper . Dik


The disk has since been replaced, so now:
k...@localhost:~$ pfexec zpool import
  pool: chronicle
id: 11592382930413748377
 state: DEGRADED
status: One or more devices contains corrupted data.
action: The pool can be imported despite missing or damaged devices.  The
fault tolerance of the pool may be compromised if imported.
   see: http://www.sun.com/msg/ZFS-8000-4J
config:

chronicle   DEGRADED
  raidz2DEGRADED
c9t2d0  FAULTED  corrupted data
c9t1d0  ONLINE
c9t0d0  ONLINE
c9t4d0  ONLINE
c9t5d0  ONLINE
c9t3d0  ONLINE

but pfexec zpool import -f chronicle still fails with the same message.


That sounds like a bug; if ZFS can recover then zpool import should
also work.

So what was wrong with the broken disk?  Not just badly plugged in?

Cas[er

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Real help

2009-09-21 Thread Mattias Pantzare
On Mon, Sep 21, 2009 at 13:34, David Magda dma...@ee.ryerson.ca wrote:
 On Sep 21, 2009, at 06:52, Chris Ridd wrote:

 Does zpool destroy prompt are you sure in any way? Some admin tools do
 (beadm destroy for example) but there's not a lot of consistency.

 No it doesn't, which I always found strange.

 Personally I always thought you should be queried for a zfs destroy, but
 add a -f option for things like scripts. Not sure if things can be changed
 now.

You can import a destroyed pool, you can find them wiht zpool import -D.

But the problem in this case was not zpool destroy, it was zpool
create. zpool create will overwrite whatever was on the partition.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] lots of zil_clean threads

2009-09-21 Thread Nils Goroll

Hi All,

out of curiosity: Can anyone come up with a good idea about why my snv_111 
laptop computer should run more than 1000 zil_clean threads?


ff0009a9dc60 fbc2c0300 tq:zil_clean
ff0009aa3c60 fbc2c0300 tq:zil_clean
ff0009aa9c60 fbc2c0300 tq:zil_clean
ff0009aafc60 fbc2c0300 tq:zil_clean
ff0009ab5c60 fbc2c0300 tq:zil_clean
ff0009abbc60 fbc2c0300 tq:zil_clean
ff0009ac1c60 fbc2c0300 tq:zil_clean
 ::threadlist!grep zil_clean| wc -l
1037

Thanks, Nils

P.S.: Please don't spend too much time on this, for me, this question is really 
academic - but I'd be grateful for any good answers.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS file disk usage

2009-09-21 Thread Andrew Deason
On Sun, 20 Sep 2009 20:31:57 -0400
Richard Elling richard.ell...@gmail.com wrote:

 If you are just building a cache, why not just make a file system and
 put a reservation on it? Turn off auto snapshots and set other
 features as per best practices for your workload? In other words,
 treat it like we
 treat dump space.
 
 I think that we are getting caught up in trying to answer the question
 you ask rather than solving the problem you have... perhaps because
 we don't understand the problem.

Yes, possibly... some of these suggestions dont quite make a lot of
sense to me. We can't just make a filesystem and put a reservation on
it; we are just an application the administrator puts on a machine for
it to access AFS. So I'm not sure when you are imagining we do that;
when the client starts up? Or part of the installation procedure?
Requiring a separate filesystem seems unnecessarily restrictive.

And I still don't see how that helps. Making an fs with a reservation
would definitely limit us to the specified space, but we still can't get
an accurate picture of the current disk usage. I already mentioned why
using statvfs is not usable with that commit delay.

But solving the general problem for me isn't necessary. If I could just
get a ballpark estimate of the max overhead for a file, I would be fine.
I haven't payed attention to it before, so I don't even have an
intuitive feel for what it is.

-- 
Andrew Deason
adea...@sinenomine.net
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] lots of zil_clean threads

2009-09-21 Thread Neil Perrin

Nils,

A zil_clean() is started for each dataset after every txg.
this includes snapshots (which is perhaps a bit inefficient).
Still, zil_clean() is fairly lightweight if there's nothing
to do (grab a non contended lock; find nothing on a list;
drop the lock  exit).

Neil.

On 09/21/09 08:08, Nils Goroll wrote:

Hi All,

out of curiosity: Can anyone come up with a good idea about why my 
snv_111 laptop computer should run more than 1000 zil_clean threads?


ff0009a9dc60 fbc2c0300 tq:zil_clean
ff0009aa3c60 fbc2c0300 tq:zil_clean
ff0009aa9c60 fbc2c0300 tq:zil_clean
ff0009aafc60 fbc2c0300 tq:zil_clean
ff0009ab5c60 fbc2c0300 tq:zil_clean
ff0009abbc60 fbc2c0300 tq:zil_clean
ff0009ac1c60 fbc2c0300 tq:zil_clean
  ::threadlist!grep zil_clean| wc -l
1037

Thanks, Nils

P.S.: Please don't spend too much time on this, for me, this question is 
really academic - but I'd be grateful for any good answers.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] resizing zpools by growing LUN

2009-09-21 Thread Sascha
Hi Darren,

sorry that it took so long before I could answer.

The good thing:
I found out what went wrong.

What I did:
After resizing a Disk on the Storage, solaris recognizes it immediately.
Everytime you resize a disk, the EVA storage updates the discription which 
contains the size. So typing echo |format shows promptly the new discription.
So that seemed to work fine.

But It turned out that format doesn't automagically update the amount of 
cylinders/sectors.
So i tried to auto configure the disk.
When I did that, the first sector of partition 0 changed from sector # 34 to # 
256. (which makes label 0 and 1 of zpool unaccessible).
The last sector changed to the new end of the disk (which makes label 3 and 4 
of the zpool unaccessible). 
Partition 8 changed accordingly and correctly.
If I then relabel the disk the zpool is destroyed.
Label 0 and 1 of the zpool are gone and, obviously  label 2 and 3 too. 

What I did to solve it:
1. Changed the size of a disk/LUN on the storage
2. Verified the position of the starting sector of partition 0
3. Did an auto configure
4. Changed starting sector of partition 0 to the formerly starting sector from 
step #2
5. labeled the disk

I then was able to impoort the zpool

Do you know why the auto configure changes the starting cylinder ?
What I also found out is that it did not happen every time.
Sometimes the first sector of partition 0 changes, sometimes not.
Up to now I can't find any correlation between when and why.

Sascha
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] resizing zpools by growing LUN

2009-09-21 Thread Richard Elling

On Sep 21, 2009, at 8:59 AM, Sascha wrote:


Hi Darren,

sorry that it took so long before I could answer.

The good thing:
I found out what went wrong.

What I did:
After resizing a Disk on the Storage, solaris recognizes it  
immediately.
Everytime you resize a disk, the EVA storage updates the discription  
which contains the size. So typing echo |format shows promptly the  
new discription.

So that seemed to work fine.

But It turned out that format doesn't automagically update the  
amount of cylinders/sectors.

So i tried to auto configure the disk.
When I did that, the first sector of partition 0 changed from sector  
# 34 to # 256. (which makes label 0 and 1 of zpool unaccessible).


This was a change made long ago, but it finally caught up with you.
You must have created the original EFI label with an older version of
[Open]Solaris. If you then relabel with automatic settings, the starting
sector will change to the 256 value, which is a much better starting  
point.

 -- richard

The last sector changed to the new end of the disk (which makes  
label 3 and 4 of the zpool unaccessible).

Partition 8 changed accordingly and correctly.
If I then relabel the disk the zpool is destroyed.
Label 0 and 1 of the zpool are gone and, obviously  label 2 and 3 too.

What I did to solve it:
1. Changed the size of a disk/LUN on the storage
2. Verified the position of the starting sector of partition 0
3. Did an auto configure
4. Changed starting sector of partition 0 to the formerly starting  
sector from step #2

5. labeled the disk

I then was able to impoort the zpool

Do you know why the auto configure changes the starting cylinder ?
What I also found out is that it did not happen every time.
Sometimes the first sector of partition 0 changes, sometimes not.
Up to now I can't find any correlation between when and why.

Sascha
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Incremental backup via zfs send / zfs receive

2009-09-21 Thread David Pacheco

Frank Middleton wrote:

The problem with the regular stream is that most of the file
system properties (such as mountpoint) are not copied as they
are with a recursive stream. This may seem an advantage to some,
(e.g., if the remote mountpoint is already in use, the mountpoint
seems to default to legacy). However, did I miss anything in the
documentation, or would it be worth submitting an RFE for an
option to send/recv properties in a non-recursive stream?



This is

6839260 want zfs send with properties

-- Dave

--
David Pacheco, Sun Microsystems Fishworks. http://blogs.sun.com/dap/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] resizing zpools by growing LUN

2009-09-21 Thread Sascha
Hej Richard.

think I'll update all our servers to the same version of zfs...
That will hopefully make sure that this doesn't happen again :-)

Darren and Richard: Thank you very much for your help !

Sascha
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] lots of zil_clean threads

2009-09-21 Thread Neil Perrin

Thinking more about this I'm confused about what you are seeing.
The function dsl_pool_zil_clean() will serialise separate calls to
zil_clean() within a pool. I don't expect you have 1037 pools on your laptop!
So I don't know what's going on. What is the typical call stack for those
zil_clean() threads?

Neil.

On 09/21/09 08:53, Neil Perrin wrote:

Nils,

A zil_clean() is started for each dataset after every txg.
this includes snapshots (which is perhaps a bit inefficient).
Still, zil_clean() is fairly lightweight if there's nothing
to do (grab a non contended lock; find nothing on a list;
drop the lock  exit).

Neil.

On 09/21/09 08:08, Nils Goroll wrote:

Hi All,

out of curiosity: Can anyone come up with a good idea about why my 
snv_111 laptop computer should run more than 1000 zil_clean threads?


ff0009a9dc60 fbc2c0300 tq:zil_clean
ff0009aa3c60 fbc2c0300 tq:zil_clean
ff0009aa9c60 fbc2c0300 tq:zil_clean
ff0009aafc60 fbc2c0300 tq:zil_clean
ff0009ab5c60 fbc2c0300 tq:zil_clean
ff0009abbc60 fbc2c0300 tq:zil_clean
ff0009ac1c60 fbc2c0300 tq:zil_clean
  ::threadlist!grep zil_clean| wc -l
1037

Thanks, Nils

P.S.: Please don't spend too much time on this, for me, this question 
is really academic - but I'd be grateful for any good answers.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS Recv slow with high CPU

2009-09-21 Thread Matthew Ahrens

Tristan Ball wrote:

Hi Everyone,

I have a couple of systems running opensolaris b118, one of which sends 
hourly snapshots to the other. This has been working well, however as of 
today, the receiving zfs process has started running extremely slowly, 
and is running at 100% CPU on one core, completely in kernel mode. A 
little bit of exploration with lockstat and dtrace seems to imply that 
the issue is around the dbuf_free_range function - or at least, that's 
what it looks like to my inexperienced eye!


This is probably RFE 6812603 zfs send can aggregate free records, which is 
currently being worked on.


--matt
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfsdle eating all resources..

2009-09-21 Thread Nilsen, Vidar
Hi,

I've got some strange problems with my serer today.
When I boot b123, it stops at reading zfs config. I've tried several
times to get past this point, but it seems to freeze there.
Then I tried single user mode, from GRUB, and it seems to get me a
little further.
After a few minutes however, the OS becomes unresponsive, and soon its
impossible to do anything.
I get messages like this File system full, swap space limit exceeded,
and some messages about shell being unable to fork.

If I'm really fast after booting single user, I can start top, and it
looks like I have tons of zfsdle processes, using all available
resources..
The rpool is 2 disks in mirror.
The other pool is 16 disks in 4x4 raidz. I've recently lost disk 12,
and was trying to replace it just before this happened.


Best Regards,
Vidar Nilsen
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Directory size value

2009-09-21 Thread Chris Banal
It appears as though zfs reports the size of a directory to be one byte per
file. Traditional file systems such as ufs or ext3 report the actual size of
the data needed to store the directory.

This causes some trouble with the default behavior of some nfs clients
(linux) to decide to to use a readdirplus call when directory contents are
small vs a readdir call when the contents are large. We've found this
particularly troublesome with Maildir style mail folders. The speedup by not
using readdirplus is a factor of 100 in this particular situation. While we
have a work around it would seem that this non-standard behavior might cause
trouble for others and in other areas. Are there any suggestions for dealing
with this difference and/or why zfs does not represent its directory sizes
in a more traditional manner?

From the linux kernel source.

./fs/nfs/inode.c:#define NFS_LIMIT_READDIRPLUS (8*PAGE_SIZE)


ie. ZFS

zfshost:~/testdir ls -1 | wc -l
330
zfshost:~/testdir stat .
  File: `.'
  Size: 332 Blocks: 486IO Block: 32768  directory
Device: 29h/41d Inode: 540058  Links: 2
Access: (0775/drwxrwxr-x)  Uid: ( 2891/   banal)   Gid: (  101/film)
Access: 2008-11-05 18:40:16.0 -0800
Modify: 2009-09-01 16:09:52.782674099 -0700
Change: 2009-09-01 16:09:52.782674099 -0700

ie. ext3

ext3host:~/testdir ls -1 | wc -l
330
ext3host:~/testdir stat .
  File: `.'
  Size: 36864   Blocks: 72 IO Block: 4096   directory
Device: 807h/2055d  Inode: 23887981Links: 2
Access: (0775/drwxrwxr-x)  Uid: ( 2891/   banal)   Gid: (  101/film)
Access: 2009-09-21 13:44:00.0 -0700
Modify: 2009-09-21 13:44:31.0 -0700
Change: 2009-09-21 13:44:31.0 -0700

Thanks,
Chris
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Directory size value

2009-09-21 Thread Tomas Ögren
On 21 September, 2009 - Chris Banal sent me these 4,4K bytes:

 It appears as though zfs reports the size of a directory to be one byte per
 file. Traditional file systems such as ufs or ext3 report the actual size of
 the data needed to store the directory.

Or rather, the size needed at some random point in time .. because it
seldom shrinks when you remove files.. At least not on UFS.. So
basically it's an uninteresting and stale piece of information. The
number of entries in the directory is much more useful.

 This causes some trouble with the default behavior of some nfs clients
 (linux) to decide to to use a readdirplus call when directory contents are
 small vs a readdir call when the contents are large. We've found this
 particularly troublesome with Maildir style mail folders. The speedup by not
 using readdirplus is a factor of 100 in this particular situation. While we
 have a work around it would seem that this non-standard behavior might cause
 trouble for others and in other areas. Are there any suggestions for dealing
 with this difference and/or why zfs does not represent its directory sizes
 in a more traditional manner?

This has been discussed before regarding software that totally broke
(segfault) with ZFS's reporting.. The POSIX spec says nothing about what
size on directories should be, iirc.. Other (Novell I think?)
filesystems has used the same size reporting as well..

 From the linux kernel source.
 
 ./fs/nfs/inode.c:#define NFS_LIMIT_READDIRPLUS (8*PAGE_SIZE)

Basically, it's an assumption given informativen given in an unspecified
format.. Since there is nothing else to go on, it's not a totally crappy
assumption, but it doesn't work all the time either..

Hrm.

 ie. ZFS
 
 zfshost:~/testdir ls -1 | wc -l
 330
 zfshost:~/testdir stat .
   File: `.'
   Size: 332 Blocks: 486IO Block: 32768  directory
 Device: 29h/41d Inode: 540058  Links: 2
 Access: (0775/drwxrwxr-x)  Uid: ( 2891/   banal)   Gid: (  101/film)
 Access: 2008-11-05 18:40:16.0 -0800
 Modify: 2009-09-01 16:09:52.782674099 -0700
 Change: 2009-09-01 16:09:52.782674099 -0700
 
 ie. ext3
 
 ext3host:~/testdir ls -1 | wc -l
 330
 ext3host:~/testdir stat .
   File: `.'
   Size: 36864   Blocks: 72 IO Block: 4096   directory
 Device: 807h/2055d  Inode: 23887981Links: 2
 Access: (0775/drwxrwxr-x)  Uid: ( 2891/   banal)   Gid: (  101/film)
 Access: 2009-09-21 13:44:00.0 -0700
 Modify: 2009-09-21 13:44:31.0 -0700
 Change: 2009-09-21 13:44:31.0 -0700
 
 Thanks,
 Chris

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



/Tomas
-- 
Tomas Ögren, st...@acc.umu.se, http://www.acc.umu.se/~stric/
|- Student at Computing Science, University of Umeå
`- Sysadmin at {cs,acc}.umu.se
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS file disk usage

2009-09-21 Thread Richard Elling

On Sep 21, 2009, at 7:11 AM, Andrew Deason wrote:


On Sun, 20 Sep 2009 20:31:57 -0400
Richard Elling richard.ell...@gmail.com wrote:


If you are just building a cache, why not just make a file system and
put a reservation on it? Turn off auto snapshots and set other
features as per best practices for your workload? In other words,
treat it like we
treat dump space.

I think that we are getting caught up in trying to answer the  
question

you ask rather than solving the problem you have... perhaps because
we don't understand the problem.


Yes, possibly... some of these suggestions dont quite make a lot of
sense to me. We can't just make a filesystem and put a reservation on
it; we are just an application the administrator puts on a machine for
it to access AFS. So I'm not sure when you are imagining we do that;
when the client starts up? Or part of the installation procedure?
Requiring a separate filesystem seems unnecessarily restrictive.

And I still don't see how that helps. Making an fs with a reservation
would definitely limit us to the specified space, but we still can't  
get

an accurate picture of the current disk usage. I already mentioned why
using statvfs is not usable with that commit delay.


OK, so the problem you are trying to solve is how much stuff can I
place in the remaining free space?  I don't think this is knowable
for a dynamic file system like ZFS where metadata is dynamically
allocated.



But solving the general problem for me isn't necessary. If I could  
just
get a ballpark estimate of the max overhead for a file, I would be  
fine.

I haven't payed attention to it before, so I don't even have an
intuitive feel for what it is.


You don't know the max overhead for the file before it is allocated.
You could guess at a max of 3x size + at least three blocks.  Since
you can't control this, it seems like the worst case is when copies=3.
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] If you have ZFS in production, willing to share some details (with me)?

2009-09-21 Thread Gary Mills
On Fri, Sep 18, 2009 at 01:51:52PM -0400, Steffen Weiberle wrote:
 I am trying to compile some deployment scenarios of ZFS.
 
 # of systems

One, our e-mail server for the entire campus.

 amount of storage

2 TB that's 58% used.

 application profile(s)

This is our Cyrus IMAP spool.  In addition to user's e-mail folders
(directories) and messages (files), it contains global, per-folder,
and per-user databases.  The latter two types are quite small.

 type of workload (low, high; random, sequential; read-only, read-write, 
 write-only)

It's quite active.  Message files arrive randomly and are deleted
randomly.  As a result, files in a directory are not located in
proximity on the storage.  Individual users often read all of their
folders and messages in one IMAP session.  Databases are quite active.
Each incoming message adds a file to a directory and reads or updates
several databases.  Most IMAP I/O is done with mmap() rather than with
read()/write().  So far, IMAP peformance is adequate.  The backup,
done by EMC Networker, is very slow because it must read thousands of
small files in directory order.

 storage type(s)

We are using an Iscsi SAN with storage on a Netapp filer.  It exports
four 500-gb LUNs that are striped into one ZFS pool.  All disk
mangement is done on the Netapp.  We have had several disk failures
and replacements on the Netapp, with no effect on the e-mail server.

 industry

A University with 35,000 enabled e-mail accounts.

 whether it is private or I can share in a summary
 anything else that might be of interest

You are welcome to share this information.

-- 
-Gary Mills--Unix Group--Computer and Network Services-
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS file disk usage

2009-09-21 Thread Andrew Deason
On Mon, 21 Sep 2009 17:13:26 -0400
Richard Elling richard.ell...@gmail.com wrote:

 OK, so the problem you are trying to solve is how much stuff can I
 place in the remaining free space?  I don't think this is knowable
 for a dynamic file system like ZFS where metadata is dynamically
 allocated.

Yes. And I acknowledge that we can't know that precisely; I'm trying for
an estimate on the bound.

 You don't know the max overhead for the file before it is allocated.
 You could guess at a max of 3x size + at least three blocks.  Since
 you can't control this, it seems like the worst case is when copies=3.

Is that max with copies=3? Assume copies=1; what is it then?

-- 
Andrew Deason
adea...@sinenomine.net
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] possibilities of AFP ever making it into ZFS like NFS an

2009-09-21 Thread Ron Mexico
I was able to get Netatalk built on OpenSolaris for my ZFS NAS at home. 
Everything is running great so far, and I'm planning on using it on the 96TB 
NAS I'm building for my office. It would be nice to have this supported out of 
the box, but there are probably licensing issues involved.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re-import RAID-Z2 with faulted disk

2009-09-21 Thread Kyle J. Aleshire
On Mon, Sep 21, 2009 at 3:37 AM, casper@sun.com wrote:


 
 The disk has since been replaced, so now:
 k...@localhost:~$ pfexec zpool import
   pool: chronicle
 id: 11592382930413748377
  state: DEGRADED
 status: One or more devices contains corrupted data.
 action: The pool can be imported despite missing or damaged devices.  The
 fault tolerance of the pool may be compromised if imported.
see: http://www.sun.com/msg/ZFS-8000-4J
 config:
 
 chronicle   DEGRADED
   raidz2DEGRADED
 c9t2d0  FAULTED  corrupted data
 c9t1d0  ONLINE
 c9t0d0  ONLINE
 c9t4d0  ONLINE
 c9t5d0  ONLINE
 c9t3d0  ONLINE
 
 but pfexec zpool import -f chronicle still fails with the same message.
 

 That sounds like a bug; if ZFS can recover then zpool import should
 also work.

 So what was wrong with the broken disk?  Not just badly plugged in?

 Cas[er


The disk physically failed. When I powered the system on, it made
frighteningly loud clicking sounds and would not be acknowledged by any
system I was running, like it wasn't even there.

- Kyle
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re-import RAID-Z2 with faulted disk

2009-09-21 Thread Kyle J. Aleshire
I'm running vanilla 2009.06 since its release. I'll definitely give it a
shot with the Live CD.

Also I tried importing with only the five good disks physically attached and
get the same message.

- Kyle

On Mon, Sep 21, 2009 at 3:50 AM, Chris Murray
chrismurra...@googlemail.comwrote:

 That really sounds like a scenario that ZFS would be able to cope with.

 What operating system are you using now? I've had good results
 'fixing' pools with problems which were preventing import in the past by
 using the Opensolaris 2009.06 Live CD. It would depend what build you're on
 now as to whether that will yield any results, of course ...

 Also, is the import any more successful if the new, empty disk just isn't
 present at all, and c9t2d0 is connected to nothing?

 Chris

 2009/9/21 Kyle J. Aleshire kjalesh...@gmail.com

 Hi all, I have a RAID-Z2 setup with 6x 500Gb SATA disks. I exported the
 array to use under a different system but during or after the export one of
 the disks failed:

 k...@localhost:~$ pfexec zpool import
   pool: chronicle
 id: 11592382930413748377
  state: DEGRADED
 status: One or more devices are missing from the system.
 action: The pool can be imported despite missing or damaged devices.  The
 fault tolerance of the pool may be compromised if imported.
see: http://www.sun.com/msg/ZFS-8000-2Q
 config:

 chronicle   DEGRADED
   raidz2DEGRADED
 c9t2d0  UNAVAIL  cannot open
 c9t1d0  ONLINE
 c9t0d0  ONLINE
 c9t4d0  ONLINE
 c9t5d0  ONLINE
 c9t3d0  ONLINE

 I have no success trying to reimport the pool:
 k...@localhost:~$ pfexec zpool import -f chronicle
 cannot import 'chronicle': one or more devices is currently unavailable

 The disk has since been replaced, so now:
 k...@localhost:~$ pfexec zpool import
   pool: chronicle
 id: 11592382930413748377
  state: DEGRADED
 status: One or more devices contains corrupted data.
 action: The pool can be imported despite missing or damaged devices.  The
 fault tolerance of the pool may be compromised if imported.
see: http://www.sun.com/msg/ZFS-8000-4J
 config:

 chronicle   DEGRADED
   raidz2DEGRADED
 c9t2d0  FAULTED  corrupted data
 c9t1d0  ONLINE
 c9t0d0  ONLINE
 c9t4d0  ONLINE
 c9t5d0  ONLINE
 c9t3d0  ONLINE

 but pfexec zpool import -f chronicle still fails with the same message.

 I've Google'd this several times for a fix but to no avail. Any assistance
 is appreciated.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] How to recover from can't open objset, cannot iterate filesystems?

2009-09-21 Thread Albert Chin
Recently upgraded a system from b98 to b114. Also replaced two 400G
Seagate Barracudea 7200.8 SATA disks with two WD 750G RE3 SATA disks
from a 6-device raidz1 pool. Replacing the first 750G went ok. While
replacing the second 750G disk, I noticed CKSUM errors on the first
disk. Once the second disk was replaced, I halted the system, upgraded
to b114, and rebooted. Both b98 and b114 gave the errors:
  WARNING: can't open objset for tww/opt/dists/cd-8.1
  cannot iterate filesystems: I/O error

How do I recover from this?

# zpool status tww
  pool: tww
 state: ONLINE
status: One or more devices has experienced an error resulting in data
corruption.  Applications may be affected.
action: Restore the file in question if possible.  Otherwise restore the
entire pool from backup.
   see: http://www.sun.com/msg/ZFS-8000-8A
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
tww ONLINE   0 0 3
  raidz1ONLINE   0 012
c4t0d0  ONLINE   0 0 0
c4t1d0  ONLINE   0 0 0
c4t4d0  ONLINE   0 0 0
c4t5d0  ONLINE   0 0 0
c4t6d0  ONLINE   0 0 0
c4t7d0  ONLINE   0 0 0

errors: 855 data errors, use '-v' for a list

-- 
albert chin (ch...@thewrittenword.com)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss