Re: [zfs-discuss] resilver that never finishes

2010-09-19 Thread Tom Bird

On 18/09/10 15:25, George Wilson wrote:

Tom Bird wrote:



In my case, other than an hourly snapshot, the data is not
significantly changing.

It'd be nice to see a response other than you're doing it wrong,
rebuilding 5x the data on a drive relative to its capacity is clearly
erratic behaviour, I am curious as to what is actually happening.



It sounds like you're hitting '6891824 7410 NAS head continually
resilvering following HDD replacement'. If you stop taking and
destroying snapshots you should see the resilver finish.


George, I think you've won the prize.  I suspended the snapshots last 
night and this morning one pool had completed, one left to go.


Thanks,

Tom
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] resilver that never finishes

2010-09-18 Thread Tom Bird

On 18/09/10 09:02, Ian Collins wrote:

On 09/18/10 06:47 PM, Carsten Aulbert wrote:



Has someone an idea how it is possible to resilver 678G of data on a 500G
drive?


I see this all the time on a troublesome Thumper. I believe this happens
because the data in the pool is continuously changing.


In my case, other than an hourly snapshot, the data is not significantly 
changing.


It'd be nice to see a response other than you're doing it wrong, 
rebuilding 5x the data on a drive relative to its capacity is clearly 
erratic behaviour, I am curious as to what is actually happening.


All said and done though, we will have to live with snv_134's bugs from 
now on, or perhaps I could try Sol 10.


Tom
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] resilver that never finishes

2010-09-18 Thread Tom Bird

On 18/09/10 13:06, Edho P Arief wrote:

On Sat, Sep 18, 2010 at 7:01 PM, Tom Birdt...@marmot.org.uk  wrote:

All said and done though, we will have to live with snv_134's bugs from now
on, or perhaps I could try Sol 10.


or OpenIllumos. Or Nexenta. Or FreeBSD. Orinsert osol distro name.


... none of which will receive ZFS code updates unless Oracle deigns to 
bestow them upon the community, this or ZFS dev is taken over by said 
community, in which case we end up with diverging code bases that would 
be a sisyphean task to try and merge.


Tom
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] resilver that never finishes

2010-09-17 Thread Tom Bird

Morning,

c7t5000CCA221F4EC54d0 is a 2T disk, how can it resilver 5.63T of it?

This is actually an old capture of the status output, it got to nearly 
10T before deciding that there was an error and not completing, reseat 
disk and it's doing it all again.


It's happened on another pool as well, looking at a load av of around 40 
on the box currently, just sitting there churning disk IO.


OS is snv_134 on x86.

# zpool status -x
  pool: content4
 state: DEGRADED
status: One or more devices could not be opened.  Sufficient replicas 
exist for

the pool to continue functioning in a degraded state.
action: Attach the missing device and online it using 'zpool online'.
   see: http://www.sun.com/msg/ZFS-8000-2Q
 scrub: resilver in progress for 147h39m, 100.00% done, 0h0m to go
config:

NAME STATE READ WRITE CKSUM
content4 DEGRADED 0 0 0
  raidz2-0   DEGRADED 0 0 0
c7t5000CCA221DE1E1Dd0ONLINE   0 0 0
c7t5000CCA221DE17BFd0ONLINE   0 0 0
c7t5000CCA221DE2229d0ONLINE   0 0 0
replacing-3  DEGRADED 0 0 0
  c7t5000CCA221DE0CC7d0  UNAVAIL  0 0 0  cannot 
open
  c7t5000CCA221F4EC54d0  ONLINE   0 0 0  5.63T 
resilvered

c7t5000CCA221DE200Ad0ONLINE   0 0 0
c7t5000CCA221DDFE6Ed0ONLINE   0 0 0
c7t5000CCA221DE0103d0ONLINE   0 0 0
c7t5000CCA221DE00C9d0ONLINE   0 0 0
c7t5000CCA221DE0D2Dd0ONLINE   0 0 0
c7t5000CCA221DE189Cd0ONLINE   0 0 0
c7t5000CCA221DE18A7d0ONLINE   0 0 0
c7t5000CCA221DE2A47d0ONLINE   0 0 0
c7t5000CCA221DE1E48d0ONLINE   0 0 0
c7t5000CCA221DE18A1d0ONLINE   0 0 0
c7t5000CCA221DE18A2d0ONLINE   0 0 0
c7t5000CCA221DE2A3Ed0ONLINE   0 0 0
c7t5000CCA221DE2A42d0ONLINE   0 0 0
c7t5000CCA221DE2225d0UNAVAIL  0 0 0  cannot 
open

c7t5000CCA221DE28A3d0ONLINE   0 0 0
c7t5000CCA221DE2A46d0ONLINE   0 0 0
c7t5000CCA221DE0789d0ONLINE   0 0 0
c7t5000CCA221DE221Dd0ONLINE   0 0 0
c7t5000CCA221DE054Fd0ONLINE   0 0 0
c7t5000CCA221DE2EBEd0ONLINE   0 0 0

errors: No known data errors

--
Tom

// www.portfast.co.uk
// hosted services, domains, virtual machines, consultancy
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] pool won't import

2010-03-29 Thread Tom Bird

r...@cs6:~# zpool import
  pool: content3
id: 14184872052409584084
 state: FAULTED
status: The pool metadata is corrupted.
action: The pool cannot be imported due to damaged devices or data.
The pool may be active on another system, but can be imported using
the '-f' flag.
   see: http://www.sun.com/msg/ZFS-8000-72
config:

content3FAULTED  corrupted data
  c6t8d0ONLINE
r...@cs6:~# uname -a
SunOS cs6.kw.bbc.co.uk 5.11 snv_116 sun4v sparc SUNW,SPARC-Enterprise-T5220
r...@cs6:~# zpool import -f content3
cannot import 'content3': I/O error


Ideas, anyone?

--
Tom

// www.portfast.co.uk -- internet services and consultancy
// hosting from 1.65 per domain
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] demise of community edition

2010-01-31 Thread Tom Bird

Afternoon,

I note to my dismay that I can't get the community edition any more 
past snv_129, this version was closest to the normal way of doing things 
that I am used to with Solaris = 10, the standard OpenSolaris releases 
seem only to have this horrible Gnome based installer that gives you 
only one option - install everything.


Am I just doing it wrong or is there another way to get OpenSolaris 
installed in a sane manner other than just sticking with community 
edition at snv_129?


--
Tom

// www.portfast.co.uk -- internet services and consultancy
// hosting from 1.65 per domain
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] demise of community edition

2010-01-31 Thread Tom Bird

Richard Elling wrote:


It is not true that there is only a horrible Gnome based installer.  Try the 
Automated
Installation (AI) version instead of the LiveCD if you've used JumpStart 
previously.

But if you just want a text-based installer and AI is overkill, then b131 is available 
with the Text Installer Project.  Downloads available on:

http://www.genunix.org
http://hub.opensolaris.org/bin/view/Project+caiman/TextInstallerProject


Thanks, this looks useful.


Nothing is sane about Solaris 10 installer, good riddance :-)


It wasn't that bad! :)

PS sorry for this being a non specifically ZFS question, but ZFS is the 
reason I use opensolaris so there's a link in there somewhere.


Tom
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS tale of woe and fail

2009-08-20 Thread Tom Bird

Ross wrote:

Yup, that one was down to a known (and fixed) bug though, so it isn't
the normal story of ZFS problems.


Got a bug ID or anything for that, just out of interest?

As an update on my storage situation, I've got some JBODs now, see how 
that goes.


--
Tom

// www.portfast.co.uk -- internet services and consultancy
// hosting from 1.65 per domain
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS tale of woe and fail

2009-08-14 Thread Tom Bird

Victor Latushkin wrote:

This issue (and previous one reported by Tom) has got some publicity 
recently - see here


http://www.uknof.org.uk/uknof13/Bird-Redux.pdf

So i feel like i need to provide a little bit more information about the 
outcome (sorry that it is delayed and not as full as previous one).


Morning,

Right, the PDF on there doesn't really give the full story of the 
presentation, unfortunate as I see it seems to have got around a bit. 
In the actual presentation I wasn't perhaps as harsh as it seems on the 
slides!


First permanent error means that root block of the filesystem named 
'content' was corrupted (all copies), so it was not possible to open it 
and access any content of that filesystem.


Fortunately enough, there were not too much activity on the pool, so we 
decided to try previous states of the pool. I do not remember exact txg 
number we tried, but it was something like hundred txg back or so. We 
checked it with zdb and discovered that that state was more or less good 
- at least filesystem content was openable and it was possible to access 
its content, so we decided to reactivate that previous state. Pool 
imported fine and contents of 'content' was there. Subsequent scrub did 
find some errors but I do not remember exactly how much. Tom may have 
exact number.


I can't remember how many errors the check found, however all the data
copied off successfully, as far as we know.

--
Tom

// www.portfast.co.uk -- internet services and consultancy
// hosting from 1.65 per domain

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] poor performance / lots of disk activity with low throughput, seems prefetch related

2009-07-07 Thread Tom Bird

Hi guys,

I've been having trouble with my archival kit, in the performance 
department rather than data loss this time (phew!).


At the point when I took these stats where was about 250 mbit of traffic 
outbound on an ixgb NIC on the thing, also about 100 mbit of new stuff 
incoming.


As you can see, the amount of disk activity far exceeds the data being 
served.  I've disabled prefetching:


echo zfs_prefetch_disable/W0t1 | mdb -kw

and now activity looks more reasonable, the machine seems able to cope 
with the load, so potentially the prefetch is a bit too aggressive.


The machine itself has 8GB of RAM and is serving lots of different large 
files to users, so any cache will be fairly useless.



Stats: (with prefetch ENABLED)

With prefetch off most of the bandwidth figures are single / low double 
digits.


# uname -a
SunOS cs0 5.11 snv_116 sun4v sparc SUNW,Sun-Fire-T200

# zpool iostat 5
   capacity operationsbandwidth
pool used  avail   read  write   read  write
--  -  -  -  -  -  -
content037.2T  2.85T237  1  29.5M  9.54K
content128.5T  11.5T300 53  37.4M  5.82M
content219.0T  1.04T183  0  22.9M  4.27K
content33.85T  11.2T312 65  39.0M  7.57M
content43.85T  11.2T338 63  42.2M  7.47M
content53.84T  11.2T312 65  39.0M  7.50M
content619.0T  1.05T108  0  13.5M  4.12K
content714.0T   991G115  0  14.4M  3.39K
--  -  -  -  -  -  -
content037.2T  2.85T129  0  16.1M  0
content128.5T  11.5T564 57  70.5M  6.62M
content219.0T  1.04T 99  0  12.5M  0
content33.85T  11.2T361 79  45.1M  8.08M
content43.85T  11.2T995  0   124M  0
content53.84T  11.2T243 46  30.3M  2.64M
content619.0T  1.05T  0  0  0  0
content714.0T   991G203  0  25.4M  0
--  -  -  -  -  -  -
content037.2T  2.85T135  0  16.9M  0
content128.5T  11.5T578 57  72.3M  5.79M
content219.0T  1.04T 96  0  12.0M  0
content33.85T  11.2T377100  47.1M  9.01M
content43.85T  11.2T980  0   122M  0
content53.84T  11.2T216106  26.9M  10.4M
content619.0T  1.05T  0  0  0  0
content714.0T   991G200  0  25.1M  0
--  -  -  -  -  -  -
content037.2T  2.85T131  0  16.4M  0
content128.5T  11.5T578 50  72.3M  4.82M
content219.0T  1.04T 99  0  12.4M  0
content33.85T  11.2T377 63  47.1M  6.40M
content43.85T  11.2T944  0   118M  0
content53.84T  11.2T243 61  30.4M  5.97M
content619.0T  1.05T  0  0  0  0
content714.0T   991G209  0  26.1M  0
--  -  -  -  -  -  -
^C

--
Tom

// www.portfast.co.uk -- internet services and consultancy
// hosting from 1.65 per domain
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS tale of woe and fail

2009-01-19 Thread Tom Bird
Toby Thain wrote:
 On 18-Jan-09, at 6:12 PM, Nathan Kroenert wrote:
 
 Hey, Tom -

 Correct me if I'm wrong here, but it seems you are not allowing ZFS any
 sort of redundancy to manage.

Every other file system out there runs fine on a single LUN, when things
go wrong you have a fsck utility that patches it up and the world keeps
on turning.

I can't find anywhere that will sell me a 48 drive SATA JBOD with all
the drives presented on a single SAS channel, so running on a single
giant LUN is a real world scenario that ZFS should be able to cope with,
as this is how the hardware I am stuck with is arranged.

 Which is particularly catastrophic when one's 'content' is organized as
 a monolithic file, as it is here - unless, of course, you have some way
 of scavenging that file based on internal structure.

No, it's not a monolithic file, the point I was making there is that no
files are showing up.

 r...@cs4:~# find /content
 /content
 r...@cs4:~# (yes that really is it)

thanks
-- 
Tom

// www.portfast.co.uk -- internet services and consultancy
// hosting from 1.65 per domain
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS tale of woe and fail

2009-01-18 Thread Tom Bird
Morning,

For those of you who remember last time, this is a different Solaris,
different disk box and different host, but the epic nature of the fail
is similar.

The RAID box that is the 63T LUN has a hardware fault and has been
crashing, up to now the box and host got restarted and both came up
fine.  However, just now as I have got replacement hardware in position
and was ready to start copying, it went bang and my data has all gone.

Ideas?


r...@cs4:~# zpool list
NAME  SIZE   USED  AVAILCAP  HEALTH  ALTROOT
content  62.5T  59.9T  2.63T95%  ONLINE  -

r...@cs4:~# zpool status -v
  pool: content
 state: ONLINE
status: One or more devices has experienced an error resulting in data
corruption.  Applications may be affected.
action: Restore the file in question if possible.  Otherwise restore the
entire pool from backup.
   see: http://www.sun.com/msg/ZFS-8000-8A
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
content ONLINE   0 032
  c2t8d0ONLINE   0 032

errors: Permanent errors have been detected in the following files:

content:0x0
content:0x2c898

r...@cs4:~# find /content
/content
r...@cs4:~# (yes that really is it)

r...@cs4:~# uname -a
SunOS cs4.kw 5.11 snv_99 sun4v sparc SUNW,Sun-Fire-T200

from format:
   2. c2t8d0 IFT-S12S-G1033-363H-62.76TB
  /p...@7c0/p...@0/p...@8/LSILogic,s...@0/s...@8,0

Also, content does not show in df output.

thanks
-- 
Tom

// www.portfast.co.uk -- internet services and consultancy
// hosting from 1.65 per domain
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS tale of woe and fail

2009-01-18 Thread Tom Bird
Tim wrote:
 On Sun, Jan 18, 2009 at 8:02 AM, Tom Bird t...@marmot.org.uk
 mailto:t...@marmot.org.uk wrote:

 errors: Permanent errors have been detected in the following files:
 
content:0x0
content:0x2c898
 
 r...@cs4:~# find /content
 /content
 r...@cs4:~# (yes that really is it)


 Those are supposedly the two inodes that are corrupt.  The 0x0 is a bit
 scary... you should be able to find out what file(s) they're tied to (if
 any) with:
 
 find /content -inum 0
 find /content -inum 182424
 
 If you can live without those files, delete them, export the pool,
 re-import, and resilver, and you should be good to go.

Hi, well one of the problems is that find doesn't find anything as it is
not presenting any files, so I can't delete anything.

I've exported the pool but on reimport, I get the same error as I was
getting last time something popped:

r...@cs4:~# zpool import content
cannot open 'content': I/O error

Last time, Victor Latushkin fixed it by modifying the file system to
point to an older copy of the data.  I've not really been following the
list of late, any more sign of a fsck.zfs...?

thanks
-- 
Tom

// www.portfast.co.uk -- internet services and consultancy
// hosting from 1.65 per domain
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] SDXC and the future of ZFS

2009-01-11 Thread Tom Bird
Bob Friesenhahn wrote:
 On Sun, 11 Jan 2009, Eric D. Mudama wrote:
 My impression is not that other OS's aren't interested in ZFS, they
 are, it's that the licensing restrictions limit native support to
 Solaris, BSD, and OS-X.

 Perhaps the philosophical issues of the other OS's (i.e. Linux) are
 more significant than the actual licensing issues.  Many/most Linux
 users could legally use a native optimized kernel implementation of
 Sun ZFS if it was offered to them to do so.  GPLv2 only adds
 restrictions when copying binaries.  A pure source based distribution
 like Gentoo has hardly any issues at all.

Nobody in their right mind is using Gentoo.

If you want it in Linux then it has to be a proper GPL compliant effort.

I for one would like this to happen.

Tom
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] more ZFS recovery

2008-08-26 Thread Tom Bird
Victor Latushkin wrote:
 Hi Tom and all,

 [EMAIL PROTECTED]:~# uname -a
 SunOS cs3.kw 5.10 Generic_127127-11 sun4v sparc SUNW,Sun-Fire-T200
 
 Btw, have you considered opening support call for this issue?

As a follow up to the whole story, with the fantastic help of Victor,
the failed pool is now imported and functional thanks to the redundancy
in the meta data.

This does however highlight the need and practical application of a
fsck-like tool.  Fine to say that if ZFS can't guarantee my data then I 
should restore from backups so I know what I've got, but in the case of 
this 42T device that would take days.

Something to think about,

Tom

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] more ZFS recovery

2008-08-11 Thread Tom Bird
Victor Latushkin wrote:
 Hi Tom and all,
 
 Tom Bird wrote:
 Hi,

 Have a problem with a ZFS on a single device, this device is 48 1T SATA
 drives presented as a 42T LUN via hardware RAID 6 on a SAS bus which had
 a ZFS on it as a single device.

 There was a problem with the SAS bus which caused various errors
 including the inevitable kernel panic, the thing came back up with 3 out
 of 4 zfs mounted.
 
 It would be nice to see a panic stack.

I'm afraid I don't have that but now have an open connection to the
terminal server logging everything in case it should happen again.

 [EMAIL PROTECTED]:~# zpool import -f content
 cannot import 'content': I/O error
 
 As long as it does not panic and just returns I/O error which is rather
 generic, you may try to dig a little bit deeper with DTrace to have a
 chance to see where this I/O error is generated first, e.g. something
 like this with the attached dtrace script:
 
 dtrace -s /path/to/script -c zpool import -f content

dtrace output was 6MB, a bit rude to post to the list so I've uploaded
it here: http://picard.portfast.net/~tom/import.txt

 It is also interesting what impact SAS bus problem had on the storage
 controller. Btw, what is storage controller in question here?

The controller is an LSI Logic PCI express with 2 external SAS ports
which runs to an eonstor 2u 12 disk RAID chassis with 3 JBOD packs daisy
chained from that.  It seems I can't run the JBODs directly to the SAS
controller when using SATA drives (may be a different story with proper
SAS) and the RAID box has no JBOD mode so the redundancy has to stay in
the box and can't be transferred to ZFS.  The entire faulted array reads
cleanly at /dev/rdsk level into /dev/null.

There are 4 such arrays connected to the server via two SAS cards with a
ZFS on each one, the supplied internal SAS card and an ixgb NIC are the
only other cards installed.  System boots from the standard internal disks.

 [EMAIL PROTECTED]:~# uname -a
 SunOS cs3.kw 5.10 Generic_127127-11 sun4v sparc SUNW,Sun-Fire-T200
 
 Btw, have you considered opening support call for this issue?

Would have thought that unless they have a secret zfsck utility there's
probably not much they can do.  It's not a Sun disk array or Sun branded
SAS card.

thanks
-- 
Tom

// www.portfast.co.uk -- internet services and consultancy
// hosting from 1.65 per domain
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] more ZFS recovery

2008-08-06 Thread Tom Bird
Hi,

Have a problem with a ZFS on a single device, this device is 48 1T SATA
drives presented as a 42T LUN via hardware RAID 6 on a SAS bus which had
a ZFS on it as a single device.

There was a problem with the SAS bus which caused various errors
including the inevitable kernel panic, the thing came back up with 3 out
of 4 zfs mounted.

I've tried reading the partition table with format, works fine, also can
dd the first 100G from the device quite happily so the communication
issue appears resolved however the device just won't mount.  Googling
around I see that ZFS does have features designed to reduce the impact
of corruption at a particular point, multiple meta data copies and so
on, however commands to help me tidy up a zfs will only run once the
thing has been mounted.

Would be grateful for any ideas, relevant output here:

[EMAIL PROTECTED]:~# zpool import
  pool: content
id: 14205780542041739352
 state: FAULTED
status: The pool metadata is corrupted.
action: The pool cannot be imported due to damaged devices or data.
The pool may be active on on another system, but can be imported
using
the '-f' flag.
   see: http://www.sun.com/msg/ZFS-8000-72
config:

content FAULTED   corrupted data
  c2t9d0ONLINE

[EMAIL PROTECTED]:~# zpool import content
cannot import 'content': pool may be in use from other system
use '-f' to import anyway

[EMAIL PROTECTED]:~# zpool import -f content
cannot import 'content': I/O error

[EMAIL PROTECTED]:~# uname -a
SunOS cs3.kw 5.10 Generic_127127-11 sun4v sparc SUNW,Sun-Fire-T200


Thanks
-- 
Tom

// www.portfast.co.uk -- internet services and consultancy
// hosting from 1.65 per domain
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss