Re: [zfs-discuss] Fwd: zpool import despite missing log [PSARC/2010/292Self Review]

2010-12-23 Thread Dmitry Sorokin
Yesterday I was able to import zpool with missing log device using
zpool import -f -m myzpool command.
I had to boot from Oracle Solaris Express Live CD. Then I just did
zpool remove myzpool logdevice
That's it. Now I've got my pool back with all the data and with ONLINE
status.
I had my zpool (with 8 x 500 GB disks) sitting for almost 6 months
unavailable.
This was my Christmas present!


Best regards,
Dmitry

Office phone:   905.625.6471 ext. 104
Cell phone:   416.529.1627

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Jim Doyle
Sent: Sunday, August 01, 2010 1:40 AM
To: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] Fwd: zpool import despite missing log
[PSARC/2010/292Self Review]

A solution to this problem would be my early Christmas present! 

Here is how I lost access to an otherwise healthy mirrored pool two
months ago:

Box running snv_130 with two disks in a mirror and an iRAM
battery-backed ZIL device was shutdown orderly and powered down
normally.  While I was away on travel, the PSU in the PC died while in
its lowest-power standby state - this caused the Li battery in the iRAM
to discharge and all of the SLOG contents in the DRAM went poof.

Powered box back up... zpool import -f tank failed to bring the pool
back online.
After much research, I found the 'logfix' tool, got it compile on
another snv_122 box and followed the directions to synthesize a forged
log device header using the guid of the original device extracted from
vdev list.  This failed to work despite the binary tool running and some
inspection of the guids using zdb -l spoofed_new_logdev.

What's intrigueing is that zpool is not even properly reporting the
'missing device'.  See the output below from zpool, then zdb - notice
that zdb shows the remnants of a vdev for a log device but with guid = 0



# zpool import
  pool: tank
id: 6218740473633775200
 state: UNAVAIL
status: One or more devices are missing from the system.
action: The pool cannot be imported. Attach the missing
devices and try again.
   see: http://www.sun.com/msg/ZFS-8000-6X
config:
[b]
tank UNAVAIL  missing device
  mirror-0   ONLINE
c0t1d0   ONLINE
c0t2d0   ONLINE
[/b]
Additional devices are known to be part of this pool, though
their




 # zdb -e tank

Configuration for import:
vdev_children: 2
version: 22
pool_guid: 6218740473633775200
name: 'tank'
state: 0
hostid: 9271202
hostname: 'eon'
vdev_tree:
type: 'root'
id: 0
guid: 6218740473633775200
children[0]:
type: 'mirror'
id: 0
guid: 5245507142600321917
metaslab_array: 23
metaslab_shift: 33
ashift: 9
asize: 1000188936192
is_log: 0
children[0]:
type: 'disk'
id: 0
guid: 15634594394239615149
phys_path: '/p...@0,0/pci1458,b...@11/d...@2,0:a'
whole_disk: 1
DTL: 55
path: '/dev/dsk/c0t1d0s0'
devid:
'id1,s...@sata_st31000333as9te1jx8c/a'
children[1]:
type: 'disk'
id: 1
guid: 3144903288495510072
phys_path: '/p...@0,0/pci1458,b...@11/d...@1,0:a'
whole_disk: 1
DTL: 54
path: '/dev/dsk/c0t2d0s0'
devid:
'id1,s...@sata_st31000528as9vp2kwam/a'
[b]
children[1]:
type: 'missing'
id: 1
guid: 0
[/b]
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Fwd: zpool import despite missing log [PSARC/2010/292Self Review]

2010-07-30 Thread Dmitry Sorokin
Thanks for the update Robert.

 

Currently I have failed zpool with slog missing, which I was unable to
recover, although I was able to find out what the GUID was for the slog
device (below is the uotput of zpool import command).

I couldn't compile logfix binary either, so I ran out of any ideas of
how I can recover this zpool.

So for now it just sits there untouched.

This proposed improvement to zfs is definetely a hope for me.

When do you think it'll be implemented (roughly - this year, early next
year) and would I be able to import this pool at it's current
version 22 (snv_129)?

 

 

[r...@storage ~]# zpool import

pool: tank

id: 1346464136813319526

state: UNAVAIL

status: The pool was last accessed by another system.

action: The pool cannot be imported due to damaged devices or data.

   see: http://www.sun.com/msg/ZFS-8000-EY

config:

 

tank UNAVAIL  missing device

  raidz2-0   ONLINE

c4t0d0   ONLINE

c4t1d0   ONLINE

c4t2d0   ONLINE

c4t3d0   ONLINE

c4t4d0   ONLINE

c4t5d0   ONLINE

c4t6d0   ONLINE

c4t7d0   ONLINE

[r...@storage ~]#

 

Bets regards,

Dmitry

 

 

From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Robert
Milkowski
Sent: Wednesday, July 28, 2010 7:12 PM
To: ZFS Discussions
Subject: [zfs-discuss] Fwd: zpool import despite missing log
[PSARC/2010/292Self Review]

 


fyi

-- 
Robert Milkowski
http://milek.blogspot.com


 Original Message  

Subject: 

zpool import despite missing log [PSARC/2010/292 Self Review]

Date: 

Mon, 26 Jul 2010 08:38:22 -0600

From: 

Tim Haley tim.ha...@oracle.com mailto:tim.ha...@oracle.com 

To: 

psarc-...@sun.com

CC: 

zfs-t...@sun.com

 

I am sponsoring the following case for George Wilson.  Requested binding

is micro/patch.  Since this is a straight-forward addition of a command 
line option, I think itqualifies for self review.  If an ARC member 
disagrees, let me know and I'll convert to a fast-track.
 
Template Version: @(#)sac_nextcase 1.70 03/30/10 SMI
This information is Copyright (c) 2010, Oracle and/or its affiliates. 
All rights reserved.
1. Introduction
1.1. Project/Component Working Name:
 zpool import despite missing log
1.2. Name of Document Author/Supplier:
 Author:  George Wilson
1.3  Date of This Document:
26 July, 2010
 
4. Technical Description
 
OVERVIEW:
 
 ZFS maintains a GUID (global unique identifier) on each device
and
 the sum of all GUIDs of a pool are stored into the ZFS
uberblock.
 This sum is used to determine the availability of all vdevs
 within a pool when a pool is imported or opened.  Pools which
 contain a separate intent log device (e.g. a slog) will fail to
 import when that device is removed or is otherwise unavailable.
 This proposal aims to address this particular issue.
 
PROPOSED SOLUTION:
 
 This fast-track introduce a new command line flag to the
 'zpool import' sub-command.  This new option, '-m', allows
 pools to import even when a log device is missing.  The
contents
 of that log device are obviously discarded and the pool will
 operate as if the log device were offlined.
 
MANPAGE DIFFS:
 
   zpool import [-o mntopts] [-p property=value] ... [-d dir | -c
cachefile]
-  [-D] [-f] [-R root] [-n] [-F] -a
+  [-D] [-f] [-m] [-R root] [-n] [-F] -a
 
 
   zpool import [-o mntopts] [-o property=value] ... [-d dir | -c
cachefile]
-  [-D] [-f] [-R root] [-n] [-F] pool |id [newpool]
+  [-D] [-f] [-m] [-R root] [-n] [-F] pool |id [newpool]
 
   zpool import [-o mntopts] [ -o property=value] ... [-d dir |
- -c cachefile] [-D] [-f] [-n] [-F] [-R root] -a
+ -c cachefile] [-D] [-f] [-m] [-n] [-F] [-R root] -a
 
   Imports all  pools  found  in  the  search  directories.
   Identical to the previous command, except that all pools
 
+ -m
+
+Allows a pool to import when there is a missing log device
 
EXAMPLES:
 
1). Configuration with a single intent log device:
 
# zpool status tank
   pool: tank
state: ONLINE
 scan: none requested
 config:
 
 NAMESTATE READ WRITE CKSUM
 tankONLINE   0 0 0
   c7t0d0ONLINE   0 0 0
 logs
   c5t0d0ONLINE   0 0 0
 
errors: No known data errors
 
# zpool import tank
The devices below are missing, use '-m' to import the pool anyway:
 c5t0d0 [log]
 
cannot import 'tank': one or more devices is currently unavailable
 
# zpool import -m tank
# zpool status tank
   pool: tank
  state: DEGRADED
status: One or more devices could not be opened.  Sufficient replicas
exist for
 the pool to continue functioning in a degraded

Re: [zfs-discuss] ZIL SSD failed

2010-07-13 Thread Dmitry Sorokin
Hi Richard,

What happened is this SSD gave some IO errors and the pool become degraded, so 
after the machine got rebooted, I found that the pool became unavailable.
The SSD drive itself is toasted, as bios now reports it as 8 GB in size and 
name is Inuem SS E Cootmoader!, so all the partitions are gone and it is 
completely unusable at the moment.
I was able to find GUID of the slog from zpool.cache file that I backed up in 
January this year. However I was unable to compile logfix binary and the one 
provided by the author (compiled on snv_111) dumps core on snv_129 and snv_134.
Does anyone have logfix binary compiled for snv_129?

Thanks,
Dmitry


-Original Message-
From: Richard Elling [mailto:richard.ell...@gmail.com] 
Sent: Tuesday, July 13, 2010 5:12 PM
To: Dmitry Sorokin
Subject: Re: [zfs-discuss] ZIL SSD failed

Is the drive still there? If so, then try removing it or temporarily zeroing 
out the s0 label. If that allows zpool import -F to work, then please let me 
know and I'll add your post to a new bug report. I have seen something like 
this recently and have been trying to reproduce. 

 -- richard

On Jul 12, 2010, at 6:22 PM, Dmitry Sorokin dmitry.soro...@bmcorp.ca wrote:

 I have/had Intel M25-E 32GB SSD drive as ZIL/cache device (2 GB ZIL 
 slice0 and the rest is cache slice1)
 
 The SSD drive has failed and zpool is unavailable anymore.
 
 Is there any way to import the pool/recover data, even with some latest 
 transactions lost?
 
 I’ve tried zdb –e -bcsvL pool name but it didn’t work.
 
  
 
 Below are the details:
 
  
 
 [r...@storage ~]# uname -a
 
 SunOS storage 5.11 snv_129 i86pc i386 i86pc
 
  
 
 [r...@storage ~]# zpool import
 
   pool: neosys
 
 id: 1346464136813319526
 
 state: UNAVAIL
 
 status: The pool was last accessed by another system.
 
 action: The pool cannot be imported due to damaged devices or data.
 
see: http://www.sun.com/msg/ZFS-8000-EY
 
 config:
 
  
 
 neosys   UNAVAIL  missing device
 
   raidz2-0   ONLINE
 
 c4t0d0   ONLINE
 
 c4t1d0   ONLINE
 
 c4t2d0   ONLINE
 
 c4t3d0   ONLINE
 
 c4t4d0   ONLINE
 
 c4t5d0   ONLINE
 
 c4t6d0   ONLINE
 
 c4t7d0   ONLINE
 
  
 
 [r...@storage ~]# zdb -e neosys
 
  
 
 Configuration for import:
 
 vdev_children: 2
 
 version: 22
 
 pool_guid: 1346464136813319526
 
 name: 'neosys'
 
 state: 0
 
 hostid: 577477
 
 hostname: 'storage'
 
 vdev_tree:
 
 type: 'root'
 
 id: 0
 
 guid: 1346464136813319526
 
 children[0]:
 
 type: 'raidz'
 
 id: 0
 
 guid: 12671265726510370964
 
 nparity: 2
 
 metaslab_array: 25
 
 metaslab_shift: 35
 
 ashift: 9
 
 asize: 4000755744768
 
 is_log: 0
 
 children[0]:
 
 type: 'disk'
 
 id: 0
 
 guid: 10831801542309994254
 
 phys_path: 
 '/p...@0,0/pci10de,3...@f/pci1000,3...@0/s...@0,0:a'
 
 whole_disk: 1
 
 DTL: 3489
 
 path: '/dev/dsk/c4t0d0s0'
 
 devid: 'id1,s...@n5000cca32cc21642/a'
 
 children[1]:
 
 type: 'disk'
 
 id: 1
 
 guid: 39402223705908332
 
 phys_path: 
 '/p...@0,0/pci10de,3...@f/pci1000,3...@0/s...@1,0:a'
 
 whole_disk: 1
 
 DTL: 3488
 
 path: '/dev/dsk/c4t1d0s0'
 
 devid: 'id1,s...@n5000cca32cc1f061/a'
 
 children[2]:
 
 type: 'disk'
 
 id: 2
 
 guid: 5642566785254158202
 
 phys_path: 
 '/p...@0,0/pci10de,3...@f/pci1000,3...@0/s...@2,0:a'
 
 whole_disk: 1
 
 DTL: 3487
 
 path: '/dev/dsk/c4t2d0s0'
 
 devid: 'id1,s...@n5000cca32cc20121/a'
 
 children[3]:
 
 type: 'disk'
 
 id: 3
 
 guid: 5006664765902732873
 
 phys_path: 
 '/p...@0,0/pci10de,3...@f/pci1000,3...@0/s...@3,0:a'
 
 whole_disk: 1
 
 DTL: 3486
 
 path: '/dev/dsk/c4t3d0s0'
 
 devid: 'id1,s...@n5000cca32cf43053/a'
 
 children[4]:
 
 type: 'disk'
 
 id: 4
 
 guid: 106648579627377843
 
 phys_path: 
 '/p...@0,0/pci10de,3...@f/pci1000,3...@0/s...@4,0:a'
 
 whole_disk: 1
 
 DTL: 3485
 
 path: '/dev/dsk

[zfs-discuss] ZIL SSD failed

2010-07-12 Thread Dmitry Sorokin
...@storage ~]# zdb -e -bcsvL neosys

zdb: can't open 'neosys': No such device or address

 

Any help would be greatly appreciated.

 

Thanks,

Dmitry

 



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS dedup, snapshots and NFS

2010-05-26 Thread Dmitry Sorokin
Hi All,

 

I was just wandering if the issue that affects NFS availability when
deleting large snapshots on ZFS data sets with dedup enabled was fixed.

 

More on the issue here:

http://www.mail-archive.com/zfs-discuss@opensolaris.org/msg37288.html

and here:

http://opensolaris.org/jive/thread.jspa?messageID=474765
http://opensolaris.org/jive/thread.jspa?messageID=474765tstart=0
tstart=0

 

I have similar problem - when destroying snapshot (few gigabytes in size) on
ZFS data set with dedup enabled (was enabled prior and at a time when
snapshot was taken, but disabled later) NFS hangs and becomes unavailable to
remote hosts (same gigabit LAN).

I have a decent hardware with quad core Intel Nehalem CPU, 12 GB RAM and
Intel SSD drives as read and write cache for this zfs pool configured. I'm
running Nexenta Community edition 3.0.1.

I've posted to Nexenta community support forum but they referred me to
OpenSolaris community mailing list or as they've explained it's not Nexenta,
but rather OpenSolaris issue.

 

Thanks,

Dmitry

 

 



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS dedup, snapshots and NFS

2010-05-26 Thread Dmitry Sorokin
Thanks a lot for heads up  Garrett. I'll be watching for an update from
Nexenta then.

 

Dmitry

 

 

From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Garrett D'Amore
Sent: Wednesday, May 26, 2010 3:57 PM
To: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] ZFS dedup, snapshots and NFS

 

On 5/26/2010 11:47 AM, Dmitry Sorokin wrote: 

Hi All,

 

I was just wandering if the issue that affects NFS availability when
deleting large snapshots on ZFS data sets with dedup enabled was fixed.


There is a fix for this in b141 of the OpenSolaris source product.  We are
looking at including this fix in a forthcoming patch update/micro update to
the Nexenta product.  So stay tuned for more on that.

- Garrett




 

More on the issue here:

http://www.mail-archive.com/zfs-discuss@opensolaris.org/msg37288.html

and here:

http://opensolaris.org/jive/thread.jspa?messageID=474765
http://opensolaris.org/jive/thread.jspa?messageID=474765tstart=0
tstart=0

 

I have similar problem - when destroying snapshot (few gigabytes in size) on
ZFS data set with dedup enabled (was enabled prior and at a time when
snapshot was taken, but disabled later) NFS hangs and becomes unavailable to
remote hosts (same gigabit LAN).

I have a decent hardware with quad core Intel Nehalem CPU, 12 GB RAM and
Intel SSD drives as read and write cache for this zfs pool configured. I'm
running Nexenta Community edition 3.0.1.

I've posted to Nexenta community support forum but they referred me to
OpenSolaris community mailing list or as they've explained it's not Nexenta,
but rather OpenSolaris issue.

 

Thanks,

Dmitry

 

 

 
 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
  

 



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Which build is the most stable, mainly for NAS (zfs)?

2010-04-15 Thread Dmitry
Thanks for the tips, 
I tried EON, but it is too minimalistic, I plan to use this server for other 
(monitoring server and etc.) 
Nexenta is a strange hybrid, and use the not commercial version, without its 
ability, i don't know...

A napp-it i'll try for sure
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Which build is the most stable, mainly for NAS (zfs)?

2010-04-15 Thread Dmitry
free nexentastor community edition = commercial edition without support,
You are opened my eyes :) 
start to download, tomorrow will look
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Which build is the most stable, mainly for NAS (zfs)?

2010-04-14 Thread Dmitry
Which build is the most stable, mainly for NAS?
I plann NAS  zfs + CIFS,iSCSI

Thanks
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Which build is the most stable, mainly for NAS (zfs)?

2010-04-14 Thread Dmitry
Yesterday I received a victim.

SuperServer 5026T-3RF 19 2U, Intel X58, 1xCPU LGA1366 8xSAS/SATA hot-swap 
drive bays, 8 ports SAS LSI 1068E, 6 ports SATA-II Intel ICH10R, 2xGigabit 
Ethernet

and i have 2 ways Openfiler vs Opensolaris :)
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS poor performance on Areca 1231ML

2009-01-10 Thread Dmitry Razguliaev
At the time of writing that post, no, I didn't run zpool iostat -v 1. However, 
I run it after that. Results for operations of iostat command has changed from 
1 for every device in raidz to something in between 20 and 400 for raidz volume 
and from 3 to something in between 200 and 450 for a single device zfs volume, 
but the final result remained the same: single disk zfs volume is only about 
twice slower, then 9 disks raidz zfs volume, which seems to be very strange. My 
expectations are in a range of 6-7 times difference in performance.

Best Regards, Dmitry
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS poor performance on Areca 1231ML

2008-12-20 Thread Dmitry Razguliaev
Hi, I faced with a similar problem, like Ross, but still have not found a 
solution. I have raidz out of 9 sata disks connected to internal and 2 external 
sata controllers. Bonnie++ gives me the following results: 
nexenta,8G,104393,43,159637,30,57855,13,77677,38,56296,7,281.8,1,16,26450,99,+,+++,29909,93,24232,99,+,+++,13912,99
while running on a single disk it gives me the following:
nexenta,8G,54382,23,49141,8,25955,5,58696,27,60815,5,270.8,1,16,19793,76,+,+++,32637,99,22958,99,+,+++,10490,99
The performance difference of between those two seems to be too small. I 
checked zpool iostat -v during bonnie++ itelligent writing and it looks it, 
every time more or less like this:

   capacity operationsbandwidth
pool used  avail   read  write   read  write
--  -  -  -  -  -  -
iTank   7.20G  2.60T  12 13  1.52M  1.58M
  raidz17.20G  2.60T  12 13  1.52M  1.58M
c8d0-  -   1  1   172K   203K
c7d1-  -   1  1   170K   203K
c6t0d0  -  -  1  1   172K   203K
c8d1-  -   1  1   173K   203K
c9d0-  -   1  1   174K   203K
c10d0   -  -  1  1   174K   203K
c6t1d0  -  -  1  1   175K   203K
c5t0d0s0  -  -   1  1   176K   203K
c5t1d0s0  -  -   1  1   176K   203K

As far as I understand it, less each vdev executes only 1 i/o in a time. time. 
however, on a single device zpool iostat -v gives me the following:


   capacity operationsbandwidth
pool used  avail   read  write   read  write
--  -  -  -  -  -  -
rpool5.47G   181G  3  3   441K   434K
  c7d0s05.47G   181G  3  3   441K   434K
--  -  -  -  -  -  -

In this case this device performs 3 i/o in a time, which gives it much higher 
bandwidth per unit.

Is there any way to increase i/o counts for my iTank zpool?
I'm running OS-11.2008 on MSI P45 Diamond with 4G of memory

Best Regards, Dmitry
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Lack of physical memory evidences

2007-10-17 Thread Dmitry Degrave
In pre-ZFS era, we had observable parameters like scan rate and anonymous 
page-in/-out counters to discover situations when a system experiences a lack 
of physical memory. With ZFS, it's difficult to use mentioned parameters to 
figure out situations like that. Has someone any idea what we can use for the 
same purpose now ?

Thanks in advance,
Dmeetry
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS quota

2007-08-07 Thread Dmitry
Hello

Is there a way to limit size of filesystem not including snapshots?
Or even better size of data on filesystem regardless of compression.
If not is it planned?
It is hard to explain to user that it is normal that after deleting his files 
he did not receive more space. Even harder to ask to use not more than half of 
given quota.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Samba and ZFS ACL Question

2007-05-16 Thread Dmitry Tourkin
May be this link could help you? 

http://www.nabble.com/VFS-module-handling-ACL-on-ZFS-t3730348.html
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss