Re: [zfs-discuss] cannot destroy, volume is busy

2013-02-21 Thread John D Groenveld
In message <7f44e458-5d27-42b6-ac81-7f4ff61d6...@gmail.com>, Richard Elling wri
tes:
>The iSCSI service is not STMF. STMF will need to be disabled, or the =
>volume no longer
>used by STMF.
>
>iSCSI service is svc:/network/iscsi/target:default
>STMF service is svc:/system/stmf:default

Thank you for the gentle nudge with the clue stick, forgot the
process I used...
http://docs.oracle.com/cd/E26502_01/html/E29007/gaypf.html>

>One would think that fuser would work, but in my experience, fuser =
>rarely does
>what I expect.

fuser(1M) came up blank.

>If you suspect STMF, then try
>   stmfadm list-lu -v

Bingo!
Deleted the LU and destroyed the volume.

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] cannot destroy, volume is busy

2013-02-21 Thread John D Groenveld
# zfs list -t vol
NAME   USED  AVAIL  REFER  MOUNTPOINT
rpool/dump4.00G  99.9G  4.00G  -
rpool/foo128  66.2M   100G16K  -
rpool/swap4.00G  99.9G  4.00G  -

# zfs destroy rpool/foo128
cannot destroy 'rpool/foo128': volume is busy

I checked that the volume is not a dump or swap device
and that iSCSI is disabled.

On Solaris 11.1, how would I determine what's busying it?

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] HELP! RPool problem

2013-02-16 Thread John D Groenveld
In message 
, Karl Wagner writes:
>The SSD was the first boot drive, and every time it tried to boot it
>panicked and rebooted, ending up in a loop. I tried to change to the second
>rpool drive, but either I forgot to install grub on it or it has become
>corrupted (probably the first, I can be that stupid at times).
>
>Can anyone give me any advice on how to get this system back? Can I trick
>grub, installed on the SSD, to boot from the HDD's rpool mirror? Is
>something more sinister going on?

Remove the broken drive, boot installation media, import the
mirror drive.
If it imports, you will be able to installgrub(1M).

>By the way, whatever the error message is when booting, it disapears so
>quickly I can't read it, so I am only guessing that this is the reason.

Boot with kernel debugger so you can see the panic.

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] pool metadata has duplicate children

2013-01-08 Thread John Giannandrea

Gregg Wonderly  wrote:
> Have you tried importing the pool with that drive completely unplugged?  

Thanks for your reply.   I just tried that.  zpool import now says:

   pool: d
 id: 13178956075737687211
  state: FAULTED
 status: The pool metadata is corrupted.
 action: The pool cannot be imported due to damaged devices or data.
The pool may be active on another system, but can be imported using
the '-f' flag.
   see: http://illumos.org/msg/ZFS-8000-72
 config:

dFAULTED  corrupted data
  raidz1-0   FAULTED  corrupted data
da1  ONLINE
3419704811362497180  OFFLINE
da2  ONLINE
da3  ONLINE
da4  ONLINE

Notice that in the absence of the faulted da2 the OS has assigned da3 to da2 
etc.  I suspect this was part of the original problem in creating a label with 
two da2s

zdb still reports that the label has two da2 children:

vdev_tree:
type: 'raidz'
id: 0
guid: 11828532517066189487
nparity: 1
metaslab_array: 23
metaslab_shift: 36
ashift: 9
asize: 920660480
is_log: 0
children[0]:
type: 'disk'
id: 0
guid: 13697627234083630557
path: '/dev/da1'
whole_disk: 0
DTL: 78
children[1]:
type: 'disk'
id: 1
guid: 3419704811362497180
path: '/dev/da2'
whole_disk: 0
DTL: 71
offline: 1
children[2]:
type: 'disk'
id: 2
guid: 6790266178760006782
path: '/dev/da2'
whole_disk: 0
DTL: 77
children[3]:
type: 'disk'
id: 3
guid: 2883571222332651955
path: '/dev/da3'
whole_disk: 0
DTL: 76
children[4]:
type: 'disk'
id: 4
guid: 16640597255468768296
path: '/dev/da4'
whole_disk: 0
DTL: 75



> Which HBA are you using?  How many of these disks are on same or separate 
> HBAs?

all the disks are on the same HBA

twa0: <3ware 9000 series Storage Controller>
twa0: INFO: (0x15: 0x1300): Controller details:: Model 9500S-8, 8 ports, 
Firmware FE9X 2.08.00.006
da0 at twa0 bus 0 scbus0 target 0 lun 0
da1 at twa0 bus 0 scbus0 target 1 lun 0
da2 at twa0 bus 0 scbus0 target 2 lun 0
da3 at twa0 bus 0 scbus0 target 3 lun 0
da4 at twa0 bus 0 scbus0 target 4 lun 0

-jg

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] pool metadata has duplicate children

2013-01-08 Thread John Giannandrea

I seem to have managed to end up with a pool that is confused abut its children 
disks.  The pool is faulted with corrupt metadata:

  pool: d
 state: FAULTED
status: The pool metadata is corrupted and the pool cannot be opened.
action: Destroy and re-create the pool from
a backup source.
   see: http://illumos.org/msg/ZFS-8000-72
  scan: none requested
config:

NAME STATE READ WRITE CKSUM
dFAULTED  0 0 1
  raidz1-0   FAULTED  0 0 6
da1  ONLINE   0 0 0
3419704811362497180  OFFLINE  0 0 0  was /dev/da2
da3  ONLINE   0 0 0
da4  ONLINE   0 0 0
da5  ONLINE   0 0 0

But if I look at the labels on all the online disks I see this:

# zdb -ul /dev/da1 | egrep '(children|path)'
children[0]:
path: '/dev/da1'
children[1]:
path: '/dev/da2'
children[2]:
path: '/dev/da2'
children[3]:
path: '/dev/da3'
children[4]:
path: '/dev/da4'
...

But the offline disk (da2) shows the older correct label:

children[0]:
path: '/dev/da1'
children[1]:
path: '/dev/da2'
children[2]:
path: '/dev/da3'
children[3]:
path: '/dev/da4'
children[4]:
path: '/dev/da5'

zpool import -F doesnt help because none of the labels on the unfaulted disks 
seem to have the right label.  And unless I can import the pool I cant replace 
the bad drive.

Also zpool seems to really not want to import a raidz1 pool with one faulted 
drive even though that should be readable.  I have read about the undocumented 
-V option but dont know if that would help.

I got into this state when i noticed the pool was DEGRADED and was trying to 
replace the bad disk.   I am debugging it under FreeBSD 9.1 

Suggestions of things to try welcome, Im more interested in learning what went 
wrong than restoring the pool.  I dont think I should have been able to go from 
one offline drive to a unrecoverable pool this easily.

-jg

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] The format command crashes on 3TB disk but zpool create ok

2012-12-13 Thread John D Groenveld
# pstack core

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] dm-crypt + ZFS on Linux

2012-11-23 Thread John Baxter
Replacing the SANs is cost prohibitive.


On Fri, Nov 23, 2012 at 10:24 AM, Tim Cook  wrote:

>
>
> On Fri, Nov 23, 2012 at 9:49 AM, John Baxter wrote:
>
>>
>> We have the need to encypt our data, approximately 30TB on three ZFS
>> volumes under Solaris 10. The volumes currently reside on iscsi sans
>> connected via 10Gb/s ethernet. We have tested Solaris 11 with ZFS encrypted
>> volumes and found the performance to be very poor and have an open bug
>> report with Oracle.
>>
>> We are a Linux shop and since performance is so poor and still no
>> resolution, we are considering ZFS on Linux with dm-crypt.
>> I have read once or twice that if we implemented ZFS + dm-crypt we would
>> loose features, however which features are not specified.
>> We currently mirror the volumes across identical iscsi sans with ZFS and
>> we use hourly ZFS snapshots to update our DR site.
>>
>> Which features of ZFS are lost if we use dm-crypt? My guess would be they
>> are related to raidz but unsure.
>>
>>
>
> Why don't you just use a SAN that supports full drive encryption?  There
> should be basically 0 performance overhead.
>
> --Tim
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] dm-crypt + ZFS on Linux

2012-11-23 Thread John Baxter
After searching for dm-crypt and ZFS on Linux and finding too little
information, I shall ask here. Please keep in mind this in the context of
running this in a production environment.

We have the need to encypt our data, approximately 30TB on three ZFS
volumes under Solaris 10. The volumes currently reside on iscsi sans
connected via 10Gb/s ethernet. We have tested Solaris 11 with ZFS encrypted
volumes and found the performance to be very poor and have an open bug
report with Oracle.

We are a Linux shop and since performance is so poor and still no
resolution, we are considering ZFS on Linux with dm-crypt.
I have read once or twice that if we implemented ZFS + dm-crypt we would
loose features, however which features are not specified.
We currently mirror the volumes across identical iscsi sans with ZFS and we
use hourly ZFS snapshots to update our DR site.

Which features of ZFS are lost if we use dm-crypt? My guess would be they
are related to raidz but unsure.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] horrible slow pool

2012-10-11 Thread Carsten John
Hello everybody,

I just wanted to share my experience with a (partially) broken SSD that was in 
use in a ZIL mirror.

We experienced a dramatic performance problem with one of our zpools, serving 
home directories. Mainly NFS clients were affected. Our SunRay infrastructure 
came to a complete halt.

Finally we were able to identify one SSD as the root caus. The SSD was still 
working, but quite slow.

The issue didn't trigger ZFS to detect the disk as faulty. FMA didn't detect 
it, too.

We identified the broken disk by issuing "iostat -en'. After replacing the SSD, 
everything went back to normal.

To prevent outages like this in the future I hacked together a "quick and 
dirty" bash script to detect disks with a given rate of total errors. The 
script might be used in conjunction with nagios.

Perhaps it's of use for others sa well:

###
#!/bin/bash
# check disk in all pools for errors.
# partially failing (or slow) disks
# may result in horribly degradded 
# performance of zpools despite the fact
# the pool is still healthy

# exit codes
# 0 OK
# 1 WARNING
# 2 CRITICAL
# 3 UNKONOWN

OUTPUT=""
WARNING="0"
CRITICAL="0"
SOFTLIMIT="5"
HARDLIMIT="20"

LIST=$(zpool status | grep "c[1-9].*d0 " | awk '{print $1}')
for DISK in $LIST 
do  
ERROR=$(iostat -enr $DISK | cut -d "," -f 4 | grep "^[0-9]")
if [[ $ERROR -gt $SOFTLIMIT ]]
then
OUTPUT="$OUTPUT, $DISK:$ERROR"
WARNING="1"
fi
if [[ $ERROR -gt $HARDLIMIT ]]
then
OUTPUT="$OUTPUT, $DISK:$ERROR"
CRITICAL="1"
fi
done

if [[ $CRITICAL -gt 0 ]]
then
echo "CRITICAL: Disks with error count >= $HARDLIMIT found: $OUTPUT"
exit 2
fi
if [[ $WARNING -gt 0 ]]
then
echo "WARNING: Disks with error count >= $SOFTLIMIT found: $OUTPUT"
exit 1
fi

echo "OK: No significant disk errors found"
exit 0

###



cu

Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Sol11 time-slider / snapshot not starting [SOLVED]

2012-09-11 Thread Carsten John
-Original message-
To: zfs-discuss@opensolaris.org; 
From:   Carsten John 
Sent:   Tue 11-09-2012 13:08
Subject:[zfs-discuss] Sol11 time-slider / snapshot not starting [again]
> Hello everybody,
> 
> my time-slider service on a Sol11 machine died. I already 
> deinstalled/installed 
> the time-slider packeage, restarted manifest-import service etc., but no 
> success.
> 
> /var/svc/log/application-time-slider:default.log:


Finally I was able to fix it:

- uninstall time-slider
- restert manifest-import service
- install time-slider
- restart manifest-import service
- enable time-slider service
- enable snapshot services


I have no clue why it has to be done exact in this order, but finally I 
succeeded.


cu

Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Sol11 time-slider / snapshot not starting [again]

2012-09-11 Thread Carsten John
Hello everybody,

my time-slider service on a Sol11 machine died. I already deinstalled/installed 
the time-slider packeage, restarted manifest-import service etc., but no 
success.

/var/svc/log/application-time-slider:default.log:



--snip--


[ Sep 11 12:40:04 Enabled. ]
[ Sep 11 12:40:04 Executing start method ("/lib/svc/method/time-slider start"). 
]
Traceback (most recent call last):
  File "/usr/lib/time-sliderd", line 10, in 
main(abspath(__file__))
  File "/usr/lib/../share/time-slider/lib/time_slider/timesliderd.py", line 
941, in main
snapshot = SnapshotManager(systemBus)
  File "/usr/lib/../share/time-slider/lib/time_slider/timesliderd.py", line 83, 
in __init__
self.refresh()
  File "/usr/lib/../share/time-slider/lib/time_slider/timesliderd.py", line 
188, in refresh
self._rebuild_schedules()
  File "/usr/lib/../share/time-slider/lib/time_slider/timesliderd.py", line 
285, in _rebuild_schedules
"Details:\n" + str(message)
RuntimeError: Error reading SMF schedule instances
Details:
['/usr/bin/svcs', '-H', '-o', 'state', 
'svc:/system/filesystem/zfs/auto-snapshot:monthly'] failed with exit code 1
svcs: Pattern 'svc:/system/filesystem/zfs/auto-snapshot:monthly' doesn't match 
any instances

Time Slider failed to start: error 95
[ Sep 11 12:40:06 Method "start" exited with status 95. ]

--snip--





Any suggestions?


thx

Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can the ZFS "copies" attribute substitute HW disk redundancy?

2012-07-30 Thread John Martin

On 07/29/12 14:52, Bob Friesenhahn wrote:


My opinion is that complete hard drive failure and block-level media
failure are two totally different things.


That would depend on the recovery behavior of the drive for
block-level media failure.  A drive whose firmware does excessive
(reports of up to 2 minutes) retries of a bad sector may be
indistinguishable from a failed drive.  See previous discussions
of the firmware differences between desktop and enterprise drives.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Very poor small-block random write performance

2012-07-19 Thread John Martin

On 07/19/12 19:27, Jim Klimov wrote:


However, if the test file was written in 128K blocks and then
is rewritten with 64K blocks, then Bob's answer is probably
valid - the block would have to be re-read once for the first
rewrite of its half; it might be taken from cache for the
second half's rewrite (if that comes soon enough), and may be
spooled to disk as a couple of 64K blocks or one 128K block
(if both changes come soon after each other - within one TXG).


What are the values for zfs_txg_synctime_ms and zfs_txg_timeout
on this system (FreeBSD, IIRC)?

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] New fast hash algorithm - is it needed?

2012-07-10 Thread John Martin

On 07/10/12 19:56, Sašo Kiselkov wrote:

Hi guys,

I'm contemplating implementing a new fast hash algorithm in Illumos' ZFS
implementation to supplant the currently utilized sha256. On modern
64-bit CPUs SHA-256 is actually much slower than SHA-512 and indeed much
slower than many of the SHA-3 candidates, so I went out and did some
testing (details attached) on a possible new hash algorithm that might
improve on this situation.


Is the intent to store the 512 bit hash or truncate to 256 bit?

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] S11 zfs snapshot crash

2012-07-10 Thread John D Groenveld
I am toying with Phil Brown's zrep script.
Does anyone have an Oracle BugID for this crashdump?

#!/bin/ksh
srcfs=rpool/testvol
destfs=rpool/destvol
snap="${srcfs}@zrep_00"
zfs destroy -r $srcfs
zfs destroy -r $destfs
zfs create -V 100M $srcfs
zfs set foo:bar=foobar $srcfs
zfs create -o readonly=on $destfs
zfs snapshot $snap
zfs send -p $snap| zfs recv -vuF $destfs

# mdb unix.1 vmcore.1
> $c
zap_leaf_lookup+0x4d(ff01dd4e1888, ff01e198d500, ff000858e130)
fzap_lookup+0x9a(ff01e198d500, 1, 4b, ff000858e430, 0, 0)
zap_lookup_norm+0x131(ff01dd48c9c0, 1, f7a99e30, 1, 4b, 
ff000858e430)
zap_lookup+0x2d(ff01dd48c9c0, 1, f7a99e30, 1, 4b, ff000858e430)
zfs_get_mlslabel+0x56(ff01dd48c9c0, ff000858e430, 4b)
zfs_mount_label_policy+0x62(ff01dd68cb60, ff01ceede200)
zfs_mount+0x499(ff01dd68cb60, ff01e191b500, ff000858ee20, 
ff01dbfb23b8)
fsop_mount+0x22(ff01dd68cb60, ff01e191b500, ff000858ee20, 
ff01dbfb23b8)
domount+0xd33(0, ff000858ee20, ff01e191b500, ff01dbfb23b8, 
ff000858ee18)
mount+0xc0(ff01ce5cf3b8, ff000858ee98)
syscall_ap+0x92()
_sys_sysenter_post_swapgs+0x149()


# pkg info entire| grep Summary
   Summary: entire incorporation including Support Repository Update 
(Oracle Solaris 11 11/11 SRU 8.5).

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Sol11 missing snapshot facility [solved]

2012-07-06 Thread Carsten John
-Original message-
To: Carsten John ; 
CC: zfs-discuss@opensolaris.org; 
From:   Ian Collins 
Sent:   Thu 05-07-2012 21:40
Subject:Re: [zfs-discuss] Sol11 missing snapshot facility
> On 07/ 5/12 11:32 PM, Carsten John wrote:
> > -Original message-
> > To: Carsten John;
> > CC: zfs-discuss@opensolaris.org;
> > From:   Ian Collins
> > Sent:   Thu 05-07-2012 11:35
> > Subject:Re: [zfs-discuss] Sol11 missing snapshot facility
> >> On 07/ 5/12 09:25 PM, Carsten John wrote:
> >>
> >>> Hi Ian,
> >>>
> >>> yes, I already checked that:
> >>>
> >>> svcs -a | grep zfs
> >>> disabled   11:50:39 svc:/application/time-slider/plugin:zfs-send
> >>>
> >>> is the only service I get listed.
> >>>
> >> Odd.
> >>
> >> How did you install?
> >>
> >> Is the manifest there
> >> (/lib/svc/manifest/system/filesystem/auto-snapshot.xml)?
> >>
> > Hi Ian,
> >
> > I installed from CD/DVD, but it might have been in a rush, as I needed to 
> replace a broken machine as quick as possible.
> >
> > The manifest is there:
> >
> >
> > ls /lib/svc/manifest/system/filesystem/
> > .  .. auto-snapshot.xml  autofs.xml 
> local-fs.xml   minimal-fs.xml rmvolmgr.xml   root-fs.xml
> ufs-quota.xml  usr-fs.xml
> >
> 
> Running "svcadm restart manifest-import" should load it, or give you 
> some idea why it won't load.
> 
> -- 
> Ian.
> 
> 

Hi Ian,

it did the trick, but I had to uninstall/install the time-slider package.


thx for the help

Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Sol11 missing snapshot facility

2012-07-05 Thread Carsten John
-Original message-
To: Carsten John ; 
CC: zfs-discuss@opensolaris.org; 
From:   Ian Collins 
Sent:   Thu 05-07-2012 11:35
Subject:Re: [zfs-discuss] Sol11 missing snapshot facility
> On 07/ 5/12 09:25 PM, Carsten John wrote:
> 
> > Hi Ian,
> >
> > yes, I already checked that:
> >
> > svcs -a | grep zfs
> > disabled   11:50:39 svc:/application/time-slider/plugin:zfs-send
> >
> > is the only service I get listed.
> >
> Odd.
> 
> How did you install?
> 
> Is the manifest there 
> (/lib/svc/manifest/system/filesystem/auto-snapshot.xml)?
> 
> -- 
> Ian.
> 
> 

Hi Ian,

I installed from CD/DVD, but it might have been in a rush, as I needed to 
replace a broken machine as quick as possible.

The manifest is there:


ls /lib/svc/manifest/system/filesystem/
.  .. auto-snapshot.xml  autofs.xml 
local-fs.xml   minimal-fs.xml rmvolmgr.xml   root-fs.xml
ufs-quota.xml  usr-fs.xml



thx


Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Sol11 missing snapshot facility

2012-07-05 Thread Carsten John
-Original message-
To: Carsten John ; 
CC: zfs-discuss@opensolaris.org; 
From:   Ian Collins 
Sent:   Thu 05-07-2012 09:59
Subject:Re: [zfs-discuss] Sol11 missing snapshot facility
> On 07/ 5/12 06:52 PM, Carsten John wrote:
> > Hello everybody,
> >
> >
> > for some reason I can not find the zfs-autosnapshot service facility any 
> more. I already reinstalles time-slider, but it refuses to start:
> >
> >
> > RuntimeError: Error reading SMF schedule instances
> > Details:
> > ['/usr/bin/svcs', '-H', '-o', 'state', 
> 'svc:/system/filesystem/zfs/auto-snapshot:monthly'] failed with exit code 1
> > svcs: Pattern 'svc:/system/filesystem/zfs/auto-snapshot:monthly' doesn't 
> match any instances
> >
> Have you looked with svcs -a?
> 
> # svcs -a | grep zfs
> disabled   Jul_02   svc:/system/filesystem/zfs/auto-snapshot:daily
> disabled   Jul_02   svc:/system/filesystem/zfs/auto-snapshot:frequent
> disabled   Jul_02   svc:/system/filesystem/zfs/auto-snapshot:hourly
> disabled   Jul_02   svc:/system/filesystem/zfs/auto-snapshot:monthly
> disabled   Jul_02   svc:/system/filesystem/zfs/auto-snapshot:weekly
> disabled   Jul_02   svc:/application/time-slider/plugin:zfs-send
> 
> -- 
> Ian.
> 
> 


Hi Ian,

yes, I already checked that:

svcs -a | grep zfs
disabled   11:50:39 svc:/application/time-slider/plugin:zfs-send

is the only service I get listed.


thx

Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Sol11 missing snapshot facility

2012-07-04 Thread Carsten John
Hello everybody,


for some reason I can not find the zfs-autosnapshot service facility any more. 
I already reinstalles time-slider, but it refuses to start:


RuntimeError: Error reading SMF schedule instances
Details:
['/usr/bin/svcs', '-H', '-o', 'state', 
'svc:/system/filesystem/zfs/auto-snapshot:monthly'] failed with exit code 1
svcs: Pattern 'svc:/system/filesystem/zfs/auto-snapshot:monthly' doesn't match 
any instances



did anybody know a way to get the services back again?


thx


Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Interaction between ZFS intent log and mmap'd files

2012-07-04 Thread John Martin

On 07/04/12 16:47, Nico Williams wrote:


I don't see that the munmap definition assures that anything is written to
"disk".  The system is free to buffer the data in RAM as long as it likes
without writing anything at all.


Oddly enough the manpages at the Open Group don't make this clear.  So
I think it may well be advisable to use msync(3C) before munmap() on
MAP_SHARED mappings.  However, I think all implementors should, and
probably all do (Linux even documents that it does) have an implied
msync(2) when doing a munmap(2).  I really makes no sense at all to
have munmap(2) not imply msync(3C).


This assumes msync() has the behavior you expect.  See:

  http://pubs.opengroup.org/onlinepubs/009695399/functions/msync.html

In particular, the paragraph starting with "For mappings to files, ...".
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snapshots slow on sol11? SOLVED

2012-06-29 Thread Carsten John
-Original message-
To: Jim Klimov ; 
CC: ZFS Discussions ; 
From:   Carsten John 
Sent:   Wed 27-06-2012 08:48
Subject:Re: [zfs-discuss] snapshots slow on sol11?
> -Original message-
> CC:   ZFS Discussions ; 
> From: Jim Klimov 
> Sent: Tue 26-06-2012 22:34
> Subject:  Re: [zfs-discuss] snapshots slow on sol11?
> > 2012-06-26 23:57, Carsten John wrote:
> > > Hello everybody,
> > >
> > > I recently migrated a file server (NFS & Samba) from OpenSolaris (Build 
> 111) 
> > to Sol11.
> >  > (After?) the move we are facing random (or random looking) outages of 
> > our Samba...
> > 
> > As for the timeouts, check if your tuning (i.e. the migrated files
> > like /etc/system) don't enforce long TXG syncs (default was 30sec)
> > or something like that.
> > 
> > Find some DTrace scripts to see if ZIL is intensively used during
> > these user-profile writes, and if these writes are synchronous -
> > maybe an SSD/DDR logging device might be useful for this scenario?
> > 
> > Regarding the zfs-auto-snapshot, it is possible to install the old
> > scripted package from OpenSolaris onto Solaris 10 at least; I did
> > not have much experience with newer releases yet (timesliderd) so
> > can't help better.
> > 
> > HTH,
> > //Jim Klimov
> > 
> > ___
> > zfs-discuss mailing list
> > zfs-discuss@opensolaris.org
> > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> > 
> 
> 
> Hi everybody,
> 
> in the meantime I was able to eliminate the snapshots. I disabled snapshot, 
> but 
> the issue still persists. I will now check Jim's suggestions.
> 
> thx so far
> 
> 
> Carsten
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> 


We finally found the reason (thanks, wireshark). It was, in fact, not zfs nor 
solaris related.

Due to a problem with a dhcp3-relay, the clients got an initial DHCP IP by 
DHCPDISCOVER, but couldn't renew their IP via DHCPREQUEST when the maximum 
lease time was reached.

This resulted in a short loss of connectivity, until the client got a new IP 
(actually the same as before) via DHCPDISCOVER.

The problem only showed up, if the client tried to write to the fileserver in 
the period of time when the IP was lost (usually when firefox or thunderbird 
tried to write to their disk cache).


thanks for the suggestions


Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snapshots slow on sol11?

2012-06-26 Thread Carsten John
-Original message-
CC: ZFS Discussions ; 
From:   Jim Klimov 
Sent:   Tue 26-06-2012 22:34
Subject:Re: [zfs-discuss] snapshots slow on sol11?
> 2012-06-26 23:57, Carsten John wrote:
> > Hello everybody,
> >
> > I recently migrated a file server (NFS & Samba) from OpenSolaris (Build 
> > 111) 
> to Sol11.
>  > (After?) the move we are facing random (or random looking) outages of 
> our Samba...
> 
> As for the timeouts, check if your tuning (i.e. the migrated files
> like /etc/system) don't enforce long TXG syncs (default was 30sec)
> or something like that.
> 
> Find some DTrace scripts to see if ZIL is intensively used during
> these user-profile writes, and if these writes are synchronous -
> maybe an SSD/DDR logging device might be useful for this scenario?
> 
> Regarding the zfs-auto-snapshot, it is possible to install the old
> scripted package from OpenSolaris onto Solaris 10 at least; I did
> not have much experience with newer releases yet (timesliderd) so
> can't help better.
> 
> HTH,
> //Jim Klimov
> 
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> 


Hi everybody,

in the meantime I was able to eliminate the snapshots. I disabled snapshot, but 
the issue still persists. I will now check Jim's suggestions.

thx so far


Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] snapshots slow on sol11?

2012-06-26 Thread Carsten John
Hello everybody,

I recently migrated a file server (NFS & Samba) from OpenSolaris (Build 111) to 
Sol11. This the move we are facing random (or random looking) outages of our 
Samba. As we have moved several folders (like Desktop and ApplicationData) out 
of the usual profile to a folder inside the users home share, the setup is 
sensible about timeouts. From time to time users are getting the infamous 
Windows "Delayed Write Failure".

After checking nearly every parameter that came to my mind in the last days, 
the zfs-auto-snapshot mechanism inside Solaris 11 came to my attention. We had 
hourly and daily snapshot enabled and discovered that the snapshots are not 
rotated as expected.

As there were known issues (if I remember correctly) with timesliderd in 
OpenIndiana and we had the old zfs-auto-snap mechanism (without timesliderd) 
running without any problems before the update, I'm wondering if the are any 
known (performance) issues with the stuff in Solaris 11.



thx


Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Migrating 512 byte block zfs root pool to 4k disks

2012-06-16 Thread John Martin

On 06/16/12 12:23, Richard Elling wrote:

On Jun 15, 2012, at 7:37 AM, Hung-Sheng Tsao Ph.D. wrote:


by the way
when you format start with cylinder 1 donot use 0


There is no requirement for skipping cylinder 0 for root on Solaris, and there
never has been.


Maybe not for core Solaris, but it is still wise advice
if you plan to use Oracle ASM.  See section 3.3.1.4, 2c:


http://docs.oracle.com/cd/E11882_01/install.112/e24616/storage.htm#CACHGBAH

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Migrating 512 byte block zfs root pool to 4k disks

2012-06-15 Thread John Martin

On 06/15/12 15:52, Cindy Swearingen wrote:


Its important to identify your OS release to determine if
booting from a 4k disk is supported.


In addition, whether the drive is really 4096p or 512e/4096p.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] (fwd) Re: ZFS NFS service hanging on Sunday morning

2012-06-14 Thread John D Groenveld
In message <201206141413.q5eedvzq017...@mklab.ph.rhul.ac.uk>, tpc...@mklab.ph.r
hul.ac.uk writes:
>Memory: 2048M phys mem, 32M free mem, 16G total swap, 16G free swap


My WAG is that your "zpool history" is hanging due to lack of
RAM.

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Is there an actual newsgroup for zfs-discuss?

2012-06-11 Thread John D Groenveld
In message <008c01cd4812$7399c180$5acd4480$@net>, David Combs writes:
>Actual newsgroup for zfs-discuss?

Did you try Gmane's interface?
http://groups.google.com/groups?selm=jo43q0%24no50%241%40tr22n12.aset.psu.edu>

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Spare drive inherited cksum errors?

2012-05-29 Thread John D Groenveld
In message <4fc509e8.8080...@jvm.de>, Stephan Budach writes:
>If now I'd only knew how to get the actual S11 release level of my box. 
>Neither uname -a nor cat /etc/release does give me a clue, since they 
>display all the same data when run on different hosts that are on 
>different updates.

$ pkg info entire

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Advanced Format HDD's - are we there yet? (or - how to buy a drive that won't be teh sux0rs on zfs)

2012-05-29 Thread John Martin

On 05/29/12 07:26, bofh wrote:


ashift:9  is that standard?


Depends on what the drive reports as physical sector size.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Advanced Format HDD's - are we there yet? (or - how to buy a drive that won't be teh sux0rs on zfs)

2012-05-29 Thread John Martin

On 05/29/12 08:35, Nathan Kroenert wrote:

Hi John,

Actually, last time I tried the whole AF (4k) thing, it's performance
was worse than woeful.

But admittedly, that was a little while ago.

The drives were the seagate green barracuda IIRC, and performance for
just about everything was 20MB/s per spindle or worse, when it should
have been closer to 100MB/s when streaming. Things were worse still when
doing random...

I'm actually looking to put in something larger than the 3*2TB drives
(triple mirror for read perf) this pool has in it - preferably 3 * 4TB
drives. (I don't want to put in more spindles - just replace the current
ones...)

I might just have to bite the bullet and try something with current SW. :).



Raw read from one of the mirrors:

#  timex dd if=/dev/rdsk/c0t2d0s2 of=/dev/null bs=1024000 count=1
1+0 records in
1+0 records out

real  49.26
user   0.01
sys0.27


filebench filemicro_seqread reports an impossibly high number (4GB/s)
so the ARC is likely handling all reads.

The label on the boxes I bought say:

  1TB 32MB INTERNAL KIT 7200
  ST310005N1A1AS-RK
  S/N: ...
  PN:9BX1A8-573

The drives in the box were really
ST1000DM003-9YN162 with 64MB of cache.
I have multiple pools on each disk so the
cache should be disabled.  The drive reports
512 byte logical sectors and 4096 physical sectors.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Advanced Format HDD's - are we there yet? (or - how to buy a drive that won't be teh sux0rs on zfs)

2012-05-29 Thread John Martin

On 05/28/12 08:48, Nathan Kroenert wrote:


Looking to get some larger drives for one of my boxes. It runs
exclusively ZFS and has been using Seagate 2TB units up until now (which
are 512 byte sector).

Anyone offer up suggestions of either 3 or preferably 4TB drives that
actually work well with ZFS out of the box? (And not perform like
rubbish)...

I'm using Oracle Solaris 11 , and would prefer not to have to use a
hacked up zpool to create something with ashift=12.


Are you replacing a failed drive or creating a new pool?

I had a drive in a mirrored pool recently fail.  Both
drives were 1TB Seagate ST310005N1A1AS-RK with 512 byte sectors.
All the 1TB Seagate boxed drives I could find with the same
part number on the box (with factory seals in place)
were really ST1000DM003-9YN1 with 512e/4196p.  Just being
cautious, I ended up migrating the pools over to a pair
of the new drives.  The pools were created with ashift=12
automatically:

  $ zdb -C | grep ashift
  ashift: 12
  ashift: 12
  ashift: 12

Resilvering the three pools concurrently went fairly quickly:

  $ zpool status
scan: resilvered 223G in 2h14m with 0 errors on Tue May 22 21:02:32 
2012
scan: resilvered 145G in 4h13m with 0 errors on Tue May 22 23:02:38 
2012
scan: resilvered 153G in 3h44m with 0 errors on Tue May 22 22:30:51 
2012


What performance problem do you expect?
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] kernel panic during zfs import [UPDATE]

2012-04-17 Thread Carsten John
Hello everybody,

just to let you know what happened in the meantime:

I was able to open a Service Request at Oracle.

The issue is a known bug (Bug 6742788 : assertion panic at: zfs:zap_deref_leaf)

The bug has bin fixed (according to Oracle support) since build 164, but there 
is no fix for Solaris 11 available so far (will be fixed in S11U7?).

There is a workaround available that works (partly), but my system crashed 
again when trying to rebuild the offending zfs within the affected zpool.

At the moment I'm waiting for a so called "interim diagnostic relief" patch


cu

Carsten

-- 
Max Planck Institut fuer marine Mikrobiologie
- Network Administration -
Celsiustr. 1
D-28359 Bremen
Tel.: +49 421 2028568
Fax.: +49 421 2028565
PGP public key:http://www.mpi-bremen.de/Carsten_John.html
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] kernel panic during zfs import [ORACLE should notice this]

2012-03-30 Thread Carsten John
-Original message-
To: zfs-discuss@opensolaris.org; 
From:   John D Groenveld 
Sent:   Fri 30-03-2012 21:47
Subject:Re: [zfs-discuss] kernel panic during zfs import [ORACLE should 
notice this]
> In message <4f735451.2020...@oracle.com>, Deepak Honnalli writes:
> > Thanks for your reply. I would love to take a look at the core
> > file. If there is a way this can somehow be transferred to
> > the internal cores server, I can work on the bug.
> >
> > I am not sure about the modalities of transferring the core
> > file though. I will ask around and see if I can help you here.
> 
> How to Upload Data to Oracle Such as Explorer and Core Files [ID 1020199.1]
> 
> John
> groenv...@acm.org
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> 

Hi John,

in the meantime I managed to open a service request at Oracle. There is a 
webportal https://supportfiles.sun.com. There you can upload the files...


cu

Carsten

-- 
Max Planck Institut fuer marine Mikrobiologie
- Network Administration -
Celsiustr. 1
D-28359 Bremen
Tel.: +49 421 2028568
Fax.: +49 421 2028565
PGP public key:http://www.mpi-bremen.de/Carsten_John.html
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] kernel panic during zfs import [ORACLE should notice this]

2012-03-30 Thread John D Groenveld
In message <4f735451.2020...@oracle.com>, Deepak Honnalli writes:
> Thanks for your reply. I would love to take a look at the core
> file. If there is a way this can somehow be transferred to
> the internal cores server, I can work on the bug.
>
> I am not sure about the modalities of transferring the core
> file though. I will ask around and see if I can help you here.

How to Upload Data to Oracle Such as Explorer and Core Files [ID 1020199.1]

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Good tower server for around 1,250 USD?

2012-03-30 Thread John D Groenveld
In message <4f7571de.7080...@netdemons.com>, Erik Trimble writes:
>Oracle (that is, if Oracle hasn't completely stopped selling support 
>contracts for Solaris for non-Oracle hardware already).

Still available on Oracle Store:
https://shop.oracle.com/pls/ostore/f?p=dstore:product:2091882785479247::NO:RP,6:P6_LPI:27242443094470222098916>

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Puzzling problem with zfs receive exit status

2012-03-29 Thread Carsten John
-Original message-
To: zfs-discuss@opensolaris.org; 
From:   Borja Marcos 
Sent:   Thu 29-03-2012 11:49
Subject:[zfs-discuss] Puzzling problem with zfs receive exit status
> 
> Hello,
> 
> I hope someone has an idea. 
> 
> I have a replication program that copies a dataset from one server to another 
> one. The replication mechanism is the obvious one, of course:
> 
>  zfs send -Ri from snapshot(n-1) snapshot(n) > file
> scp file remote machine (I do it this way instead of using a pipeline so that 
> a 
> network error won't interrupt a receive data stream)
> and on the remote machine,
> zfs receive -Fd pool
> 
> It's been working perfectly for months, no issues. However, yesterday we 
> began 
> to see something weird: the zfs receive being executed on the remote machine 
> is 
> exiting with an exit status of 1, even though the replication is finished, 
> and 
> I see the copied snapshots on the remote machine. 
> 
> Any ideas? It's really puzzling. It seems that the replication is working (a 
> zfs list -t snapshot shows the new snapshots correctly applied to the 
> dataset) 
> but I'm afraid there's some kind of corruption.
> 
> The OS is Solaris, SunOS  5.10 Generic_141445-09 i86pc i386 i86pc.
> 
> Any ideas?
> 
> 
> 
> Thanks in advance,
> 
> 
> 
> 
> 
> Borja.
> 
> 
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> 


Hi Borja,


did you try to check the snapshot file with zstreamdump? It will validate the 
checksums.

Perhaps the information here

http://blog.richardelling.com/2009/10/check-integrity-of-zfs-send-streams.html

might be useful for you.



Carsten

-- 
Max Planck Institut fuer marine Mikrobiologie
- Network Administration -
Celsiustr. 1
D-28359 Bremen
Tel.: +49 421 2028568
Fax.: +49 421 2028565
PGP public key:http://www.mpi-bremen.de/Carsten_John.html
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] kernel panic during zfs import [ORACLE should notice this]

2012-03-28 Thread John D Groenveld
In message , =?utf-
8?Q?Carsten_John?= writes:
>I just spent about an hour (or two) trying to file a bug report regarding the 
>issue without success.
>
>Seems to me, that I'm too stupid to use this "MyOracleSupport" portal.
>
>So, as I'm getting paid for keeping systems running and not clicking through f
>lash overloaded support portals searching for CSIs, I'm giving the relevant in
>formation to the list now.

If the Flash interface is broken, try the non-Flash MOS site:
http://SupportHTML.Oracle.COM/>

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] kernel panic during zfs import [ORACLE should notice this]

2012-03-28 Thread Carsten John
-Original message-
To: zfs-discuss@opensolaris.org; 
From:   Deepak Honnalli 
Sent:   Wed 28-03-2012 09:12
Subject:Re: [zfs-discuss] kernel panic during zfs import
> Hi Carsten,
> 
>  This was supposed to be fixed in build 164 of Nevada (6742788). If 
> you are still seeing this
>  issue in S11, I think you should raise a bug with relevant details. 
> As Paul has suggested,
>  this could also be due to incomplete snapshot.
> 
>  I have seen interrupted zfs recv's causing weired bugs.
> 
> Thanks,
> Deepak.


Hi Deepak,

I just spent about an hour (or two) trying to file a bug report regarding the 
issue without success.

Seems to me, that I'm too stupid to use this "MyOracleSupport" portal.

So, as I'm getting paid for keeping systems running and not clicking through 
flash overloaded support portals searching for CSIs, I'm giving the relevant 
information to the list now.

Perhaps, someone at Oracle, reading the list, is able to file a bug report, or 
contact me off list.



Background:

Machine A
- Sun X4270 
- Opensolaris Build 111b
- zpool version 14
- primary file server
- sending snapshots via zfs send
- direct attached Sun J4400 SAS JBODs with totally 40 TB storage

Machine B
- Sun X4270
- Solaris 11
- zpool version 33
- mirror server
- receiving snapshots via zfs receive
- FC attached Storagetek FLX280 storage 


Incident:

After a zfs send/receive run machine B had a hanging zfs receive process. To 
get rid of the process, I rebooted the machine. During reboot the kernel 
panics, resulting in a reboot loop.

To bring up the system, I rebooted single user, removed /etc/zfs/zpool.cache 
and rebooted again.

The damaged pool can imported readonly, giving a warning:

   $>zpool import -o readonly=on san_pool
   cannot set property for 'san_pool/home/someuser': dataset is read-only
   cannot set property for 'san_pool/home/someotheruser': dataset is read-only

The ZFS debugger zdb does not give any additional information:

   $>zdb -d -e san_pool
   Dataset san_pool [ZPL], ID 18, cr_txg 1, 36.0K, 11 objects


The issue can reproduced by trying to import the pool r/w, resulting in a 
kernel panic.


The fmdump utility gives the following information for the relevant UUID:

   $>fmdump -Vp -u 91da1503-74c5-67c2-b7c1-d4e245e4d968
   TIME   UUID 
SUNW-MSG-ID
   Mar 28 2012 12:54:26.563203000 91da1503-74c5-67c2-b7c1-d4e245e4d968 
SUNOS-8000-KL

 TIME CLASS ENA
 Mar 28 12:54:24.2698 ireport.os.sunos.panic.dump_available 
0x
 Mar 28 12:54:05.9826 ireport.os.sunos.panic.dump_pending_on_device 
0x

   nvlist version: 0
version = 0x0
class = list.suspect
uuid = 91da1503-74c5-67c2-b7c1-d4e245e4d968
code = SUNOS-8000-KL
diag-time = 1332932066 541092
de = fmd:///module/software-diagnosis
fault-list-sz = 0x1
__case_state = 0x1
topo-uuid = 3b4117e0-0ac7-cde5-b434-b9735176d591
fault-list = (array of embedded nvlists)
(start fault-list[0])
nvlist version: 0
version = 0x0
class = defect.sunos.kernel.panic
certainty = 0x64
asru = 
sw:///:path=/var/crash/.91da1503-74c5-67c2-b7c1-d4e245e4d968
resource = 
sw:///:path=/var/crash/.91da1503-74c5-67c2-b7c1-d4e245e4d968
savecore-succcess = 1
dump-dir = /var/crash
dump-files = vmdump.0
os-instance-uuid = 91da1503-74c5-67c2-b7c1-d4e245e4d968
panicstr = BAD TRAP: type=e (#pf Page fault) 
rp=ff002f6dcc50 addr=20 occurred in module "zfs" due to a NULL pointer 
dereference
panicstack = unix:die+d8 () | unix:trap+152b () | 
unix:cmntrap+e6 () | zfs:zap_leaf_lookup_closest+45 () | 
zfs:fzap_cursor_retrieve+cd () | zfs:zap_cursor_retrieve+195 () | 
zfs:zfs_purgedir+4d () |   zfs:zfs_rmnode+57 () | zfs:zfs_zinactive+b4 () | 
zfs:zfs_inactive+1a3 () | genunix:fop_inactive+b1 () | genunix:vn_rele+58 () | 
zfs:zfs_unlinked_drain+a7 () | zfs:zfsvfs_setup+f1 () | zfs:zfs_domount+152 () 
| zfs:zfs_mount+4e3 () | genunix:fsop_mount+22 () | genunix:domount+d2f () | 
genunix:mount+c0 () | genunix:syscall_ap+92 () | unix:brand_sys_sysenter+1cf () 
| 
crashtime = 1332931339
panic-time = March 28, 2012 12:42:19 PM CEST CEST
(end fault-list[0])

fault-status = 0x1
severity = Major
__ttl = 0x1
__tod = 0x4f72ede2 0x2191cbb8


The 'first view' debugger output looks like:

   mdb unix.0 vmcore.0 
   Loading modules: [ unix genunix specfs dtrace mac cpu.generic uppc pcplusmp 
scsi_vhci zfs mpt sd ip hook neti arp usba uhci sockfs qlc fctl s1394 kssl lofs 
random idm sppp crypto sata fcip cpc fcp ufs logindmux ptm ]
   > $c
   zap_leaf_lookup_closest+0x45(ff0728eac588, 0, 0, ff

Re: [zfs-discuss] kernel panic during zfs import

2012-03-28 Thread Carsten John
-Original message-
To: ZFS Discussions ; 
From:   Paul Kraus 
Sent:   Tue 27-03-2012 15:05
Subject:Re: [zfs-discuss] kernel panic during zfs import
> On Tue, Mar 27, 2012 at 3:14 AM, Carsten John  wrote:
> > Hallo everybody,
> >
> > I have a Solaris 11 box here (Sun X4270) that crashes with a kernel panic 
> during the import of a zpool (some 30TB) containing ~500 zfs filesystems 
> after 
> reboot. This causes a reboot loop, until booted single user and removed 
> /etc/zfs/zpool.cache.
> >
> >
> > From /var/adm/messages:
> >
> > savecore: [ID 570001 auth.error] reboot after panic: BAD TRAP: type=e (#pf 
> Page fault) rp=ff002f9cec50 addr=20 occurred in module "zfs" due to a 
> NULL 
> pointer dereference
> > savecore: [ID 882351 auth.error] Saving compressed system crash dump in 
> /var/crash/vmdump.2
> >
> 
> I ran into a very similar problem with Solaris 10U9 and the
> replica (zfs send | zfs recv destination) of a zpool of about 25 TB of
> data. The problem was an incomplete snapshot (the zfs send | zfs recv
> had been interrupted). On boot the system was trying to import the
> zpool and as part of that it was trying to destroy the offending
> (incomplete) snapshot. This was zpool version 22 and destruction of
> snapshots is handled as a single TXG. The problem was that the
> operation was running the system out of RAM (32 GB worth). There is a
> fix for this and it is in zpool 26 (or newer), but any snapshots
> created while the zpool is at a version prior to 26 will have the
> problem on-disk. We have support with Oracle and were able to get a
> loaner system with 128 GB RAM to clean up the zpool (it took about 75
> GB RAM to do so).
> 
> If you are at zpool 26 or later this is not your problem. If you
> are at zpool < 26, then test for an incomplete snapshot by importing
> the pool read only, then `zdb -d  | grep '%'` as the incomplete
> snapshot will have a '%' instead of a '@' as the dataset / snapshot
> separator. You can also run the zdb against the _un_imported_ zpool
> using the -e option to zdb.
> 
> See the following Oracle Bugs for more information.
> 
> CR# 6876953
> CR# 6910767
> CR# 7082249
> 
> CR#7082249 has been marked as a duplicate of CR# 6948890
> 
> P.S. I have a suspect that the incomplete snapshot was also corrupt in
> some strange way, but could never make a solid determination of that.
> We think what caused the zfs send | zfs recv to be interrupted was
> hitting an e1000g Ethernet device driver bug.
> 
> -- 
> {1-2-3-4-5-6-7-}
> Paul Kraus
> -> Senior Systems Architect, Garnet River ( http://www.garnetriver.com/ )
> -> Sound Coordinator, Schenectady Light Opera Company (
> http://www.sloctheater.org/ )
> -> Technical Advisor, Troy Civic Theatre Company
> -> Technical Advisor, RPI Players
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> 

Hi,


this scenario seems to fit. The machine that was sending the snapshot is on 
OpenSolaris Build 111b (which is running zpool version 14).

I rebooted the receiving machine due to a hanging "zfs receive" that couldn't 
be killed.

zdb -d -e  does not give any useful information:

zdb -d -e san_pool   
Dataset san_pool [ZPL], ID 18, cr_txg 1, 36.0K, 11 objects


When importing the pool readonly, I get an error about two datasets:

zpool import -o readonly=on san_pool
cannot set property for 'san_pool/home/someuser': dataset is read-only
cannot set property for 'san_pool/home/someotheruser': dataset is read-only

As this is a mirror machine, I still have the option to destroy the pool and 
copy over the stuff via send/receive from the primary. But nobody knows how 
long this will work until I'm hit again

If an interrupted send/receive can screw up a 30TB target pool, then 
send/receive isn't an option for replication data at all, furthermore it should 
be flagged as "don't use it if your target pool might contain any valuable data"

I wil reproduce the crash once more and try to file a bug report for S11 as 
recommended by Deepak (not so easy these days...).



thanks



Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] kernel panic during zfs import

2012-03-27 Thread Carsten John
Hallo everybody,

I have a Solaris 11 box here (Sun X4270) that crashes with a kernel panic 
during the import of a zpool (some 30TB) containing ~500 zfs filesystems after 
reboot. This causes a reboot loop, until booted single user and removed 
/etc/zfs/zpool.cache.


>From /var/adm/messages:

savecore: [ID 570001 auth.error] reboot after panic: BAD TRAP: type=e (#pf Page 
fault) rp=ff002f9cec50 addr=20 occurred in module "zfs" due to a NULL 
pointer dereference
savecore: [ID 882351 auth.error] Saving compressed system crash dump in 
/var/crash/vmdump.2

This is what mdb tells:

mdb unix.2 vmcore.2
Loading modules: [ unix genunix specfs dtrace mac cpu.generic uppc pcplusmp 
scsi_vhci zfs mpt sd ip hook neti arp usba uhci sockfs qlc fctl s1394 kssl lofs 
random fcp idm sata fcip cpc crypto ufs logindmux ptm sppp ]
$c
zap_leaf_lookup_closest+0x45(ff0700ca2a98, 0, 0, ff002f9cedb0)
fzap_cursor_retrieve+0xcd(ff0700ca2a98, ff002f9ceed0, ff002f9cef10)
zap_cursor_retrieve+0x195(ff002f9ceed0, ff002f9cef10)
zfs_purgedir+0x4d(ff0721d32c20)
zfs_rmnode+0x57(ff0721d32c20)
zfs_zinactive+0xb4(ff0721d32c20)
zfs_inactive+0x1a3(ff0721d3a700, ff07149dc1a0, 0)
fop_inactive+0xb1(ff0721d3a700, ff07149dc1a0, 0)
vn_rele+0x58(ff0721d3a700)
zfs_unlinked_drain+0xa7(ff07022dab40)
zfsvfs_setup+0xf1(ff07022dab40, 1)
zfs_domount+0x152(ff07223e3c70, ff0717830080)
zfs_mount+0x4e3(ff07223e3c70, ff07223e5900, ff002f9cfe20, 
ff07149dc1a0)
fsop_mount+0x22(ff07223e3c70, ff07223e5900, ff002f9cfe20, 
ff07149dc1a0)
domount+0xd2f(0, ff002f9cfe20, ff07223e5900, ff07149dc1a0, 
ff002f9cfe18)
mount+0xc0(ff0713612c78, ff002f9cfe98)
syscall_ap+0x92()
_sys_sysenter_post_swapgs+0x149()


I can import the pool readonly.

The server is a mirror for our primary file server and is synced via zfs 
send/receive.

I saw a similar effect some time ago on a opensolaris box (build 111b). That 
time my final solution was to copy over the read only mounted stuff to a newly 
created pool. As it is the second time this failure occures (on different 
machines) I'm really concerned about overall reliability 



Any suggestions?


thx

Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Good tower server for around 1,250 USD?

2012-03-26 Thread John D Groenveld
In message , Bob 
Friesenhahn writes:
>Almost all of the systems listed on the HCL are defunct and no longer 
>purchasable except for on the used market.  Obtaining an "approved" 
>system seems very difficult. In spite of this, Solaris runs very well 
>on many non-approved modern systems.

http://www.oracle.com/webfolder/technetwork/hcl/data/s11ga/systems/views/nonoracle_systems_all_results.mfg.page1.html>

>I don't know what that means as far as the ability to purchase Solaris 
>"support".

I believe it must pass the HCTS before Oracle will support
Solaris running on third-party hardware.
http://www.oracle.com/webfolder/technetwork/hcl/hcts/index.html>

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Good tower server for around 1,250 USD?

2012-03-24 Thread The Honorable Senator and Mrs . John Blutarsky
laotsu said:

> well check  this link
>
> https://shop.oracle.com/pls/ostore/product?p1=3DSunFireX4270M2server&p2=3D&p=
> 3=3D&p4=3D&sc=3Docom_x86_SunFireX4270M2server&tz=3D-4:00
>
> you may not like the price

Hahahah! Thanks for the laugh. The dual 10Gbe PCI card breaks my budget. I'm
not going to try to configure a server and see how much it costs...

I can't even get to the site from my country btw. I had to use a proxy
through my company in America to get pricing. Oracle doesn't want to sell
certain things everywhere or they don't know how to run a website.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Good tower server for around 1,250 USD?

2012-03-23 Thread The Honorable Senator and Mrs . John Blutarsky
On Fri Mar 23 at 10:06:12 2012 laot...@gmail.com wrote:

> well
> use component of x4170m2 as example you will be ok
> intel cpu
> lsi sas controller non raid
> sas 72rpm hdd
> my 2c

That sounds too vague to be useful unless I could afford an X4170M2. I
can't build a custom box and I don't have the resources to go over the parts
list and order something with the same components. Thanks though.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Good tower server for around 1,250 USD?

2012-03-23 Thread The Honorable Senator and Mrs . John Blutarsky
Bob Friesenhahn  wrote:

> On Thu, 22 Mar 2012, The Honorable Senator and Mrs. John Blutarsky wrote:
> >
> > This will be a do-everything machine. I will use it for development, hosting
> > various apps in zones (web, file server, mail server etc.) and running other
> > systems (like a Solaris 11 test system) in VirtualBox. Ultimately I would
> > like to put it under Solaris support so I am looking for something
> > officially approved. The problem is there are so many systems on the HCL I
> > don't know where to begin. One of the Supermicro super workstations looks
> 
> Almost all of the systems listed on the HCL are defunct and no longer 
> purchasable except for on the used market.

In my third world country some of these are still found new in box and sold
at "just released" prices. I'm surprised to hear about the state of the HCL,
but that is good info to be aware of. Why aren't they maintaining it? If you
can find a system can you at least depend on their statement that they
support it? Or is even that unknown? I would think if somebody buys a 5 year
old new server based on them showing Premier support available and then they
refuse to support it there would be the possibility of interesting legal
action.

> Obtaining an "approved" system seems very difficult.

Because of the list being out of date and so the systems are no longer
available, or because systems available now don't show up on the list?

> In spite of this, Solaris runs very well on many non-approved modern
> systems.

Yes, I have entitlements from Sun so I am running an Update 8 box on a
custom build with no problems. It wasn't built for Solaris or ZFS but works
great. After reading the horror stories on the list I don't want to take a
chance and buy the wrong machine and then have ZFS fail or Oracle tell me
they don't support the machine.

Thanks for your post.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Good tower server for around 1,250 USD?

2012-03-22 Thread The Honorable Senator and Mrs . John Blutarsky
Ladies and Gentlemen,

I'm thinking about spending around 1,250 USD for a tower format (desk side)
server with RAM but without disks. I'd like to have 16G ECC RAM as a
minimum and ideally 2 or 3 times that amount and I'd like for the case to
have room for at least 6 drives, more would be better but not essential. I
want to run Solaris 10 and possibly upgrade to Solaris 11 if I like
it. Right now I have nothing to run Solaris 11 on and I know Solaris 10 well
enough to know it will do what I want.

This will be a do-everything machine. I will use it for development, hosting
various apps in zones (web, file server, mail server etc.) and running other
systems (like a Solaris 11 test system) in VirtualBox. Ultimately I would
like to put it under Solaris support so I am looking for something
officially approved. The problem is there are so many systems on the HCL I
don't know where to begin. One of the Supermicro super workstations looks
good and I know they have good a reputation but Dell has better sales
channels where I live and I could get one of those or even an HP more easily
than a Supermicro as far as I know. I will be checking more on this.

I have a bunch of white box systems but I don't know anybody capable of
building a server grade box so I'm probably going to buy off the shelf.

Can anybody tell me is what I am looking for going to be available at this
price point and if not, what should I expect to pay? If you have experience
with any of the commodity server towers good or bad with Solaris and ZFS I'd
like to hear your opinions. I am refraining for asking for advice on drives
because I see the list has a few thousand posts archived on this topic and
until I go over some of those I don't want to ask about that subject just
yet. Thanks for the help.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Any recommendations on Perc H700 controller on Dell Rx10 ?

2012-03-10 Thread John D Groenveld
In message 
, Sriram Narayanan writes:
>At work, I have an R510, and R610 and an R710 - all with the H700 PERC
>controller.

BTW with some effort, Dell's sales critter will sell you the 
H200 LSI SAS HBA as a replacement for the H700 LSI MegaRAID
controller for those boxes.

>Based on experiments, it seems like there is no way to bypass the PERC
>controller - it seems like one can only access the individual disks if
>they are set up in RAID0 each.

Did you try deleting all of the RAID0 Virtual Disks and
then enabling JBOD?
# MegaCli -AdpSetProp -EnableJBOD -1 -aALL
# MegaCli -PDMakeJBOD -PhysDrv[E0:S0,E1:S1,...] -aALL

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs send/receive script

2012-03-06 Thread Carsten John
Hello everybody,

I set up a script to replicate all zfs filesystems (some 300 user home 
directories in this case) within a given pool to a "mirror" machine. The basic 
idea is to send the snapshots incremental if the corresponding snapshot exists 
on the remote side or send a complete snapshot if no corresponding previous 
snapshot is available

Thee setup basically works, but form time to time (within a run over all 
filesystems) I get error messages like:

"cannot receive new filesystem stream: dataset is busy" or

"cannot receive incremental filesystem stream: dataset is busy"

The complete script is available under:

http://pastebin.com/AWevkGAd


does anybody have a suggestion what might cause the dataset to be busy?



thx


Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs import from i86 to sparc

2012-03-06 Thread Carsten John
Hi everybody,

are there any problems to expect if we try to export/import a zfs pool from 
opensolaris (intel) (zpool version 14) to solaris 10 (sparc) (zpool version 19)?


thanks


Carsten
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] 4k sector support in Solaris 11?

2012-02-21 Thread John D Groenveld
In message <4f435ca9.8010...@tuneunix.com>, nathank writes:
>Is there actually a fix to allow manual setting of ashift now that I 

No.
http://docs.oracle.com/cd/E23824_01/html/821-1462/zpool-1m.html>

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] 4k sector support in Solaris 11?

2012-02-16 Thread John D Groenveld
In message , Dave Pooser writes:
>If I want to use a batch of new Seagate 3TB Barracudas with Solaris 11,

I'm using the Seagate's 3TB external drive with S11.
Seagate Expansion 3 TB USB 3.0 Desktop External Hard Drive STAY3000102

>will zpool let me create a new pool with ashift=12 out of the box or will

Yes.

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] What is your data error rate?

2012-01-25 Thread John Martin

On 01/25/12 09:08, Edward Ned Harvey wrote:


Assuming the failure rate of drives is not linear, but skewed toward higher 
failure rate after some period of time (say, 3 yrs)  ...


See section 3.1 of the Google study:

  http://research.google.com/archive/disk_failures.pdf

although section 4.2 of the Carnegie Mellon study
is much more supportive of the assumption.

  http://www.usenix.org/events/fast07/tech/schroeder/schroeder.pdf
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] What is your data error rate?

2012-01-24 Thread John Martin

On 01/24/12 17:06, Gregg Wonderly wrote:

What I've noticed, is that when I have my drives in a situation of small
airflow, and hence hotter operating temperatures, my disks will drop
quite quickly.


While I *believe* the same thing and thus have over provisioned
airflow in my cases (for both drives and memory), there
are studies which failed to find a strong correlation between
drive temperature and failure rates:

  http://research.google.com/archive/disk_failures.pdf

  http://www.usenix.org/events/fast07/tech/schroeder.html

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Data loss by memory corruption?

2012-01-16 Thread John Martin

On 01/16/12 11:08, David Magda wrote:


The conclusions are hardly unreasonable:


While the reliability mechanisms in ZFS are able to provide reasonable
robustness against disk corruptions, memory corruptions still remain a
serious problem to data integrity.


I've heard the same thing said ("use ECC!") on this list many times over
the years.


I believe the whole paragraph quoted from the USENIX paper above is
important:

  While the reliability mechanisms in ZFS are able to
  provide reasonable robustness against disk corruptions,
  memory corruptions still remain a serious problem to
  data integrity. Our results for memory corruptions in-
  dicate cases where bad data is returned to the user, oper-
  ations silently fail, and the whole system crashes. Our
  probability analysis shows that one single bit flip has
  small but non-negligible chances to cause failures such
  as reading/writing corrupt data and system crashing.

The authors provide probability calculations in section 6.3
for single bit flips.  ECC provides detection and correction
of single bit flips.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs read-ahead and L2ARC

2012-01-09 Thread John Martin

On 01/08/12 10:15, John Martin wrote:


I believe Joerg Moellenkamp published a discussion
several years ago on how L1ARC attempt to deal with the pollution
of the cache by large streaming reads, but I don't have
a bookmark handy (nor the knowledge of whether the
behavior is still accurate).


http://www.c0t0d0s0.org/archives/5329-Some-insight-into-the-read-cache-of-ZFS-or-The-ARC.html
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs read-ahead and L2ARC

2012-01-09 Thread John Martin

On 01/08/12 20:10, Jim Klimov wrote:


Is it true or false that: ZFS might skip the cache and
go to disks for "streaming" reads?


I don't believe this was ever suggested.  Instead, if
data is not already in the file system cache and a
large read is made from disk should the file system
put this data into the cache?

BTW, I chose the term streaming to be a subset
of sequential where the access pattern is sequential but
at what appears to be artificial time intervals.
The suggested pre-read of the entire file would
be a simple sequential read done as quickly
as the hardware allows.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs read-ahead and L2ARC

2012-01-08 Thread John Martin

On 01/08/12 11:30, Jim Klimov wrote:


However for smaller servers, such as home NASes which have
about one user overall, pre-reading and caching files even
for a single use might be an objective per se - just to let
the hard-disks spin down. Say, if I sit down to watch a
movie from my NAS, it is likely that for 90 or 120 minutes
there will be no other IO initiated by me. The movie file
can be pre-read in a few seconds, and then most of the
storage system can go to sleep.


Isn't this just a more extreme case of prediction?
In addition to the file system knowing there will only
be one client reading 90-120 minutes of (HD?) video
that will fit in the memory of a small(er) server,
now the hard drive power management code also knows there
won't be another access for 90-120 minutes so it is OK
to spin down the hard drive(s).
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs read-ahead and L2ARC

2012-01-08 Thread John Martin

On 01/08/12 09:30, Edward Ned Harvey wrote:


In the case of your MP3 collection...  Probably the only thing you can do is
to write a script which will simply go read all the files you predict will
be read soon.  The key here is the prediction - There's no way ZFS or
solaris, or any other OS in the present day is going to intelligently
predict which files you'll be requesting soon.



The other prediction is whether the blocks will be reused.
If the blocks of a streaming read are only used once, then
it may be wasteful for a file system to allow these blocks
to placed in the cache.  If a file system purposely
chooses to not cache streaming reads, manually scheduling a
"pre-read" of particular files may simply cause the file to be read
from disk twice: on the manual pre-read and when it is read again
by the actual application.

I believe Joerg Moellenkamp published a discussion
several years ago on how L1ARC attempt to deal with the pollution
of the cache by large streaming reads, but I don't have
a bookmark handy (nor the knowledge of whether the
behavior is still accurate).
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can I create a mirror for a root rpool?

2011-12-15 Thread John D Groenveld
In message , "Anonymous R
emailer (austria)" writes:
>On Solaris 10 If I install using ZFS root on only one drive is there a way
>to add another drive as a mirror later? Sorry if this was discussed
>already. I searched the archives and couldn't find the answer. Thank you.

http://docs.oracle.com/cd/E23823_01/html/819-5461/ggset.html#gkdep>
| How to Create a Mirrored ZFS Root Pool (Postinstallation)

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-12-05 Thread John D Groenveld
In message <4ebbfb5...@oracle.com>, Cindy Swearingen writes:
>CR 7102272:
>
>  ZFS storage pool created on a 3 TB USB 3.0 device has device label 
>problems
>
>Let us know if this is still a problem in the OS11 FCS release.

I finally got upgraded from Solaris 11 Express SRU 12 to S11 FCS.

Solaris 11 11/11 still spews the "I/O request is not aligned with
4096 disk sector size" warnings but zpool(1M) create's label
persists and I can export and import between systems.

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-11-10 Thread John D Groenveld
In message <4e9db04b.80...@oracle.com>, Cindy Swearingen writes:
>This is CR 7102272.

What is the title of this BugId?
I'm trying to attach my Oracle CSI to it but Chuck Rozwat
and company's support engineer can't seem to find it.

Once I get upgraded from S11x SRU12 to S11, I'll reproduce
on a more recent kernel build.

Thanks,
John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-10-18 Thread John D Groenveld
In message <4e9db04b.80...@oracle.com>, Cindy Swearingen writes:
>This is CR 7102272.

Anyone out there have Western Digital's competing 3TB Passport
drive handy to duplicate this bug?

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-10-18 Thread John D Groenveld
In message <4e9da8b1.7020...@oracle.com>, Cindy Swearingen writes:
>1. If you re-create the pool on the whole disk, like this:
>
># zpool create foo c1t0d0
>
>Then, resend the prtvtoc output for c1t0d0s0.

# zpool create snafu c1t0d0
# zpool status snafu
  pool: snafu
 state: ONLINE
 scan: none requested
config:

NAMESTATE READ WRITE CKSUM
snafu   ONLINE   0 0 0
  c1t0d0ONLINE   0 0 0

errors: No known data errors
# prtvtoc /dev/rdsk/c1t0d0s0
* /dev/rdsk/c1t0d0s0 partition map
*
* Dimensions:
*4096 bytes/sector
* 732566642 sectors
* 732566631 accessible sectors
*
* Flags:
*   1: unmountable
*  10: read-only
*
* Unallocated space:
*   First SectorLast
*   Sector CountSector
*   6   250   255
*
*  First SectorLast
* Partition  Tag  FlagsSector CountSector  Mount Directory
   0  400256 732549997 732550252
   8 1100  732550253 16384 732566636

>We should be able to tell if format is creating a dummy label,
>which means the ZFS data is never getting written to this disk.
>This would be a bug.

# zdb -l /dev/dsk/c1t0d0s0

LABEL 0

failed to unpack label 0

LABEL 1

failed to unpack label 1

LABEL 2

failed to unpack label 2

LABEL 3

failed to unpack label 3

>2. You are running this early S11 release:
>
>SunOS 5.11 151.0.1.12 i386
>
>You might retry this on more recent bits, like the EA release,
>which I think is b 171.

Doubtful I'll find time to install EA before S11 FCS's
November launch.

>I'll still file the CR.

Thank you.

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-10-18 Thread John D Groenveld
In message <4e9d98b1.8040...@oracle.com>, Cindy Swearingen writes:
>I'm going to file a CR to get this issue reviewed by the USB team
>first, but if you could humor me with another test:
>
>Can you run newfs to create a UFS file system on this device
>and mount it?

# uname -srvp
SunOS 5.11 151.0.1.12 i386
# zpool destroy foobar
# newfs /dev/rdsk/c1t0d0s0
newfs: construct a new file system /dev/rdsk/c1t0d0s0: (y/n)? y
The device sector size 4096 is not supported by ufs!

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-10-18 Thread John D Groenveld
In message <201110150202.p9f22w2n000...@elvis.arl.psu.edu>, John D Groenveld 
writes:
>I'm baffled why zpool import is unable to find the pool on the
>drive, but the drive is definitely functional.

Per Richard Elling, it looks like ZFS is unable to find
the requisite labels for importing.

John
groenv...@acm.org

# prtvtoc /dev/rdsk/c1t0d0s2
* /dev/rdsk/c1t0d0s2 partition map
*
* Dimensions:
*4096 bytes/sector
*  63 sectors/track
* 255 tracks/cylinder
*   16065 sectors/cylinder
*   45599 cylinders
*   45597 accessible cylinders
*
* Flags:
*   1: unmountable
*  10: read-only
*
* Unallocated space:
*   First SectorLast
*   Sector CountSector
*   0 16065 16064
*
*  First SectorLast
* Partition  Tag  FlagsSector CountSector  Mount Directory
   0  200  16065 732483675 732499739
   2  501  0 732515805 732515804
   8  101  0 16065 16064
# zpool create -f foobar c1t0d0s0
# zpool status foobar
  pool: foobar
 state: ONLINE
 scan: none requested
config:

NAMESTATE READ WRITE CKSUM
foobar  ONLINE   0 0 0
  c1t0d0s0  ONLINE   0 0 0

errors: No known data errors
# zdb -l /dev/dsk/c1t0d0s0

LABEL 0

failed to unpack label 0

LABEL 1

failed to unpack label 1

LABEL 2

failed to unpack label 2

LABEL 3

failed to unpack label 3

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-10-14 Thread John D Groenveld
As a sanity check, I connected the drive to a Windows 7 installation.
I was able to partition, create an NTFS volume on it, eject and
remount it.

I also tried creating the zpool on my Solaris 10 system, exporting
and trying to import the pool on my Solaris 11X system and again
no love.

I'm baffled why zpool import is unable to find the pool on the
drive, but the drive is definitely functional.

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-10-13 Thread John D Groenveld
In message <4e970387.3040...@oracle.com>, Cindy Swearingen writes:
>Any USB-related messages in /var/adm/messages for this device?

Negative.
cfgadm(1M) shows the drive and format->fdisk->analyze->read
runs merrily.

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-10-13 Thread John D Groenveld
In message <201110131150.p9dbo8yk011...@acsinet22.oracle.com>, Casper.Dik@oracl
e.com writes:
>What is the partition table?

I thought about that so I reproduced with the legacy SMI label
and a Solaris fdisk partition with ZFS on slice 0.
Same result as EFI; once I export the pool I cannot import it.

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-10-12 Thread John D Groenveld
In message <4e95e10f.9070...@oracle.com>, Cindy Swearingen writes:
>In the steps below, you're missing a zpool import step.
>I would like to see the error message when the zpool import
>step fails.

"zpool import" returns nothing.
The truss shows it poking around c1t0d0 fdisk partitions and
Solaris slices presumably hunting for pools.

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-10-12 Thread John D Groenveld
In message <4e95cb2a.30...@oracle.com>, Cindy Swearingen writes:
>What is the error when you attempt to import this pool?

"cannot import 'foo': no such pool available"
John
groenv...@acm.org

# format -e
Searching for disks...done


AVAILABLE DISK SELECTIONS:
   0. c1t0d0 
  /pci@0,0/pci108e,6676@2,1/hub@7/storage@2/disk@0,0
   1. c8t0d0 
  /pci@0,0/pci108e,6676@5/disk@0,0
   2. c8t1d0 
  /pci@0,0/pci108e,6676@5/disk@1,0
Specify disk (enter its number): ^C
# zpool create foo c1t0d0
# zfs create foo/bar
# zfs list -r foo
NAME  USED  AVAIL  REFER  MOUNTPOINT
foo   126K  2.68T32K  /foo
foo/bar31K  2.68T31K  /foo/bar
# zpool export foo
# zfs list -r foo
cannot open 'foo': dataset does not exist
# truss -t open zpool import foo
open("/var/ld/ld.config", O_RDONLY) Err#2 ENOENT
open("/lib/libumem.so.1", O_RDONLY) = 3
open("/lib/libc.so.1", O_RDONLY)= 3
open("/lib/libzfs.so.1", O_RDONLY)  = 3
open("/usr/lib/fm//libtopo.so", O_RDONLY)   = 3
open("/lib/libxml2.so.2", O_RDONLY) = 3
open("/lib/libpthread.so.1", O_RDONLY)  = 3
open("/lib/libz.so.1", O_RDONLY)= 3
open("/lib/libm.so.2", O_RDONLY)= 3
open("/lib/libsocket.so.1", O_RDONLY)   = 3
open("/lib/libnsl.so.1", O_RDONLY)  = 3
open("/usr/lib//libshare.so.1", O_RDONLY)   = 3
open("/usr/lib/locale/en_US.UTF-8/LC_MESSAGES/SUNW_OST_SGS.mo", O_RDONLY) Err#2 
ENOENT
open("/usr/lib/locale/en_US.UTF-8/LC_MESSAGES/SUNW_OST_OSLIB.mo", O_RDONLY) 
Err#2 ENOENT
open("/usr/lib/locale/en_US.UTF-8/en_US.UTF-8.so.3", O_RDONLY) = 3
open("/usr/lib/locale/en_US.UTF-8/methods_unicode.so.3", O_RDONLY) = 3
open("/dev/zfs", O_RDWR)= 3
open("/etc/mnttab", O_RDONLY)   = 4
open("/etc/dfs/sharetab", O_RDONLY) = 5
open("/lib/libavl.so.1", O_RDONLY)  = 6
open("/lib/libnvpair.so.1", O_RDONLY)   = 6
open("/lib/libuutil.so.1", O_RDONLY)= 6
open64("/dev/rdsk/", O_RDONLY)  = 6
/3: openat64(6, "c8t0d0s0", O_RDONLY)   = 9
/3: open("/lib/libadm.so.1", O_RDONLY)  = 15
/9: openat64(6, "c8t0d0s2", O_RDONLY)   = 13
/5: openat64(6, "c8t1d0s0", O_RDONLY)   = 10
/7: openat64(6, "c8t1d0s2", O_RDONLY)   = 14
/8: openat64(6, "c1t0d0s0", O_RDONLY)   = 7
/4: openat64(6, "c1t0d0s2", O_RDONLY)   Err#5 EIO
/8: open("/lib/libefi.so.1", O_RDONLY)  = 15
/3: openat64(6, "c1t0d0", O_RDONLY) = 9
/5: openat64(6, "c1t0d0p0", O_RDONLY)   = 10
/9: openat64(6, "c1t0d0p1", O_RDONLY)   = 13
/7: openat64(6, "c1t0d0p2", O_RDONLY)   Err#5 EIO
/4: openat64(6, "c1t0d0p3", O_RDONLY)   Err#5 EIO
/7: openat64(6, "c1t0d0s8", O_RDONLY)   = 14
/2: openat64(6, "c7t0d0s0", O_RDONLY)   = 8
/6: openat64(6, "c7t0d0s2", O_RDONLY)   = 12
/1: Received signal #20, SIGWINCH, in lwp_park() [default]
/3: openat64(6, "c7t0d0p0", O_RDONLY)   = 9
/4: openat64(6, "c7t0d0p1", O_RDONLY)   = 11
/5: openat64(6, "c7t0d0p2", O_RDONLY)   = 10
/6: openat64(6, "c8t0d0p0", O_RDONLY)   = 12
/6: openat64(6, "c8t0d0p1", O_RDONLY)   = 12
/6: openat64(6, "c8t0d0p2", O_RDONLY)   Err#5 EIO
/6: openat64(6, "c8t0d0p3", O_RDONLY)   Err#5 EIO
/6: openat64(6, "c8t0d0p4", O_RDONLY)   Err#5 EIO
/6: openat64(6, "c8t1d0p0", O_RDONLY)   = 12
/8: openat64(6, "c7t0d0p3", O_RDONLY)   = 7
/6: openat64(6, "c8t1d0p1", O_RDONLY)   = 12
/6: openat64(6, "c8t1d0p2", O_RDONLY)   Err#5 EIO
/6: openat64(6, "c8t1d0p3", O_RDONLY)   Err#5 EIO
/6: openat64(6, "c8t1d0p4", O_RDONLY)   Err#5 EIO
/9: openat64(6, "c7t0d0p4", O_RDONLY)   = 13
/7: openat64(6, "c7t0d0s1", O_RDONLY)   = 14
/1: open("/usr/share/locale/en_US.UTF-8/LC_MESSAGES/SUNW_OST_OSCMD.cat", 
O_RDONLY) Err#2 ENOENT
open("/usr/lib/locale/en_US.UTF-8/LC_MESSAGES/SUNW_OST_OSCMD.mo", O_RDONLY) 
Err#2 ENOENT
cannot import 'foo': no such pool available
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] weird bug with Seagate 3TB USB3 drive

2011-10-11 Thread John D Groenveld
Banging my head against a Seagate 3TB USB3 drive.
Its marketing name is:
Seagate Expansion 3 TB USB 3.0 Desktop External Hard Drive STAY3000102
format(1M) shows it identify itself as:
Seagate-External-SG11-2.73TB

Under both Solaris 10 and Solaris 11x, I receive the evil message:
| I/O request is not aligned with 4096 disk sector size.
| It is handled through Read Modify Write but the performance is very low.

However, that's not my big issue as I will use the zpool-12 hack.

My big issue is that once I zpool(1M) export the pool from
my W2100z running S10 or my Ultra 40 running S11x, I can't 
import it.

I thought weird USB connectivity issue, but I can run
"format -> analyze -> read" merrily.

Anyone seen this bug?

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] bad seagate drive?

2011-09-12 Thread John Martin

On 09/12/11 10:33, Jens Elkner wrote:


Hmmm, at least if S11x, ZFS mirror, ICH10 and cmdk (IDE) driver is involved,
I'm 99.9% confident, that "a while" turns out to be some days or weeks, only
- no matter what Platinium-Enterprise-HDDs you use ;-)


On Solaris 11 Express with a dual drive mirror, ICH10 and the AHCI
driver (not sure why you would purposely choose to run in IDE mode)
resilvering a 1TB drive (Seagate ST310005N1A1AS-RK) went at a rate of
3.2GB/min.  Deduplication was not enabled.  Only hours for a 55%
full mirror, not days or weeks.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] BAD WD drives - defective by design?

2011-09-06 Thread John Martin

http://wdc.custhelp.com/app/answers/detail/a_id/1397/~/difference-between-desktop-edition-and-raid-%28enterprise%29-edition-drives
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Kernel panic on zpool import. 200G of data inaccessible!

2011-08-19 Thread John D Groenveld
In message <1313687977.77375.yahoomail...@web121903.mail.ne1.yahoo.com>, Stu Wh
itefish writes:
>Nope, not a clue how to do that and I have installed Windows on this box inste
>ad of Solaris since I can't get my data back from ZFS.
>I have my two drives the pool is on disconnected so if this ever gets resolved
> I can reinstall Solaris and start learning again.

I believe you can configure VirtualBox for Windows to pass thru
the disk with your unimportable rpool to guest OSs.
Can OpenIndiana or FreeBSD guest import the pool?
Does Solaris 11X crash at the same place when run from within
VirtualBox?

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Kernel panic on zpool import. 200G of data inaccessible!

2011-08-15 Thread John D Groenveld
In message <1313431448.5331.yahoomail...@web121911.mail.ne1.yahoo.com>, Stu Whi
tefish writes:
>I'm sorry, I don't understand this suggestion.
>
>The pool that won't import is a mirror on two drives.

Disconnect all but the two mirrored drives that you must import
and try to import from a S11X LiveUSB.

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] matching zpool versions to development builds

2011-08-08 Thread John Martin

Is there a list of zpool versions for development builds?

I found:

  http://blogs.oracle.com/stw/entry/zfs_zpool_and_file_system

where it says Solaris 11 Express is zpool version 31, but my
system has BEs back to build 139 and I have not done a zpool upgrade
since installing this system but it reports on the current
development build:

  # zpool upgrade -v
  This system is currently running ZFS pool version 33.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] snapshots in solaris11 express

2011-07-26 Thread Carsten John
Hello everybody,

is there any known way to configure the point-in-time *when* the time-slider 
will snapshot/rotate?

With hundreds of zfs filesystems, the daily snapshot rotation slows down a big 
file server significantly, so it would be better to  have the snapshots rotated 
outside the usual workhours.

As as I found out so far, the first snapshot is taken when the service is 
restartet and then the next occurs 24 hour later (as supposed). Do I need to 
restart the service at 2:00 AM to get the desired result (not a big deal deal 
with /usr/bin/at, but not as straight forward as I would exspect).

Any suggestions?


thx


Carsten

-- 
Max Planck Institut fuer marine Mikrobiologie
- Network Administration -
Celsiustr. 1
D-28359 Bremen
Tel.: +49 421 2028568
Fax.: +49 421 2028565
PGP public key:http://www.mpi-bremen.de/Carsten_John.html
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] The open sourced ZFS test suite

2011-07-19 Thread John Kennedy
Hi,

To the best of my knowledge, the ZFS test suite (SUNWstc-fs-zfs,
as opposed to ztest) hasn't seen any development outside of
Oracle since it was open sourced in late '09. Is there anyone in
the community actively doing development with this test suite?
Is anyone using it or a derivative version for testing?

-- 
John Wren Kennedy
Delphix

275 Middlefield Road, Suite 50
Menlo Park, CA 94025
http://www.delphix.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] replace zil drive [SOLVED}

2011-06-28 Thread Carsten John
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

On 06/28/11 02:55, Edward Ned Harvey wrote:
>> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
>> boun...@opensolaris.org] On Behalf Of Carsten John
>>
>> Now I'm wondering about the best option to replace the HDD with the SSD:
> 
> What version of zpool are you running?  If it's >= 19, then you could
> actually survive a complete ZIL device failure.  So you should simply
> offline or detach or whatever the HDD and then either attach or add the new
> SDD.  Attach would be mirror, add would be two separate non-mirrored
> devices.  Maybe better performance, maybe not.
> 
> If it's zpool < 19, then you absolutely do not want to degrade to
> non-mirrored status.  First attach the new SSD, then when it's done, detach
> the HDD.
> 


Worked like a charm.

Detached the HDD, physically replaced the HDD with the new SSD and added
the new SSD to the pool's log.


thx for the suggestions


Carsten

- -- 
Max Planck Institut fuer marine Mikrobiologie
- - Network Administration -
Celsiustr. 1
D-28359 Bremen
Tel.: +49 421 2028568
Fax.: +49 421 2028565
PGP public key:http://www.mpi-bremen.de/Carsten_John.html
-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.10 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/

iEYEARECAAYFAk4JilAACgkQsRCwZeehufvfhACdF0yae2NGDKrNEswRmW4NVEFv
K5sAn0EnuAyOa9Z7ytNQQZF9QPJPjgeZ
=fVxi
-END PGP SIGNATURE-
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] replace zil drive

2011-06-28 Thread Carsten John
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

On 06/28/11 02:55, Edward Ned Harvey wrote:
>> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
>> boun...@opensolaris.org] On Behalf Of Carsten John
>>
>> Now I'm wondering about the best option to replace the HDD with the SSD:
>
> What version of zpool are you running?  If it's >= 19, then you could
> actually survive a complete ZIL device failure.  So you should simply
> offline or detach or whatever the HDD and then either attach or add the new
> SDD.  Attach would be mirror, add would be two separate non-mirrored
> devices.  Maybe better performance, maybe not.
>
> If it's zpool < 19, then you absolutely do not want to degrade to
> non-mirrored status.  First attach the new SSD, then when it's done, detach
> the HDD.
>


Sorry, sent encrypted before

I'm currently running:

zpool upgrade -v
This system is currently running ZFS pool version 31

So, detaching the HDD seems to be a safe option.


thx


Carsten


- -- 
Max Planck Institut fuer marine Mikrobiologie
- - Network Administration -
Celsiustr. 1
D-28359 Bremen
Tel.: +49 421 2028568
Fax.: +49 421 2028565
PGP public key:http://www.mpi-bremen.de/Carsten_John.html
-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.10 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/

iEYEARECAAYFAk4Je/EACgkQsRCwZeehufuHwwCglTkFJPT54dUhyyh/rqWMhFLy
sIQAn3VaJw5uDKCdhI917PDzKLb81zfY
=wtA6
-END PGP SIGNATURE-
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] replace zil drive

2011-06-27 Thread Carsten John


bin65MpxTCk5V.bin
Description: PGP/MIME version identification


encrypted.asc
Description: OpenPGP encrypted message
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] replace zil drive

2011-06-27 Thread Carsten John
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Hello everybody,

some time ago a SSD within a ZIL mirror died. As I had no SSD available
to replace it, I dropped in a normal SAS harddisk to rebuild the mirror.

In the meantime I got the warranty replacement SSD.

Now I'm wondering about the best option to replace the HDD with the SSD:

1. Remove the log mirror, put the new disk in place, add log mirror

2. Pull the HDD, forcing the mirror to fail, replace the HDD with the SSD

Unfortunately I have no free slot in the JBOD available (want to keep
the ZIL in the same JBAD as the rest of the pool):

3. Put additional temporary SAS HDD in free slot of different JBOD,
replace the HDD in the ZIL mirror with temporary HDD, pull now unused
HDD, use free slot for SSD, replace temporary HDD with SSD.



Any suggestions?


thx



Carsten





- -- 
Max Planck Institut fuer marine Mikrobiologie
- - Network Administration -
Celsiustr. 1
D-28359 Bremen
Tel.: +49 421 2028568
Fax.: +49 421 2028565
PGP public key:http://www.mpi-bremen.de/Carsten_John.html
-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.10 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/

iEYEARECAAYFAk4ISy8ACgkQsRCwZeehufs9MQCfetuYQwjbqH2Rb7qyY8G4vxaQ
TvUAoNcHPnHED1Ykat8VHF8EJIRiPmct
=jwZQ
-END PGP SIGNATURE-
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Resizing ZFS partition, shrinking NTFS?

2011-06-16 Thread John D Groenveld
In message <444915109.61308252125289.JavaMail.Twebapp@sf-app1>, Clive Meredith 
writes:
>I currently run a duel boot machine with a 45Gb partition for Win7 Ultimate an
>d a 25Gb partition for OpenSolaris 10 (134).  I need to shrink NTFS to 20Gb an
>d increase the ZFS partion to 45Gb.  Is this possible please?  I have looked a
>t using the partition tool in OpenSolaris but both partition are locked, even 
>under admin.  Win7 won't allow me to shrink the dynamic volume, as the Finsh b
>utton is always greyed out, so no luck in that direction.

Shrink the NTFS filesystem first.
I've used the Knoppix LiveCD against a defragmented NTFS.

Then use beadm(1M) to duplicate your OpenSolaris BE to
a USB drive and also send snapshots of any other rpool ZFS
there.

Then I would boot the USB drive, run format, fdisk and recreate
the Solaris fdisk partition on your system, recreate the rpool
on slice 0 of that fdisk partition, use beadm(1M) to copy
your BE back to your new rpool, and then restore any other ZFS
from those snapshots.

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Compatibility between Sun-Oracle Fishworks appliance zfs and other zfs implementations

2011-05-26 Thread John D Groenveld
In message <4dddc270.6060...@u.washington.edu>, Matt Weatherford writes:
>amount of $ on. This is a great box and we love it, although the EDU 
>discounts that Sun used to provide for hardware and support contracts 
>seem to have dried up so the cost of supporting it moving forward is 
>still unknown.

Ask Keith Block and company's sales critter about "Hardware from Oracle
- Pricing for Education (HOPE)":
http://www.oracle.com/ocom/groups/public/@ocom/documents/webcontent/364419.pdf>

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] 350TB+ storage solution

2011-05-16 Thread John Doe
following are some thoughts if it's not too late:

> 1 SuperMicro  847E1-R1400LPB
I guess you meant the 847E1[b]6[/b]-R1400LPB, the SAS1 version makes no sense

> 1 SuperMicro  H8DG6-F
not the best choice, see below why

> 171   Hitachi 7K3000 3TB
I'd go for the more environmentally friendly Ultrastar 5K3000 version - with 
that many drives you wont mind the slower rotation but WILL notice a difference 
in power and cooling cost

> 1 LSI SAS 9202-16e
this is really only a very expensive gadget to be honest, there's really no 
point to it - especially true when you start looking for the necessary cables 
that use a connector who's still in "draft" specification...

stick to the excellent LSI SAS9200-8e, of which you will need at least 3 in 
your setup, one to connect each of the 3 JBODS - with them filled with fast 
drives like you chose, you will need two links (one for the front and one for 
the back backplane) as daisychainig the backplanes together would oversaturate 
a single link. 

if you'd want to take advantage of the dual expanders on your JBOD backplanes 
for additional redundancy in case of expander or controller failure, you will 
need 6 of those  LSI SAS9200-8e - this is where your board isn't ideal as it 
has a 3/1/2 PCIe x16/x8/x4 configuration while you'd need 6 PCIe x8 - something 
the X8DTH-6F will provide, as well as the onboard LSI SAS2008 based HBA for the 
two backplanes in the server case.

> 1 LSI SAS 9211-4i
> 2 OCZ 64GB SSD Vertex 3
> 2 OCZ 256GB SSD Vertex 3
if these are meant to be connected together and used as ZIL+L2ARC, then I'd 
STRONGLY urge you to get the following instead:
1x LSI MegaRAID SAS 9265-8i 
1x LSI FastPath licence
4-8x 120GB or 240GB Vertex 3 Max IOPS Edition, whatever suits the budget

this solution allows you to push around 400k IOPS to the cache, more than 
likely way more than the stated appication of the system will need

> 1 NeterionX3120SR0001
I don't know this card personally but since it's not listed as supported 
(http://www.sun.com/io_technologies/nic/NIC1.html) I'd be careful

> My question is what is the optimum way of dividing
> these drives across vdevs?
I would do 14 x 12 drive raidz2 + 3 spare = 140*3TB = ~382TiB usable
this would allow for a logical mapping of drives to vdevs, giving you in each 
case 2 vdevs in the front and 1 in the back with the 9 drive blocks in the back 
of the JBODs used as 3 x 4/4/1, giving the remaining 2 x 12 drive vdevs plus 
one spare per case

> I could also go with 2TB drives and add an extra 45
> JBOD chassis. This would significantly decrease cost,
> but I'm running a gauntlet by getting very close to
> minimum useable space.
> 
> 12 x 18 drive raidz2
I would never do vdevs that large, it's just an accident waiting to happen!


hopefully these recommendations help you with your project. in any case, it's 
huge - the biggest system I worked on (which I actually have at home, go 
figure) only has a bit over 100TB in the following configuration:
6 x 12 drive raidz2 of Hitachi 5K3000 2TB
3 Norco 4224 with a HP SAS Expander in each
Supermicro X8DTi-LN4F with 3x LSI SAS9200-8e

so yeah, I based my thoughts on my own system but considering that it's been 
running smoothly for a while now (and that I had a very similar setup with 
smaller drives and older controllers before), I'm confident in my suggestions

Regards from Switzerland,
voyman
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Changed ACL behavior in snv_151 ?

2011-01-26 Thread Ryan John
> -Original Message-
> From: Frank Lahm [mailto:frankl...@googlemail.com] 
> Sent: 25 January 2011 14:50
> To: Ryan John
> Cc: zfs-discuss@opensolaris.org
> Subject: Re: [zfs-discuss] Changed ACL behavior in snv_151 ?

> John,

> welcome onboard!

> 2011/1/25 Ryan  John :
>> I’m sharing file systems using a smb and nfs, and since I’ve upgraded to
>> snv_151, when I do a chmod from an NFS client, I lose all the NFSv4 ACLs.

> <http://opensolaris.org/jive/thread.jspa?threadID=134162>

> I'd summarize as follows:
> in order to play nice with Windows ACL semantics via builtin CIFS,
> they choose the approach of throwing away ACLs on chmod(). Makes
> Windows happy, others not so.

> -f
Hi Frank,

This really breaks our whole setup.
Under snv_134 our users were happy with Windows ACLs, and NFSv3 and NFSv4 Linux 
clients.
They all worked very well together. The only problem we had with the deny ACLs, 
was when using the MacOS "Finder"

I don't think there's a way we can tell our users not to do a chmod.

Was it a result of PSARC/2009/029 ? 
http://arc.opensolaris.org/caselog/PSARC/2010/029/20100126_mark.shellenbaum
If so, I think that was implemented around snv_137.
This would also mean it's the same in Illumos.

Regards
John

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Changed ACL behavior in snv_151 ?

2011-01-25 Thread Ryan John
I’m using /usr/bin/chmod

From: phil.har...@gmail.com [mailto:phil.har...@gmail.com]
Sent: 25 January 2011 14:50
To: Ryan John; zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] Changed ACL behavior in snv_151 ?

Which chmod are you using? (check your PATH)
- Reply message -
From: "Ryan John" 
To: "zfs-discuss@opensolaris.org" 
Subject: [zfs-discuss] Changed ACL behavior in snv_151 ?
Date: Tue, Jan 25, 2011 13:31

Hi,

I’m sharing file systems using a smb and nfs, and since I’ve upgraded to 
snv_151, when I do a chmod from an NFS client, I lose all the NFSv4 ACLs.
Here’s what I see on a Solaris nfs client:

$ ls -lVd ACLtest/
drwxrwx---+  4 root bsse-it5 Nov 19 14:03 ACLtest/
 user:ryanj:rwxpdDaARWcCos:fd-:allow
 user:noddy:rwxpdDaARWcCos:fd-:allow
 owner@:rwxp--aARWcCos:---:allow
 group@:rwxp--a-R-c--s:---:allow
  everyone@:--a-R-c--s:---:allow
$ chmod 770 ACLtest/
$ ls -lVd ACLtest/
drwxrwx---   4 root bsse-it5 Nov 19 14:03 ACLtest/
 owner@:rwxp--aARWcCos:---:allow
 group@:rwxp--a-R-c--s:---:allow
  everyone@:--a-R-c--s:---:allow

Same happens from a Linux or Solaris client
I have aclinherit set to  passthrough

Anyone any ideas?
On a snv_134 system, the ACLs are retained.

Regards
John
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Changed ACL behavior in snv_151 ?

2011-01-25 Thread Ryan John
Hi,

I'm sharing file systems using a smb and nfs, and since I've upgraded to 
snv_151, when I do a chmod from an NFS client, I lose all the NFSv4 ACLs.
Here's what I see on a Solaris nfs client:

$ ls -lVd ACLtest/
drwxrwx---+  4 root bsse-it5 Nov 19 14:03 ACLtest/
 user:ryanj:rwxpdDaARWcCos:fd-:allow
 user:noddy:rwxpdDaARWcCos:fd-:allow
 owner@:rwxp--aARWcCos:---:allow
 group@:rwxp--a-R-c--s:---:allow
  everyone@:--a-R-c--s:---:allow
$ chmod 770 ACLtest/
$ ls -lVd ACLtest/
drwxrwx---   4 root bsse-it5 Nov 19 14:03 ACLtest/
 owner@:rwxp--aARWcCos:---:allow
 group@:rwxp--a-R-c--s:---:allow
  everyone@:--a-R-c--s:---:allow

Same happens from a Linux or Solaris client
I have aclinherit set to  passthrough

Anyone any ideas?
On a snv_134 system, the ACLs are retained.

Regards
John
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs roolback

2010-11-09 Thread John Goolsby
I'm trying to rollback from a bad patch install on Solaris 10.  From the 
failsafe BE I tried to rollback, but zfs is asking me to provide allow rollback 
permissions.  It's hard for me to tell exactly because the messages are 
scrolling off the screen before I can read them. Any help would be appreciated.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How do you use >1 partition on x86?

2010-10-25 Thread John D Groenveld
In message <1026474698.131288014974214.javamail.tweb...@sf-app1>, Bill Werner w
rites:
>So when I built my new workstation last year, I partitioned the one and only d
>isk in half, 50% for Windows, 50% for 2009.06.   Now, I'm not using Windows, s
>o I'd like to use the other half for another ZFS pool, but I can't figure out 
>how to access it.

Use beadm(1M) to duplicate your BE to a USB disk, then boot it,
then format/fdisk your workstation disk, then use beadm(1M) to
duplicate your BE back to your workstation disk.

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] BugID 6961707

2010-10-07 Thread John D Groenveld
In message <201008112022.o7bkmc2j028...@elvis.arl.psu.edu>, John D Groenveld wr
ites:
>I'm stumbling over BugID 6961707 on build 134.

I see the bug has been stomped in build 150. Awesome!
http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6961707>

In which build did it first arrive?

Thanks,
John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] pool died during scrub

2010-09-01 Thread Carsten John
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Jeff Bacon wrote:
> I have a bunch of sol10U8 boxes with ZFS pools, most all raidz2 8-disk
> stripe. They're all supermicro-based with retail LSI cards.
> 
> I've noticed a tendency for things to go a little bonkers during the
> weekly scrub (they all scrub over the weekend), and that's when I'll
> lose a disk here and there. OK, fine, that's sort of the point, and
> they're SATA drives so things happen. 
> 
> I've never lost a pool though, until now. This is Not Fun. 
> 
>> ::status
> debugging crash dump vmcore.0 (64-bit) from ny-fs4
> operating system: 5.10 Generic_142901-10 (i86pc)
> panic message:
> BAD TRAP: type=e (#pf Page fault) rp=fe80007cb850 addr=28 occurred
> in module "zfs" due to a NULL pointer dereference
> dump content: kernel pages only
>> $C
> fe80007cb960 vdev_is_dead+2()
> fe80007cb9a0 vdev_mirror_child_select+0x65()
> fe80007cba00 vdev_mirror_io_start+0x44()
> fe80007cba30 zio_vdev_io_start+0x159()
> fe80007cba60 zio_execute+0x6f()
> fe80007cba90 zio_wait+0x2d()
> fe80007cbb40 arc_read_nolock+0x668()
> fe80007cbbd0 dmu_objset_open_impl+0xcf()
> fe80007cbc20 dsl_pool_open+0x4e()
> fe80007cbcc0 spa_load+0x307()
> fe80007cbd00 spa_open_common+0xf7()
> fe80007cbd10 spa_open+0xb()
> fe80007cbd30 pool_status_check+0x19()
> fe80007cbd80 zfsdev_ioctl+0x1b1()
> fe80007cbd90 cdev_ioctl+0x1d()
> fe80007cbdb0 spec_ioctl+0x50()
> fe80007cbde0 fop_ioctl+0x25()
> fe80007cbec0 ioctl+0xac()
> fe80007cbf10 _sys_sysenter_post_swapgs+0x14b()
> 
>   pool: srv
> id: 9515618289022845993
>  state: UNAVAIL
> status: One or more devices are missing from the system.
> action: The pool cannot be imported. Attach the missing
> devices and try again.
>see: http://www.sun.com/msg/ZFS-8000-6X
> config:
> 
> srvUNAVAIL  missing device
>   raidz2   ONLINE
> c2t5000C5001F2CCE1Fd0  ONLINE
> c2t5000C5001F34F5FAd0  ONLINE
> c2t5000C5001F48D399d0  ONLINE
> c2t5000C5001F485EC3d0  ONLINE
> c2t5000C5001F492E42d0  ONLINE
> c2t5000C5001F48549Bd0  ONLINE
> c2t5000C5001F370919d0  ONLINE
> c2t5000C5001F484245d0  ONLINE
>   raidz2   ONLINE
> c2t5F000B5C8187d0  ONLINE
> c2t5F000B5C8157d0  ONLINE
> c2t5F000B5C9101d0  ONLINE
> c2t5F000B5C8167d0  ONLINE
> c2t5F000B5C9120d0  ONLINE
> c2t5F000B5C9151d0  ONLINE
> c2t5F000B5C9170d0  ONLINE
> c2t5F000B5C9180d0  ONLINE
>   raidz2   ONLINE
> c2t5000C50010A88E76d0  ONLINE
> c2t5000C5000DCD308Cd0  ONLINE
> c2t5000C5001F1F456Dd0  ONLINE
> c2t5000C50010920E06d0  ONLINE
> c2t5000C5001F20C81Fd0  ONLINE
> c2t5000C5001F3C7735d0  ONLINE
> c2t5000C500113BC008d0  ONLINE
> c2t5000C50014CD416Ad0  ONLINE
> 
> Additional devices are known to be part of this pool, though
> their
> exact configuration cannot be determined.
> 
> 
> All of this would be ok... except THOSE ARE THE ONLY DEVICES THAT WERE
> PART OF THE POOL. How can it be missing a device that didn't exist? 
> 
> A "zpool import -fF" results in the above kernel panic. This also
> creates /etc/zfs/zpool.cache.tmp, which then results in the pool being
> imported, which leads to a continuous reboot/panic cycle. 
> 
> I can't obviously use b134 to import the pool without logs, since that
> would imply upgrading the pool first, which is hard to do if it's not
> imported. 
> 
> My zdb skills are lacking - zdb -l gets you about so far and that's it.
> (where the heck are the other options to zdb even written down, besides
> in the code?)
> 
> OK, so this isn't the end of the world, but it's 15TB of data I'd really
> rather not have to re-copy across a 100Mbit line. It really more
> concerns me that ZFS would do this in the first place - it's not
> supposed to corrupt itself!!
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Hi Jeff,



looks similar to a crash I had here at our site a few month ago. Same
symptoms, no actual solution. We had to recover from a rsync backup server.


We had the logs on an mirrored SSD and an additional SSD as cache.

The machine (SUN 4270 with SUN J4400 JBODS and SUN SAS disks) crashed in
the same manner (core dumping while trying to import the pool). After
booting into single user mode we found the log pool mirror corrupted
(one disk unavailbale). Even after replacing the disk and resilvering
the log mirror we were not able to import the pool.

I suggest that it may has been related to memory (perhaps a lack of memory).


all the best

Re: [zfs-discuss] VM's on ZFS - 7210

2010-08-27 Thread John
Wouldn't it be possible to saturate the SSD ZIL with enough backlogged sync 
writes? 

What I mean is, doesn't the ZIL eventually need to make it to the pool, and if 
the pool as a whole (spinning disks) can't keep up with 30+ vm's of write 
requests, couldn't you fill up the ZIL that way?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Opensolaris is apparently dead

2010-08-18 Thread John D Groenveld
In message <4c6c4e30.7060...@ianshome.com>, Ian Collins writes:
>If you count Monday this week as lately, we have never had to wait more 
>than 24 hours for replacement drives for our 45x0 or 7000 series 

Same here, but two weeks ago for a failed drive in an X4150.

Last week SunSolve was sending my service order requests to
/dev/null, but someone manually entered after I submitted
web feedback.

John
groenv...@acm.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] BugID 6961707

2010-08-11 Thread John D Groenveld
I'm stumbling over BugID 6961707 on build 134.
OpenSolaris Development snv_134 X86

Via the b134 Live CD, when I try to "zpool import -f -F -n rpool"
I get this helpful panic.

panic[cpu4]/thread=ff006cd06c60: zfs: allocating allocated 
segment(offset=95698377728 size=16384)


ff006cd06580 genunix:vcmn_err+2c ()
ff006cd06670 zfs:zfs_panic_recover+ae ()
ff006cd06710 zfs:space_map_add+d3 ()
ff006cd067c0 zfs:space_map_load+470 ()
ff006cd06820 zfs:metaslab_activate+95 ()
ff006cd068e0 zfs:metaslab_group_alloc+246 ()
ff006cd069a0 zfs:metaslab_alloc_dva+2aa ()
ff006cd06a40 zfs:metaslab_alloc+9c ()
ff006cd06a80 zfs:zio_dva_allocate+57 ()
ff006cd06ab0 zfs:zio_execute+8d ()
ff006cd06b50 genunix:taskq_thread+248 ()
ff006cd06b60 unix:thread_start+8 ()

While I wait for SunSolve to forward my service request to
the right engineer, has anyone here hit this and gotten it
resolved?

Is the pool corrupted on disk?

John
groenv...@acm.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Optimal Disk configuration

2010-07-21 Thread John Andrunas
I know this is potentially a loaded question, but what is generally
considered the optimal disk configuration for ZFS.  I have 48 disks on
2 RAID controllers (2x24).  The RAID controller can do RAID
0/1/5/6/10/50/60 or JBOD.  What is generally considered the optimal
configuration for the disks.  JBOD, RAID zvols on both controllers.

-- 
John
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Need ZFS master!

2010-07-12 Thread john
Hello all. I am new...very new to opensolaris and I am having an issue and have 
no idea what is going wrong. So I have 5 drives in my machine. all 500gb. I 
installed open solaris on the first drive and rebooted. . Now what I want to do 
is ad a second drive so they are mirrored. How does one do this!!! I am getting 
no where and need some help.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] mount zfs boot disk on another server?

2010-06-17 Thread Seaman, John
Could you import it back on the original server with

Zpool import -f newpool rpool?

Jay

-Original Message-
From: Brandon High [mailto:bh...@freaks.com] 
Sent: Wednesday, June 16, 2010 2:19 PM
To: Seaman, John
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] mount zfs boot disk on another server?

On Wed, Jun 16, 2010 at 10:24 AM, Jay Seaman
 wrote:
> zpool import
>
> to get the id number of the non-native rpool

You'll get the names and IDs for any un-imported pools.

> then use
> zpool import -f  -R /mnt newpool

That should work. You may have to put the pool id after the -R argument
though.

You're renaming to pool in the process, so it may not work in the
original host afterwards, since it expects the pool name to be rpool.
You could boot a system off the LiveCD (which doesn't have an rpool)
and import it without changing the name.

-B

-- 
Brandon High : bh...@freaks.com


The information in this Internet e-mail is confidential and may be legally 
privileged. It is intended solely for the addressee(s). Access to this Internet 
e-mail by anyone else is unauthorized.

If you are not the intended recipient of this e-mail, any disclosure, copying, 
or distribution of it is prohibited and may be unlawful. If you have received 
this e-mail in error, please notify the sender and immediately and permanently 
delete it and destroy any copies of it that were printed out.  When addressed 
to our clients any opinions or advice contained in this Internet e-mail is 
subject to the terms and conditions expressed in any applicable governing ING 
terms of business or client engagement letter.

Visit us at www.ing.com


 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Dedup performance hit

2010-06-13 Thread John J Balestrini
Howdy all,

I too dabbled with dedup and found the performance poor with only 4gb ram. I've 
since disabled dedup and find the performance better but "zpool list" still 
shows a 1.15x dedup ratio. Is this still a hit on disk io performance? Aside 
from copying the data off and back onto the filesystem, is there another way to 
de-dedup the pool?

Thanks,

John



On Jun 13, 2010, at 10:17 PM, Erik Trimble wrote:

> Hernan F wrote:
>> Hello, I tried enabling dedup on a filesystem, and moved files into it to 
>> take advantage of it. I had about 700GB of files and left it for some hours. 
>> When I returned, only 70GB were moved.
>> 
>> I checked zpool iostat, and it showed about 8MB/s R/W performance (the old 
>> and new zfs filesystems are in the same pool). So I disabled dedup for a few 
>> seconds and instantly the performance jumped to 80MB/s
>> 
>> It's Athlon64 x2 machine with 4GB RAM, it's only a fileserver (4x1TB SATA 
>> for ZFS). arcstat.pl shows 2G for arcsz, top shows 13% CPU during the 8MB/s 
>> transfers. 
>> Is this normal behavior? Should I always expect such low performance, or is 
>> there anything wrong with my setup? 
>> Thanks in advance,
>> Hernan
>>  
> You are severely RAM limited.  In order to do dedup, ZFS has to maintain a 
> catalog of every single block it writes and the checksum for that block. This 
> is called the Dedup Table (DDT for short).  
> So, during the copy, ZFS has to (a) read a block from the old filesystem, (b) 
> check the current DDT to see if that block exists and (c) either write the 
> block to the new filesytem (and add an appropriate DDT entry for it), or 
> write a metadata update with the dedup reference block reference.
> 
> Likely, you have two problems:
> 
> (1) I suspect your source filesystem has lots of blocks (that is, it's likely 
> made up smaller-sized files).  Lots of blocks means lots of seeking back and 
> forth to read all those blocks.
> 
> (2) Lots of blocks also means lots of entries in the DDT.  It's trivial to 
> overwhelm a 4GB system with a large DDT.  If the DDT can't fit in RAM, then 
> it has to get partially refreshed from disk.
> 
> Thus, here's what's likely going on:
> 
> (1)  ZFS reads a block and it's checksum from the old filesystem
> (2)  it checks the DDT to see if that checksum exists
> (3) finding that the entire DDT isn't resident in RAM, it starts a cycle to 
> read the rest of the (potential) entries from the new filesystems' metadata.  
> That is, it tries to reconstruct the DDT from disk.  Which involves a HUGE 
> amount of random seek reads on the new filesystem.
> 
> In essence, since you likely can't fit the DDT in RAM, each block read from 
> the old filesystem forces a flurry of reads from the new filesystem. Which 
> eats up the IOPS that your single pool can provide.  It thrashes the disks.  
> Your solution is to either buy more RAM, or find something you can use as an 
> L2ARC cache device for your pool.  Ideally, it would be an SSD.  However, in 
> this case, a plain hard drive would do OK (NOT one already in a pool).To 
> add such a device, you would do:  'zpool add tank mycachedevice'
> 
> 
> 
> 
> -- 
> Erik Trimble
> Java System Support
> Mailstop:  usca22-123
> Phone:  x17195
> Santa Clara, CA
> Timezone: US/Pacific (GMT-0800)
> 
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] don't mount a zpool on boot

2010-05-20 Thread John Andrunas
Can I make a pool not mount on boot?  I seem to recall reading
somewhere how to do it, but can't seem to find it now.


-- 
John
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs mount -a kernel panic

2010-05-19 Thread John Andrunas
OK, I got a core dump, what do I do with it now?

It is 1.2G in size.


On Wed, May 19, 2010 at 10:54 AM, John Andrunas  wrote:
> Hmmm... no coredump even though I configured it.
>
> Here is the trace though  I will see what I can do about the coredump
>
> r...@cluster:/export/home/admin# zfs mount vol2/vm2
>
> panic[cpu3]/thread=ff001f45ec60: BAD TRAP: type=e (#pf Page fault)
> rp=ff001f45e950 addr=30 occurred in module "zfs" due to a NULL
> pointer deree
>
> zpool-vol2: #pf Page fault
> Bad kernel fault at addr=0x30
> pid=1469, pc=0xf795d054, sp=0xff001f45ea48, eflags=0x10296
> cr0: 8005003b cr4: 6f8
> cr2: 30cr3: 500cr8: c
>
>        rdi:                0 rsi: ff05208b2388 rdx: ff001f45e888
>        rcx:                0  r8:        3000900ff  r9:         198f5ff6
>        rax:                0 rbx:              200 rbp: ff001f45ea50
>        r10:         c0130803 r11: ff001f45ec60 r12: ff05208b2388
>        r13: ff0521fc4000 r14: ff050c0167e0 r15: ff050c0167e8
>        fsb:                0 gsb: ff04eb9b8080  ds:               4b
>         es:               4b  fs:                0  gs:              1c3
>        trp:                e err:                2 rip: f795d054
>         cs:               30 rfl:            10296 rsp: ff001f45ea48
>         ss:               38
>
> ff001f45e830 unix:die+dd ()
> ff001f45e940 unix:trap+177b ()
> ff001f45e950 unix:cmntrap+e6 ()
> ff001f45ea50 zfs:ddt_phys_decref+c ()
> ff001f45ea80 zfs:zio_ddt_free+55 ()
> ff001f45eab0 zfs:zio_execute+8d ()
> ff001f45eb50 genunix:taskq_thread+248 ()
> ff001f45eb60 unix:thread_start+8 ()
>
> syncing file systems... done
> skipping system dump - no dump device configured
> rebooting...
>
>
> On Wed, May 19, 2010 at 8:55 AM, Michael Schuster
>  wrote:
>> On 19.05.10 17:53, John Andrunas wrote:
>>>
>>> Not to my knowledge, how would I go about getting one?  (CC'ing discuss)
>>
>> man savecore and dumpadm.
>>
>> Michael
>>>
>>>
>>> On Wed, May 19, 2010 at 8:46 AM, Mark J Musante
>>>  wrote:
>>>>
>>>> Do you have a coredump?  Or a stack trace of the panic?
>>>>
>>>> On Wed, 19 May 2010, John Andrunas wrote:
>>>>
>>>>> Running ZFS on a Nexenta box, I had a mirror get broken and apparently
>>>>> the metadata is corrupt now.  If I try and mount vol2 it works but if
>>>>> I try and mount -a or mount vol2/vm2 is instantly kernel panics and
>>>>> reboots.  Is it possible to recover from this?  I don't care if I lose
>>>>> the file listed below, but the other data in the volume would be
>>>>> really nice to get back.  I have scrubbed the volume to no avail.  Any
>>>>> other thoughts.
>>>>>
>>>>>
>>>>> zpool status -xv vol2
>>>>>  pool: vol2
>>>>> state: ONLINE
>>>>> status: One or more devices has experienced an error resulting in data
>>>>>       corruption.  Applications may be affected.
>>>>> action: Restore the file in question if possible.  Otherwise restore the
>>>>>       entire pool from backup.
>>>>>  see: http://www.sun.com/msg/ZFS-8000-8A
>>>>> scrub: none requested
>>>>> config:
>>>>>
>>>>>       NAME        STATE     READ WRITE CKSUM
>>>>>       vol2        ONLINE       0     0     0
>>>>>         mirror-0  ONLINE       0     0     0
>>>>>           c3t3d0  ONLINE       0     0     0
>>>>>           c3t2d0  ONLINE       0     0     0
>>>>>
>>>>> errors: Permanent errors have been detected in the following files:
>>>>>
>>>>>       vol2/v...@snap-daily-1-2010-05-06-:/as5/as5-flat.vmdk
>>>>>
>>>>> --
>>>>> John
>>>>> ___
>>>>> zfs-discuss mailing list
>>>>> zfs-discuss@opensolaris.org
>>>>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>>>>>
>>>>
>>>>
>>>> Regards,
>>>> markm
>>>>
>>>
>>>
>>>
>>
>>
>> --
>> michael.schus...@oracle.com     http://blogs.sun.com/recursion
>> Recursion, n.: see 'Recursion'
>>
>
>
>
> --
> John
>



-- 
John
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


  1   2   3   >