I want to express my thanks. My gratitude. I am not easily impressed
by technology anymore and ZFS impressed me this morning.

Sometime late last night a primary server of mine had a critical
fault. One of the PCI cards in a V480 was the cause and for whatever
reasons this destroyed the DC-DC power convertors that powered the
primary internal disks. It also dropped the whole machine and 12
zones.

I feared the worst and made the call for service at about midnight
last night. A Sun service tech said he could be there in 2 hours
or so but he asked me to check this and check that. The people at
the datacenter were happy to tell me there was a wrench light on
but other than that, they knew nothing.

This machine, like all critical systems I have, uses mirrored disks
in ZPools with multiple links of fibre to arrays.  I dreaded what
would happen when we tried to boot this box after all the dust was
blown out and hardware swapped.

Early this morning ... I watched the detailed diags run and finally
a nice clean ok prompt.

<*>
Hardware Power On

@(#)OBP 4.22.34 2007/07/23 13:01 Sun Fire 4XX
System is initializing with diag-switch? overrides.
Online: CPU0 CPU1 CPU2 CPU3*
Validating JTAG integrity...Done
.
.
.
CPU0: System POST Completed
    Pass/Fail Status  = 0000.0000.0000.0000
    ESB Overall Status  = ffff.ffff.ffff.ffff

<*>
POST Reset
.
.
.

{3} ok show-post-results
System POST Results
Component:    Results

CPU/Memory:    Passed
IO-Bridge8:    Passed
IO-Bridge9:    Passed
GPTwo Slots:   Passed
Onboard FCAL:  Passed
Onboard Net1:  Passed
Onboard Net0:  Passed
Onboard IDE:   Passed
PCI Slots:     Passed
BBC0:          Passed
RIO:           Passed
USB:           Passed
RSC:           Passed
POST Message:  POST PASS
{3} ok boot -s

Eventually I saw my login prompt. There were no warnings about data
corruption. No data loss. No noise at all in fact.   :-O

# zpool list
NAME     SIZE   USED  AVAIL    CAP  HEALTH  ALTROOT
fibre0   680G   654G  25.8G    96%  ONLINE  -
z0      40.2G   103K  40.2G     0%  ONLINE  -

# zpool status fibre0
  pool: fibre0
 state: ONLINE
status: The pool is formatted using an older on-disk format.  The pool can
        still be used, but some features are unavailable.
action: Upgrade the pool using 'zpool upgrade'.  Once this is done, the
        pool will no longer be accessible on older software versions.
 scrub: none requested
config:

        NAME         STATE     READ WRITE CKSUM
        fibre0       ONLINE       0     0     0
          mirror     ONLINE       0     0     0
            c2t16d0  ONLINE       0     0     0
            c5t0d0   ONLINE       0     0     0
          mirror     ONLINE       0     0     0
            c5t1d0   ONLINE       0     0     0
            c2t17d0  ONLINE       0     0     0
          mirror     ONLINE       0     0     0
            c5t2d0   ONLINE       0     0     0
            c2t18d0  ONLINE       0     0     0
          mirror     ONLINE       0     0     0
            c2t20d0  ONLINE       0     0     0
            c5t4d0   ONLINE       0     0     0
          mirror     ONLINE       0     0     0
            c2t21d0  ONLINE       0     0     0
            c5t6d0   ONLINE       0     0     0
        spares
          c2t22d0    AVAIL

errors: No known data errors
#

Not one error. No message about resilver this or inode that.

Everything booted flawlessly and I was able to see all my zones :

# bin/lz
---------------------------------------------------------------------
NAME   ID  STATUS      PATH          HOSTNAME      BRAND     IP
---------------------------------------------------------------------
z_001  4   running     /zone/z_001   pluto         solaris8  excl
z_002  -   installed   /zone/z_002   ldap01        native    shared
z_003  -   installed   /zone/z_003   openfor       solaris9  shared
z_004  6   running     /zone/z_004   gaspra        native    shared
z_005  5   running     /zone/z_005   ibisprd       native    shared
z_006  7   running     /zone/z_006   io            native    shared
z_007  1   running     /zone/z_007   nis           native    shared
z_008  3   running     /zone/z_008   callistoz     native    shared
z_009  2   running     /zone/z_009   loginz        native    shared
z_010  -   installed   /zone/z_010   venus         solaris8  shared
z_011  -   installed   /zone/z_011   adbs          solaris9  shared
z_012  -   installed   /zone/z_012   auroraux      native    shared
z_013  8   running     /zone/z_013   osiris        native    excl
z_014  -   installed   /zone/z_014   jira          native    shared

People love to complain. I see it all the time.

I downloaded this OS for free and I run it in production.
I have support and I am fine with paying for support contracts.
But someone somewhere needs to buy the ZFS guys some keg(s) of
whatever beer they want. Or maybe new Porsche Cayman S toys.

That would be gratitude as something more than just words.

Thank you.

-- 
Dennis Clarke

ps: the one funny thing is that I had to get a few things swapped
    out and I guess that resets the system clock. It now reports :

# uptime
  8:19pm  up 3483 day(s), 19:07,  1 user,  load average: 0.24, 0.21, 0.18

I don't think that is accurate :-)


_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to