Hi Venku, 

thanks for your input. Meanwhile i cloned the systems with lucreate and updated 
it to the ce 09/08 release. I removed and recreated the cluster. The failure 
that both nodes mount the zfs concurrently happens after i trigger a fail over 
of the HASP resource (by rebooting the active-node). This leads to a kernel 
panic, thereafter i import manually another pool and reboot. Finally both nodes 
mount the HASP-zfs. Here are the commands and logs involved:   

Remove old HASP resource:

siegfried# clresource disable vb1-storage
siegfried# clresource delete vb1-storage
siegfried# clresourcegroup  delete vb1
siegfried# zpool destroy vb1
siegfried# zpool create UNUSED___ /dev/did/dsk/d5s0 # overwrite the zfs labels
siegfried# rm /etc/zfs/zpool.cache 
voelsung# rm /etc/zfs/zpool.cache 
voelsung# scshutdown -y -g0 # end episode 1


Re-create the resource:

siegfried# zpool create -f vb1 /dev/did/dsk/d5s0
siegfried# zfs create vb1/vb1
siegfried# zpool  status vb1
  pool: vb1
 state: ONLINE
 scrub: none requested
config:
        NAME                 STATE     READ WRITE CKSUM
        vb1                  ONLINE       0     0     0
          /dev/did/dsk/d5s0  ONLINE       0     0     0
errors: No known data errors
siegfried# zfs create vb1/vb1
siegfried# clresourcegroup create vb1
siegfried# clresourcegroup manage vb1
siegfried# clrg online vb1
siegfried# clresource create -t SUNW.HAStoragePlus \
> -g vb1 \
> -p Zpools=vb1 \
> -p AffinityOn=True vb1-storage 
siegfried# clresource status vb1-storage
=== Cluster Resources ===
Resource Name       Node Name      State        Status Message
-------------       ---------      -----        --------------
vb1-storage         voelsung       Online       Online
                    siegfried      Offline      Offline

voelsung# mount| grep vb1
/vb1 on vb1 read/write/setuid/devices/nonbmand/exec/xattr/atime/dev=2d90002 on 
Wed Oct 29 13:30:57 2008
/vb1/vb1 on vb1/vb1 
read/write/setuid/devices/nonbmand/exec/xattr/atime/dev=2d90003 on Wed Oct 29 
13:30:57 2008
siegfried# mount| grep vb1
siegfried#                              
#....no problem until here...


Things go wrong (voelsung is rebooted, siegfried panics):
voelsung# reboot 
Oct 29 13:38:47 siegfried genunix: NOTICE: clcomm: Path siegfried:e1000g2 - 
voelsung:e1000g2 being drained
Oct 29 13:38:47 siegfried genunix: NOTICE: clcomm: Path siegfried:e1000g1 - 
voelsung:e1000g1 being drained
Oct 29 13:38:47 siegfried ip: TCP_IOC_ABORT_CONN: aborted 0 connection
Oct 29 13:38:49 siegfried genunix: NOTICE: CMM: Node voelsung (nodeid = 1) is 
down.
Oct 29 13:38:49 siegfried genunix: NOTICE: CMM: Cluster members: siegfried.
Oct 29 13:38:49 siegfried genuniNx: NOTICE: CMM:onode reconfigurttion #3 
completifying cluster that this node is panicking

panic[cpu3]/thread=ffffff000f631c80: Reservation Conflict
Disk: /scsi_vhci/disk at g600d02300000000000888275cd1cc500

ffffff000f631a00 sd:sd_panic_for_res_conflict+4f ()
ffffff000f631a40 sd:sd_pkt_status_reservation_conflict+a8 ()
ffffff000f631a90 sd:sdintr+44e ()
ffffff000f631b30 scsi_vhci:vhci_intr+6ac ()
ffffff000f631b50 fcp:fcp_post_callback+1e ()
ffffff000f631b90 fcp:fcp_cmd_callback+4b ()
ffffff000f631bd0 emlxs:emlxs_iodone+b1 ()
ffffff000f631c20 emlxs:emlxs_iodone_server+15d ()
ffffff000f631c60 emlxs:emlxs_thread+15e ()
ffffff000f631c70 unix:thread_start+8 ()

syncing file systems... 6 4 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 done (not 
all i/o completed)
dumping to /dev/dsk/c4t600D0230000000000088824BC4228803d0s1, offset 1719074816, 
content: kernel
100% done: 178044 pages dumped, compression ratio 5.30, dump succeeded
rebooting...

I import a pool "dummypool" manually on siegfried (this zpool is not under HASP 
control and is manually import only on node siegfried)
siegfried# zpool  import  dummypool   
siegfried# zpool  status  dummypool                                             
                                                              
  pool: dummypool
 state: ONLINE
 scrub: none requested
config:

        NAME                                       STATE     READ WRITE CKSUM
        dummypool                                  ONLINE       0     0     0
          c4t600D02300000000000888275CD1CC500d0s0  ONLINE       0     0     0

errors: No known data errors
siegfried# reboot  
Kernel errors while booting siegfried:
Hostname: siegfried
WARNING: /scsi_vhci/disk at g600d02300000000000888275cd1cc500 (sd15):
        reservation conflict
WARNING: /scsi_vhci/disk at g600d0230000000000088824bc4228807 (sd13):
        reservation conflict

Here (finally) the error: Both nodes mount the zfs concurrently:
siegfried# zpool  status     
  pool: dummypool
 state: ONLINE
 scrub: none requested
config:

        NAME                                       STATE     READ WRITE CKSUM
        dummypool                                  ONLINE       0     0     0
          c4t600D02300000000000888275CD1CC500d0s0  ONLINE       0     0     0

errors: No known data errors

  pool: vb1
 state: ONLINE
 scrub: none requested
config:

        NAME                                       STATE     READ WRITE CKSUM
        vb1                                        ONLINE       0     0     0
          c4t600D0230000000000088824BC4228807d0s0  ONLINE       0     0     0

errors: No known data errors

voelsung# zpool  status 
  pool: vb1
 state: ONLINE
 scrub: none requested
config:

        NAME                                       STATE     READ WRITE CKSUM
        vb1                                        ONLINE       0     0     0
          c4t600D0230000000000088824BC4228807d0s0  ONLINE       0     0     0

errors: No known data errors
voelsung# clresource status vb1-storage

=== Cluster Resources ===

Resource Name       Node Name      State        Status Message
-------------       ---------      -----        --------------
vb1-storage         voelsung       Online       Online
                    siegfried      Offline      Offline


Best wishes,
 Armin
-- 
This message posted from opensolaris.org

Reply via email to