Hello,

I built one of my servers (V120) with a zfs root to see how it works, now when 
I try to rebuild and re-use the two internal disks using slices/sds and ufs I 
can still see that there are references to the old zpool called rpool. When I 
try to destroy the old pool I get core dups. Does anyone have any idea how I 
resolve this?

Many thanks,

John


ok boot rootmirror
Resetti
LOM event: +6d+5h2m18s host reset
ng ...

ChassisSerialNumber UNDEFINED
Sun Fire V120 (UltraSPARC-IIe 648MHz), No Keyboard
OpenBoot 4.0, 3072 MB memory installed, Serial #55470796.
Ethernet address 0:3:ba:4e:6a:cc, Host ID: 834e6acc.



Executing last command: boot rootmirror
Boot device: /p...@1f,0/p...@1/s...@8/d...@1,0  File and args:
SunOS Release 5.10 Version Generic_137137-09 64-bit
Copyright 1983-2008 Sun Microsystems, Inc.  All rights reserved.
Use is subject to license terms.
Hostname: ted
SUNW,eri0 : 100 Mbps full duplex link up
/dev/md/rdsk/d40 is clean
/dev/md/rdsk/d50 is clean
Reading ZFS config: done.

ted console login: root
Password:
Nov  9 16:56:49 ted login: ROOT LOGIN /dev/console
Last login: Sun Nov  9 16:14:35 on console
Sun Microsystems Inc.   SunOS 5.10      Generic January 2005
# zpool list
NAME    SIZE   USED  AVAIL    CAP  HEALTH  ALTROOT
rpool      -      -      -      -  FAULTED  -
# zpool status rpool
  pool: rpool
 state: UNAVAIL
 scrub: none requested
config:

        NAME          STATE     READ WRITE CKSUM
        rpool         UNAVAIL      0     0     0  insufficient replicas
          mirror      UNAVAIL      0     0     0  corrupted data
            c1t0d0s0  ONLINE       0     0     0
            c1t1d0s0  ONLINE       0     0     0
# df -h
Filesystem             size   used  avail capacity  Mounted on
/dev/md/dsk/d10        9.8G   4.3G   5.5G    44%    /
/devices                 0K     0K     0K     0%    /devices
ctfs                     0K     0K     0K     0%    /system/contract
proc                     0K     0K     0K     0%    /proc
mnttab                   0K     0K     0K     0%    /etc/mnttab
swap                   4.2G   1.5M   4.2G     1%    /etc/svc/volatile
objfs                    0K     0K     0K     0%    /system/object
sharefs                  0K     0K     0K     0%    /etc/dfs/sharetab
fd                       0K     0K     0K     0%    /dev/fd
swap                   4.2G    32K   4.2G     1%    /tmp
swap                   4.2G    48K   4.2G     1%    /var/run
/dev/md/dsk/d50        480M   1.0M   431M     1%    /globaldevices
/dev/md/dsk/d40         11G    11M    11G     1%    /export/home
# metastat -c
d50              m  512MB d51
    d51          s  512MB c1t1d0s5
d40              m   11GB d41
    d41          s   11GB c1t1d0s4
d30              m   10GB d31
    d31          s   10GB c1t1d0s3
d20              m  2.0GB d21
    d21          s  2.0GB c1t1d0s1
d10              m   10GB d11
    d11          s   10GB c1t1d0s0
# zpool export rpool
internal error: Invalid argument
Abort - core dumped
# zfs destroy -f rpool
internal error: Invalid argument
Abort - core dumped
#
-- 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to