This is still bugging me.

At home:

Upgrading b103 -> b115.

jana:~> zpool status
  pool: jana
 state: ONLINE
 scrub: none requested
config:

        NAME        STATE     READ WRITE CKSUM
        jana        ONLINE       0     0     0
          c1d0s0    ONLINE       0     0     0

errors: No known data errors

  pool: raid
 state: ONLINE
status: The pool is formatted using an older on-disk format.  The pool can
        still be used, but some features are unavailable.
action: Upgrade the pool using 'zpool upgrade'.  Once this is done, the
        pool will no longer be accessible on older software versions.
 scrub: scrub completed after 1h40m with 0 errors on Mon May 18 17:50:12 2009
config:

        NAME        STATE     READ WRITE CKSUM
        raid        ONLINE       0     0     0
          raidz1    ONLINE       0     0     0
            c0t0d0  ONLINE       0     0     0
            c0t1d0  ONLINE       0     0     0
            c0t2d0  ONLINE       0     0     0
            c0t3d0  ONLINE       0     0     0
            c0t4d0  ONLINE       0     0     0

errors: No known data errors
jana:~> zfs list 
NAME                     USED  AVAIL  REFER  MOUNTPOINT
jana                    10.1G  62.7G    40K  /jana
jana/ROOT               7.10G  62.7G    18K  legacy
jana/ROOT/snv_103       6.99G  62.7G  6.95G  /
jana/dump               1.00G  62.7G  1.00G  -
jana/swap                  2G  64.7G    16K  -
raid                     878G   952G  67.1K  /raid
raid/applications       17.7G   952G  9.41G  /raid/applications
raid/backup             75.4G   952G  36.7K  /raid/backup
raid/backup/bender      32.1G   952G  32.1G  /raid/backup/bender
raid/backup/betty       8.15G   952G  8.15G  /raid/backup/betty
raid/backup/holly       8.34G   952G  8.34G  /raid/backup/holly
raid/backup/jana        2.18G   952G   331M  /raid/backup/jana
raid/backup/zoe         24.7G   952G  24.7G  /raid/backup/zoe
raid/database           11.7G   952G  11.7G  /raid/database
raid/drivers            2.31G   952G  1.54G  /raid/drivers
raid/dvds               2.87G   952G  2.87G  /raid/dvds
raid/ebooks             3.70G   952G  2.75G  /raid/ebooks
raid/emulators          1.72G   952G  1.72G  /raid/emulators
raid/fonts               729K   952G   729K  /raid/fonts
raid/forex              1.73G   952G  1.69G  /raid/forex
raid/games              24.4G   952G  22.6G  /raid/games
raid/home               61.2G   952G  32.0K  /raid/home
raid/home/bridget       9.28G   952G  8.05G  /raid/home/bridget
raid/home/martin        52.0G   952G  34.0G  /raid/home/martin
raid/management          375K   952G  37.5K  /raid/management
raid/movies             15.5G   952G  15.3G  /raid/movies
raid/music               169M   952G   169M  /raid/music
raid/operating_systems  80.7G   952G  38.0G  /raid/operating_systems
raid/people             91.7M   952G  87.4M  /raid/people
raid/photos             6.51G   952G  6.39G  /raid/photos
raid/pictures           9.05G   952G  8.93G  /raid/pictures
raid/software           28.7G   952G  22.1G  /raid/software
raid/temp               8.34G   952G  6.72G  /raid/temp
raid/tv_shows            451G   952G   451G  /raid/tv_shows
jana:~> 


jana:/etc# lofiadm -a 
/raid/operating_systems/Solaris/SXCE/sol-nv-b115-x86-dvd.iso 
/dev/lofi/1
jana:/etc# mount -o ro -F hsfs -o ro /dev/lofi/1 /mnt
jana:/etc# lustatus
Boot Environment           Is       Active Active    Can    Copy      
Name                       Complete Now    On Reboot Delete Status    
-------------------------- -------- ------ --------- ------ ----------
snv_103                    yes      yes    yes       no     -         
jana:/etc#

jana:/etc# zfs snapshot -r jana/ROOT/snv_103 at preupgrade
jana:/etc# zfs list -t all -r jana
NAME                           USED  AVAIL  REFER  MOUNTPOINT
jana                          9.96G  62.9G    40K  /jana
jana/ROOT                     6.95G  62.9G    18K  legacy
jana/ROOT/snv_103             6.95G  62.9G  6.95G  /
jana/ROOT/snv_103 at preupgrade      0      -  6.95G  -
jana/dump                     1.00G  62.9G  1.00G  -
jana/swap                        2G  64.9G    16K  -
jana:/etc# /mnt/Solaris_11/Tools/Installers/
liveupgrade20* solarisn*      
jana:/etc# /mnt/Solaris_11/Tools/Installers/liveupgrade20 

...

jana:/etc# lucreate -n snv_115
Checking GRUB menu...
System has findroot enabled GRUB
Analyzing system configuration.
Comparing source boot environment <snv_103> file systems with the file 
system(s) you specified for the new boot environment. Determining which 
file systems should be in the new boot environment.
Updating boot environment description database on all BEs.
Updating system configuration files.
Creating configuration for boot environment <snv_115>.
Source boot environment is <snv_103>.
Creating boot environment <snv_115>.
Cloning file systems from boot environment <snv_103> to create boot environment 
<snv_115>.
Creating snapshot for <jana/ROOT/snv_103> on <jana/ROOT/snv_103 at snv_115>.
Creating clone for <jana/ROOT/snv_103 at snv_115> on <jana/ROOT/snv_115>.
Setting canmount=noauto for </> in zone <global> on <jana/ROOT/snv_115>.
Saving existing file </boot/grub/menu.lst> in top level dataset for BE 
<snv_115> as <mount-point>//boot/grub/menu.lst.prev.
File </boot/grub/menu.lst> propagation successful
Copied GRUB menu from PBE to ABE
No entry for BE <snv_115> in GRUB menu
Population of boot environment <snv_115> successful.
Creation of boot environment <snv_115> successful.
jana:/etc# lustatus
Boot Environment           Is       Active Active    Can    Copy      
Name                       Complete Now    On Reboot Delete Status    
-------------------------- -------- ------ --------- ------ ----------
snv_103                    yes      yes    yes       no     -         
snv_115                    yes      no     no        yes    -         
jana:/etc# zfs list -t all -r jana
NAME                           USED  AVAIL  REFER  MOUNTPOINT
jana                          9.99G  62.9G    40K  /jana
jana/ROOT                     6.98G  62.9G    18K  legacy
jana/ROOT/snv_103             6.98G  62.9G  6.95G  /
jana/ROOT/snv_103 at preupgrade  33.9M      -  6.95G  -
jana/ROOT/snv_103 at snv_115     83.5K      -  6.95G  -
jana/ROOT/snv_115              169K  62.9G  6.95G  /tmp/.alt.luupdall.22680
jana/dump                     1.00G  62.9G  1.00G  -
jana/swap                        2G  64.9G    16K  -
jana:/etc#
jana:/etc# luupgrade -u -n snv_115 -s /mnt

System has findroot enabled GRUB
No entry for BE <snv_115> in GRUB menu
Uncompressing miniroot
Copying failsafe kernel from media.
52479 blocks
miniroot filesystem is <lofs>
Mounting miniroot at </mnt/Solaris_11/Tools/Boot>
Validating the contents of the media </mnt>.
The media is a standard Solaris media.
The media contains an operating system upgrade image.
The media contains <Solaris> version <11>.
Constructing upgrade profile to use.
Locating the operating system upgrade program.
Checking for existence of previously scheduled Live Upgrade requests.
Creating upgrade profile for BE <snv_115>.
Checking for GRUB menu on ABE <snv_115>.
Saving GRUB menu on ABE <snv_115>.
Checking for x86 boot partition on ABE.
Determining packages to install or upgrade for BE <snv_115>.
Performing the operating system upgrade of the BE <snv_115>.
CAUTION: Interrupting this process may leave the boot environment unstable 
or unbootable.
Upgrading Solaris: 2% completed
...

And its running fine. Without needing to do a 'zpool export raid' to export my 
data pool.

Yet at work, on ANY of our servers, it fails. Every time.

This output is for our production server zeus, though the problem also happens 
to athena and artemis which I posted about in the past. I upgraded them 
yesterday using the 'zpool export' work around I discovered so I can't provide 
that output now, but here it is from our production server.

(I'm masking the pool name as it is a very unique name and is the name of our 
company and I do not want any of this showing up in searches :).

zeus:~# lustatus
Boot Environment           Is       Active Active    Can    Copy      
Name                       Complete Now    On Reboot Delete Status    
-------------------------- -------- ------ --------- ------ ----------
snv_101                    yes      yes    yes       no     -         
snv_115                    yes      no     no        yes    -         
zeus:~# zpool status
  pool: **datapool
 state: ONLINE
 scrub: resilver completed after 42h18m with 0 errors on Thu Jun 18 05:28:49 
2009
config:

        NAME         STATE     READ WRITE CKSUM
        **datapool   ONLINE       0     0     0
          raidz2     ONLINE       0     0     0
            c0t0d0   ONLINE       0     0     0  2.05G resilvered
            c1t8d0   ONLINE       0     0     0  1.98G resilvered
            c0t1d0   ONLINE       0     0     0  610G resilvered
            c1t9d0   ONLINE       0     0     0  1.78G resilvered
            c0t2d0   ONLINE       0     0     0  2.05G resilvered
            c1t10d0  ONLINE       0     0     0  1.98G resilvered
            c0t3d0   ONLINE       0     0     0  1.90G resilvered
            c1t11d0  ONLINE       0     0     0  1.78G resilvered
            c0t4d0   ONLINE       0     0     0  2.05G resilvered
            c1t12d0  ONLINE       0     0     0  1.98G resilvered
            c0t5d0   ONLINE       0     0     0  1.90G resilvered
            c1t13d0  ONLINE       0     0     0  1.78G resilvered
            c0t6d0   ONLINE       0     0     0  2.06G resilvered
            c1t14d0  ONLINE       0     0     0  1.97G resilvered
            c0t7d0   ONLINE       0     0     0  1.89G resilvered
            c1t15d0  ONLINE       0     0     0  1.77G resilvered

errors: No known data errors

  pool: zeus
 state: ONLINE
 scrub: scrub completed after 0h12m with 0 errors on Tue Jun 16 11:21:18 2009
config:

        NAME        STATE     READ WRITE CKSUM
        zeus        ONLINE       0     0     0
          mirror    ONLINE       0     0     0
            c4d0s0  ONLINE       0     0     0
            c5d0s0  ONLINE       0     0     0

errors: No known data errors
zeus:~# 

So lets try to upgrade:

zeus:~# luupgrade -u -n snv_115 -s /mnt

System has findroot enabled GRUB
No entry for BE <snv_115> in GRUB menu
Uncompressing miniroot
Copying failsafe kernel from media.
52479 blocks
miniroot filesystem is <lofs>
Mounting miniroot at </mnt/Solaris_11/Tools/Boot>
Validating the contents of the media </mnt>.
The media is a standard Solaris media.
The media contains an operating system upgrade image.
The media contains <Solaris> version <11>.
Constructing upgrade profile to use.
Locating the operating system upgrade program.
Checking for existence of previously scheduled Live Upgrade requests.
Creating upgrade profile for BE <snv_115>.
Checking for GRUB menu on ABE <snv_115>.
Saving GRUB menu on ABE <snv_115>.
Checking for x86 boot partition on ABE.
Determining packages to install or upgrade for BE <snv_115>.
Performing the operating system upgrade of the BE <snv_115>.
CAUTION: Interrupting this process may leave the boot environment unstable 
or unbootable.
ERROR: Installation of the packages from this media of the media failed; 
pfinstall returned these diagnostics:

Processing profile

Loading local environment and services

Generating upgrade actions
WARNING: SUNWlang-en depends on SUNWlang-enUS, which is not selected

ERROR: No upgradeable file systems found at specified mount point.
Restoring GRUB menu on ABE <snv_115>.
ABE boot partition backing deleted.
PBE GRUB has no capability information.
PBE GRUB has no versioning information.
ABE GRUB is newer than PBE GRUB. Updating GRUB.
GRUB update was successful.
Configuring failsafe for system.
Failsafe configuration is complete.
The Solaris upgrade of the boot environment <snv_115> failed.
Installing failsafe
Failsafe install is complete.
zeus:~# 

The problem is clearly "ERROR: No upgradeable file systems found at specified 
mount point."

But WHY ? What is it about the datapool that is confusing it? The large number 
of filesystems we have (over 200) and snapshots (over a million) ??

The ONLY work around that I know of is to 'zpool export datapool' being 
whatever the datapool is on the server in question. This worked to upgrade both 
athena and artemis from snv_101 to snv_103, and yesterday from snv_103 to 
snv_115.

This bug has been here a while now, should it be reported? Am I simply doing 
something wrong? Exporting the pool on the backup hosts is fine but NOT fine on 
our production server...

Note: I AM CCing this posting to the 'install' forums as someone suggested they 
might better know what the problem is.
-- 
This message posted from opensolaris.org

Reply via email to