On 06/18/09 23:23, Turanga Leela wrote: > This is still bugging me. > > At home: > > Upgrading b103 -> b115. > > jana:~> zpool status > pool: jana > state: ONLINE > scrub: none requested > config: > > NAME STATE READ WRITE CKSUM > jana ONLINE 0 0 0 > c1d0s0 ONLINE 0 0 0 > > errors: No known data errors > > pool: raid > state: ONLINE > status: The pool is formatted using an older on-disk format. The pool can > still be used, but some features are unavailable. > action: Upgrade the pool using 'zpool upgrade'. Once this is done, the > pool will no longer be accessible on older software versions. > scrub: scrub completed after 1h40m with 0 errors on Mon May 18 17:50:12 2009 > config: > > NAME STATE READ WRITE CKSUM > raid ONLINE 0 0 0 > raidz1 ONLINE 0 0 0 > c0t0d0 ONLINE 0 0 0 > c0t1d0 ONLINE 0 0 0 > c0t2d0 ONLINE 0 0 0 > c0t3d0 ONLINE 0 0 0 > c0t4d0 ONLINE 0 0 0 > > errors: No known data errors > jana:~> zfs list > NAME USED AVAIL REFER MOUNTPOINT > jana 10.1G 62.7G 40K /jana > jana/ROOT 7.10G 62.7G 18K legacy > jana/ROOT/snv_103 6.99G 62.7G 6.95G / > jana/dump 1.00G 62.7G 1.00G - > jana/swap 2G 64.7G 16K - > raid 878G 952G 67.1K /raid > raid/applications 17.7G 952G 9.41G /raid/applications > raid/backup 75.4G 952G 36.7K /raid/backup > raid/backup/bender 32.1G 952G 32.1G /raid/backup/bender > raid/backup/betty 8.15G 952G 8.15G /raid/backup/betty > raid/backup/holly 8.34G 952G 8.34G /raid/backup/holly > raid/backup/jana 2.18G 952G 331M /raid/backup/jana > raid/backup/zoe 24.7G 952G 24.7G /raid/backup/zoe > raid/database 11.7G 952G 11.7G /raid/database > raid/drivers 2.31G 952G 1.54G /raid/drivers > raid/dvds 2.87G 952G 2.87G /raid/dvds > raid/ebooks 3.70G 952G 2.75G /raid/ebooks > raid/emulators 1.72G 952G 1.72G /raid/emulators > raid/fonts 729K 952G 729K /raid/fonts > raid/forex 1.73G 952G 1.69G /raid/forex > raid/games 24.4G 952G 22.6G /raid/games > raid/home 61.2G 952G 32.0K /raid/home > raid/home/bridget 9.28G 952G 8.05G /raid/home/bridget > raid/home/martin 52.0G 952G 34.0G /raid/home/martin > raid/management 375K 952G 37.5K /raid/management > raid/movies 15.5G 952G 15.3G /raid/movies > raid/music 169M 952G 169M /raid/music > raid/operating_systems 80.7G 952G 38.0G /raid/operating_systems > raid/people 91.7M 952G 87.4M /raid/people > raid/photos 6.51G 952G 6.39G /raid/photos > raid/pictures 9.05G 952G 8.93G /raid/pictures > raid/software 28.7G 952G 22.1G /raid/software > raid/temp 8.34G 952G 6.72G /raid/temp > raid/tv_shows 451G 952G 451G /raid/tv_shows > jana:~> > > > jana:/etc# lofiadm -a > /raid/operating_systems/Solaris/SXCE/sol-nv-b115-x86-dvd.iso > /dev/lofi/1 > jana:/etc# mount -o ro -F hsfs -o ro /dev/lofi/1 /mnt > jana:/etc# lustatus > Boot Environment Is Active Active Can Copy > Name Complete Now On Reboot Delete Status > -------------------------- -------- ------ --------- ------ ---------- > snv_103 yes yes yes no - > jana:/etc# > > jana:/etc# zfs snapshot -r jana/ROOT/snv_103 at preupgrade > jana:/etc# zfs list -t all -r jana > NAME USED AVAIL REFER MOUNTPOINT > jana 9.96G 62.9G 40K /jana > jana/ROOT 6.95G 62.9G 18K legacy > jana/ROOT/snv_103 6.95G 62.9G 6.95G / > jana/ROOT/snv_103 at preupgrade 0 - 6.95G - > jana/dump 1.00G 62.9G 1.00G - > jana/swap 2G 64.9G 16K - > jana:/etc# /mnt/Solaris_11/Tools/Installers/ > liveupgrade20* solarisn* > jana:/etc# /mnt/Solaris_11/Tools/Installers/liveupgrade20 > > ... > > jana:/etc# lucreate -n snv_115 > Checking GRUB menu... > System has findroot enabled GRUB > Analyzing system configuration. > Comparing source boot environment <snv_103> file systems with the file > system(s) you specified for the new boot environment. Determining which > file systems should be in the new boot environment. > Updating boot environment description database on all BEs. > Updating system configuration files. > Creating configuration for boot environment <snv_115>. > Source boot environment is <snv_103>. > Creating boot environment <snv_115>. > Cloning file systems from boot environment <snv_103> to create boot > environment <snv_115>. > Creating snapshot for <jana/ROOT/snv_103> on <jana/ROOT/snv_103 at snv_115>. > Creating clone for <jana/ROOT/snv_103 at snv_115> on <jana/ROOT/snv_115>. > Setting canmount=noauto for </> in zone <global> on <jana/ROOT/snv_115>. > Saving existing file </boot/grub/menu.lst> in top level dataset for BE > <snv_115> as <mount-point>//boot/grub/menu.lst.prev. > File </boot/grub/menu.lst> propagation successful > Copied GRUB menu from PBE to ABE > No entry for BE <snv_115> in GRUB menu > Population of boot environment <snv_115> successful. > Creation of boot environment <snv_115> successful. > jana:/etc# lustatus > Boot Environment Is Active Active Can Copy > Name Complete Now On Reboot Delete Status > -------------------------- -------- ------ --------- ------ ---------- > snv_103 yes yes yes no - > snv_115 yes no no yes - > jana:/etc# zfs list -t all -r jana > NAME USED AVAIL REFER MOUNTPOINT > jana 9.99G 62.9G 40K /jana > jana/ROOT 6.98G 62.9G 18K legacy > jana/ROOT/snv_103 6.98G 62.9G 6.95G / > jana/ROOT/snv_103 at preupgrade 33.9M - 6.95G - > jana/ROOT/snv_103 at snv_115 83.5K - 6.95G - > jana/ROOT/snv_115 169K 62.9G 6.95G /tmp/.alt.luupdall.22680 > jana/dump 1.00G 62.9G 1.00G - > jana/swap 2G 64.9G 16K - > jana:/etc# > jana:/etc# luupgrade -u -n snv_115 -s /mnt > > System has findroot enabled GRUB > No entry for BE <snv_115> in GRUB menu > Uncompressing miniroot > Copying failsafe kernel from media. > 52479 blocks > miniroot filesystem is <lofs> > Mounting miniroot at </mnt/Solaris_11/Tools/Boot> > Validating the contents of the media </mnt>. > The media is a standard Solaris media. > The media contains an operating system upgrade image. > The media contains <Solaris> version <11>. > Constructing upgrade profile to use. > Locating the operating system upgrade program. > Checking for existence of previously scheduled Live Upgrade requests. > Creating upgrade profile for BE <snv_115>. > Checking for GRUB menu on ABE <snv_115>. > Saving GRUB menu on ABE <snv_115>. > Checking for x86 boot partition on ABE. > Determining packages to install or upgrade for BE <snv_115>. > Performing the operating system upgrade of the BE <snv_115>. > CAUTION: Interrupting this process may leave the boot environment unstable > or unbootable. > Upgrading Solaris: 2% completed > ... > > And its running fine. Without needing to do a 'zpool export raid' to export > my data pool. > > Yet at work, on ANY of our servers, it fails. Every time. > > This output is for our production server zeus, though the problem also > happens to athena and artemis which I posted about in the past. I upgraded > them yesterday using the 'zpool export' work around I discovered so I can't > provide that output now, but here it is from our production server. > > (I'm masking the pool name as it is a very unique name and is the name of our > company and I do not want any of this showing up in searches :). > > zeus:~# lustatus > Boot Environment Is Active Active Can Copy > Name Complete Now On Reboot Delete Status > -------------------------- -------- ------ --------- ------ ---------- > snv_101 yes yes yes no - > snv_115 yes no no yes - > zeus:~# zpool status > pool: **datapool > state: ONLINE > scrub: resilver completed after 42h18m with 0 errors on Thu Jun 18 05:28:49 > 2009 > config: > > NAME STATE READ WRITE CKSUM > **datapool ONLINE 0 0 0 > raidz2 ONLINE 0 0 0 > c0t0d0 ONLINE 0 0 0 2.05G resilvered > c1t8d0 ONLINE 0 0 0 1.98G resilvered > c0t1d0 ONLINE 0 0 0 610G resilvered > c1t9d0 ONLINE 0 0 0 1.78G resilvered > c0t2d0 ONLINE 0 0 0 2.05G resilvered > c1t10d0 ONLINE 0 0 0 1.98G resilvered > c0t3d0 ONLINE 0 0 0 1.90G resilvered > c1t11d0 ONLINE 0 0 0 1.78G resilvered > c0t4d0 ONLINE 0 0 0 2.05G resilvered > c1t12d0 ONLINE 0 0 0 1.98G resilvered > c0t5d0 ONLINE 0 0 0 1.90G resilvered > c1t13d0 ONLINE 0 0 0 1.78G resilvered > c0t6d0 ONLINE 0 0 0 2.06G resilvered > c1t14d0 ONLINE 0 0 0 1.97G resilvered > c0t7d0 ONLINE 0 0 0 1.89G resilvered > c1t15d0 ONLINE 0 0 0 1.77G resilvered > > errors: No known data errors > > pool: zeus > state: ONLINE > scrub: scrub completed after 0h12m with 0 errors on Tue Jun 16 11:21:18 2009 > config: > > NAME STATE READ WRITE CKSUM > zeus ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c4d0s0 ONLINE 0 0 0 > c5d0s0 ONLINE 0 0 0 > > errors: No known data errors > zeus:~# > > So lets try to upgrade: > > zeus:~# luupgrade -u -n snv_115 -s /mnt > > System has findroot enabled GRUB > No entry for BE <snv_115> in GRUB menu > Uncompressing miniroot > Copying failsafe kernel from media. > 52479 blocks > miniroot filesystem is <lofs> > Mounting miniroot at </mnt/Solaris_11/Tools/Boot> > Validating the contents of the media </mnt>. > The media is a standard Solaris media. > The media contains an operating system upgrade image. > The media contains <Solaris> version <11>. > Constructing upgrade profile to use. > Locating the operating system upgrade program. > Checking for existence of previously scheduled Live Upgrade requests. > Creating upgrade profile for BE <snv_115>. > Checking for GRUB menu on ABE <snv_115>. > Saving GRUB menu on ABE <snv_115>. > Checking for x86 boot partition on ABE. > Determining packages to install or upgrade for BE <snv_115>. > Performing the operating system upgrade of the BE <snv_115>. > CAUTION: Interrupting this process may leave the boot environment unstable > or unbootable. > ERROR: Installation of the packages from this media of the media failed; > pfinstall returned these diagnostics: > > Processing profile > > Loading local environment and services > > Generating upgrade actions > WARNING: SUNWlang-en depends on SUNWlang-enUS, which is not selected > > ERROR: No upgradeable file systems found at specified mount point. > Restoring GRUB menu on ABE <snv_115>. > ABE boot partition backing deleted. > PBE GRUB has no capability information. > PBE GRUB has no versioning information. > ABE GRUB is newer than PBE GRUB. Updating GRUB. > GRUB update was successful. > Configuring failsafe for system. > Failsafe configuration is complete. > The Solaris upgrade of the boot environment <snv_115> failed. > Installing failsafe > Failsafe install is complete. > zeus:~# > > The problem is clearly "ERROR: No upgradeable file systems found at specified > mount point." > Is it possible that the directory /var/sadm in the BE being cloned was destroyed or modified in some way?
One of the ways that the installer recognizes an ungradeable file system is by the contents of the directory /var/sadm. For example, there must be a /var/sadm/system/admin/INST_RELEASE file and it must have contents like this: OS=Solaris VERSION=11 REV=0 There must also be a package database, which is composed of /var/sadm/install/contents and the directories in /var/sadm/pkg. If the system was ever BFU'd for example, the system is no longer upgradeable, since BFU invalidates the contents of the package database. Lori > But WHY ? What is it about the datapool that is confusing it? The large > number of filesystems we have (over 200) and snapshots (over a million) ?? > > The ONLY work around that I know of is to 'zpool export datapool' being > whatever the datapool is on the server in question. This worked to upgrade > both athena and artemis from snv_101 to snv_103, and yesterday from snv_103 > to snv_115. > > This bug has been here a while now, should it be reported? Am I simply doing > something wrong? Exporting the pool on the backup hosts is fine but NOT fine > on our production server... > > Note: I AM CCing this posting to the 'install' forums as someone suggested > they might better know what the problem is. >