Hi Tundra, I see two problems in your configuration 1. Keeping the dependencies. And you have answered it already. common_zone -> personal pool -> common pool This ensure proper start and stop and does the mounting and unmounting of the file system in ZFS pool. 2. > <zone name="common" zonepath="/common_pool0/common_zone" autoboot="false" brand="ipkg" limitpriv="default,sys_smb"> > <dataset name="personal_pool0/personal"/> > </zone>
In general, it is not recommended to add the a ZFS pool dataset to zone using zonecfg(1M), that is being controlled by HAStoragePlus. The reason is when a pool is imported on another physical cluster node as part failover/switchover, the booting of zone on the current node will have a problem as the dataset is not available. I suggest to remove the dataset from zonecfg(1M) and also tunn off the zoned property of that file system. Thanks -Venku On 12/04/09 21:19, Tundra Slosek wrote: >> Hi, >> What is the mountpoint of the pool file systems which >> are set to legacy. >> The following output after the resources online might >> help. >> # clrg online + >> # zpool list >> # zfs list >> # df -F zfs >> > > Venku, I think you are onto it - but I'm not sure how to proceed from here. > > So the zone 'common_shares' has TWO pools within it, 'common_pool0' and > 'personal_pool0'. common_pool0 is relatively small and is the root to the > zone, personal_pool0 is relatively large and is the place that all user data > will end up in. > > I think the problem is that I don't have a dependancy between the 'common' > zone and the 'personal_pool0' zpool anywhere. Since the zfs > personal_pool0/personal is an added dataset in the common zone, should I > change it so that personal_pool depends on common_zpool and then common_zone > depends on personal_pool (instead of depending on common_zpool as it does > now)? Would that get the startup and shutdown sequence dependancies correct? > > I've moved all other resource groups off the node with common_shares, just to > keep these listings down to a managable length. > > --- START of output --- > root at mltproc0:~# zpool list > NAME SIZE USED AVAIL CAP HEALTH ALTROOT > common_pool0 1.98G 1.25G 756M 62% ONLINE / > personal_pool0 31.8G 2.52M 31.7G 0% ONLINE / > rpool 139G 15.2G 124G 10% ONLINE - > root at mltproc0:~# zfs list > NAME USED AVAIL REFER > MOUNTPOINT > common_pool0 1.25G 724M 21K > /common_pool0 > common_pool0/common_zone 1.24G 724M 27.5K > /common_pool0/common_zone > common_pool0/common_zone/ROOT 1.24G 724M 19K > legacy > common_pool0/common_zone/ROOT/zbe 1.24G 724M 1.24G > legacy > personal_pool0 2.50M 31.3G 19K > /personal_pool0 > personal_pool0/personal 426K 31.3G 26K > /personal_pool0/personal > personal_pool0/personal/windows_home 248K 31.3G 46K > /personal_pool0/personal/windows_home > personal_pool0/personal/windows_home/tslosek_admin 34K 31.3G 19K > /personal_pool0/personal/windows_home/tslosek_admin > rpool 55.1G 81.7G 84.5K > /rpool > rpool/ROOT 4.96G 81.7G 19K > legacy > rpool/ROOT/2009-10-01 6.35M 81.7G 3.10G / > rpool/ROOT/opensolaris 11.4M 81.7G 2.87G / > rpool/ROOT/snv_111b-2009-10-06-cluster 35.2M 81.7G 3.60G / > rpool/ROOT/snv_111b-2009-10-12_clusterxen 4.90G 81.7G 4.12G / > rpool/dump 4.00G 81.7G 4.00G - > rpool/export 2.19G 81.7G 21K > /export > rpool/export/home 2.19G 81.7G 21K > /export/home > rpool/export/home/tslosek 2.19G 81.7G 2.19G > /export/home/tslosek > rpool/export/home/tslosek/bonnie 542K 81.7G 542K > /export/home/tslosek/bonnie > rpool/perftest0 20G 98.0G 3.73G - > rpool/perftest1 20G 102G 16K - > rpool/swap 4.00G 85.4G 321M - > rpool/xendumps 19K 81.7G 19K > /var/xen/dump,quota=4G > root at mltproc0:~# df -F zfs > Filesystem 1K-blocks Used Available Use% Mounted on > rpool/ROOT/snv_111b-2009-10-12_clusterxen > 89972381 4320736 85651645 5% / > rpool/export 85651666 21 85651645 1% /export > rpool/export/home 85651666 21 85651645 1% /export/home > rpool/export/home/tslosek > 87944494 2292850 85651645 3% /export/home/tslosek > rpool/export/home/tslosek/bonnie > 85652187 542 85651645 1% > /export/home/tslosek/bonnie > rpool 85651729 85 85651645 1% /rpool > rpool/xendumps 85651664 19 85651645 1% /var/xen/dump,quota=4G > personal_pool0 32769550 19 32769531 1% /personal_pool0 > common_pool0 741848 21 741827 1% /common_pool0 > common_pool0/common_zone > 741854 28 741827 1% > /common_pool0/common_zone > common_pool0/common_zone/ROOT/zbe > 2047207 1305380 741827 64% > /common_pool0/common_zone/root > personal_pool0/personal > 32769557 26 32769531 1% > /common_pool0/common_zone/root/personal_pool0/personal > personal_pool0/personal/windows_home > 32769577 46 32769531 1% > /common_pool0/common_zone/root/personal_pool0/personal/windows_home > personal_pool0/personal/windows_home/tslosek_admin > 32769550 19 32769531 1% > /common_pool0/common_zone/root/personal_pool0/personal/windows_home/tslosek_admin > > root at mltproc0:~# cat /etc/zones/common.xml > <?xml version="1.0" encoding="UTF-8"?> > <!DOCTYPE zone PUBLIC "-//Sun Microsystems Inc//DTD Zones//EN" > "file:///usr/share/lib/xml/dtd/zonecfg.dtd.1"> > <!-- > DO NOT EDIT THIS FILE. Use zonecfg(1M) instead. > --> > <zone name="common" zonepath="/common_pool0/common_zone" autoboot="false" > brand="ipkg" limitpriv="default,sys_smb"> > <dataset name="personal_pool0/personal"/> > </zone> > root at mltproc0:~# /usr/cluster/bin/clrg show common_shares | grep Resource: > Resource: personal_pool > Resource: common_lhname > Resource: common_zpool > Resource: common_zone > root at mltproc0:~# /usr/cluster/bin/clresource show -y Resource_dependencies > common_zone > > === Resources === > > Resource: common_zone > Resource_dependencies: common_lhname common_zpool > > --- END of output --- > the other resources in the common_shares group do not have any > Resource_dependencies