Gaurav,

Jun 29 20:26:08 opensolaris gdm-binary[6634]: [ID 702911 daemon.warning] 
WARNING: Failed to start X server several times in a short time period; 
disabling display :0

Sadly, there's no indication, as to WHY that fails...

And: It seems, you have a single-node cluster, right?

Jun 23 12:51:47 opensolaris genunix: [ID 965873 kern.notice] NOTICE: CMM: Node 
opensolaris (nodeid = 1) with votecount = 1 added.
Jun 23 12:51:47 opensolaris genunix: [ID 843983 kern.notice] NOTICE: CMM: Node 
opensolaris: attempting to join cluster.
Jun 23 12:51:47 opensolaris genunix: [ID 525628 kern.notice] NOTICE: CMM: 
Cluster has reached quorum.
Jun 23 12:51:47 opensolaris genunix: [ID 377347 kern.notice] NOTICE: CMM: Node 
opensolaris (nodeid = 1) is up; new incarnation number = 1277277706.
Jun 23 12:51:47 opensolaris genunix: [ID 108990 kern.notice] NOTICE: CMM: 
Cluster members: opensolaris.
Jun 23 12:51:47 opensolaris genunix: [ID 279084 kern.notice] NOTICE: CMM: node 
reconfiguration #1 completed.
Jun 23 12:51:50 opensolaris genunix: [ID 499756 kern.notice] NOTICE: CMM: Node 
opensolaris: joined cluster.

And you have same errors in your config of the zones for the cluster, as far
as the storage setup is concerned:

Jun 23 13:09:52 opensolaris Cluster.RGM.global.rgmd: [ID 224900 daemon.notice] 
launching method <hastorageplus_validate> for resource <clusterresource>, 
resource group <mygroup>, node <opensolaris>, timeout <1800> seconds
Jun 23 13:09:52 opensolaris 
SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: [ID 
148650 daemon.notice] Started searching for devices in '/dev/dsk' to find the 
importable pools.
Jun 23 13:09:53 opensolaris 
SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: [ID 
547433 daemon.notice] Completed searching the devices in '/dev/dsk' to find the 
importable pools.
Jun 23 13:09:53 opensolaris 
SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: [ID 
792255 daemon.warning] Failed to update the cachefile contents in 
/var/cluster/run/HAStoragePlus/zfs/rpool.cachefile to CCR table rpool.cachefile 
for pool rpool : file /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile open 
failed: No such file or directory.
Jun 23 13:09:54 opensolaris 
SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: [ID 
471757 daemon.error] cannot unmount '/' : Device busy
Jun 23 13:09:54 opensolaris 
SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: [ID 
316615 daemon.error] Failed to export :rpool

And a bit later:

Jun 23 13:11:15 opensolaris 
SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_init]: [ID 
792255 daemon.warning] Failed to update the cachefile contents in 
/var/cluster/run/HAStoragePlus/zfs/rpool.cachefile to CCR table rpool.cachefile 
for pool rpool : file /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile open 
failed: No such file or directory.
Jun 23 13:11:15 opensolaris 
SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_init]: [ID 
471757 daemon.error] cannot unmount '/' : Device busy
Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 515159 daemon.notice] 
method <hastorageplus_init> completed successfully for resource 
<clusterresource>, resource group <mygroup>, node <opensolaris>, time used: 0% 
of timeout <1800 seconds>
Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 443746 daemon.notice] 
resource clusterresource state on node opensolaris change to R_OFFLINE
Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 529407 daemon.notice] 
resource group mygroup state on node opensolaris change to RG_OFFLINE
Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 529407 daemon.notice] 
resource group mygroup state on node opensolaris change to RG_PENDING_ONLINE
Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 784560 daemon.notice] 
resource clusterresource status on node opensolaris change to R_FM_UNKNOWN

So, some of these problems (X server start et.al.) might be related to these
errros.

First: The rpool is a local filesystem (better: a local pool), as it contains
the root-filesystem This can not be made as an fail-over filesystem. But it
seems, you did configure that way.

So, if I were you, I would delete all the CONFIGURATIONS for things in the
cluster, so that you get a fresh OpenSolaris 2009.06 with the OHAC packages
installed.

Then, I would reboot the server (are you sure, you want only a single node
cluster? Or am I missing something here?), and see, if the X-server comes up
again. Once that running again, I would recommend re-reading the basic cluster
setup manuals for how to setup a failover zone and the storage needed to host
the filesystems for those zones...

    Matthias

You (gaurav saxena) wrote:
> Hello
> 
> 2010/6/29 Matthias Pfützner <matth...@pfuetzner.de>
> 
> > Can you just send me the last 50 lines or so?
> >
> >
> I have attached a text file containing complete output. :)
> 
> -- 
> Thanks and Regards ,
> Gaurav Saxena
-- 
    Matthias Pfützner    | mailto:pfu...@germany | Am Anfang wäre das Stück
 @work: +49 6103 752-394 | @home: +49 6151 75717 | schnell zu Ende.
  SunCS, Ampèrestraße 6  | Lichtenbergstraße 73  | (21.10.93)
    63225 Langen, FRG    | 64289 Darmstadt, FRG  | Hellmuth Karasek
_______________________________________________
zones-discuss mailing list
zones-discuss@opensolaris.org

Reply via email to