Hello

2010/6/30 Matthias Pfützner <matth...@pfuetzner.de>

> Gaurav,
>
> Jun 29 20:26:08 opensolaris gdm-binary[6634]: [ID 702911 daemon.warning]
> WARNING: Failed to start X server several times in a short time period;
> disabling display :0
>
> Sadly, there's no indication, as to WHY that fails...
>
>
Ok. yes its really sad :(.


> And: It seems, you have a single-node cluster, right?
>
>
Yes I am having a single node cluster.


> Jun 23 12:51:47 opensolaris genunix: [ID 965873 kern.notice] NOTICE: CMM:
> Node opensolaris (nodeid = 1) with votecount = 1 added.
> Jun 23 12:51:47 opensolaris genunix: [ID 843983 kern.notice] NOTICE: CMM:
> Node opensolaris: attempting to join cluster.
> Jun 23 12:51:47 opensolaris genunix: [ID 525628 kern.notice] NOTICE: CMM:
> Cluster has reached quorum.
> Jun 23 12:51:47 opensolaris genunix: [ID 377347 kern.notice] NOTICE: CMM:
> Node opensolaris (nodeid = 1) is up; new incarnation number = 1277277706.
> Jun 23 12:51:47 opensolaris genunix: [ID 108990 kern.notice] NOTICE: CMM:
> Cluster members: opensolaris.
> Jun 23 12:51:47 opensolaris genunix: [ID 279084 kern.notice] NOTICE: CMM:
> node reconfiguration #1 completed.
> Jun 23 12:51:50 opensolaris genunix: [ID 499756 kern.notice] NOTICE: CMM:
> Node opensolaris: joined cluster.
>
> And you have same errors in your config of the zones for the cluster, as
> far
> as the storage setup is concerned:
>
>
I see these messages at boot time. Are these errors ?


> Jun 23 13:09:52 opensolaris Cluster.RGM.global.rgmd: [ID 224900
> daemon.notice] launching method <hastorageplus_validate> for resource
> <clusterresource>, resource group <mygroup>, node <opensolaris>, timeout
> <1800> seconds
> Jun 23 13:09:52 opensolaris
> SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]:
> [ID 148650 daemon.notice] Started searching for devices in '/dev/dsk' to
> find the importable pools.
> Jun 23 13:09:53 opensolaris
> SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]:
> [ID 547433 daemon.notice] Completed searching the devices in '/dev/dsk' to
> find the importable pools.
> Jun 23 13:09:53 opensolaris
> SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]:
> [ID 792255 daemon.warning] Failed to update the cachefile contents in
> /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile to CCR table
> rpool.cachefile for pool rpool : file
> /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile open failed: No such file
> or directory.
> Jun 23 13:09:54 opensolaris
> SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]:
> [ID 471757 daemon.error] cannot unmount '/' : Device busy
> Jun 23 13:09:54 opensolaris
> SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]:
> [ID 316615 daemon.error] Failed to export :rpool
>
> And a bit later:
>
> Jun 23 13:11:15 opensolaris
> SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_init]: [ID
> 792255 daemon.warning] Failed to update the cachefile contents in
> /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile to CCR table
> rpool.cachefile for pool rpool : file
> /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile open failed: No such file
> or directory.
> Jun 23 13:11:15 opensolaris
> SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_init]: [ID
> 471757 daemon.error] cannot unmount '/' : Device busy
> Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 515159
> daemon.notice] method <hastorageplus_init> completed successfully for
> resource <clusterresource>, resource group <mygroup>, node <opensolaris>,
> time used: 0% of timeout <1800 seconds>
> Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 443746
> daemon.notice] resource clusterresource state on node opensolaris change to
> R_OFFLINE
> Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 529407
> daemon.notice] resource group mygroup state on node opensolaris change to
> RG_OFFLINE
> Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 529407
> daemon.notice] resource group mygroup state on node opensolaris change to
> RG_PENDING_ONLINE
> Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 784560
> daemon.notice] resource clusterresource status on node opensolaris change to
> R_FM_UNKNOWN
>
> So, some of these problems (X server start et.al.) might be related to
> these
> errros.
>
> First: The rpool is a local filesystem (better: a local pool), as it
> contains
> the root-filesystem This can not be made as an fail-over filesystem. But it
> seems, you did configure that way.
>
>
No actually i just created two zfs filesystems on rpool as I have told
earlier.
Is it wrong ?


> So, if I were you, I would delete all the CONFIGURATIONS for things in the
> cluster, so that you get a fresh OpenSolaris 2009.06 with the OHAC packages
> installed.
>
> I think I can reinstall opensolaris only and install OHAC on it again but I
have reinstalled this for the second time only. :(


> Then, I would reboot the server (are you sure, you want only a single node
> cluster? Or am I missing something here?), and see, if the X-server comes
> up
> again. Once that running again, I would recommend re-reading the basic
> cluster
> setup manuals for how to setup a failover zone and the storage needed to
> host
> the filesystems for those zones...
>
>    Matthias--
>    Matthias Pfützner    | mailto:pfu...@germany | Am Anfang wäre das Stück
>  @work: +49 6103 752-394 | @home: +49 6151 75717 | schnell zu Ende.
>  SunCS, Ampčrestraße 6  | Lichtenbergstraße 73  | (21.10.93)
>    63225 Langen, FRG    | 64289 Darmstadt, FRG  | Hellmuth Karasek
>
Actually in the manual it was suggested to create a mirror zpool of the root
filesystem and then install zfs on it or create a pool on file but I thought
I could use the original rpool only.
Is this the reason for all problems ?
But as far as I remember display is not working since I ran scinstall and I
did this all zone work from the console only so I don't think this could be
the reason.


-- 
Thanks and Regards ,
Gaurav Saxena
_______________________________________________
zones-discuss mailing list
zones-discuss@opensolaris.org

Reply via email to