Re: [zones-discuss] Duplicating zones ??
On 06/30/10 04:34, Warren Zeeman wrote: Hello, IHAC who wants to duplicate a global zone, as a zone on another server !!! Does anybody have any thoughts on the easiest way to achieve this ? We call this p2v (physical to virtual). Its been in opensolaris for quite a while now, so if you running a fairly recent build you already have this. I blogged about this early in 2009. http://blogs.sun.com/jerrysblog/entry/zones_p2v Jerry ___ zones-discuss mailing list zones-discuss@opensolaris.org
Re: [zones-discuss] OHAC :Apache failover between zones
Gaurav, Jun 29 20:26:08 opensolaris gdm-binary[6634]: [ID 702911 daemon.warning] WARNING: Failed to start X server several times in a short time period; disabling display :0 Sadly, there's no indication, as to WHY that fails... And: It seems, you have a single-node cluster, right? Jun 23 12:51:47 opensolaris genunix: [ID 965873 kern.notice] NOTICE: CMM: Node opensolaris (nodeid = 1) with votecount = 1 added. Jun 23 12:51:47 opensolaris genunix: [ID 843983 kern.notice] NOTICE: CMM: Node opensolaris: attempting to join cluster. Jun 23 12:51:47 opensolaris genunix: [ID 525628 kern.notice] NOTICE: CMM: Cluster has reached quorum. Jun 23 12:51:47 opensolaris genunix: [ID 377347 kern.notice] NOTICE: CMM: Node opensolaris (nodeid = 1) is up; new incarnation number = 1277277706. Jun 23 12:51:47 opensolaris genunix: [ID 108990 kern.notice] NOTICE: CMM: Cluster members: opensolaris. Jun 23 12:51:47 opensolaris genunix: [ID 279084 kern.notice] NOTICE: CMM: node reconfiguration #1 completed. Jun 23 12:51:50 opensolaris genunix: [ID 499756 kern.notice] NOTICE: CMM: Node opensolaris: joined cluster. And you have same errors in your config of the zones for the cluster, as far as the storage setup is concerned: Jun 23 13:09:52 opensolaris Cluster.RGM.global.rgmd: [ID 224900 daemon.notice] launching method for resource , resource group , node , timeout <1800> seconds Jun 23 13:09:52 opensolaris SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: [ID 148650 daemon.notice] Started searching for devices in '/dev/dsk' to find the importable pools. Jun 23 13:09:53 opensolaris SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: [ID 547433 daemon.notice] Completed searching the devices in '/dev/dsk' to find the importable pools. Jun 23 13:09:53 opensolaris SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: [ID 792255 daemon.warning] Failed to update the cachefile contents in /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile to CCR table rpool.cachefile for pool rpool : file /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile open failed: No such file or directory. Jun 23 13:09:54 opensolaris SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: [ID 471757 daemon.error] cannot unmount '/' : Device busy Jun 23 13:09:54 opensolaris SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: [ID 316615 daemon.error] Failed to export :rpool And a bit later: Jun 23 13:11:15 opensolaris SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_init]: [ID 792255 daemon.warning] Failed to update the cachefile contents in /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile to CCR table rpool.cachefile for pool rpool : file /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile open failed: No such file or directory. Jun 23 13:11:15 opensolaris SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_init]: [ID 471757 daemon.error] cannot unmount '/' : Device busy Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 515159 daemon.notice] method completed successfully for resource , resource group , node , time used: 0% of timeout <1800 seconds> Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 443746 daemon.notice] resource clusterresource state on node opensolaris change to R_OFFLINE Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 529407 daemon.notice] resource group mygroup state on node opensolaris change to RG_OFFLINE Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 529407 daemon.notice] resource group mygroup state on node opensolaris change to RG_PENDING_ONLINE Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 784560 daemon.notice] resource clusterresource status on node opensolaris change to R_FM_UNKNOWN So, some of these problems (X server start et.al.) might be related to these errros. First: The rpool is a local filesystem (better: a local pool), as it contains the root-filesystem This can not be made as an fail-over filesystem. But it seems, you did configure that way. So, if I were you, I would delete all the CONFIGURATIONS for things in the cluster, so that you get a fresh OpenSolaris 2009.06 with the OHAC packages installed. Then, I would reboot the server (are you sure, you want only a single node cluster? Or am I missing something here?), and see, if the X-server comes up again. Once that running again, I would recommend re-reading the basic cluster setup manuals for how to setup a failover zone and the storage needed to host the filesystems for those zones... Matthias You (gaurav saxena) wrote: > Hello > > 2010/6/29 Matthias Pfützner > > > Can you just send me the last 50 lines or so? > > > > > I have attached a text file containing complete output. :) > > -- > Thanks and Regards , > Gaurav Saxena -- Matthias Pfützner| mailto:pfu...@germany | Am Anfang wäre das Stück @work: +49 6103 752-394 | @home: +49 6151 75
Re: [zones-discuss] OHAC :Apache failover between zones
Hello 2010/6/30 Matthias Pfützner > Gaurav, > > Jun 29 20:26:08 opensolaris gdm-binary[6634]: [ID 702911 daemon.warning] > WARNING: Failed to start X server several times in a short time period; > disabling display :0 > > Sadly, there's no indication, as to WHY that fails... > > Ok. yes its really sad :(. > And: It seems, you have a single-node cluster, right? > > Yes I am having a single node cluster. > Jun 23 12:51:47 opensolaris genunix: [ID 965873 kern.notice] NOTICE: CMM: > Node opensolaris (nodeid = 1) with votecount = 1 added. > Jun 23 12:51:47 opensolaris genunix: [ID 843983 kern.notice] NOTICE: CMM: > Node opensolaris: attempting to join cluster. > Jun 23 12:51:47 opensolaris genunix: [ID 525628 kern.notice] NOTICE: CMM: > Cluster has reached quorum. > Jun 23 12:51:47 opensolaris genunix: [ID 377347 kern.notice] NOTICE: CMM: > Node opensolaris (nodeid = 1) is up; new incarnation number = 1277277706. > Jun 23 12:51:47 opensolaris genunix: [ID 108990 kern.notice] NOTICE: CMM: > Cluster members: opensolaris. > Jun 23 12:51:47 opensolaris genunix: [ID 279084 kern.notice] NOTICE: CMM: > node reconfiguration #1 completed. > Jun 23 12:51:50 opensolaris genunix: [ID 499756 kern.notice] NOTICE: CMM: > Node opensolaris: joined cluster. > > And you have same errors in your config of the zones for the cluster, as > far > as the storage setup is concerned: > > I see these messages at boot time. Are these errors ? > Jun 23 13:09:52 opensolaris Cluster.RGM.global.rgmd: [ID 224900 > daemon.notice] launching method for resource > , resource group , node , timeout > <1800> seconds > Jun 23 13:09:52 opensolaris > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: > [ID 148650 daemon.notice] Started searching for devices in '/dev/dsk' to > find the importable pools. > Jun 23 13:09:53 opensolaris > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: > [ID 547433 daemon.notice] Completed searching the devices in '/dev/dsk' to > find the importable pools. > Jun 23 13:09:53 opensolaris > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: > [ID 792255 daemon.warning] Failed to update the cachefile contents in > /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile to CCR table > rpool.cachefile for pool rpool : file > /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile open failed: No such file > or directory. > Jun 23 13:09:54 opensolaris > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: > [ID 471757 daemon.error] cannot unmount '/' : Device busy > Jun 23 13:09:54 opensolaris > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: > [ID 316615 daemon.error] Failed to export :rpool > > And a bit later: > > Jun 23 13:11:15 opensolaris > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_init]: [ID > 792255 daemon.warning] Failed to update the cachefile contents in > /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile to CCR table > rpool.cachefile for pool rpool : file > /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile open failed: No such file > or directory. > Jun 23 13:11:15 opensolaris > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_init]: [ID > 471757 daemon.error] cannot unmount '/' : Device busy > Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 515159 > daemon.notice] method completed successfully for > resource , resource group , node , > time used: 0% of timeout <1800 seconds> > Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 443746 > daemon.notice] resource clusterresource state on node opensolaris change to > R_OFFLINE > Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 529407 > daemon.notice] resource group mygroup state on node opensolaris change to > RG_OFFLINE > Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 529407 > daemon.notice] resource group mygroup state on node opensolaris change to > RG_PENDING_ONLINE > Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 784560 > daemon.notice] resource clusterresource status on node opensolaris change to > R_FM_UNKNOWN > > So, some of these problems (X server start et.al.) might be related to > these > errros. > > First: The rpool is a local filesystem (better: a local pool), as it > contains > the root-filesystem This can not be made as an fail-over filesystem. But it > seems, you did configure that way. > > No actually i just created two zfs filesystems on rpool as I have told earlier. Is it wrong ? > So, if I were you, I would delete all the CONFIGURATIONS for things in the > cluster, so that you get a fresh OpenSolaris 2009.06 with the OHAC packages > installed. > > I think I can reinstall opensolaris only and install OHAC on it again but I have reinstalled this for the second time only. :( > Then, I would reboot the server (are you sure, you want only a single node > cluster? Or am I missing something here?), and see, if the X-server comes > up > again. Once that runnin
Re: [zones-discuss] OHAC :Apache failover between zones
You (gaurav saxena) wrote: > Hello > > 2010/6/30 Matthias Pfützner > > > Gaurav, > > > > Jun 29 20:26:08 opensolaris gdm-binary[6634]: [ID 702911 daemon.warning] > > WARNING: Failed to start X server several times in a short time period; > > disabling display :0 > > > > Sadly, there's no indication, as to WHY that fails... > > > > > Ok. yes its really sad :(. I assume, it has something to do with the below error messages regarding the rpool. But, sadly, I'm not sure... > > And: It seems, you have a single-node cluster, right? > > > > > Yes I am having a single node cluster. Which is OK, but for learning purposes might be a bit to lightwight... A two-node cluster is better... > > Jun 23 12:51:47 opensolaris genunix: [ID 965873 kern.notice] NOTICE: CMM: > > Node opensolaris (nodeid = 1) with votecount = 1 added. > > Jun 23 12:51:47 opensolaris genunix: [ID 843983 kern.notice] NOTICE: CMM: > > Node opensolaris: attempting to join cluster. > > Jun 23 12:51:47 opensolaris genunix: [ID 525628 kern.notice] NOTICE: CMM: > > Cluster has reached quorum. > > Jun 23 12:51:47 opensolaris genunix: [ID 377347 kern.notice] NOTICE: CMM: > > Node opensolaris (nodeid = 1) is up; new incarnation number = 1277277706. > > Jun 23 12:51:47 opensolaris genunix: [ID 108990 kern.notice] NOTICE: CMM: > > Cluster members: opensolaris. > > Jun 23 12:51:47 opensolaris genunix: [ID 279084 kern.notice] NOTICE: CMM: > > node reconfiguration #1 completed. > > Jun 23 12:51:50 opensolaris genunix: [ID 499756 kern.notice] NOTICE: CMM: > > Node opensolaris: joined cluster. > > > > And you have same errors in your config of the zones for the cluster, as > > far > > as the storage setup is concerned: > > > > > I see these messages at boot time. Are these errors ? No, that's OK! > > Jun 23 13:09:52 opensolaris Cluster.RGM.global.rgmd: [ID 224900 > > daemon.notice] launching method for resource > > , resource group , node , timeout > > <1800> seconds > > Jun 23 13:09:52 opensolaris > > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: > > [ID 148650 daemon.notice] Started searching for devices in '/dev/dsk' to > > find the importable pools. > > Jun 23 13:09:53 opensolaris > > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: > > [ID 547433 daemon.notice] Completed searching the devices in '/dev/dsk' to > > find the importable pools. > > Jun 23 13:09:53 opensolaris > > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: > > [ID 792255 daemon.warning] Failed to update the cachefile contents in > > /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile to CCR table > > rpool.cachefile for pool rpool : file > > /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile open failed: No such file > > or directory. > > Jun 23 13:09:54 opensolaris > > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: > > [ID 471757 daemon.error] cannot unmount '/' : Device busy > > Jun 23 13:09:54 opensolaris > > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_validate]: > > [ID 316615 daemon.error] Failed to export :rpool > > > > And a bit later: > > > > Jun 23 13:11:15 opensolaris > > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_init]: [ID > > 792255 daemon.warning] Failed to update the cachefile contents in > > /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile to CCR table > > rpool.cachefile for pool rpool : file > > /var/cluster/run/HAStoragePlus/zfs/rpool.cachefile open failed: No such file > > or directory. > > Jun 23 13:11:15 opensolaris > > SC[,SUNW.HAStoragePlus:8,mygroup,clusterresource,hastorageplus_init]: [ID > > 471757 daemon.error] cannot unmount '/' : Device busy > > Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 515159 > > daemon.notice] method completed successfully for > > resource , resource group , node , > > time used: 0% of timeout <1800 seconds> > > Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 443746 > > daemon.notice] resource clusterresource state on node opensolaris change to > > R_OFFLINE > > Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 529407 > > daemon.notice] resource group mygroup state on node opensolaris change to > > RG_OFFLINE > > Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 529407 > > daemon.notice] resource group mygroup state on node opensolaris change to > > RG_PENDING_ONLINE > > Jun 23 13:11:15 opensolaris Cluster.RGM.global.rgmd: [ID 784560 > > daemon.notice] resource clusterresource status on node opensolaris change to > > R_FM_UNKNOWN > > > > So, some of these problems (X server start et.al.) might be related to > > these > > errros. > > > > First: The rpool is a local filesystem (better: a local pool), as it > > contains > > the root-filesystem This can not be made as an fail-over filesystem. But it > > seems, you did configure that way. > > > > > No actually i just created two zfs filesystems on rpool as I have told > earlier. > Is it wrong ? I assume so, y
Re: [zones-discuss] OHAC :Apache failover between zones
Hello 2010/6/30 Matthias Pfützner > I assume, it has something to do with the below error messages regarding > the > rpool. But, sadly, I'm not sure... > > Ok. > Which is OK, but for learning purposes might be a bit to lightwight... A > two-node cluster is better... > > Ok. But what I really want to do is something which can be done using single node cluster only.So I can just now do that with single node cluster and then try two-node cluster. > No, that's OK! > > Ok. > > > No actually i just created two zfs filesystems on rpool as I have told > > earlier. > > Is it wrong ? > > I assume so, yes... As the thing, that can be moved are the POOLS, not the > zfs > Filesystems. Therefore, you would have needed a new pool, which can be > moved > from system to system... you can't MOVE the rpool, as that's needed for the > basic underlying OS... ;-) > > Ok. Actually I could not get much after reading about pools so I tried the shortcut way. Ok. I will try doing something else for it. Also one more thing I will tell you is that I have not created a separate "/globaldevices" as it was written in the manual. so scinstall asked me to use lofi device (I don't know what it means) and I accepted that option. Could it be a problem ? > > I think I can reinstall opensolaris only and install OHAC on it again but > I > > have reinstalled this for the second time only. :( > > I guess, you would NOT need to re-install, but to just DECONFIGURE all the > resources, storage et.al. that you did in the cluster config. > > Ok. I will do that once I figure out everything correctly because I don't want to commit this mistake again. :( > > Actually in the manual it was suggested to create a mirror zpool of the > root > > filesystem and then install zfs on it or create a pool on file but I > thought > > I could use the original rpool only. > > For testing, you don't need the mirror. > Yes, as stated above, the "thing" that will be "failed-over" is the > POOL. Therefore you need a pool. And you simply can't use the rpool... > > Ok. Could you suggest me how do I do that ? Or I will be consult manual again zpool create -m mountpoint pool mirror /dev/rdsk/cNtXdY \ /dev/rdsk/cNtXdZ this is the command given in the manual could you suggest me what shoud I give in place of cNtXdY and I have virtual hard disk named c7d0 > > Is this the reason for all problems ? > > It might be! > > > But as far as I remember display is not working since I ran scinstall and > I > > did this all zone work from the console only so I don't think this could > be > > the reason. > > That's the strange thing... But, in order to be sure, when that really > started, I would propose to really de-configure the actual cluster setup, > and > staring all over again (not with INSTALL, only with configuring!) > > Yes that's the most confusing thing. Ok I try doing that only now. Thank you very much for your help. :) > > -- > > Thanks and Regards , > > Gaurav Saxena > > Matthias > -- >Matthias Pfützner| mailto:pfu...@germany | Am Anfang wäre das Stück > @work: +49 6103 752-394 | @home: +49 6151 75717 | schnell zu Ende. > SunCS, Ampčrestraße 6 | Lichtenbergstraße 73 | (21.10.93) >63225 Langen, FRG| 64289 Darmstadt, FRG | Hellmuth Karasek > -- Thanks and Regards , Gaurav Saxena ___ zones-discuss mailing list zones-discuss@opensolaris.org