On Wed, Apr 20, 2016 at 1:42 PM, Wee Sritippho <we...@forest.go.th> wrote: > Hi Didi & Martin, > > I followed your instructions and are able to add the 2nd host. Thank you :) > > This is what I've done: > > [root@host01 ~]# hosted-engine --set-maintenance --mode=global > > [root@host01 ~]# systemctl stop ovirt-ha-agent > > [root@host01 ~]# systemctl stop ovirt-ha-broker > > [root@host01 ~]# find /rhev -name hosted-engine.metadata > /rhev/data-center/mnt/blockSD/47e3e4ac-534a-4e11-b14e-27ecb4585431/ha_agent/hosted-engine.metadata > /rhev/data-center/mnt/blockSD/336dc4a3-f65c-4a67-bc42-1f73597564cf/ha_agent/hosted-engine.metadata > > [root@host01 ~]# ls -al > /rhev/data-center/mnt/blockSD/47e3e4ac-534a-4e11-b14e-27ecb4585431/ha_agent/hosted-engine.metadata > lrwxrwxrwx. 1 vdsm kvm 132 Apr 20 02:56 > /rhev/data-center/mnt/blockSD/47e3e4ac-534a-4e11-b14e-27ecb4585431/ha_agent/hosted-engine.metadata > -> > /var/run/vdsm/storage/47e3e4ac-534a-4e11-b14e-27ecb4585431/d92632bf-8c15-44ba-9aa8-4a39dcb81e8d/4761bb8d-779e-4378-8b13-7b12f96f5c56 > > [root@host01 ~]# ls -al > /rhev/data-center/mnt/blockSD/336dc4a3-f65c-4a67-bc42-1f73597564cf/ha_agent/hosted-engine.metadata > lrwxrwxrwx. 1 vdsm kvm 132 Apr 21 03:40 > /rhev/data-center/mnt/blockSD/336dc4a3-f65c-4a67-bc42-1f73597564cf/ha_agent/hosted-engine.metadata > -> > /var/run/vdsm/storage/336dc4a3-f65c-4a67-bc42-1f73597564cf/49d6ee16-cfa0-47f2-b461-125bc6f614db/89ee314d-33ce-43fb-9a66-0852c5f675d3 > > [root@host01 ~]# dd if=/dev/zero > of=/var/run/vdsm/storage/336dc4a3-f65c-4a67-bc42-1f73597564cf/49d6ee16-cfa0-47f2-b461-125bc6f614db/89ee314d-33ce-43fb-9a66-0852c5f675d3 > bs=1M > dd: error writing > ‘/var/run/vdsm/storage/336dc4a3-f65c-4a67-bc42-1f73597564cf/49d6ee16-cfa0-47f2-b461-125bc6f614db/89ee314d-33ce-43fb-9a66-0852c5f675d3’: > No space left on device > 129+0 records in > 128+0 records out > 134217728 bytes (134 MB) copied, 0.246691 s, 544 MB/s > > [root@host01 ~]# systemctl start ovirt-ha-broker > > [root@host01 ~]# systemctl start ovirt-ha-agent > > [root@host01 ~]# hosted-engine --set-maintenance --mode=none > > (Found 2 metadata files but the first one is red when I used 'ls -al' so I > assume it is a leftover from the previous failed installation and didn't > touch it) > > BTW, how to properly clean the FC storage before using it with oVirt? I used > "parted /dev/mapper/wwid mklabel msdos" to destroy the partition table. > Isn't that enough?
Even this should not be needed in 3.6. Did you start with 3.6? Or upgraded from a previous version? Also please verify that output of 'hosted-engine --vm-status' makes sense. Thanks, > > > On 20/4/2559 15:11, Martin Sivak wrote: >>> >>> Assuming you never deployed a host with ID 52, this is likely a result of >>> a >>> corruption or dirt or something like that. >>> I see that you use FC storage. In previous versions, we did not clean >>> such >>> storage, so you might have dirt left. >> >> This is the exact reason for an error like yours. Using dirty block >> storage. Please stop all hosted engine tooling (both agent and broker) >> and fill the metadata drive with zeros. >> >> You will have to find the proper hosted-engine.metadata file (which >> will be a symlink) under /rhev: >> >> Example: >> >> [root@dev-03 rhev]# find . -name hosted-engine.metadata >> >> >> ./data-center/mnt/str-01.rhev.lab.eng.brq.redhat.com:_mnt_export_nfs_lv2_msivak/868a1a4e-9f94-42f5-af23-8f884b3c53d5/ha_agent/hosted-engine.metadata >> >> [root@dev-03 rhev]# ls -al >> >> ./data-center/mnt/str-01:_mnt_export_nfs_lv2_msivak/868a1a4e-9f94-42f5-af23-8f884b3c53d5/ha_agent/hosted-engine.metadata >> >> lrwxrwxrwx. 1 vdsm kvm 201 Mar 15 15:00 >> >> ./data-center/mnt/str-01:_mnt_export_nfs_lv2_msivak/868a1a4e-9f94-42f5-af23-8f884b3c53d5/ha_agent/hosted-engine.metadata >> -> >> /rhev/data-center/mnt/str-01:_mnt_export_nfs_lv2_msivak/868a1a4e-9f94-42f5-af23-8f884b3c53d5/images/6ab3f215-f234-4cd4-b9d4-8680767c3d99/dcbfa48d-8543-42d1-93dc-aa40855c4855 >> >> And use (for example) dd if=/dev/zero of=/path/to/metadata bs=1M to >> clean it - But be CAREFUL to not touch any other file or disk you >> might find. >> >> Then restart the hosted engine tools and all should be fine. >> >> >> >> Martin >> >> >> On Wed, Apr 20, 2016 at 8:20 AM, Yedidyah Bar David <d...@redhat.com> >> wrote: >>> >>> On Wed, Apr 20, 2016 at 7:15 AM, Wee Sritippho <we...@forest.go.th> >>> wrote: >>>> >>>> Hi, >>>> >>>> I used CentOS-7-x86_64-Minimal-1511.iso to install the hosts and the >>>> engine. >>>> >>>> The 1st host and the hosted-engine were installed successfully, but the >>>> 2nd >>>> host failed with this error message: >>>> >>>> "Failed to execute stage 'Setup validation': Metadata version 2 from >>>> host 52 >>>> too new for this agent (highest compatible version: 1)" >>> >>> Assuming you never deployed a host with ID 52, this is likely a result of >>> a >>> corruption or dirt or something like that. >>> >>> What do you get on host 1 running 'hosted-engine --vm-status'? >>> >>> I see that you use FC storage. In previous versions, we did not clean >>> such >>> storage, so you might have dirt left. See also [1]. You can try cleaning >>> using [2]. >>> >>> [1] https://bugzilla.redhat.com/show_bug.cgi?id=1238823 >>> [2] >>> https://www.ovirt.org/documentation/how-to/hosted-engine/#lockspace-corrupted-recovery-procedure >>> >>>> Here is the package versions: >>>> >>>> [root@host02 ~]# rpm -qa | grep ovirt >>>> libgovirt-0.3.3-1.el7_2.1.x86_64 >>>> ovirt-vmconsole-1.0.0-1.el7.centos.noarch >>>> ovirt-vmconsole-host-1.0.0-1.el7.centos.noarch >>>> ovirt-host-deploy-1.4.1-1.el7.centos.noarch >>>> ovirt-hosted-engine-ha-1.3.5.1-1.el7.centos.noarch >>>> ovirt-hosted-engine-setup-1.3.4.0-1.el7.centos.noarch >>>> ovirt-release36-007-1.noarch >>>> ovirt-engine-sdk-python-3.6.3.0-1.el7.centos.noarch >>>> ovirt-setup-lib-1.0.1-1.el7.centos.noarch >>>> >>>> [root@engine ~]# rpm -qa | grep ovirt >>>> ovirt-engine-setup-base-3.6.4.1-1.el7.centos.noarch >>>> >>>> ovirt-engine-setup-plugin-ovirt-engine-common-3.6.4.1-1.el7.centos.noarch >>>> ovirt-vmconsole-proxy-1.0.0-1.el7.centos.noarch >>>> ovirt-engine-tools-3.6.4.1-1.el7.centos.noarch >>>> ovirt-engine-vmconsole-proxy-helper-3.6.4.1-1.el7.centos.noarch >>>> ovirt-host-deploy-1.4.1-1.el7.centos.noarch >>>> ovirt-release36-007-1.noarch >>>> ovirt-engine-sdk-python-3.6.3.0-1.el7.centos.noarch >>>> ovirt-iso-uploader-3.6.0-1.el7.centos.noarch >>>> ovirt-engine-extensions-api-impl-3.6.4.1-1.el7.centos.noarch >>>> ovirt-setup-lib-1.0.1-1.el7.centos.noarch >>>> ovirt-host-deploy-java-1.4.1-1.el7.centos.noarch >>>> ovirt-engine-cli-3.6.2.0-1.el7.centos.noarch >>>> ovirt-engine-setup-plugin-websocket-proxy-3.6.4.1-1.el7.centos.noarch >>>> ovirt-vmconsole-1.0.0-1.el7.centos.noarch >>>> ovirt-engine-backend-3.6.4.1-1.el7.centos.noarch >>>> ovirt-engine-dbscripts-3.6.4.1-1.el7.centos.noarch >>>> ovirt-engine-webadmin-portal-3.6.4.1-1.el7.centos.noarch >>>> ovirt-engine-setup-3.6.4.1-1.el7.centos.noarch >>>> ovirt-engine-3.6.4.1-1.el7.centos.noarch >>>> >>>> ovirt-engine-setup-plugin-vmconsole-proxy-helper-3.6.4.1-1.el7.centos.noarch >>>> ovirt-guest-agent-common-1.0.11-1.el7.noarch >>>> ovirt-engine-wildfly-8.2.1-1.el7.x86_64 >>>> ovirt-engine-wildfly-overlay-8.0.5-1.el7.noarch >>>> ovirt-engine-websocket-proxy-3.6.4.1-1.el7.centos.noarch >>>> ovirt-engine-restapi-3.6.4.1-1.el7.centos.noarch >>>> ovirt-engine-userportal-3.6.4.1-1.el7.centos.noarch >>>> ovirt-engine-setup-plugin-ovirt-engine-3.6.4.1-1.el7.centos.noarch >>>> ovirt-image-uploader-3.6.0-1.el7.centos.noarch >>>> ovirt-engine-extension-aaa-jdbc-1.0.6-1.el7.noarch >>>> ovirt-engine-lib-3.6.4.1-1.el7.centos.noarch >>>> >>>> >>>> Here are the log files: >>>> https://gist.github.com/weeix/1743f88d3afe1f405889a67ed4011141 >>>> >>>> -- >>>> Wee >>>> >>>> _______________________________________________ >>>> Users mailing list >>>> Users@ovirt.org >>>> http://lists.ovirt.org/mailman/listinfo/users >>> >>> >>> >>> -- >>> Didi >>> _______________________________________________ >>> Users mailing list >>> Users@ovirt.org >>> http://lists.ovirt.org/mailman/listinfo/users > > > -- > วีร์ ศรีทิพโพธิ์ > นักวิชาการคอมพิวเตอร์ปฏิบัติการ > ศูนย์สารสนเทศ กรมป่าไม้ > โทร. 025614292-3 ต่อ 5621 > มือถือ. 0864678919 > -- Didi _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users