What is the output of 'lsblk -t' on all nodes ? Best Regards, Strahil NIkolov692371
В 11:19 +0100 на 17.01.2021 (нд), Christian Reiss написа: > Hey folks, > > quick (I hope) question: On my 3-node cluster I am swapping out all > the > SSDs with fewer but higher capacity ones. So I took one node down > (maintenance, stop), then removed all SSDs, set up a new RAID, set > up > lvm and gluster, let it resync. Gluster health status shows no > unsynced > entries. > > Uppon going from maintenance to online from ovirt management It goes > into non-operational status, vdsm log on the node shows: > > 2021-01-17 11:13:29,051+0100 INFO (jsonrpc/6) [api.host] START > getAllVmStats() from=::1,48580 (api:48) > 2021-01-17 11:13:29,051+0100 INFO (jsonrpc/6) [api.host] FINISH > getAllVmStats return={'status': {'message': 'Done', 'code': 0}, > 'statsList': (suppressed)} from=::1,48580 (api:54) > 2021-01-17 11:13:29,052+0100 INFO (jsonrpc/6) > [jsonrpc.JsonRpcServer] > RPC call Host.getAllVmStats succeeded in 0.00 seconds (__init__:312) > 2021-01-17 11:13:30,420+0100 WARN (monitor/4a62cdb) [storage.LVM] > Reloading VGs failed (vgs=[u'4a62cdb4-b314-4c7f-804e-8e7275518a7f'] > rc=5 > out=[] err=[' Volume group "4a62cdb4-b314-4c7f-804e-8e7275518a7f" > not > found', ' Cannot process volume group > 4a62cdb4-b314-4c7f-804e-8e7275518a7f']) (lvm:470) > 2021-01-17 11:13:30,424+0100 ERROR (monitor/4a62cdb) > [storage.Monitor] > Setting up monitor for 4a62cdb4-b314-4c7f-804e-8e7275518a7f failed > (monitor:330) > Traceback (most recent call last): > File "/usr/lib/python2.7/site-packages/vdsm/storage/monitor.py", > line > 327, in _setupLoop > self._setupMonitor() > File "/usr/lib/python2.7/site-packages/vdsm/storage/monitor.py", > line > 349, in _setupMonitor > self._produceDomain() > File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 159, > in > wrapper > value = meth(self, *a, **kw) > File "/usr/lib/python2.7/site-packages/vdsm/storage/monitor.py", > line > 367, in _produceDomain > self.domain = sdCache.produce(self.sdUUID) > File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line > 110, in produce > domain.getRealDomain() > File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line > 51, > in getRealDomain > return self._cache._realProduce(self._sdUUID) > File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line > 134, in _realProduce > domain = self._findDomain(sdUUID) > File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line > 151, in _findDomain > return findMethod(sdUUID) > File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line > 176, in _findUnfetchedDomain > raise se.StorageDomainDoesNotExist(sdUUID) > > > I assume due to the changed LVM UUID it fails, right? Can I someone > fix/change the UUID and get the node back up again? It does not seem > to > be a major issue, to be honest. > > I can see the gluster mount (what ovirt mounts when it onlines a > node) > already, and gluster is happy too. > > Any help is appreciated! > > -Chris. > > _______________________________________________ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/privacy-policy.html > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/LS4OEXZOE6UA4CDQYXFKC3TZCCO42SU4/ _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/3IYV67D7JBE3CK6SQ7SEUUNRVWU5QNYK/