Status: New
Owner: ----
New issue 1161 by [email protected]: gnt-instance verify disks did not detect
that I needed to run activate-disks
https://code.google.com/p/ganeti/issues/detail?id=1161
What software version are you running? Please provide the output of "gnt-
cluster --version", "gnt-cluster version", and "hspace --version".
gnt-cluster (ganeti 2.15.1-1~bpo8+1) 2.15.1
hspace (ganeti) version 2.15.1-1~bpo8+1
compiled with ghc 7.6
running on linux x86_64
What distribution are you using?
Debian Jessie.
I had to do some reboots of the nodes in my cluster, so I shuffled
things around to do them, and now that the nodes are back, I'm trying to
shuffle things back to being in a good state.
I'm having a problem with migrating an instance to its secondary, i'm
trying to do a 'gnt-instance migrate' but it fails with:
root@minla-pn:~# gnt-instance migrate mx2
Instance mx2 will be migrated. Note that migration might impact the
instance if anything goes wrong (e.g. due to bugs in the hypervisor).
Continue?
y/[n]/?: y
Wed Dec 16 23:45:00 2015 Migrating instance mx2
Wed Dec 16 23:45:03 2015 * checking disk consistency between source and
target
Wed Dec 16 23:45:04 2015 - WARNING: Can't find disk on node martin-pn
Failure: command execution error:
Disk 0 is degraded or not fully synchronized on target node, aborting
migration
root@minla-pn:~#
Wed Dec 16 22:26:33 2015 - WARNING: Can't find disk on node martin-pn
Disk 0 is degraded or not fully synchronized on target node, aborting
migration
/proc/drbd on the node doesn't show it syncing or disconnected:
root@martin-pn:/var/log/ganeti# cat /proc/drbd
version: 8.4.3 (api:1/proto:86-101)
srcversion: 1A9F77B1CA5FF92235C2213
0: cs:Connected ro:Secondary/Primary ds:UpToDate/UpToDate C r-----
ns:0 nr:441716 dw:441716 dr:0 al:0 bm:17 lo:0 pe:0 ua:0 ap:0 ep:1 wo:f
oos:0
2: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r-----
ns:670300 nr:0 dw:699436 dr:101164 al:156 bm:12 lo:0 pe:0 ua:0 ap:0
ep:1 wo:f oos:0
root@martin-pn:/var/log/ganeti#
it seems to run these commands on the node I'm trying to migrate to:
2015-12-16 22:29:33,164: ganeti-noded pid=11969 INFO RunCmd lvs
--noheadings '--separator=|' --units=k --nosuffix
-olv_attr,lv_kernel_major,lv_kernel_minor,vg_extent_size,stripes,devices
/dev/ganetivg0/a8c902d8-b3c1-4b29-8b01-de1737017597.disk0_data
2015-12-16 22:29:33,304: ganeti-noded pid=11969 INFO RunCmd lvs
--noheadings '--separator=|' --units=k --nosuffix
-olv_attr,lv_kernel_major,lv_kernel_minor,vg_extent_size,stripes,devices
/dev/ganetivg0/a8c902d8-b3c1-4b29-8b01-de1737017597.disk0_meta
2015-12-16 22:29:33,461: ganeti-noded pid=11969 INFO 10.0.1.105:45024 POST
/blockdev_find HTTP/1.1 200
2015-12-16 22:29:33,744: ganeti-noded pid=11972 INFO 10.0.1.105:45025 POST
/jobqueue_update HTTP/1.1 200
^C
root@martin-pn:/var/log/ganeti# lvs --noheadings '--separator=|' --units=k
--nosuffix
-olv_attr,lv_kernel_major,lv_kernel_minor,vg_extent_size,stripes,devices
/dev/ganetivg0/a8c902d8-b3c1-4b29-8b01-de1737017597.disk0_data
-wi-a-----|253|11|4096.00|1|/dev/mapper/md3_crypt(3936)
root@martin-pn:/var/log/ganeti# lvs --noheadings '--separator=|' --units=k
--nosuffix
-olv_attr,lv_kernel_major,lv_kernel_minor,vg_extent_size,stripes,devices
/dev/ganetivg0/a8c902d8-b3c1-4b29-8b01-de1737017597.disk0_meta
-wi-a-----|253|12|4096.00|1|/dev/mapper/md3_crypt(5216)
root@martin-pn:/var/log/ganeti#
gnt-cluster verify had no errors, and gnt-cluster verify-disks was also fine
In the end, I found that I needed to do 'activate-disks' on the
node... then things worked.
Why doesn't verify-disks, or even just 'verify' find this case and give
you a clue?
--
You received this message because this project is configured to send all
issue notifications to this address.
You may adjust your notification preferences at:
https://code.google.com/hosting/settings