Status: New
Owner: ----

New issue 1161 by [email protected]: gnt-instance verify disks did not detect that I needed to run activate-disks
https://code.google.com/p/ganeti/issues/detail?id=1161

What software version are you running? Please provide the output of "gnt-
cluster --version", "gnt-cluster version", and "hspace --version".

gnt-cluster (ganeti 2.15.1-1~bpo8+1) 2.15.1
hspace (ganeti) version 2.15.1-1~bpo8+1
compiled with ghc 7.6
running on linux x86_64

What distribution are you using?

Debian Jessie.

I had to do some reboots of the nodes in my cluster, so I shuffled
things around to do them, and now that the nodes are back, I'm trying to
shuffle things back to being in a good state.

I'm having a problem with migrating an instance to its secondary, i'm
trying to do a 'gnt-instance migrate' but it fails with:

root@minla-pn:~# gnt-instance migrate mx2
Instance mx2 will be migrated. Note that migration might impact the
instance if anything goes wrong (e.g. due to bugs in the hypervisor).
Continue?
y/[n]/?: y
Wed Dec 16 23:45:00 2015 Migrating instance mx2
Wed Dec 16 23:45:03 2015 * checking disk consistency between source and target
Wed Dec 16 23:45:04 2015  - WARNING: Can't find disk on node martin-pn
Failure: command execution error:
Disk 0 is degraded or not fully synchronized on target node, aborting migration
root@minla-pn:~#
Wed Dec 16 22:26:33 2015 - WARNING: Can't find disk on node martin-pn
Disk 0 is degraded or not fully synchronized on target node, aborting migration

/proc/drbd on the node doesn't show it syncing or disconnected:

root@martin-pn:/var/log/ganeti# cat /proc/drbd
version: 8.4.3 (api:1/proto:86-101)
srcversion: 1A9F77B1CA5FF92235C2213
 0: cs:Connected ro:Secondary/Primary ds:UpToDate/UpToDate C r-----
ns:0 nr:441716 dw:441716 dr:0 al:0 bm:17 lo:0 pe:0 ua:0 ap:0 ep:1 wo:f oos:0

 2: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r-----
ns:670300 nr:0 dw:699436 dr:101164 al:156 bm:12 lo:0 pe:0 ua:0 ap:0 ep:1 wo:f oos:0
root@martin-pn:/var/log/ganeti#

it seems to run these commands on the node I'm trying to migrate to:

2015-12-16 22:29:33,164: ganeti-noded pid=11969 INFO RunCmd lvs --noheadings '--separator=|' --units=k --nosuffix -olv_attr,lv_kernel_major,lv_kernel_minor,vg_extent_size,stripes,devices /dev/ganetivg0/a8c902d8-b3c1-4b29-8b01-de1737017597.disk0_data 2015-12-16 22:29:33,304: ganeti-noded pid=11969 INFO RunCmd lvs --noheadings '--separator=|' --units=k --nosuffix -olv_attr,lv_kernel_major,lv_kernel_minor,vg_extent_size,stripes,devices /dev/ganetivg0/a8c902d8-b3c1-4b29-8b01-de1737017597.disk0_meta 2015-12-16 22:29:33,461: ganeti-noded pid=11969 INFO 10.0.1.105:45024 POST /blockdev_find HTTP/1.1 200 2015-12-16 22:29:33,744: ganeti-noded pid=11972 INFO 10.0.1.105:45025 POST /jobqueue_update HTTP/1.1 200
^C
root@martin-pn:/var/log/ganeti# lvs --noheadings '--separator=|' --units=k --nosuffix -olv_attr,lv_kernel_major,lv_kernel_minor,vg_extent_size,stripes,devices /dev/ganetivg0/a8c902d8-b3c1-4b29-8b01-de1737017597.disk0_data
  -wi-a-----|253|11|4096.00|1|/dev/mapper/md3_crypt(3936)
root@martin-pn:/var/log/ganeti# lvs --noheadings '--separator=|' --units=k --nosuffix -olv_attr,lv_kernel_major,lv_kernel_minor,vg_extent_size,stripes,devices /dev/ganetivg0/a8c902d8-b3c1-4b29-8b01-de1737017597.disk0_meta
  -wi-a-----|253|12|4096.00|1|/dev/mapper/md3_crypt(5216)
root@martin-pn:/var/log/ganeti#

gnt-cluster verify had no errors, and gnt-cluster verify-disks was also fine

In the end, I found that I needed to do 'activate-disks' on the
node... then things worked.

Why doesn't verify-disks, or even just 'verify' find this case and give
you a clue?

--
You received this message because this project is configured to send all issue notifications to this address.
You may adjust your notification preferences at:
https://code.google.com/hosting/settings

Reply via email to