On 17.01.2014 16:45:20, Samuli Heinonen wrote: > Hello Bernhard, > Can you test if setting option network.remote-dio to enable allows you to use > cache=none? >
Hi Samuli,nope network.remote-dio enable didn't changed it.got a too unclean system meanwhile,will do a full reinstall of the hostsand than come back to this topicthnx Bernhard > -samuli > > Bernhard Glomm <> bernhard.gl...@ecologic.eu> > kirjoitti 17.1.2014 kello > 16.41: > > > Pranith, > > > > I stopped the volume > > > > started it again, > > > > mounted it on both hosts > > > > started the VM > > > > did the livemigration > > > > and collected the logs: > > > > - etc-glusterfs-glusterd.vol.log > > > > - glustershd.log > > > > - srv-vm_infrastructure-vm-atom01.log > > > > - cli.log > > > > from the beginning of the gluster volume start. > > > > You can found them here (part 1 to 3): > > > > http://pastebin.com/mnATm2BE > > > > http://pastebin.com/RYZFP3E9 > > > > http://pastebin.com/HAXEGd54 > > > > further more: > > > > gluster --version: glusterfs 3.4.2 built on Jan 11 2014 03:21:47 > > > > ubuntu: raring > > > > filesystem on the gluster bricks: zfs-0.6.2 > > > > gluster volume info fs_vm_atom01 > > > > Volume Name: fs_vm_atom01 > > > > Type: Replicate > > > > Volume ID: fea9bdcf-783e-442a-831d-f564f8dbe551 > > > > Status: Started > > > > Number of Bricks: 1 x 2 = 2 > > > > Transport-type: tcp > > > > Bricks: > > > > Brick1: 172.24.1.11:/zp_ping_1/fs_vm_atom01 > > > > Brick2: 172.24.1.13:/zp_pong_1/fs_vm_atom01 > > > > Options Reconfigured: > > > > diagnostics.client-log-level: DEBUG > > > > server.allow-insecure: on > > > > > > disk part of VM configuration > > > > <emulator>/usr/bin/kvm-spice</emulator> > > > > <disk type='file' device='disk'> > > > > <driver name='qemu' type='raw' cache='writethrough'/> > > > > <source file='/srv/vm_infrastructure/vm/atom01/atom01.img'/> > > > > <target dev='vda' bus='virtio'/> > > > > <address type='pci' domain='0x0000' bus='0x00' slot='0x04' > > > > function='0x0'/> > > > > </disk> > > > > > > can't use <source protocol='gluster' ... > > > > as josh suggested because couldn't get > > > > my qemu recompiled with gluster enabled yet. > > > > Are there other special tuning parameter for kvm/qemu/ to set on gluster? > > > > as mentioned: all works except the livemigration (disk image file > > > > becomes read only) > > > > and I have to use something different than cache=none... > > > > TIA > > > > Bernhard > > > > On 17.01.2014 05:04:52, Pranith Kumar Karampuri wrote: > > > Bernhard, > > > Configuration seems ok. Could you please give the log files of the bricks > > > and mount please. If you think it is not a big procedure to do this live > > > migration, could you set client-log-level to DEBUG and provide the log > > > files of that run. > > > > > > Pranith > > > > > > ----- Original Message ----- > > > > From: "Bernhard Glomm" <> > > > bernhard.gl...@ecologic.eu> > > > > > > > > To:> > > > > > > > pkara...@redhat.com > > > > Cc:> > > > > > > > gluster-users@gluster.org > > > > Sent: Thursday, January 16, 2014 5:58:17 PM > > > > Subject: Re: [Gluster-users] gluster and kvm livemigration > > > > > > > > > > > > hi Pranith > > > > > > > > # gluster volume info fs_vm_atom01 > > > > > > > > Volume Name: fs_vm_atom01 > > > > Type: Replicate > > > > Volume ID: fea9bdcf-783e-442a-831d-f564f8dbe551 > > > > Status: Started > > > > Number of Bricks: 1 x 2 = 2 > > > > Transport-type: tcp > > > > Bricks: > > > > Brick1: 172.24.1.11:/zp_ping_1/fs_vm_atom01 > > > > Brick2: 172.24.1.13:/zp_pong_1/fs_vm_atom01 > > > > Options Reconfigured: > > > > diagnostics.client-log-level: ERROR > > > > > > > > > > > > TIA > > > > Bernhard > > > > > > > > > > > > On 16.01.2014 13:05:12, Pranith Kumar Karampuri wrote: > > > > > hi Bernhard, > > > > > Could you give gluster volume info output? > > > > > > > > > > Pranith > > > > > > > > > > ----- Original Message ----- > > > > > > From: "Bernhard Glomm" <> >> > > > > > > > > > > > > > > > > > bernhard.gl...@ecologic.eu> > > > > > > > > > > > > > > To: > >> > > > > > > > > > > > gluster-users@gluster.org > > > > > > Sent: Thursday, January 16, 2014 4:22:36 PM > > > > > > Subject: [Gluster-users] gluster and kvm livemigration > > > > > > > > > > > > I experienced a strange behavior of glusterfs during livemigration > > > > > > of a qemu-kvm guest > > > > > > using a 10GB file on a mirrored gluster 3.4.2 volume > > > > > > (both on ubuntu 13.04) > > > > > > I run > > > > > > virsh migrate --verbose --live --unsafe --p2p --domain atom01 > > > > > > --desturi > > > > > > qemu+ssh://<target_ip>/system > > > > > > and the migration works, > > > > > > the running machine is pingable and keeps sending pings. > > > > > > nevertheless, when I let the machine touch a file during migration > > > > > > it stops, complaining that it's filesystem is read only (from that > > > > > > moment > > > > > > that > > > > > > migration finished) > > > > > > A reboot from inside the machine failes, > > > > > > machine goes down and comes up with an error > > > > > > unable to write to sector xxxxxx on hd0 > > > > > > (than falling into the initrd). > > > > > > a > > > > > > virsh destroy VM && virsh start VM > > > > > > leads to a perfect running VM again, > > > > > > no matter on which of the two hosts I start the machine > > > > > > anybody better experience with livemigration? > > > > > > any hint on a procedure how to debug that? > > > > > > TIA > > > > > > Bernhard > > > > > > > > > > > > -- > > > > > > > > > > > > > > > > > > Bernhard Glomm > > > > > > IT Administration > > > > > > > > > > > > Phone: +49 (30) 86880 134 > > > > > > Fax: +49 (30) 86880 100 > > > > > > Skype: bernhard.glomm.ecologic > > > > > > > > > > > > Ecologic Institut gemeinnützige GmbH | Pfalzburger Str. 43/44 | > > > > > > 10717 > > > > > > Berlin > > > > > > | Germany > > > > > > GF: R. Andreas Kraemer | AG: Charlottenburg HRB 57947 | > > > > > > USt/VAT-IdNr.: > > > > > > DE811963464 > > > > > > Ecologic is a Trade Mark (TM) of Ecologic Institut gemeinnützige > > > > > > GmbH > > > > > > > > > > > > > > > > > > _______________________________________________ > > > > > > Gluster-users mailing list > > > > > > Gluster-users@gluster.org > > > > > > http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users