Hello Bernhard, Can you test if setting option network.remote-dio to enable allows you to use cache=none?
-samuli Bernhard Glomm <[email protected]> kirjoitti 17.1.2014 kello 16.41: > Pranith, > I stopped the volume > started it again, > mounted it on both hosts > started the VM > did the livemigration > and collected the logs: > - etc-glusterfs-glusterd.vol.log > - glustershd.log > - srv-vm_infrastructure-vm-atom01.log > - cli.log > from the beginning of the gluster volume start. > You can found them here (part 1 to 3): > http://pastebin.com/mnATm2BE > http://pastebin.com/RYZFP3E9 > http://pastebin.com/HAXEGd54 > > further more: > gluster --version: glusterfs 3.4.2 built on Jan 11 2014 03:21:47 > ubuntu: raring > filesystem on the gluster bricks: zfs-0.6.2 > > gluster volume info fs_vm_atom01 > Volume Name: fs_vm_atom01 > Type: Replicate > Volume ID: fea9bdcf-783e-442a-831d-f564f8dbe551 > Status: Started > Number of Bricks: 1 x 2 = 2 > Transport-type: tcp > Bricks: > Brick1: 172.24.1.11:/zp_ping_1/fs_vm_atom01 > Brick2: 172.24.1.13:/zp_pong_1/fs_vm_atom01 > Options Reconfigured: > diagnostics.client-log-level: DEBUG > server.allow-insecure: on > > disk part of VM configuration > > <emulator>/usr/bin/kvm-spice</emulator> > <disk type='file' device='disk'> > <driver name='qemu' type='raw' cache='writethrough'/> > <source file='/srv/vm_infrastructure/vm/atom01/atom01.img'/> > <target dev='vda' bus='virtio'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x04' > function='0x0'/> > </disk> > > can't use <source protocol='gluster' ... > as josh suggested because couldn't get > my qemu recompiled with gluster enabled yet. > > Are there other special tuning parameter for kvm/qemu/ to set on gluster? > as mentioned: all works except the livemigration (disk image file becomes > read only) > and I have to use something different than cache=none... > > TIA > > Bernhard > > > On 17.01.2014 05:04:52, Pranith Kumar Karampuri wrote: > Bernhard, > Configuration seems ok. Could you please give the log files of the bricks and > mount please. If you think it is not a big procedure to do this live > migration, could you set client-log-level to DEBUG and provide the log files > of that run. > > Pranith > > ----- Original Message ----- > From: "Bernhard Glomm" <[email protected]> > To: [email protected] > Cc: [email protected] > Sent: Thursday, January 16, 2014 5:58:17 PM > Subject: Re: [Gluster-users] gluster and kvm livemigration > > > hi Pranith > > # gluster volume info fs_vm_atom01 > > Volume Name: fs_vm_atom01 > Type: Replicate > Volume ID: fea9bdcf-783e-442a-831d-f564f8dbe551 > Status: Started > Number of Bricks: 1 x 2 = 2 > Transport-type: tcp > Bricks: > Brick1: 172.24.1.11:/zp_ping_1/fs_vm_atom01 > Brick2: 172.24.1.13:/zp_pong_1/fs_vm_atom01 > Options Reconfigured: > diagnostics.client-log-level: ERROR > > > TIA > Bernhard > > > On 16.01.2014 13:05:12, Pranith Kumar Karampuri wrote: > hi Bernhard, > Could you give gluster volume info output? > > Pranith > > ----- Original Message ----- > From: "Bernhard Glomm" <> > [email protected]> > > > To: > > [email protected] > Sent: Thursday, January 16, 2014 4:22:36 PM > Subject: [Gluster-users] gluster and kvm livemigration > > I experienced a strange behavior of glusterfs during livemigration > of a qemu-kvm guest > using a 10GB file on a mirrored gluster 3.4.2 volume > (both on ubuntu 13.04) > I run > virsh migrate --verbose --live --unsafe --p2p --domain atom01 --desturi > qemu+ssh://<target_ip>/system > and the migration works, > the running machine is pingable and keeps sending pings. > nevertheless, when I let the machine touch a file during migration > it stops, complaining that it's filesystem is read only (from that moment > that > migration finished) > A reboot from inside the machine failes, > machine goes down and comes up with an error > unable to write to sector xxxxxx on hd0 > (than falling into the initrd). > a > virsh destroy VM && virsh start VM > leads to a perfect running VM again, > no matter on which of the two hosts I start the machine > anybody better experience with livemigration? > any hint on a procedure how to debug that? > TIA > Bernhard > > -- > > > Bernhard Glomm > IT Administration > > Phone: +49 (30) 86880 134 > Fax: +49 (30) 86880 100 > Skype: bernhard.glomm.ecologic > > Ecologic Institut gemeinnützige GmbH | Pfalzburger Str. 43/44 | 10717 > Berlin > | Germany > GF: R. Andreas Kraemer | AG: Charlottenburg HRB 57947 | USt/VAT-IdNr.: > DE811963464 > Ecologic™ is a Trade Mark (TM) of Ecologic Institut gemeinnützige GmbH > > > _______________________________________________ > Gluster-users mailing list > [email protected] > http://supercolony.gluster.org/mailman/listinfo/gluster-users > > > > > -- > > > > > > > > > > Bernhard Glomm > > IT Administration > > > > > Phone: > > > +49 (30) 86880 134 > > > Fax: > > > +49 (30) 86880 100 > > > Skype: > > > bernhard.glomm.ecologic > > > > > > > > > > > > > > > > > Ecologic Institut gemeinnützige GmbH | Pfalzburger Str. 43/44 | > 10717 Berlin | Germany > > GF: R. Andreas Kraemer | AG: Charlottenburg HRB 57947 | > USt/VAT-IdNr.: DE811963464 > > Ecologic™ is a Trade Mark (TM) of Ecologic Institut gemeinnützige > GmbH > > > > > > > > > > -- > Bernhard Glomm > IT Administration > > Phone: +49 (30) 86880 134 > Fax: +49 (30) 86880 100 > Skype: bernhard.glomm.ecologic > > Ecologic Institut gemeinnützige GmbH | Pfalzburger Str. 43/44 | 10717 Berlin > | Germany > GF: R. Andreas Kraemer | AG: Charlottenburg HRB 57947 | USt/VAT-IdNr.: > DE811963464 > Ecologic™ is a Trade Mark (TM) of Ecologic Institut gemeinnützige GmbH > _______________________________________________ > Gluster-users mailing list > [email protected] > http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list [email protected] http://supercolony.gluster.org/mailman/listinfo/gluster-users
