Still, now I cannot start none of the 2 machines! I get ID 119 VM proxy2 is down. Exit message: Child quit during startup handshake: Input/output error.""
Something similar to bug https://bugzilla.redhat.com/show_bug.cgi?id=1033064, except that in my case selinux is permissive! On Wed, Dec 18, 2013 at 2:10 PM, Gabi C <[email protected]> wrote: > in my case $brick_path =/data > > > getfattr -d /data return NOTHING on both nodes!!! > > > > On Wed, Dec 18, 2013 at 1:46 PM, Fabian Deutsch <[email protected]>wrote: > >> Am Mittwoch, den 18.12.2013, 13:26 +0200 schrieb Gabi C: >> > Update on Glusterfs issue >> > >> > >> > I manage to recover lost volume after recretaing the same volume name >> > with same bricks, whisch raised an error message, resolved by, on both >> > nodes: >> > >> > setfattr -x trusted.glusterfs.volume-id $brick_path >> > setfattr -x trusted.gfid $brick_path >> >> Hey, >> >> good that you could recover them. >> >> Could you please provide $brick_path and getfattr -d $brick_path >> >> The question is if and/or why the fattrs are not stored. >> >> - fabian >> >> > >> > >> > >> > On Wed, Dec 18, 2013 at 12:12 PM, Gabi C <[email protected]> wrote: >> > node 1: >> > >> > [root@virtual5 admin]# cat /config/files >> > /etc/fstab >> > /etc/shadow >> > /etc/default/ovirt >> > /etc/ssh/ssh_host_key >> > /etc/ssh/ssh_host_key.pub >> > /etc/ssh/ssh_host_dsa_key >> > /etc/ssh/ssh_host_dsa_key.pub >> > /etc/ssh/ssh_host_rsa_key >> > /etc/ssh/ssh_host_rsa_key.pub >> > /etc/rsyslog.conf >> > /etc/libvirt/libvirtd.conf >> > /etc/libvirt/passwd.db >> > /etc/passwd >> > /etc/sysconfig/network >> > /etc/collectd.conf >> > /etc/libvirt/qemu/networks >> > /etc/ssh/sshd_config >> > /etc/pki >> > /etc/logrotate.d/ovirt-node >> > /var/lib/random-seed >> > /etc/iscsi/initiatorname.iscsi >> > /etc/libvirt/qemu.conf >> > /etc/sysconfig/libvirtd >> > /etc/logrotate.d/libvirtd >> > /etc/multipath.conf >> > /etc/hosts >> > /etc/sysconfig/network-scripts/ifcfg-enp3s0 >> > /etc/sysconfig/network-scripts/ifcfg-lo >> > /etc/ntp.conf >> > /etc/shadow >> > /etc/vdsm-reg/vdsm-reg.conf >> > /etc/shadow >> > /etc/shadow >> > /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt >> > /etc/sysconfig/network-scripts/route-ovirtmgmt >> > /etc/sysconfig/network-scripts/rule-ovirtmgmt >> > /root/.ssh/authorized_keys >> > /etc/vdsm/vdsm.id >> > /etc/udev/rules.d/12-ovirt-iosched.rules >> > /etc/vdsm/vdsm.conf >> > /etc/sysconfig/iptables >> > /etc/resolv.conf >> > /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY >> > /etc/sysconfig/network-scripts/ifcfg-enp6s0 >> > /etc/sysconfig/network-scripts/ifcfg-enp6s0.50 >> > /etc/glusterfs/glusterd.vol >> > /etc/selinux/config >> > >> > >> > >> > >> > >> > >> > >> > >> > node 2: >> > >> > >> > [root@virtual4 ~]# cat /config/files >> > /etc/fstab >> > /etc/shadow >> > /etc/default/ovirt >> > /etc/ssh/ssh_host_key >> > /etc/ssh/ssh_host_key.pub >> > /etc/ssh/ssh_host_dsa_key >> > /etc/ssh/ssh_host_dsa_key.pub >> > /etc/ssh/ssh_host_rsa_key >> > /etc/ssh/ssh_host_rsa_key.pub >> > /etc/rsyslog.conf >> > /etc/libvirt/libvirtd.conf >> > /etc/libvirt/passwd.db >> > /etc/passwd >> > /etc/sysconfig/network >> > /etc/collectd.conf >> > /etc/libvirt/qemu/networks >> > /etc/ssh/sshd_config >> > /etc/pki >> > /etc/logrotate.d/ovirt-node >> > /var/lib/random-seed >> > /etc/iscsi/initiatorname.iscsi >> > /etc/libvirt/qemu.conf >> > /etc/sysconfig/libvirtd >> > /etc/logrotate.d/libvirtd >> > /etc/multipath.conf >> > /etc/hosts >> > /etc/sysconfig/network-scripts/ifcfg-enp3s0 >> > /etc/sysconfig/network-scripts/ifcfg-lo >> > /etc/shadow >> > /etc/shadow >> > /etc/vdsm-reg/vdsm-reg.conf >> > /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt >> > /etc/sysconfig/network-scripts/route-ovirtmgmt >> > /etc/sysconfig/network-scripts/rule-ovirtmgmt >> > /root/.ssh/authorized_keys >> > /etc/shadow >> > /etc/shadow >> > /etc/vdsm/vdsm.id >> > /etc/udev/rules.d/12-ovirt-iosched.rules >> > /etc/sysconfig/iptables >> > /etc/vdsm/vdsm.conf >> > /etc/shadow >> > /etc/resolv.conf >> > /etc/ntp.conf >> > /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY >> > /etc/sysconfig/network-scripts/ifcfg-enp6s0 >> > /etc/sysconfig/network-scripts/ifcfg-enp6s0.50 >> > /etc/glusterfs/glusterd.vol >> > /etc/selinux/config >> > >> > >> > >> > >> > On Wed, Dec 18, 2013 at 12:07 PM, Fabian Deutsch >> > <[email protected]> wrote: >> > Am Mittwoch, den 18.12.2013, 12:03 +0200 schrieb Gabi >> > C: >> > > So here it is: >> > > >> > > >> > > in tab volumes add new volume - Replicated, then >> > added storage - >> > > data/glusterfs. Then I impoerted Vm, ran them and at >> > some point, >> > > needing some space for a Redhat Satellite instance >> > I decided to put >> > > both node in maintenace stop them add new disk >> > devices and restart, >> > > but after restart the gluster volume defined under >> > Volumes Tab >> > > vanished! >> > >> > >> > Antoni, >> > >> > can you tell what log files to look at to find out why >> > that storage >> > domain vanished - from a Engine side? >> > >> > And do you know what files related to gluster are >> > changed on the Node >> > side? >> > >> > Gabi, >> > >> > could you please provide the contents of /config/files >> > on the Node. >> > >> > > Glusterfs data goes under /data directory which was >> > automatically >> > > configured when I installed the node. >> > >> > >> > Yep, /data is on the Data LV - that should be good. >> > >> > - fabian >> > >> > > >> > > >> > > On Wed, Dec 18, 2013 at 11:45 AM, Fabian Deutsch >> > <[email protected]> >> > > wrote: >> > > Am Mittwoch, den 18.12.2013, 11:42 +0200 >> > schrieb Gabi C: >> > > > Yes, it is the VM part..I just run into an >> > issue. My setup >> > > consist in >> > > > 2 nodes with glusterfs and after adding >> > supplemental hard >> > > disk, after >> > > > reboot I've lost glusterfs volumes! >> > > >> > > >> > > Could you exactly explain what you >> > configured? >> > > >> > > > >> > > > How can I persist any configuration on >> > node and I refer here >> > > to >> > > > ''setenforce 0'' - for ssh login to work- >> > and further >> > > >> > > >> > > How changes can be persisted on Node can be >> > found here: >> > > >> > >> http://www.ovirt.org/Node_Troubleshooting#Making_changes_on_the_host >> > > >> > > Do you know into what path the glusterfs >> > data goes? Or is it >> > > written >> > > directly onto a disk/LV? >> > > >> > > - fabian >> > > >> > > > "" >> > http://www.ovirt.org/Features/GlusterFS_Storage_Domain >> > > > * option rpc-auth-allow-insecure on >> > ==> in >> > > glusterd.vol (ensure >> > > > u restart glusterd service... for >> > this to take >> > > effect) >> > > >> > > > * volume set <volname> >> > server.allow-insecure on ==> >> > > (ensure u >> > > > stop and start the volume.. for >> > this to take >> > > effect)'' >> > > > >> > > > >> > > > Thanks! >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > On Wed, Dec 18, 2013 at 11:35 AM, Fabian >> > Deutsch >> > > <[email protected]> >> > > > wrote: >> > > > Am Mittwoch, den 18.12.2013, 08:34 >> > +0200 schrieb >> > > Gabi C: >> > > > > Hello! >> > > > > >> > > > > >> > > > > In order to increase disk space >> > I want to add a >> > > new disk >> > > > drive to >> > > > > ovirt node. After adding this >> > should I proceed as >> > > "normal" - >> > > > pvcreate, >> > > > > vgcreate, lvcreate and so on - >> > or these >> > > configuration will >> > > > not >> > > > > persist? >> > > > >> > > > >> > > > Hey Gabi, >> > > > >> > > > basically plain LVM is used in >> > Node - so yes >> > > pvcreate and >> > > > lvextend can >> > > > be used. >> > > > What storage part do you want to >> > extend? The part >> > > where the >> > > > VMs reside? >> > > > You will also need to take care to >> > extend the >> > > filesystem. >> > > > >> > > > - fabian >> > > > >> > > > >> > > > >> > > >> > > >> > > >> > > >> > >> > >> > >> > >> > >> > >> >> >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

