Hey Matthew! Can you please provide me with the following to help you debug the issue that you are facing? 1. oVirt and gdeploy version 2. /var/log/messages file 3. /root/.gdeploy file
On Mon, Feb 25, 2019 at 1:23 PM Parth Dhanjal <[email protected]> wrote: > Hey Matthew! > > Can you please provide which oVirt and gdeploy version have you installed? > > Regards > Parth Dhanjal > > On Mon, Feb 25, 2019 at 12:56 PM Sahina Bose <[email protected]> wrote: > >> +Gobinda Das +Dhanjal Parth can you please check? >> >> On Fri, Feb 22, 2019 at 11:52 PM Matthew Roth <[email protected]> wrote: >> > >> > I have 3 servers, Node 1 is 3tb /dev/sda, Node 2, 3tb /dev/sdb, node3 >> 3tb /dev/sdb >> > >> > I start the process for gluster deployment. I change node 1 to sda and >> all the other ones to sdb. I get no errors however, >> > >> > when I get to >> > Creating physical Volume all it does is spin forever . doesnt get any >> further. I can leave it there for 5 hours and doesn't go anywhere. >> > >> > #gdeploy configuration generated by cockpit-gluster plugin >> > [hosts] >> > cmdnode1.cmd911.com >> > cmdnode2.cmd911.com >> > cmdnode3.cmd911.com >> > >> > [script1:cmdnode1.cmd911.com] >> > action=execute >> > ignore_script_errors=no >> > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sda -h >> cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com >> > >> > [script1:cmdnode2.cmd911.com] >> > action=execute >> > ignore_script_errors=no >> > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h >> cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com >> > >> > [script1:cmdnode3.cmd911.com] >> > action=execute >> > ignore_script_errors=no >> > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h >> cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com >> > >> > [disktype] >> > raid6 >> > >> > [diskcount] >> > 12 >> > >> > [stripesize] >> > 256 >> > >> > [service1] >> > action=enable >> > service=chronyd >> > >> > [service2] >> > action=restart >> > service=chronyd >> > >> > [shell2] >> > action=execute >> > command=vdsm-tool configure --force >> > >> > [script3] >> > action=execute >> > file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh >> > ignore_script_errors=no >> > >> > [pv1:cmdnode1.cmd911.com] >> > action=create >> > devices=sda >> > ignore_pv_errors=no >> > >> > [pv1:cmdnode2.cmd911.com] >> > action=create >> > devices=sdb >> > ignore_pv_errors=no >> > >> > [pv1:cmdnode3.cmd911.com] >> > action=create >> > devices=sdb >> > ignore_pv_errors=no >> > >> > [vg1:cmdnode1.cmd911.com] >> > action=create >> > vgname=gluster_vg_sda >> > pvname=sda >> > ignore_vg_errors=no >> > >> > [vg1:cmdnode2.cmd911.com] >> > action=create >> > vgname=gluster_vg_sdb >> > pvname=sdb >> > ignore_vg_errors=no >> > >> > [vg1:cmdnode3.cmd911.com] >> > action=create >> > vgname=gluster_vg_sdb >> > pvname=sdb >> > ignore_vg_errors=no >> > >> > [lv1:cmdnode1.cmd911.com] >> > action=create >> > poolname=gluster_thinpool_sda >> > ignore_lv_errors=no >> > vgname=gluster_vg_sda >> > lvtype=thinpool >> > size=1005GB >> > poolmetadatasize=5GB >> > >> > [lv2:cmdnode2.cmd911.com] >> > action=create >> > poolname=gluster_thinpool_sdb >> > ignore_lv_errors=no >> > vgname=gluster_vg_sdb >> > lvtype=thinpool >> > size=1005GB >> > poolmetadatasize=5GB >> > >> > [lv3:cmdnode3.cmd911.com] >> > action=create >> > poolname=gluster_thinpool_sdb >> > ignore_lv_errors=no >> > vgname=gluster_vg_sdb >> > lvtype=thinpool >> > size=41GB >> > poolmetadatasize=1GB >> > >> > [lv4:cmdnode1.cmd911.com] >> > action=create >> > lvname=gluster_lv_engine >> > ignore_lv_errors=no >> > vgname=gluster_vg_sda >> > mount=/gluster_bricks/engine >> > size=100GB >> > lvtype=thick >> > >> > [lv5:cmdnode1.cmd911.com] >> > action=create >> > lvname=gluster_lv_data >> > ignore_lv_errors=no >> > vgname=gluster_vg_sda >> > mount=/gluster_bricks/data >> > lvtype=thinlv >> > poolname=gluster_thinpool_sda >> > virtualsize=500GB >> > >> > [lv6:cmdnode1.cmd911.com] >> > action=create >> > lvname=gluster_lv_vmstore >> > ignore_lv_errors=no >> > vgname=gluster_vg_sda >> > mount=/gluster_bricks/vmstore >> > lvtype=thinlv >> > poolname=gluster_thinpool_sda >> > virtualsize=500GB >> > >> > [lv7:cmdnode2.cmd911.com] >> > action=create >> > lvname=gluster_lv_engine >> > ignore_lv_errors=no >> > vgname=gluster_vg_sdb >> > mount=/gluster_bricks/engine >> > size=100GB >> > lvtype=thick >> > >> > [lv8:cmdnode2.cmd911.com] >> > action=create >> > lvname=gluster_lv_data >> > ignore_lv_errors=no >> > vgname=gluster_vg_sdb >> > mount=/gluster_bricks/data >> > lvtype=thinlv >> > poolname=gluster_thinpool_sdb >> > virtualsize=500GB >> > >> > [lv9:cmdnode2.cmd911.com] >> > action=create >> > lvname=gluster_lv_vmstore >> > ignore_lv_errors=no >> > vgname=gluster_vg_sdb >> > mount=/gluster_bricks/vmstore >> > lvtype=thinlv >> > poolname=gluster_thinpool_sdb >> > virtualsize=500GB >> > >> > [lv10:cmdnode3.cmd911.com] >> > action=create >> > lvname=gluster_lv_engine >> > ignore_lv_errors=no >> > vgname=gluster_vg_sdb >> > mount=/gluster_bricks/engine >> > size=20GB >> > lvtype=thick >> > >> > [lv11:cmdnode3.cmd911.com] >> > action=create >> > lvname=gluster_lv_data >> > ignore_lv_errors=no >> > vgname=gluster_vg_sdb >> > mount=/gluster_bricks/data >> > lvtype=thinlv >> > poolname=gluster_thinpool_sdb >> > virtualsize=20GB >> > >> > [lv12:cmdnode3.cmd911.com] >> > action=create >> > lvname=gluster_lv_vmstore >> > ignore_lv_errors=no >> > vgname=gluster_vg_sdb >> > mount=/gluster_bricks/vmstore >> > lvtype=thinlv >> > poolname=gluster_thinpool_sdb >> > virtualsize=20GB >> > >> > [selinux] >> > yes >> > >> > [service3] >> > action=restart >> > service=glusterd >> > slice_setup=yes >> > >> > [firewalld] >> > action=add >> > >> ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp >> > services=glusterfs >> > >> > [script2] >> > action=execute >> > file=/usr/share/gdeploy/scripts/disable-gluster-hooks.sh >> > >> > [shell3] >> > action=execute >> > command=usermod -a -G gluster qemu >> > >> > [volume1] >> > action=create >> > volname=engine >> > transport=tcp >> > replica=yes >> > replica_count=3 >> > >> key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal >> > value=virt,36,36,30,on,off,enable >> > brick_dirs=cmdnode1.cmd911.com: >> /gluster_bricks/engine/engine,cmdnode2.cmd911.com: >> /gluster_bricks/engine/engine,cmdnode3.cmd911.com: >> /gluster_bricks/engine/engine >> > ignore_volume_errors=no >> > arbiter_count=1 >> > >> > [volume2] >> > action=create >> > volname=data >> > transport=tcp >> > replica=yes >> > replica_count=3 >> > >> key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal >> > value=virt,36,36,30,on,off,enable >> > brick_dirs=cmdnode1.cmd911.com: >> /gluster_bricks/data/data,cmdnode2.cmd911.com: >> /gluster_bricks/data/data,cmdnode3.cmd911.com:/gluster_bricks/data/data >> > ignore_volume_errors=no >> > arbiter_count=1 >> > >> > [volume3] >> > action=create >> > volname=vmstore >> > transport=tcp >> > replica=yes >> > replica_count=3 >> > >> key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal >> > value=virt,36,36,30,on,off,enable >> > brick_dirs=cmdnode1.cmd911.com: >> /gluster_bricks/vmstore/vmstore,cmdnode2.cmd911.com: >> /gluster_bricks/vmstore/vmstore,cmdnode3.cmd911.com: >> /gluster_bricks/vmstore/vmstore >> > ignore_volume_errors=no >> > arbiter_count=1 >> > _______________________________________________ >> > Users mailing list -- [email protected] >> > To unsubscribe send an email to [email protected] >> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> > oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> > List Archives: >> https://lists.ovirt.org/archives/list/[email protected]/message/5UA6OAQ2I7QJPCM6KQTLXG2NXVTYOTHD/ >> >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/AHFDUVDGQ2UKRC25BYMHCCCAP3RYNSN4/

