On Mon, May 20, 2019 at 9:55 PM Adrian Quintero <adrianquint...@gmail.com>
wrote:

> Sahina,
> Yesterday I started with a fresh install, I completely wiped clean all the
> disks, recreated the arrays from within my controller of our DL380 Gen 9's.
>
> OS: RAID 1 (2x600GB HDDs): /dev/sda    // Using ovirt node 4.3.3.1 iso.
> engine and VMSTORE1: JBOD (1x3TB HDD):/dev/sdb
> DATA1: JBOD (1x3TB HDD): /dev/sdc
> DATA2: JBOD (1x3TB HDD): /dev/sdd
> Caching disk: JOBD (1x440GB SDD): /dev/sde
>
> *After the OS install on the first 3 servers and setting up ssh keys,  I
> started the Hyperconverged deploy process:*
> 1.-Logged int to the first server http://host1.example.com:9090
> 2.-Selected Hyperconverged, clicked on "Run Gluster Wizard"
> 3.-Followed the wizard steps (Hosts, FQDNs, Packages, Volumes, Bricks,
> Review)
> *Hosts/FQDNs:*
> host1.example.com
> host2.example.com
> host3.example.com
> *Packages:*
> *Volumes:*
> engine:replicate:/gluster_bricks/engine/engine
> vmstore1:replicate:/gluster_bricks/vmstore1/vmstore1
> data1:replicate:/gluster_bricks/data1/data1
> data2:replicate:/gluster_bricks/data2/data2
> *Bricks:*
> engine:/dev/sdb:100GB:/gluster_bricks/engine
> vmstore1:/dev/sdb:2600GB:/gluster_bricks/vmstrore1
> data1:/dev/sdc:2700GB:/gluster_bricks/data1
> data2:/dev/sdd:2700GB:/gluster_bricks/data2
> LV Cache:
> /dev/sde:400GB:writethrough
> 4.-After I hit deploy on the last step of the "Wizard" that is when I get
> the disk filter error.
> TASK [gluster.infra/roles/backend_setup : Create volume groups]
> ****************
> failed: [vmm10.virt.iad3p] (item={u'vgname': u'gluster_vg_sdb', u'pvname':
> u'/dev/sdb'}) => {"changed": false, "err": "  Device /dev/sdb excluded by a
> filter.\n", "item": {"pvname": "/dev/sdb", "vgname": "gluster_vg_sdb"},
> "msg": "Creating physical volume '/dev/sdb' failed", "rc": 5}
> failed: [vmm12.virt.iad3p] (item={u'vgname': u'gluster_vg_sdb', u'pvname':
> u'/dev/sdb'}) => {"changed": false, "err": "  Device /dev/sdb excluded by a
> filter.\n", "item": {"pvname": "/dev/sdb", "vgname": "gluster_vg_sdb"},
> "msg": "Creating physical volume '/dev/sdb' failed", "rc": 5}
> failed: [vmm11.virt.iad3p] (item={u'vgname': u'gluster_vg_sdb', u'pvname':
> u'/dev/sdb'}) => {"changed": false, "err": "  Device /dev/sdb excluded by a
> filter.\n", "item": {"pvname": "/dev/sdb", "vgname": "gluster_vg_sdb"},
> "msg": "Creating physical volume '/dev/sdb' failed", "rc": 5}
> failed: [vmm12.virt.iad3p] (item={u'vgname': u'gluster_vg_sdc', u'pvname':
> u'/dev/sdc'}) => {"changed": false, "err": "  Device /dev/sdc excluded by a
> filter.\n", "item": {"pvname": "/dev/sdc", "vgname": "gluster_vg_sdc"},
> "msg": "Creating physical volume '/dev/sdc' failed", "rc": 5}
> failed: [vmm10.virt.iad3p] (item={u'vgname': u'gluster_vg_sdc', u'pvname':
> u'/dev/sdc'}) => {"changed": false, "err": "  Device /dev/sdc excluded by a
> filter.\n", "item": {"pvname": "/dev/sdc", "vgname": "gluster_vg_sdc"},
> "msg": "Creating physical volume '/dev/sdc' failed", "rc": 5}
> failed: [vmm11.virt.iad3p] (item={u'vgname': u'gluster_vg_sdc', u'pvname':
> u'/dev/sdc'}) => {"changed": false, "err": "  Device /dev/sdc excluded by a
> filter.\n", "item": {"pvname": "/dev/sdc", "vgname": "gluster_vg_sdc"},
> "msg": "Creating physical volume '/dev/sdc' failed", "rc": 5}
> failed: [vmm10.virt.iad3p] (item={u'vgname': u'gluster_vg_sdd', u'pvname':
> u'/dev/sdd'}) => {"changed": false, "err": "  Device /dev/sdd excluded by a
> filter.\n", "item": {"pvname": "/dev/sdd", "vgname": "gluster_vg_sdd"},
> "msg": "Creating physical volume '/dev/sdd' failed", "rc": 5}
> failed: [vmm12.virt.iad3p] (item={u'vgname': u'gluster_vg_sdd', u'pvname':
> u'/dev/sdd'}) => {"changed": false, "err": "  Device /dev/sdd excluded by a
> filter.\n", "item": {"pvname": "/dev/sdd", "vgname": "gluster_vg_sdd"},
> "msg": "Creating physical volume '/dev/sdd' failed", "rc": 5}
> failed: [vmm11.virt.iad3p] (item={u'vgname': u'gluster_vg_sdd', u'pvname':
> u'/dev/sdd'}) => {"changed": false, "err": "  Device /dev/sdd excluded by a
> filter.\n", "item": {"pvname": "/dev/sdd", "vgname": "gluster_vg_sdd"},
> "msg": "Creating physical volume '/dev/sdd' failed", "rc": 5}
>
> Attached is the generated yml file ( /etc/ansible/hc_wizard_inventory.yml)
> and the "Deployment Failed" file
>
>
>
>
> Also wondering if I hit this bug?
> https://bugzilla.redhat.com/show_bug.cgi?id=1635614
>
>
+Sachidananda URS <s...@redhat.com> +Gobinda Das <go...@redhat.com> to
review the inventory file and failures


>
> Thanks for looking into this.
>
> *Adrian Quintero*
> *adrianquint...@gmail.com <adrianquint...@gmail.com> |
> adrian.quint...@rackspace.com <adrian.quint...@rackspace.com>*
>
>
> On Mon, May 20, 2019 at 7:56 AM Sahina Bose <sab...@redhat.com> wrote:
>
>> To scale existing volumes - you need to add bricks and run rebalance on
>> the gluster volume so that data is correctly redistributed as Alex
>> mentioned.
>> We do support expanding existing volumes as the bug
>> https://bugzilla.redhat.com/show_bug.cgi?id=1471031 has been fixed
>>
>> As to procedure to expand volumes:
>> 1. Create bricks from UI - select Host -> Storage Device -> Storage
>> device. Click on "Create Brick"
>> If the device is shown as locked, make sure there's no signature on
>> device.  If multipath entries have been created for local devices, you can
>> blacklist those devices in multipath.conf and restart multipath.
>> (If you see device as locked even after you do this -please report back).
>> 2. Expand volume using Volume -> Bricks -> Add Bricks, and select the 3
>> bricks created in previous step
>> 3. Run Rebalance on the volume. Volume -> Rebalance.
>>
>>
>> On Thu, May 16, 2019 at 2:48 PM Fred Rolland <froll...@redhat.com> wrote:
>>
>>> Sahina,
>>> Can someone from your team review the steps done by Adrian?
>>> Thanks,
>>> Freddy
>>>
>>> On Thu, Apr 25, 2019 at 5:14 PM Adrian Quintero <
>>> adrianquint...@gmail.com> wrote:
>>>
>>>> Ok, I will remove the extra 3 hosts, rebuild them from scratch and
>>>> re-attach them to clear any possible issues and try out the suggestions
>>>> provided.
>>>>
>>>> thank you!
>>>>
>>>> On Thu, Apr 25, 2019 at 9:22 AM Strahil Nikolov <hunter86...@yahoo.com>
>>>> wrote:
>>>>
>>>>> I have the same locks , despite I have blacklisted all local disks:
>>>>>
>>>>> # VDSM PRIVATE
>>>>> blacklist {
>>>>>         devnode "*"
>>>>>         wwid Crucial_CT256MX100SSD1_14390D52DCF5
>>>>>         wwid WDC_WD5000AZRX-00A8LB0_WD-WCC1U0056126
>>>>>         wwid WDC_WD5003ABYX-01WERA0_WD-WMAYP2335378
>>>>>         wwid
>>>>> nvme.1cc1-324a31313230303131353936-414441544120535838323030504e50-00000001
>>>>> }
>>>>>
>>>>> If you have multipath reconfigured, do not forget to rebuild the
>>>>> initramfs (dracut -f). It's a linux issue , and not oVirt one.
>>>>>
>>>>> In your case you had something like this:
>>>>>                /dev/VG/LV
>>>>>           /dev/disk/by-id/pvuuid
>>>>>      /dev/mapper/multipath-uuid
>>>>> /dev/sdb
>>>>>
>>>>> Linux will not allow you to work with /dev/sdb , when multipath is
>>>>> locking the block device.
>>>>>
>>>>> Best Regards,
>>>>> Strahil Nikolov
>>>>>
>>>>> В четвъртък, 25 април 2019 г., 8:30:16 ч. Гринуич-4, Adrian Quintero <
>>>>> adrianquint...@gmail.com> написа:
>>>>>
>>>>>
>>>>> under Compute, hosts, select the host that has the locks on /dev/sdb,
>>>>> /dev/sdc, etc.., select storage devices and in here is where you see a
>>>>> small column with a bunch of lock images showing for each row.
>>>>>
>>>>>
>>>>> However as a work around, on the newly added hosts (3 total), I had to
>>>>> manually modify /etc/multipath.conf and add the following at the end as
>>>>> this is what I noticed from the original 3 node setup.
>>>>>
>>>>> -------------------------------------------------------------
>>>>> # VDSM REVISION 1.3
>>>>> # VDSM PRIVATE
>>>>> # BEGIN Added by gluster_hci role
>>>>>
>>>>> blacklist {
>>>>>         devnode "*"
>>>>> }
>>>>> # END Added by gluster_hci role
>>>>> ----------------------------------------------------------
>>>>> After this I restarted multipath and the lock went away and was able
>>>>> to configure the new bricks thru the UI, however my concern is what will
>>>>> happen if I reboot the server will the disks be read the same way by the 
>>>>> OS?
>>>>>
>>>>> Also now able to expand the gluster with a new replicate 3 volume if
>>>>> needed using http://host4.mydomain.com:9090.
>>>>>
>>>>>
>>>>> thanks again
>>>>>
>>>>> On Thu, Apr 25, 2019 at 8:00 AM Strahil Nikolov <hunter86...@yahoo.com>
>>>>> wrote:
>>>>>
>>>>> In which menu do you see it this way ?
>>>>>
>>>>> Best Regards,
>>>>> Strahil Nikolov
>>>>>
>>>>> В сряда, 24 април 2019 г., 8:55:22 ч. Гринуич-4, Adrian Quintero <
>>>>> adrianquint...@gmail.com> написа:
>>>>>
>>>>>
>>>>> Strahil,
>>>>> this is the issue I am seeing now
>>>>>
>>>>> [image: image.png]
>>>>>
>>>>> The is thru the UI when I try to create a new brick.
>>>>>
>>>>> So my concern is if I modify the filters on the OS what impact will
>>>>> that have after server reboots?
>>>>>
>>>>> thanks,
>>>>>
>>>>>
>>>>>
>>>>> On Mon, Apr 22, 2019 at 11:39 PM Strahil <hunter86...@yahoo.com>
>>>>> wrote:
>>>>>
>>>>> I have edited my multipath.conf to exclude local disks , but you need
>>>>> to set '#VDSM private' as per the comments in the header of the file.
>>>>> Otherwise, use the /dev/mapper/multipath-device notation - as you
>>>>> would do with any linux.
>>>>>
>>>>> Best Regards,
>>>>> Strahil NikolovOn Apr 23, 2019 01:07, adrianquint...@gmail.com wrote:
>>>>> >
>>>>> > Thanks Alex, that makes more sense now  while trying to follow the
>>>>> instructions provided I see that all my disks /dev/sdb, /dev/sdc, /dev/sdd
>>>>> are locked and inidicating " multpath_member" hence not letting me create
>>>>> new bricks. And on the logs I see
>>>>> >
>>>>> > Device /dev/sdb excluded by a filter.\n", "item": {"pvname":
>>>>> "/dev/sdb", "vgname": "gluster_vg_sdb"}, "msg": "Creating physical volume
>>>>> '/dev/sdb' failed", "rc": 5}
>>>>> > Same thing for sdc, sdd
>>>>> >
>>>>> > Should I manually edit the filters inside the OS, what will be the
>>>>> impact?
>>>>> >
>>>>> > thanks again.
>>>>> > _______________________________________________
>>>>> > Users mailing list -- users@ovirt.org
>>>>> > To unsubscribe send an email to users-le...@ovirt.org
>>>>> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>> > oVirt Code of Conduct:
>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>> > List Archives:
>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/FW3IR3NMQTYZLXBT2VLOCLBKOYJS3MYF/
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Adrian Quintero
>>>>> _______________________________________________
>>>>> Users mailing list -- users@ovirt.org
>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>> oVirt Code of Conduct:
>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>> List Archives:
>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/EW7NKT76JR3TLPP63M7DTDF2TLSMX556/
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Adrian Quintero
>>>>> _______________________________________________
>>>>> Users mailing list -- users@ovirt.org
>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>> oVirt Code of Conduct:
>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>> List Archives:
>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/KRKR5LFARNRHFRHVQUA5IUFAHLVG2ENK/
>>>>>
>>>>
>>>>
>>>> --
>>>> Adrian Quintero
>>>> _______________________________________________
>>>> Users mailing list -- users@ovirt.org
>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>> oVirt Code of Conduct:
>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>> List Archives:
>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/5EQPZTGK6PWG427OSDBIDCQPT4RDY4ZC/
>>>>
>>>
>
> --
> Adrian Quintero
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MSARMX67N4BFMNXJYOYEBQO3NZDLI5TE/

Reply via email to