[ovirt-users] Gluster setup for oVirt

2022-09-12 Thread Jonas

Hello all

I tried to setup Gluster volumes in cockpit using the wizard. Based on 
Red Hat's recommendations I wanted to put the Volume for the oVirt 
Engine on a thick provisioned logical volume [1] and therefore removed 
the line thinpoolname and corresponding configuration from the yml file 
(see below). Unfortunately, this approach was not successful. My 
solution is now to only create a data volume and manually create a thick 
provisioned gluster volume manually. What would you recommend doing?


Thanks your any input :)

Regards,
Jonas

[1]: 
https://access.redhat.com/documentation/en-us/red_hat_hyperconverged_infrastructure_for_virtualization/1.8/html/deploying_red_hat_hyperconverged_infrastructure_for_virtualization/rhhi-requirements#rhhi-req-lv


hc_nodes:
  hosts:
    server-005.storage.int.rabe.ch:
  gluster_infra_volume_groups:
    - vgname: vg_tier1_01
  pvname: /dev/md/raid_tier1_gluster
  gluster_infra_mount_devices:
    - path: /gluster_bricks/tier1-ovirt-engine-01/gb-01
  lvname: lv_tier1_ovirt_engine_01
  vgname: vg_tier1_01
    - path: /gluster_bricks/tier1-ovirt-data-01/gb-01
  lvname: lv_tier1_ovirt_data_01
  vgname: vg_tier1_01
  blacklist_mpath_devices:
    - raid_tier1_gluster
  gluster_infra_thinpools:
    - vgname: vg_tier1_01
  thinpoolname: lv_tier1_ovirt_data_01_tp
  poolmetadatasize: 16G
  gluster_infra_lv_logicalvols:
    - vgname: vg_tier1_01
  lvname: lv_tier1_ovirt_engine_01
  lvsize: 100G
    - vgname: vg_tier1_01
  thinpool: lv_tier1_ovirt_data_01_tp
  lvname: lv_tier1_ovirt_data_01
  lvsize: 16000G
    server-006.storage.int.rabe.ch:
  gluster_infra_volume_groups:
    - vgname: vg_tier1_01
  pvname: /dev/md/raid_tier1_gluster
  gluster_infra_mount_devices:
    - path: /gluster_bricks/tier1-ovirt-engine-01/gb-01
  lvname: lv_tier1_ovirt_engine_01
  vgname: vg_tier1_01
    - path: /gluster_bricks/tier1-ovirt-data-01/gb-01
  lvname: lv_tier1_ovirt_data_01
  vgname: vg_tier1_01
  blacklist_mpath_devices:
    - raid_tier1_gluster
  gluster_infra_thinpools:
    - vgname: vg_tier1_01
  thinpoolname: lv_tier1_ovirt_data_01_tp
  poolmetadatasize: 16G
  gluster_infra_lv_logicalvols:
    - vgname: vg_tier1_01
  lvname: lv_tier1_ovirt_engine_01
  lvsize: 100G
    - vgname: vg_tier1_01
  thinpool: lv_tier1_ovirt_data_01_tp
  lvname: lv_tier1_ovirt_data_01
  lvsize: 16000G
    server-007.storage.int.rabe.ch:
  gluster_infra_volume_groups:
    - vgname: vg_tier0_01
  pvname: /dev/md/raid_tier0_gluster
  gluster_infra_mount_devices:
    - path: /gluster_bricks/tier1-ovirt-engine-01/gb-01
  lvname: lv_tier1_ovirt_engine_01
  vgname: vg_tier0_01
    - path: /gluster_bricks/tier1-ovirt-data-01/gb-01
  lvname: lv_tier1_ovirt_data_01
  vgname: vg_tier0_01
  blacklist_mpath_devices:
    - raid_tier0_gluster
  gluster_infra_thinpools:
    - vgname: vg_tier0_01
  thinpoolname: lv_tier1_ovirt_data_01_tp
  poolmetadatasize: 1G
  gluster_infra_lv_logicalvols:
    - vgname: vg_tier0_01
  lvname: lv_tier1_ovirt_engine_01
  lvsize: 20G
    - vgname: vg_tier0_01
  thinpool: lv_tier1_ovirt_data_01_tp
  lvname: lv_tier1_ovirt_data_01
  lvsize: 32G
  vars:
    gluster_infra_disktype: JBOD
    gluster_infra_daling: 1024K
    gluster_set_selinux_labels: true
    gluster_infra_fw_ports:
  - 2049/tcp
  - 54321/tcp
  - 5900/tcp
  - 5900-6923/tcp
  - 5666/tcp
  - 16514/tcp
    gluster_infra_fw_permanent: true
    gluster_infra_fw_state: enabled
    gluster_infra_fw_zone: public
    gluster_infra_fw_services:
  - glusterfs
    gluster_features_force_varlogsizecheck: false
    cluster_nodes:
  - server-005.storage.int.rabe.ch
  - server-006.storage.int.rabe.ch
  - server-007.storage.int.rabe.ch
    gluster_features_hci_cluster: '{{ cluster_nodes }}'
    gluster_features_hci_volumes:
  - volname: tier1-ovirt-engine-01
    brick: /gluster_bricks/tier1-ovirt-engine-01/gb-01
    arbiter: 1
  - volname: tier1-ovirt-data-01
    brick: /gluster_bricks/tier1-ovirt-data-01/gb-01
    arbiter: 1
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WX2CIKP43KOCG76UUWIOVI4QPXYNKQNX/


Re: [ovirt-users] gluster setup in ovirt

2016-10-27 Thread Sahina Bose
On Thu, Oct 27, 2016 at 4:28 AM, Thing  wrote:

> Hi,
>
> I have 3 machines imported into ovirt 4.0.4 just to do storage.  I have no
> storage setup.  I am a bit confused, can I add new storage from scratch via
> ovirt? ie picking mount points (I have /gv1 so set on each) or do I create
> the gluster replicated setup manually on each of the three nodes first? and
> import "ready made"?
>

You can setup gluster volume from scratch via oVirt
Once you add 3 nodes to the cluster - create bricks using the Storage
Devices sub-tab under each host. This will provision and mount the brick
from an unused raw block device on host.
Once the brick directories are mounted - use the New volume dialog to
create the volume



>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users