Well, personally I know ZFS and I don’t know VDO. Going to have to check it out 
now that I know it exists, sounds interesting to have it at the dm layer. I 
don’t use it exclusively either, but have been finding it useful for 
compression.

What is your source for the statement that COW filesystems have downsides over 
time for VM workloads?

> On Nov 16, 2018, at 3:11 PM, Donny Davis <[email protected]> wrote:
> 
> Why not just use the built in stuff like VDO. What benefits does ZFS bring 
> for the use case? 
> For most vm based workloads ZFS is the opposite of ideal over the lifecycle 
> of a VM. COW filesystems have downsides over time. 
> 
> 
> 
> 
> On Thu, Nov 15, 2018 at 6:09 PM Darrell Budic <[email protected] 
> <mailto:[email protected]>> wrote:
> I did this in the past and didn’t have any trouble with gluster/ZFS, but 
> 4.2.x probably does more validation.
> 
> I recommend these settings on your zfs volumes, I set mine at the root(v0 
> here) and let them inherit:
> 
> required:
> v0    xattr                 sa                     local
> v0    acltype               posixacl               local
> optional but I recommend them:
> v0    relatime              on                     local
> v0    compression           lz4                    local
> 
> For gluster, I think it checks that it’s been “optimized for virt storage”. 
> Either apply the virt group or set the options you’ll find in 
> /var/lib/glusterd/groups/virt.
> 
> Note that I don’t recommend the default settings for cluster.shd-max-threads 
> & cluster.shd-wait-qlength. They can swamp your machines during heals unless 
> you have a lot of cores and ram. You get a slightly faster heal, but often 
> have VMs pausing for storage or other odd ball storage related errors. I 
> prefer max-threads = 1 or maybe 2, and wait-qlength=1024 or 2048. These are 
> per volume, so they hit harder than you think they will if you have a lot of 
> volumes running.
> 
> Also make sure the gluster volumes themselves got set to 36.36 for 
> owner.group, doesn’t matter for the bricks. Can do it with volume settings or 
> mount the volume and set it manually.
> 
> Hope it helps!
> 
>   -Darrell
> 
> 
>> On Nov 15, 2018, at 6:28 AM, Thomas Simmons <[email protected] 
>> <mailto:[email protected]>> wrote:
>> 
>> Hello All,
>> 
>> I recently took a new job in a RedHat shop and I'd like to move all of my 
>> homelab systems to RedHat upstream products to better align with what I 
>> manage at work. I had a "custom" (aka - hacked together) 3-node 
>> Hyperconverged XenServer cluster and would like to get this moved over to 
>> Ovirt (I'm currently testing with 4.2.7). Unfortunately, my storage is 
>> limited to software RAID with a 128GB SSD for cache. If at all possible, I 
>> would prefer to use ZFS (RAIDZ+ZIL+L2ARC) instead of MD RAID + lvmcache, 
>> however I'm not able to get this working and I'm not sure why. My ZFS and 
>> Gluster configuration is working - at least where I can manually mount all 
>> of my gluster volumes from all of my nodes, however, hosted-engine --deploy 
>> fails. I understand this isn't an out of the box configuration for Ovirt, 
>> however I see no reason why this shouldn't work. I would think this would be 
>> no different than using any other Gluster volume for the engine datastore, 
>> Am I missing something that would prevent this from working?
>> 
>> [ INFO  ] TASK [Add glusterfs storage domain]
>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is 
>> "[Storage Domain target is unsupported]". HTTP response code is 400.
>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "deprecations": 
>> [{"msg": "The 'ovirt_storage_domains' module is being renamed 
>> 'ovirt_storage_domain'", "version": 2.8}], "msg": "Fault reason is 
>> \"Operation Failed\". Fault detail is \"[Storage Domain target is 
>> unsupported]\". HTTP response code is 400."}
>> 
>> Even though it fails, it appears to have mounted and written 
>> __DIRECT_IO_TEST__ to my Gluster volume:
>> 
>> [root@vmh1 ~]# mount -t glusterfs localhost:/engine /mnt/engine/
>> [root@vmh1 ~]# ls /mnt/engine/
>> __DIRECT_IO_TEST__
>> 
>> If I cancel and try to run the deploy again, I get a different failure:
>> 
>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is "[Error 
>> creating a storage domain]". HTTP response code is 400.
>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "deprecations": 
>> [{"msg": "The 'ovirt_storage_domains' module is being renamed 
>> 'ovirt_storage_domain'", "version": 2.8}], "msg": "Fault reason is 
>> \"Operation Failed\". Fault detail is \"[Error creating a storage domain]\". 
>> HTTP response code is 400."}
>> 
>> Gluster seems ok...
>> 
>> [root@vmh1 /]# gluster volume info engine
>>  
>> Volume Name: engine
>> Type: Replicate
>> Volume ID: 2e34f8f5-0129-4ba5-983f-1eb5178deadc
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 1 x 3 = 3
>> Transport-type: tcp
>> Bricks:
>> Brick1: vmh1-ib:/zpool1/engine
>> Brick2: vmh2-ib:/zpool1/engine
>> Brick3: vmh3-ib:/zpool1/engine
>> Options Reconfigured:
>> transport.address-family: inet
>> nfs.disable: on
>> performance.client-io-threads: off
>> 
>> ZFS looks good too...
>> 
>> [root@vmh1 ~]# ansible ovirthosts -m shell -a 'zpool status' -b
>> vmh1 | CHANGED | rc=0 >>
>>   pool: zpool1
>>  state: ONLINE
>>   scan: none requested
>> config:
>> 
>>      NAME        STATE     READ WRITE CKSUM
>>      zpool1      ONLINE       0     0     0
>>        sdc       ONLINE       0     0     0
>>        sdd       ONLINE       0     0     0
>>        sde       ONLINE       0     0     0
>>      logs
>>        sdb2      ONLINE       0     0     0
>>      cache
>>        sdb1      ONLINE       0     0     0
>> 
>> errors: No known data errors
>> 
>> vmh3 | CHANGED | rc=0 >>
>>   pool: zpool1
>>  state: ONLINE
>>   scan: none requested
>> config:
>> 
>>      NAME        STATE     READ WRITE CKSUM
>>      zpool1      ONLINE       0     0     0
>>        sdc       ONLINE       0     0     0
>>        sdd       ONLINE       0     0     0
>>        sde       ONLINE       0     0     0
>>      logs
>>        sdb2      ONLINE       0     0     0
>>      cache
>>        sdb1      ONLINE       0     0     0
>> 
>> errors: No known data errors
>> 
>> vmh2 | CHANGED | rc=0 >>
>>   pool: zpool1
>>  state: ONLINE
>>   scan: none requested
>> config:
>> 
>>      NAME        STATE     READ WRITE CKSUM
>>      zpool1      ONLINE       0     0     0
>>        sdc       ONLINE       0     0     0
>>        sdd       ONLINE       0     0     0
>>        sde       ONLINE       0     0     0
>>      logs
>>        sdb2      ONLINE       0     0     0
>>      cache
>>        sdb1      ONLINE       0     0     0
>> 
>> errors: No known data errors
>> 
>> Permissions seem ok
>> 
>> [root@vmh1 ~]# ansible ovirthosts -m shell -a 'ls -n /zpool1' -b
>> vmh3 | CHANGED | rc=0 >>
>> total 2
>> drwxr-xr-x. 3 36 36 3 Nov 15 04:56 data
>> drwxr-xr-x. 3 36 36 4 Nov 15 06:31 engine
>> drwxr-xr-x. 3 36 36 3 Nov 15 04:56 iso
>> 
>> vmh1 | CHANGED | rc=0 >>
>> total 2
>> drwxr-xr-x. 3 36 36 3 Nov 15 04:56 data
>> drwxr-xr-x. 3 36 36 4 Nov 15 06:31 engine
>> drwxr-xr-x. 3 36 36 3 Nov 15 04:56 iso
>> 
>> vmh2 | CHANGED | rc=0 >>
>> total 2
>> drwxr-xr-x. 3 36 36 3 Nov 15 04:56 data
>> drwxr-xr-x. 3 36 36 4 Nov 15 06:31 engine
>> drwxr-xr-x. 3 36 36 3 Nov 15 04:56 iso
>> 
>> _______________________________________________
>> Users mailing list -- [email protected] <mailto:[email protected]>
>> To unsubscribe send an email to [email protected] 
>> <mailto:[email protected]>
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ 
>> <https://www.ovirt.org/site/privacy-policy/>
>> oVirt Code of Conduct: 
>> https://www.ovirt.org/community/about/community-guidelines/ 
>> <https://www.ovirt.org/community/about/community-guidelines/>
>> List Archives: 
>> https://lists.ovirt.org/archives/list/[email protected]/message/SEZLAUIUSLR4UFHGQNS27KQ5Y5TIZDL6/
>>  
>> <https://lists.ovirt.org/archives/list/[email protected]/message/SEZLAUIUSLR4UFHGQNS27KQ5Y5TIZDL6/>
> 
> _______________________________________________
> Users mailing list -- [email protected] <mailto:[email protected]>
> To unsubscribe send an email to [email protected] 
> <mailto:[email protected]>
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ 
> <https://www.ovirt.org/site/privacy-policy/>
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/ 
> <https://www.ovirt.org/community/about/community-guidelines/>
> List Archives: 
> https://lists.ovirt.org/archives/list/[email protected]/message/ZFZ7P6RFZ7XB3YMRIIIVOLE4H4DGYTWY/
>  
> <https://lists.ovirt.org/archives/list/[email protected]/message/ZFZ7P6RFZ7XB3YMRIIIVOLE4H4DGYTWY/>

_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/XZSGDKKPLWUWRIY6A2PODSICYSO3HC7X/

Reply via email to