I tried adding a new storage domain on my hyper converged test cluster running 
Ovirt 4.3.3.7 and gluster 6.1. I was able to create the new gluster volume 
fine, but it’s not able to add the gluster storage domain (as either a managed 
gluster volume or directly entering values). The created gluster volume mounts 
and looks fine from the CLI. Errors in VDSM log:

2019-05-16 10:25:08,158-0500 INFO  (jsonrpc/1) [vdsm.api] START 
connectStorageServer(domType=7, spUUID=u'00000000-0000-0000-0000-000000000000', 
conList=[{u'mnt_options': u'backup-volfile-servers=10.50.3.11:10.50.3.10', 
u'id': u'00000000-0000-0000-0000-000000000000', u'connection': 
u'10.50.3.12:/test', u'iqn': u'', u'user': u'', u'tpgt': u'1', u'ipv6_enabled': 
u'false', u'vfs_type': u'glusterfs', u'password': '********', u'port': u''}], 
options=None) from=::ffff:10.100.90.5,44732, 
flow_id=fcde45c4-3b03-4a85-818a-06be560edee4, 
task_id=0582219d-ce68-4951-8fbd-3dce6d102fca (api:48)
2019-05-16 10:25:08,306-0500 INFO  (jsonrpc/1) 
[storage.StorageServer.MountConnection] Creating directory 
u'/rhev/data-center/mnt/glusterSD/10.50.3.12:_test' (storageServer:168)
2019-05-16 10:25:08,306-0500 INFO  (jsonrpc/1) [storage.fileUtils] Creating 
directory: /rhev/data-center/mnt/glusterSD/10.50.3.12:_test mode: None 
(fileUtils:199)
2019-05-16 10:25:08,306-0500 WARN  (jsonrpc/1) 
[storage.StorageServer.MountConnection] Using user specified 
backup-volfile-servers option (storageServer:275)
2019-05-16 10:25:08,306-0500 INFO  (jsonrpc/1) [storage.Mount] mounting 
10.50.3.12:/test at /rhev/data-center/mnt/glusterSD/10.50.3.12:_test (mount:204)
2019-05-16 10:25:08,453-0500 INFO  (jsonrpc/1) [IOProcessClient] (Global) 
Starting client (__init__:308)
2019-05-16 10:25:08,460-0500 INFO  (ioprocess/5389) [IOProcess] (Global) 
Starting ioprocess (__init__:434)
2019-05-16 10:25:08,473-0500 INFO  (itmap/0) [IOProcessClient] 
(/glusterSD/10.50.3.12:_test) Starting client (__init__:308)
2019-05-16 10:25:08,481-0500 INFO  (ioprocess/5401) [IOProcess] 
(/glusterSD/10.50.3.12:_test) Starting ioprocess (__init__:434)
2019-05-16 10:25:08,484-0500 INFO  (jsonrpc/1) [vdsm.api] FINISH 
connectStorageServer return={'statuslist': [{'status': 0, 'id': 
u'00000000-0000-0000-0000-000000000000'}]} from=::ffff:10.100.90.5,44732, 
flow_id=fcde45c4-3b03-4a85-818a-06be560edee4, 
task_id=0582219d-ce68-4951-8fbd-3dce6d102fca (api:54)
2019-05-16 10:25:08,484-0500 INFO  (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call 
StoragePool.connectStorageServer succeeded in 0.33 seconds (__init__:312)

2019-05-16 10:25:09,169-0500 INFO  (jsonrpc/7) [vdsm.api] START 
connectStorageServer(domType=7, spUUID=u'00000000-0000-0000-0000-000000000000', 
conList=[{u'mnt_options': u'backup-volfile-servers=10.50.3.11:10.50.3.10', 
u'id': u'd0ab6b05-2486-40f0-9b15-7f150017ec12', u'connection': 
u'10.50.3.12:/test', u'iqn': u'', u'user': u'', u'tpgt': u'1', u'ipv6_enabled': 
u'false', u'vfs_type': u'glusterfs', u'password': '********', u'port': u''}], 
options=None) from=::ffff:10.100.90.5,44732, flow_id=31d993dd, 
task_id=9eb2f42c-852d-4af6-ae4e-f65d8283d6e0 (api:48)
2019-05-16 10:25:09,180-0500 INFO  (jsonrpc/7) [vdsm.api] FINISH 
connectStorageServer return={'statuslist': [{'status': 0, 'id': 
u'd0ab6b05-2486-40f0-9b15-7f150017ec12'}]} from=::ffff:10.100.90.5,44732, 
flow_id=31d993dd, task_id=9eb2f42c-852d-4af6-ae4e-f65d8283d6e0 (api:54)
2019-05-16 10:25:09,180-0500 INFO  (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC call 
StoragePool.connectStorageServer succeeded in 0.01 seconds (__init__:312)
2019-05-16 10:25:09,186-0500 INFO  (jsonrpc/5) [vdsm.api] START 
createStorageDomain(storageType=7, 
sdUUID=u'4037f461-2b6d-452f-8156-fcdca820a8a1', domainName=u'gTest', 
typeSpecificArg=u'10.50.3.12:/test', domClass=1, domVersion=u'4', 
block_size=512, max_hosts=250, options=None) from=::ffff:10.100.90.5,44732, 
flow_id=31d993dd, task_id=ecea28f3-60d4-476d-9ba8-b753b7c9940d (api:48)
2019-05-16 10:25:09,492-0500 WARN  (jsonrpc/5) [storage.LVM] Reloading VGs 
failed (vgs=[u'4037f461-2b6d-452f-8156-fcdca820a8a1'] rc=5 out=[] err=['  
Volume group "4037f461-2b6d-452f-8156-fcdca820a8a1" not found', '  Cannot 
process volume group 4037f461-2b6d-452f-8156-fcdca820a8a1']) (lvm:442)
2019-05-16 10:25:09,507-0500 INFO  (jsonrpc/5) [storage.StorageDomain] 
sdUUID=4037f461-2b6d-452f-8156-fcdca820a8a1 domainName=gTest 
remotePath=10.50.3.12:/test domClass=1, block_size=512, alignment=1048576 
(nfsSD:86)
2019-05-16 10:25:09,521-0500 INFO  (jsonrpc/5) [IOProcessClient] 
(4037f461-2b6d-452f-8156-fcdca820a8a1) Starting client (__init__:308)
2019-05-16 10:25:09,528-0500 INFO  (ioprocess/5437) [IOProcess] 
(4037f461-2b6d-452f-8156-fcdca820a8a1) Starting ioprocess (__init__:434)
2019-05-16 10:25:09,584-0500 ERROR (jsonrpc/5) [storage.fileSD] Underlying file 
system doesn't supportdirect IO (fileSD:110)
2019-05-16 10:25:09,584-0500 INFO  (jsonrpc/5) [vdsm.api] FINISH 
createStorageDomain error=Storage Domain target is unsupported: () 
from=::ffff:10.100.90.5,44732, flow_id=31d993dd, 
task_id=ecea28f3-60d4-476d-9ba8-b753b7c9940d (api:52)
2019-05-16 10:25:09,584-0500 ERROR (jsonrpc/5) [storage.TaskManager.Task] 
(Task='ecea28f3-60d4-476d-9ba8-b753b7c9940d') Unexpected error (task:875)
2019-05-16 10:25:09,584-0500 INFO  (jsonrpc/5) [storage.TaskManager.Task] 
(Task='ecea28f3-60d4-476d-9ba8-b753b7c9940d') aborting: Task is aborted: 
'Storage Domain target is unsupported: ()' - code 399 (task:1181)
2019-05-16 10:25:09,584-0500 ERROR (jsonrpc/5) [storage.Dispatcher] FINISH 
createStorageDomain error=Storage Domain target is unsupported: () 
(dispatcher:83)
2019-05-16 10:25:09,584-0500 INFO  (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC call 
StorageDomain.create failed (error 399) in 0.40 seconds (__init__:312)


> On May 16, 2019, at 11:55 AM, Nir Soffer <[email protected]> wrote:
> 
> On Thu, May 16, 2019 at 7:42 PM Strahil <[email protected] 
> <mailto:[email protected]>> wrote:
> Hi Sandro,
> 
> Thanks for the update.
> 
> I have just upgraded to RC1 (using gluster v6 here)  and the issue  I 
> detected in 4.3.3.7 <http://4.3.3.7/> - where gluster Storage domain fails 
> creation - is still present.
> 
> 
> What is is this issue? can you provide a link to the bug/mail about it?
> 
> Can you check if the 'dd' command executed during creation has been recently 
> modified ?
> 
> I've received update from Darrell  (also gluster v6) , but haven't received 
> an update from anyone who is using gluster v5 -> thus I haven't opened a bug 
> yet.
> 
> Best Regards,
> Strahil Nikolov
> 
> On May 16, 2019 11:21, Sandro Bonazzola <[email protected] 
> <mailto:[email protected]>> wrote:
> The oVirt Project is pleased to announce the availability of the oVirt 4.3.4 
> First Release Candidate, as of May 16th, 2019.
> 
> This update is a release candidate of the fourth in a series of stabilization 
> updates to the 4.3 series.
> This is pre-release software. This pre-release should not to be used 
> inproduction.
> 
> This release is available now on x86_64 architecture for:
> * Red Hat Enterprise Linux 7.6 or later
> * CentOS Linux (or similar) 7.6 or later
> 
> This release supports Hypervisor Hosts on x86_64 and ppc64le architectures 
> for:
> * Red Hat Enterprise Linux 7.6 or later
> * CentOS Linux (or similar) 7.6 or later
> * oVirt Node 4.3 (available for x86_64 only)
> 
> Experimental tech preview for x86_64 and s390x architectures for Fedora 28 is 
> also included.
> 
> See the release notes [1] for installation / upgrade instructions and a list 
> of new features and bugs fixed.
> 
> Notes:
> - oVirt Appliance is already available
> - oVirt Node is already available[2]
> 
> Additional Resources:
> * Read more about the oVirt 4.3.4 release 
> highlights:http://www.ovirt.org/release/4.3.4/ 
> <http://www.ovirt.org/release/4.3.4/>
> * Get more oVirt Project updates on Twitter: https://twitter.com/ovirt 
> <https://twitter.com/ovirt>
> * Check out the latest project news on the oVirt 
> blog:http://www.ovirt.org/blog/ <http://www.ovirt.org/blog/>
> 
> [1] http://www.ovirt.org/release/4.3.4/ <http://www.ovirt.org/release/4.3.4/>
> [2] http://resources.ovirt.org/pub/ovirt-4.3-pre/iso/ 
> <http://resources.ovirt.org/pub/ovirt-4.3-pre/iso/>
> 
> -- 
> Sandro Bonazzola
> 
> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
> Red Hat EMEA <https://www.redhat.com/>
> [email protected] <mailto:[email protected]>   
>  <https://red.ht/sig> 
>  <https://redhat.com/summit>_______________________________________________
> Users mailing list -- [email protected] <mailto:[email protected]>
> To unsubscribe send an email to [email protected] 
> <mailto:[email protected]>
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ 
> <https://www.ovirt.org/site/privacy-policy/>
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/ 
> <https://www.ovirt.org/community/about/community-guidelines/>
> List Archives: 
> https://lists.ovirt.org/archives/list/[email protected]/message/REDV54BH7CIIDRCRUPCUYN4TX5Z3SL6R/
>  
> <https://lists.ovirt.org/archives/list/[email protected]/message/REDV54BH7CIIDRCRUPCUYN4TX5Z3SL6R/>
> _______________________________________________
> Announce mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives: 
> https://lists.ovirt.org/archives/list/[email protected]/message/ABFECS5ES4MVL3UZC34GLIDN5PNDTNOR/

_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/RO6PQQ4XQ6KZXRUOHKMM2J2HFN2LTEGB/

Reply via email to