[ovirt-devel] Re: execution failed: javax.net.ssl.SSLPeerUnverifiedException (was: vdsm.storage.exception.UnknownTask: Task id unknown (was: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master -

2020-07-21 Thread Martin Perina
On Tue, Jul 21, 2020 at 3:44 PM Yedidyah Bar David  wrote:

> We had several different threads about this failure. I'll use current
> to send a summary:
>
> - Our wildfly build was patched to include newer httpclient/core:
> https://gerrit.ovirt.org/110324
>
> - The appliance was patched to use master snapshot:
> https://gerrit.ovirt.org/110397
>
> - Now he-basic-suite succeeded, first time after many months:
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/1680/


So that's great news and from now on we should make sure that it stays
green (meaning everybody need to pay the same attention as to the basic
suite)


>
> Thanks to everyone involved!
>
> Best regards,
>
> On Tue, Jul 7, 2020 at 3:41 PM Martin Perina  wrote:
> >
> > Hi,
> >
> > I'm not aware of change regarding certificates recently. So is this
> error reproducible outside Jenkins? Or even better is it reproducible on
> some easier flow other than HE installation so we can debug what
> certificate is loaded in VDSM?
> >
> > Thanks,
> > Martin
> >
> > On Tue, Jul 7, 2020 at 2:07 PM Yedidyah Bar David 
> wrote:
> >>
> >> On Tue, Jul 7, 2020 at 12:50 PM Yedidyah Bar David 
> wrote:
> >> >
> >> > On Wed, Jun 24, 2020 at 2:14 PM Evgeny Slutsky 
> wrote:
> >> > >
> >> > > Hi,
> >> > > changing the hostname to include also the domain name fixed the
> cert deployment issue:
> >> > > https://gerrit.ovirt.org/#/c/109842/
> >> > >
> >> > > not sure how it affects the engine certificate content.
> >> > > from my offline discussion with @Martin Perina  this was that
> change that could cause it:
> >> > > https://gerrit.ovirt.org/#/c/109636/
> >> > >
> >> > > any thoughts?
> >> >
> >> > Above two patches are merged, but we still fail the same way:
> >> >
> >> >
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/1664/
> >> >
> >> >
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/1664/artifact/exported-artifacts/test_logs/he-basic-suite-master/post-he_deploy/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-setup/engine-logs-2020-07-07T03%3A15%3A01Z/ovirt-engine/engine.log
> >> >
> >> > 2020-07-06 23:04:25,555-04 ERROR
> >> > [org.ovirt.engine.core.vdsbroker.irsbroker.UploadStreamVDSCommand]
> >> >
> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-38)
> >> > [fb28ce9] Command 'UploadStreamVDSCommand(HostName =
> >> > lago-he-basic-suite-master-host-0.lago.local,
> >> >
> UploadStreamVDSCommandParameters:{hostId='e096650f-a7d6-4383-b1bb-f2e61327aac0'})'
> >> > execution failed: javax.net.ssl.SSLPeerUnverifiedException:
> >> > Certificate for  doesn't
> >> > match any of the subject alternative names:
> >> > [lago-he-basic-suite-master-host-0.lago.local]
> >> >
> >> > Any idea?
> >>
> >> And I now see this is indeed what's failing hosted-engine deploy at:
> >>
> >> 2020-07-07 05:51:58,573-0400 INFO ansible task start {'status': 'OK',
> >> 'ansible_type': 'task', 'ansible_playbook':
> >> '/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml',
> >> 'ansible_task': 'ovirt.hosted_engine_setup : Check OVF_STORE volume
> >> status'}
> >>
> >> (See other thread: [oVirt Jenkins]
> >> ovirt-system-tests_he-basic-suite-master - Build # 1655 - Still
> >> Failing! )
> >>
> >> On a successful run, engine.log has:
> >>
> >> 2020-07-02 18:01:55,527+03 INFO
> >>
> [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand]
> >> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-84)
> >>  [2b0721d8] Running command: ProcessOvfUpdateForStorageDomainCommand
> >> internal: true. Entities affected :  ID:
> >> e102d7b5-1a37-490f-a3e7-20e56c37791f Type: StorageAction group
> >> MANIPULATE_STORAG
> >> E_DOMAIN with role type ADMIN
> >> 2020-07-02 18:01:55,607+03 INFO
> >>
> [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand]
> >> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-84)
> >> [2b0721d8
> >> ] START, SetVolumeDescriptionVDSCommand(
> >>
> SetVolumeDescriptionVDSCommandParameters:{storagePoolId='b9dccefe-bc61-11ea-8ebe-001a4a231728',
> >> ignoreFailoverLimit='false', storageDomainId='e102d7b
> >> 5-1a37-490f-a3e7-20e56c37791f',
> >> imageGroupId='db934a98-4111-4faf-8cb9-6b36928cd61c',
> >> imageId='f898c40e-1f88-48db-b59b-f2c73162ddb7'}), log id: e203e51
> >> 2020-07-02 18:01:55,609+03 INFO
> >>
> [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand]
> >> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-84)
> >> [2b0721d8
> >> ] -- executeIrsBrokerCommand: calling 'setVolumeDescription',
> parameters:
> >> 2020-07-02 18:01:55,609+03 INFO
> >>
> [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand]
> >> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-84)
> >> [2b0721d8
> >> ] ++ spUUID=b9dccefe-bc61-11ea-8ebe-001a4a231728
> >> 2020-07-02 18:01:55,609+03 INFO
> >>
> 

[ovirt-devel] Re: execution failed: javax.net.ssl.SSLPeerUnverifiedException (was: vdsm.storage.exception.UnknownTask: Task id unknown (was: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master -

2020-07-21 Thread Yedidyah Bar David
We had several different threads about this failure. I'll use current
to send a summary:

- Our wildfly build was patched to include newer httpclient/core:
https://gerrit.ovirt.org/110324

- The appliance was patched to use master snapshot:
https://gerrit.ovirt.org/110397

- Now he-basic-suite succeeded, first time after many months:
https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/1680/

Thanks to everyone involved!

Best regards,

On Tue, Jul 7, 2020 at 3:41 PM Martin Perina  wrote:
>
> Hi,
>
> I'm not aware of change regarding certificates recently. So is this error 
> reproducible outside Jenkins? Or even better is it reproducible on some 
> easier flow other than HE installation so we can debug what certificate is 
> loaded in VDSM?
>
> Thanks,
> Martin
>
> On Tue, Jul 7, 2020 at 2:07 PM Yedidyah Bar David  wrote:
>>
>> On Tue, Jul 7, 2020 at 12:50 PM Yedidyah Bar David  wrote:
>> >
>> > On Wed, Jun 24, 2020 at 2:14 PM Evgeny Slutsky  wrote:
>> > >
>> > > Hi,
>> > > changing the hostname to include also the domain name fixed the  cert 
>> > > deployment issue:
>> > > https://gerrit.ovirt.org/#/c/109842/
>> > >
>> > > not sure how it affects the engine certificate content.
>> > > from my offline discussion with @Martin Perina  this was that change 
>> > > that could cause it:
>> > > https://gerrit.ovirt.org/#/c/109636/
>> > >
>> > > any thoughts?
>> >
>> > Above two patches are merged, but we still fail the same way:
>> >
>> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/1664/
>> >
>> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/1664/artifact/exported-artifacts/test_logs/he-basic-suite-master/post-he_deploy/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-setup/engine-logs-2020-07-07T03%3A15%3A01Z/ovirt-engine/engine.log
>> >
>> > 2020-07-06 23:04:25,555-04 ERROR
>> > [org.ovirt.engine.core.vdsbroker.irsbroker.UploadStreamVDSCommand]
>> > (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-38)
>> > [fb28ce9] Command 'UploadStreamVDSCommand(HostName =
>> > lago-he-basic-suite-master-host-0.lago.local,
>> > UploadStreamVDSCommandParameters:{hostId='e096650f-a7d6-4383-b1bb-f2e61327aac0'})'
>> > execution failed: javax.net.ssl.SSLPeerUnverifiedException:
>> > Certificate for  doesn't
>> > match any of the subject alternative names:
>> > [lago-he-basic-suite-master-host-0.lago.local]
>> >
>> > Any idea?
>>
>> And I now see this is indeed what's failing hosted-engine deploy at:
>>
>> 2020-07-07 05:51:58,573-0400 INFO ansible task start {'status': 'OK',
>> 'ansible_type': 'task', 'ansible_playbook':
>> '/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml',
>> 'ansible_task': 'ovirt.hosted_engine_setup : Check OVF_STORE volume
>> status'}
>>
>> (See other thread: [oVirt Jenkins]
>> ovirt-system-tests_he-basic-suite-master - Build # 1655 - Still
>> Failing! )
>>
>> On a successful run, engine.log has:
>>
>> 2020-07-02 18:01:55,527+03 INFO
>> [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand]
>> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-84)
>>  [2b0721d8] Running command: ProcessOvfUpdateForStorageDomainCommand
>> internal: true. Entities affected :  ID:
>> e102d7b5-1a37-490f-a3e7-20e56c37791f Type: StorageAction group
>> MANIPULATE_STORAG
>> E_DOMAIN with role type ADMIN
>> 2020-07-02 18:01:55,607+03 INFO
>> [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand]
>> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-84)
>> [2b0721d8
>> ] START, SetVolumeDescriptionVDSCommand(
>> SetVolumeDescriptionVDSCommandParameters:{storagePoolId='b9dccefe-bc61-11ea-8ebe-001a4a231728',
>> ignoreFailoverLimit='false', storageDomainId='e102d7b
>> 5-1a37-490f-a3e7-20e56c37791f',
>> imageGroupId='db934a98-4111-4faf-8cb9-6b36928cd61c',
>> imageId='f898c40e-1f88-48db-b59b-f2c73162ddb7'}), log id: e203e51
>> 2020-07-02 18:01:55,609+03 INFO
>> [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand]
>> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-84)
>> [2b0721d8
>> ] -- executeIrsBrokerCommand: calling 'setVolumeDescription', parameters:
>> 2020-07-02 18:01:55,609+03 INFO
>> [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand]
>> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-84)
>> [2b0721d8
>> ] ++ spUUID=b9dccefe-bc61-11ea-8ebe-001a4a231728
>> 2020-07-02 18:01:55,609+03 INFO
>> [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand]
>> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-84)
>> [2b0721d8
>> ] ++ sdUUID=e102d7b5-1a37-490f-a3e7-20e56c37791f
>> 2020-07-02 18:01:55,609+03 INFO
>> [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand]
>> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-84)
>> [2b0721d8
>> ] ++ 

[ovirt-devel] Re: Check OVF_STORE volume status task failures

2020-07-21 Thread Yedidyah Bar David
On Tue, Jul 21, 2020 at 12:56 PM Marcin Sobczyk  wrote:
>
> Hi,
>
> this is fixed since some time with [1], but I think that the appliance still 
> uses old version of ovirt-engine-wildfly-overlay.
> The fixed version is 19.1.0-2 [2]. Nir, could you take a look at it and bump 
> the version in the appliance if necessary?

Should be fixed by https://gerrit.ovirt.org/110397

>
> Thanks, Marcin
>
> [1] https://gerrit.ovirt.org/#/c/110324/
> [2] https://gerrit.ovirt.org/#/c/110324/1/automation/build-artifacts.sh
>
> On 7/20/20 3:53 PM, Artem Hrechanychenko wrote:
>
> Hi all,
> maybe I miss some information, but I'm still have troubles with he 
> installation using ost CI.
>
> Is that already fixed ?
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/10415/
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/10415/artifact/check-patch.he-basic_suite_master.el8.x86_64/test_logs/he-basic-suite-master/post-he_deploy/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20200720061238-soz41s.log
>
> 2020-07-20 06:36:04,455-0400 INFO 
> otopi.ovirt_hosted_engine_setup.ansible_utils 
> ansible_utils._process_output:111 TASK [ovirt.hosted_engine_setup : Check 
> OVF_STORE volume status]
> 2020-07-20 06:40:22,815-0400 DEBUG 
> otopi.ovirt_hosted_engine_setup.ansible_utils 
> ansible_utils._process_output:105 {'results': [{'cmd': ['vdsm-client', 
> 'Volume', 'getInfo', 'storagepoolID=03a466fe-ca73-11ea-b77e-5452c0a8c863', 
> 'storagedomainID=6685cca5-f0e1-4831-acdf-6f7b50596142', 
> 'imageID=4d2f7009-5b79-4b44-b0ef-e152bc51649f', 
> 'volumeID=7b953d0e-662d-4e72-9fdc-823ea867262b'], 'stdout': '{\n
> "apparentsize": "134217728",\n"capacity": "134217728",\n"children": 
> [],\n"ctime": "1595241259",\n"description": 
> "{\\"Updated\\":false,\\"Last Updated\\":null,\\"Storage 
> Domains\\":[{\\"uuid\\":\\"6685cca5-f0e1-4831-acdf-6f7b50596142\\"}],\\"Disk 
> Description\\":\\"OVF_STORE\\"}",\n"disktype": "OVFS",\n"domain": 
> "6685cca5-f0e1-4831-acdf-6f7b50596142",\n"format": "RAW",\n
> "generation": 0,\n"image": "4d2f7009-5b79-4b44-b0ef-e152bc51649f",\n
> "lease": {\n"offset": 0,\n"owners": [],\n"path": 
> "/rhev/data-center/mnt/lago-he-basic-suite-master-storage:_exports_nfs__he/6685cca5-f0e1-4831-acdf-6f7b50596142/images/4d2f7009-5b79-4b44-b0ef-e152bc51649f/7b953d0e-662d-4e72-9fdc-823ea867262b.lease",\n
> "version": null\n},\n"legality": "LEGAL",\n"mtime": 
> "0",\n"parent": "----",\n"pool": 
> "",\n"status": "OK",\n"truesize": "134217728",\n"type": 
> "PREALLOCATED",\n"uuid": "7b953d0e-662d-4e72-9fdc-823ea867262b",\n
> "voltype": "LEAF"\n}', 'stderr': '', 'rc': 0, 'start': '2020-07-20 
> 06:38:13.456845', 'end': '2020-07-20 06:38:13.897280', 'delta': 
> '0:00:00.440435', 'changed': True, 'invocation': {'module_args': 
> {'_raw_params': 'vdsm-client Volume getInfo 
> storagepoolID=03a466fe-ca73-11ea-b77e-5452c0a8c863 
> storagedomainID=6685cca5-f0e1-4831-acdf-6f7b50596142 
> imageID=4d2f7009-5b79-4b44-b0ef-e152bc51649f 
> volumeID=7b953d0e-662d-4e72-9fdc-823ea867262b', 'warn': True, '_uses_shell': 
> False, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 
> 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': 
> None}}, 'stdout_lines': ['{', '"apparentsize": "134217728",', '
> "capacity": "134217728",', '"children": [],', '"ctime": 
> "1595241259",', '"description": "{\\"Updated\\":false,\\"Last 
> Updated\\":null,\\"Storage 
> Domains\\":[{\\"uuid\\":\\"6685cca5-f0e1-4831-acdf-6f7b50596142\\"}],\\"Disk 
> Description\\":\\"OVF_STORE\\"}",', '"disktype": "OVFS",', '"domain": 
> "6685cca5-f0e1-4831-acdf-6f7b50596142",', '"format": "RAW",', '
> "generation": 0,', '"image": "4d2f7009-5b79-4b44-b0ef-e152bc51649f",', '  
>   "lease": {', '"offset": 0,', '"owners": [],', '
> "path": 
> "/rhev/data-center/mnt/lago-he-basic-suite-master-storage:_exports_nfs__he/6685cca5-f0e1-4831-acdf-6f7b50596142/images/4d2f7009-5b79-4b44-b0ef-e152bc51649f/7b953d0e-662d-4e72-9fdc-823ea867262b.lease",',
>  '"version": null', '},', '"legality": "LEGAL",', '
> "mtime": "0",', '"parent": "----",', '
> "pool": "",', '"status": "OK",', '"truesize": "134217728",', '
> "type": "PREALLOCATED",', '"uuid": 
> "7b953d0e-662d-4e72-9fdc-823ea867262b",', '"voltype": "LEAF"', '}'], 
> 'stderr_lines': [], '_ansible_no_log': False, 'failed': True, 'attempts': 12, 
> 'item': {'name': 'OVF_STORE', 'image_id': 
> '7b953d0e-662d-4e72-9fdc-823ea867262b', 'id': 
> '4d2f7009-5b79-4b44-b0ef-e152bc51649f'}, 'ansible_loop_var': 'item', 
> '_ansible_item_label': {'name': 'OVF_STORE', 'image_id': 
> '7b953d0e-662d-4e72-9fdc-823ea867262b', 'id': 
> 

[ovirt-devel] Re: Check OVF_STORE volume status task failures

2020-07-21 Thread Marcin Sobczyk

Hi,

this is fixed since some time with [1], but I think that the appliance 
still uses old version of ovirt-engine-wildfly-overlay.
The fixed version is 19.1.0-2 [2]. Nir, could you take a look at it and 
bump the version in the appliance if necessary?


Thanks, Marcin

[1] https://gerrit.ovirt.org/#/c/110324/
[2] https://gerrit.ovirt.org/#/c/110324/1/automation/build-artifacts.sh

On 7/20/20 3:53 PM, Artem Hrechanychenko wrote:

Hi all,
maybe I miss some information, but I'm still have troubles with he 
installation using ost CI.


Is that already fixed ?

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/10415/

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/10415/artifact/check-patch.he-basic_suite_master.el8.x86_64/test_logs/he-basic-suite-master/post-he_deploy/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20200720061238-soz41s.log

2020-07-20 06:36:04,455-0400 INFO otopi.ovirt_hosted_engine_setup.ansible_utils 
ansible_utils._process_output:111 TASK [ovirt.hosted_engine_setup : Check 
OVF_STORE volume status]
2020-07-20 06:40:22,815-0400 DEBUG otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:105 {'results': [{'cmd': ['vdsm-client', 'Volume', 'getInfo', 'storagepoolID=03a466fe-ca73-11ea-b77e-5452c0a8c863', 'storagedomainID=6685cca5-f0e1-4831-acdf-6f7b50596142', 'imageID=4d2f7009-5b79-4b44-b0ef-e152bc51649f', 'volumeID=7b953d0e-662d-4e72-9fdc-823ea867262b'], 'stdout': '{\n"apparentsize": "134217728",\n"capacity": "134217728",\n"children": [],\n"ctime": "1595241259",\n"description": "{\\"Updated\\":false,\\"Last Updated\\":null,\\"Storage Domains\\":[{\\"uuid\\":\\"6685cca5-f0e1-4831-acdf-6f7b50596142\\"}],\\"Disk Description\\":\\"OVF_STORE\\"}",\n"disktype": "OVFS",\n"domain": "6685cca5-f0e1-4831-acdf-6f7b50596142",\n"format": "RAW",\n"generation": 0,\n"image": "4d2f7009-5b79-4b44-b0ef-e152bc51649f",\n"lease": {\n"offset": 0,\n"owners": [],\n"path": "/rhev/data-center/mnt/lago-he-basic-suite-master-storage:_exports_nfs__he/6685cca5-f0e1-4831-acdf-6f7b50596142/images/4d2f7009-5b79-4b44-b0ef-e152bc51649f/7b953d0e-662d-4e72-9fdc-823ea867262b.lease",\n"version": null\n},\n"legality": "LEGAL",\n"mtime": "0",\n"parent": "----",\n"pool": "",\n"status": "OK",\n"truesize": "134217728",\n"type": "PREALLOCATED",\n"uuid": "7b953d0e-662d-4e72-9fdc-823ea867262b",\n"voltype": "LEAF"\n}', 'stderr': '', 'rc': 0, 'start': '2020-07-20 06:38:13.456845', 'end': '2020-07-20 
06:38:13.897280', 'delta': '0:00:00.440435', 'changed': True, 'invocation': {'module_args': {'_raw_params': 'vdsm-client Volume getInfo storagepoolID=03a466fe-ca73-11ea-b77e-5452c0a8c863 storagedomainID=6685cca5-f0e1-4831-acdf-6f7b50596142 imageID=4d2f7009-5b79-4b44-b0ef-e152bc51649f volumeID=7b953d0e-662d-4e72-9fdc-823ea867262b', 'warn': True, '_uses_shell': False, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['{', '"apparentsize": "134217728",', '"capacity": "134217728",', '"children": [],', '"ctime": "1595241259",', '"description": "{\\"Updated\\":false,\\"Last Updated\\":null,\\"Storage Domains\\":[{\\"uuid\\":\\"6685cca5-f0e1-4831-acdf-6f7b50596142\\"}],\\"Disk Description\\":\\"OVF_STORE\\"}",', '"disktype": "OVFS",', '"domain": "6685cca5-f0e1-4831-acdf-6f7b50596142",', '"format": "RAW",', '"generation": 0,', '"image": "4d2f7009-5b79-4b44-b0ef-e152bc51649f",', '"lease": {', '"offset": 0,', '"owners": [],', '"path": "/rhev/data-center/mnt/lago-he-basic-suite-master-storage:_exports_nfs__he/6685cca5-f0e1-4831-acdf-6f7b50596142/images/4d2f7009-5b79-4b44-b0ef-e152bc51649f/7b953d0e-662d-4e72-9fdc-823ea867262b.lease",', '"version": null', '},', '"legality": "LEGAL",', '"mtime": "0",', '"parent": "----",', '"pool": "",', '"status": "OK",', '"truesize": "134217728",', '"type": "PREALLOCATED",', '
"uuid": "7b953d0e-662d-4e72-9fdc-823ea867262b",', '"voltype": "LEAF"', '}'], 'stderr_lines': [], '_ansible_no_log': False, 'failed': True, 'attempts': 12, 'item': {'name': 'OVF_STORE', 'image_id': '7b953d0e-662d-4e72-9fdc-823ea867262b', 'id': '4d2f7009-5b79-4b44-b0ef-e152bc51649f'}, 'ansible_loop_var': 'item', '_ansible_item_label': {'name': 'OVF_STORE', 'image_id': '7b953d0e-662d-4e72-9fdc-823ea867262b', 'id': '4d2f7009-5b79-4b44-b0ef-e152bc51649f'}}, {'cmd': ['vdsm-client', 'Volume', 'getInfo', 'storagepoolID=03a466fe-ca73-11ea-b77e-5452c0a8c863', 'storagedomainID=6685cca5-f0e1-4831-acdf-6f7b50596142', 'imageID=044e384a-dedf-4589-8dfb-beca170138ee', 'volumeID=033d64fd-6f93-42be-84bc-082b03095ef3'], 'stdout': '{\n

[ovirt-devel] ovirt-engine has been tagged (ovirt-engine-4.4.1.10)

2020-07-21 Thread Tal Nisan

___
Devel mailing list -- devel@ovirt.org
To unsubscribe send an email to devel-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/devel@ovirt.org/message/7VWGN5AMWNKTDR5BBVU4GF3JXWGDTB6N/


[ovirt-devel] Re: device compatibility interface for live migration with assigned devices

2020-07-21 Thread Jason Wang


On 2020/7/20 下午6:39, Sean Mooney wrote:

On Mon, 2020-07-20 at 11:41 +0800, Jason Wang wrote:

On 2020/7/18 上午12:12, Alex Williamson wrote:

On Thu, 16 Jul 2020 16:32:30 +0800
Yan Zhao  wrote:


On Thu, Jul 16, 2020 at 12:16:26PM +0800, Jason Wang wrote:

On 2020/7/14 上午7:29, Yan Zhao wrote:

hi folks,
we are defining a device migration compatibility interface that helps upper
layer stack like openstack/ovirt/libvirt to check if two devices are
live migration compatible.
The "devices" here could be MDEVs, physical devices, or hybrid of the two.
e.g. we could use it to check whether
- a src MDEV can migrate to a target MDEV,
- a src VF in SRIOV can migrate to a target VF in SRIOV,
- a src MDEV can migration to a target VF in SRIOV.
 (e.g. SIOV/SRIOV backward compatibility case)

The upper layer stack could use this interface as the last step to check
if one device is able to migrate to another device before triggering a real
live migration procedure.
we are not sure if this interface is of value or help to you. please don't
hesitate to drop your valuable comments.


(1) interface definition
The interface is defined in below way:

__userspace
 /\  \
/ \write
   / read  \
  /__   ___\|/_
 | migration_version | | migration_version |-->check migration
 - -   compatibility
device Adevice B


a device attribute named migration_version is defined under each device's
sysfs node. e.g. 
(/sys/bus/pci/devices/\:00\:02.0/$mdev_UUID/migration_version).

Are you aware of the devlink based device management interface that is
proposed upstream? I think it has many advantages over sysfs, do you
consider to switch to that?

Advantages, such as?


My understanding for devlink(netlink) over sysfs (some are mentioned at
the time of vDPA sysfs mgmt API discussion) are:

i tought netlink was used more a as a configuration protocoal to qurry and 
confire nic and i guess
other devices in its devlink form requireint a tool to be witten that can speak 
the protocal to interact with.
the primary advantate of sysfs is that everything is just a file. there are no 
addtional depleenceis
needed



Well, if you try to build logic like introspection on top for a 
sophisticated hardware, you probably need to have library on top. And 
it's attribute per file is pretty inefficient.




  and unlike netlink there are not interoperatblity issues in a coanitnerised 
env. if you are using diffrenet
version of libc and gcc in the contaienr vs the host my understanding is tools 
like ethtool from ubuntu deployed
in a container on a centos host can have issue communicating with the host 
kernel.



Kernel provides stable ABI for userspace, so it's not something that we 
can't fix.




if its jsut a file unless
the format the data is returnin in chagnes or the layout of sysfs changes its 
compatiable regardless of what you
use to read it.



I believe you can't change sysfs layout which is part of uABI. But as I 
mentioned below, sysfs has several drawbacks. It's not harm to compare 
between different approach when you start a new device management API.


Thanks



- existing users (NIC, crypto, SCSI, ib), mature and stable
- much better error reporting (ext_ack other than string or errno)
- namespace aware
- do not couple with kobject

Thanks


___
Devel mailing list -- devel@ovirt.org
To unsubscribe send an email to devel-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/devel@ovirt.org/message/N525IDOKAXQ3IGK6LUQJ65AARPAEDIBV/


[ovirt-devel] Re: device compatibility interface for live migration with assigned devices

2020-07-21 Thread Yan Zhao
On Fri, Jul 17, 2020 at 10:12:58AM -0600, Alex Williamson wrote:
<...>
> > yes, in another reply, Alex proposed to use an interface in json format.
> > I guess we can define something like
> > 
> > { "self" :
> >   [
> > { "pciid" : "8086591d",
> >   "driver" : "i915",
> >   "gvt-version" : "v1",
> >   "mdev_type"   : "i915-GVTg_V5_2",
> >   "aggregator"  : "1",
> >   "pv-mode" : "none",
> > }
> >   ],
> >   "compatible" :
> >   [
> > { "pciid" : "8086591d",
> >   "driver" : "i915",
> >   "gvt-version" : "v1",
> >   "mdev_type"   : "i915-GVTg_V5_2",
> >   "aggregator"  : "1"
> >   "pv-mode" : "none",
> > },
> > { "pciid" : "8086591d",
> >   "driver" : "i915",
> >   "gvt-version" : "v1",
> >   "mdev_type"   : "i915-GVTg_V5_4",
> >   "aggregator"  : "2"
> >   "pv-mode" : "none",
> > },
> > { "pciid" : "8086591d",
> >   "driver" : "i915",
> >   "gvt-version" : "v2",
> >   "mdev_type"   : "i915-GVTg_V5_4",
> >   "aggregator"  : "2"
> >   "pv-mode" : "none, ppgtt, context",
> > }
> > ...
> >   ]
> > }
> > 
> > But as those fields are mostly vendor specific, the userspace can
> > only do simple string comparing, I guess the list would be very long as
> > it needs to enumerate all possible targets.
> 
> 
> This ignores so much of what I tried to achieve in my example :(
> 
sorry, I just was eager to show and confirm the way to list all compatible
combination of mdev_type and mdev attributes.

> 
> > also, in some fileds like "gvt-version", is there a simple way to express
> > things like v2+?
> 
> 
> That's not a reasonable thing to express anyway, how can you be certain
> that v3 won't break compatibility with v2?  Sean proposed a versioning
> scheme that accounts for this, using an x.y.z version expressing the
> major, minor, and bugfix versions, where there is no compatibility
> across major versions, minor versions have forward compatibility (ex. 1
> -> 2 is ok, 2 -> 1 is not) and bugfix version number indicates some
> degree of internal improvement that is not visible to the user in terms
> of features or compatibility, but provides a basis for preferring
> equally compatible candidates.
>
right. if self version is v1, it can't know its compatible version is
v2. it can only be done in reverse. i.e.
when self version is v2, it can list its compatible version is v1 and
v2.
and maybe later when self version is v3, there's no v1 in its compatible
list.

In this way, do you think we still need the complex x.y.z versioning scheme?

>  
> > If the userspace can read this interface both in src and target and
> > check whether both src and target are in corresponding compatible list, I
> > think it will work for us.
> > 
> > But still, kernel should not rely on userspace's choice, the opaque
> > compatibility string is still required in kernel. No matter whether
> > it would be exposed to userspace as an compatibility checking interface,
> > vendor driver would keep this part of code and embed the string into the
> > migration stream. so exposing it as an interface to be used by libvirt to
> > do a safety check before a real live migration is only about enabling
> > the kernel part of check to happen ahead.
> 
> As you indicate, the vendor driver is responsible for checking version
> information embedded within the migration stream.  Therefore a
> migration should fail early if the devices are incompatible.  Is it
but as I know, currently in VFIO migration protocol, we have no way to
get vendor specific compatibility checking string in migration setup stage
(i.e. .save_setup stage) before the device is set to _SAVING state.
In this way, for devices who does not save device data in precopy stage,
the migration compatibility checking is as late as in stop-and-copy
stage, which is too late.
do you think we need to add the getting/checking of vendor specific
compatibility string early in save_setup stage?

> really libvirt's place to second guess what it has been directed to do?
if libvirt uses the scheme of reading compatibility string at source and
writing for checking at the target, it can not be called "a second guess".
It's not a guess, but a confirmation.

> Why would we even proceed to design a user parse-able version interface
> if we still have a dependency on an opaque interface?  Thanks,
one reason is that libvirt can't trust the parsing result from
openstack.
Another reason is that libvirt can use this opaque interface easier than
another parsing by itself, in the fact that it would not introduce more
burden to kernel who would write this part of code anyway, no matter
libvirt uses it or not.
 
Thanks
Yan
___
Devel mailing list -- devel@ovirt.org
To unsubscribe send an email to devel-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 

[ovirt-devel] Re: device compatibility interface for live migration with assigned devices

2020-07-21 Thread Sean Mooney
On Mon, 2020-07-20 at 11:41 +0800, Jason Wang wrote:
> On 2020/7/18 上午12:12, Alex Williamson wrote:
> > On Thu, 16 Jul 2020 16:32:30 +0800
> > Yan Zhao  wrote:
> > 
> > > On Thu, Jul 16, 2020 at 12:16:26PM +0800, Jason Wang wrote:
> > > > On 2020/7/14 上午7:29, Yan Zhao wrote:
> > > > > hi folks,
> > > > > we are defining a device migration compatibility interface that helps 
> > > > > upper
> > > > > layer stack like openstack/ovirt/libvirt to check if two devices are
> > > > > live migration compatible.
> > > > > The "devices" here could be MDEVs, physical devices, or hybrid of the 
> > > > > two.
> > > > > e.g. we could use it to check whether
> > > > > - a src MDEV can migrate to a target MDEV,
> > > > > - a src VF in SRIOV can migrate to a target VF in SRIOV,
> > > > > - a src MDEV can migration to a target VF in SRIOV.
> > > > > (e.g. SIOV/SRIOV backward compatibility case)
> > > > > 
> > > > > The upper layer stack could use this interface as the last step to 
> > > > > check
> > > > > if one device is able to migrate to another device before triggering 
> > > > > a real
> > > > > live migration procedure.
> > > > > we are not sure if this interface is of value or help to you. please 
> > > > > don't
> > > > > hesitate to drop your valuable comments.
> > > > > 
> > > > > 
> > > > > (1) interface definition
> > > > > The interface is defined in below way:
> > > > > 
> > > > >__userspace
> > > > > /\  \
> > > > >/ \write
> > > > >   / read  \
> > > > >  /__   ___\|/_
> > > > > | migration_version | | migration_version |-->check migration
> > > > > - -   compatibility
> > > > >device Adevice B
> > > > > 
> > > > > 
> > > > > a device attribute named migration_version is defined under each 
> > > > > device's
> > > > > sysfs node. e.g. 
> > > > > (/sys/bus/pci/devices/\:00\:02.0/$mdev_UUID/migration_version).
> > > > 
> > > > Are you aware of the devlink based device management interface that is
> > > > proposed upstream? I think it has many advantages over sysfs, do you
> > > > consider to switch to that?
> > 
> > Advantages, such as?
> 
> 
> My understanding for devlink(netlink) over sysfs (some are mentioned at 
> the time of vDPA sysfs mgmt API discussion) are:
i tought netlink was used more a as a configuration protocoal to qurry and 
confire nic and i guess
other devices in its devlink form requireint a tool to be witten that can speak 
the protocal to interact with.
the primary advantate of sysfs is that everything is just a file. there are no 
addtional depleenceis
needed and unlike netlink there are not interoperatblity issues in a 
coanitnerised env. if you are using diffrenet
version of libc and gcc in the contaienr vs the host my understanding is tools 
like ethtool from ubuntu deployed
in a container on a centos host can have issue communicating with the host 
kernel. if its jsut a file unless
the format the data is returnin in chagnes or the layout of sysfs changes its 
compatiable regardless of what you
use to read it.
> 
> - existing users (NIC, crypto, SCSI, ib), mature and stable
> - much better error reporting (ext_ack other than string or errno)
> - namespace aware
> - do not couple with kobject
> 
> Thanks
> 
___
Devel mailing list -- devel@ovirt.org
To unsubscribe send an email to devel-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/devel@ovirt.org/message/NZGZCNEUJIZWVQCLTDBRAN5WCVF7TDVD/


[ovirt-devel] Re: device compatibility interface for live migration with assigned devices

2020-07-21 Thread Jason Wang


On 2020/7/18 上午12:12, Alex Williamson wrote:

On Thu, 16 Jul 2020 16:32:30 +0800
Yan Zhao  wrote:


On Thu, Jul 16, 2020 at 12:16:26PM +0800, Jason Wang wrote:

On 2020/7/14 上午7:29, Yan Zhao wrote:

hi folks,
we are defining a device migration compatibility interface that helps upper
layer stack like openstack/ovirt/libvirt to check if two devices are
live migration compatible.
The "devices" here could be MDEVs, physical devices, or hybrid of the two.
e.g. we could use it to check whether
- a src MDEV can migrate to a target MDEV,
- a src VF in SRIOV can migrate to a target VF in SRIOV,
- a src MDEV can migration to a target VF in SRIOV.
(e.g. SIOV/SRIOV backward compatibility case)

The upper layer stack could use this interface as the last step to check
if one device is able to migrate to another device before triggering a real
live migration procedure.
we are not sure if this interface is of value or help to you. please don't
hesitate to drop your valuable comments.


(1) interface definition
The interface is defined in below way:

   __userspace
/\  \
   / \write
  / read  \
 /__   ___\|/_
| migration_version | | migration_version |-->check migration
- -   compatibility
   device Adevice B


a device attribute named migration_version is defined under each device's
sysfs node. e.g. 
(/sys/bus/pci/devices/\:00\:02.0/$mdev_UUID/migration_version).


Are you aware of the devlink based device management interface that is
proposed upstream? I think it has many advantages over sysfs, do you
consider to switch to that?


Advantages, such as?



My understanding for devlink(netlink) over sysfs (some are mentioned at 
the time of vDPA sysfs mgmt API discussion) are:


- existing users (NIC, crypto, SCSI, ib), mature and stable
- much better error reporting (ext_ack other than string or errno)
- namespace aware
- do not couple with kobject

Thanks
___
Devel mailing list -- devel@ovirt.org
To unsubscribe send an email to devel-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/devel@ovirt.org/message/PKAGNMWYGBZ6LEJ3ULVZUJOLU6EULLJG/


[ovirt-devel] Re: [ARM64] Possiblity to support oVirt on ARM64

2020-07-21 Thread Zhenyu Zheng
Hi Nir,

Thanks alot for the quick response and information, actually I also  worked
a little bit on libvirt and other from our
team has experiences in qemu, those projects works OK on ARM platform, it
is just the capability might be less
when comparing to other platforms, for example, myself just added the
getHost() capability for ARM64 in libvirt and
now planning to work on compareCPU() API, after that we will have better
migration workflow on ARM platform.

For resources, actually, yes, I think we can provide those for oVirt
community, could you tell me more details about
what kind of resources are more suitable, our current resources are mostly
VMs from public cloud, will that be OK?

BR,

Zhenyu

On Mon, Jul 20, 2020 at 2:07 AM Nir Soffer  wrote:

> On Sun, Jul 19, 2020 at 5:04 PM Zhenyu Zheng 
> wrote:
> >
> > Hi oVirt,
> >
> > We are currently trying to make oVirt work on ARM64 platform, since I'm
> quite new to oVirt community, I'm wondering what is the current status
> about ARM64 support in the oVirt upstream, as I saw the oVirt Wikipedia
> page mentioned there is an ongoing efforts to support ARM platform. We have
> a small team here and we are willing to also help to make this work.
>
> Hi Zhenyu,
>
> I think this is a great idea, both supporting more hardware, and
> enlarging the oVirt
> community.
>
> Regarding hardware support we depend mostly on libvirt and qemu, and I
> don't know
> that is the status. Adding relevant lists and people.
>
> I don't know about any effort  on oVirt side, but last week I added
> arm builds for
> ovirt-imageio and it works:
>
> https://copr.fedorainfracloud.org/coprs/nsoffer/ovirt-imageio-preview/build/1555705/
>
> We have many dependendencies, but oVirt itself is mostly python and java,
> with
> tiny bits in C or using ctypes, so it should not be too hard.
>
> I think the first thing is getting some hardware for testing. Do you
> have such hardware,
> or have some contacts that can help to get hardware contribution for this?
>
> Nir
>
>
___
Devel mailing list -- devel@ovirt.org
To unsubscribe send an email to devel-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/devel@ovirt.org/message/MUYF2KTQVSI4LANQORMLCYQHKT4UUTUX/