Re: Change in ovirt-engine[master]: packaging: engine-setup: Use otopi's checkForSafeUpdate

2022-01-12 Thread Yedidyah Bar David
On Wed, Jan 12, 2022 at 12:06 PM Code Review  wrote:
>
> From Jenkins CI :
>
> Jenkins CI has posted comments on this change. ( 
> https://gerrit.ovirt.org/c/ovirt-engine/+/116920 )
>
> Change subject: packaging: engine-setup: Use otopi's checkForSafeUpdate
> ..
>
>
> Patch Set 6:
>
> Build Failed
>
> https://jenkins.ovirt.org/job/standard-enqueue/36418/ :
> This change was not submitted to any change queues for system testing. You 
> will need to create some 'build-artifacts' jobs if you want changes to be 
> submitted to change queues, take part in the system tests and be deployed to 
> the nightly snapshot repositories. If your project uses STDCI V2 and you have 
> release branches configured, you may disregard this message.
>
>
> https://jenkins.ovirt.org/job/ovirt-engine_standard-on-merge/2762/ : UNSTABLE

I think it's because of:

[2022-01-12T09:42:30.703Z] Traceback (most recent call last):
[2022-01-12T09:42:30.703Z]   File
"/home/jenkins/agent/workspace/ovirt-engine_standard-on-merge@tmp/durable-ecea2c02/script.sh",
line 10, in 
[2022-01-12T09:42:30.703Z] change = GerritMergedChange.from_jenkins_env()
[2022-01-12T09:42:30.703Z]   File
"/home/jenkins/agent/workspace/ovirt-engine_standard-on-merge/jenkins/stdci_libs/change_queue/changes/__init__.py",
line 290, in from_jenkins_env
[2022-01-12T09:42:30.703Z] o =
cls(gerrit_patchset=GerritPatchset.from_jenkins_env())
[2022-01-12T09:42:30.703Z]   File
"/home/jenkins/agent/workspace/ovirt-engine_standard-on-merge/jenkins/stdci_libs/gerrit.py",
line 157, in from_jenkins_env
[2022-01-12T09:42:30.703Z] 'GERRIT_PATCHSET_UPLOADER', env
[2022-01-12T09:42:30.703Z]   File
"/home/jenkins/agent/workspace/ovirt-engine_standard-on-merge/jenkins/stdci_libs/gerrit.py",
line 97, in from_jenkins_env
[2022-01-12T09:42:30.703Z] return cls(env[prefix + '_NAME'],
env[prefix + '_EMAIL'])
[2022-01-12T09:42:30.703Z]   File "/usr/lib64/python3.6/os.py", line
669, in __getitem__
[2022-01-12T09:42:30.703Z] raise KeyError(key) from None
[2022-01-12T09:42:30.703Z] KeyError: 'GERRIT_PATCHSET_UPLOADER_NAME'

and I *think* it's because I commented 'ost check' on two patches,
which were one above the other, and CI merged the upper one before
finishing the gating of the lower one (current).

Current case can most probably be ignored, but I wonder if we want to
do something in the general case - perhaps the code already does
everything well, didn't check - e.g. checking that all pending patches
in the stack have all acks, etc.

Best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/6RBUEGF3TAHJKVQLN564U2A4B32EAVWC/


Re: resources.ovirt.org's cert expired

2022-01-03 Thread Yedidyah Bar David
On Mon, Jan 3, 2022 at 11:46 AM Galit Rosenthal  wrote:
>
> Hi Didi,
>
> Denis fixed it
> Please check

Looks good to me. Thanks!

>
> Regards,
> Galit
>
> On Sun, Jan 2, 2022 at 9:30 AM Yedidyah Bar David  wrote:
>>
>> Hi all,
>>
>> $Subject:
>>
>> Not After Sun, 02 Jan 2022 05:06:16 GMT
>>
>> Do we have the renewal automated?
>>
>> This currently breaks some CI flows.
>>
>> Thanks and best regards,
>> --
>> Didi
>> ___
>> Infra mailing list -- infra@ovirt.org
>> To unsubscribe send an email to infra-le...@ovirt.org
>> Privacy Statement: https://www.ovirt.org/privacy-policy.html
>> oVirt Code of Conduct: 
>> https://www.ovirt.org/community/about/community-guidelines/
>> List Archives: 
>> https://lists.ovirt.org/archives/list/infra@ovirt.org/message/FLWXAMNPHKVY2XTMBAFHC6YEL25OVLOL/
>
>
>
> --
>
> Galit Rosenthal
>
> SOFTWARE ENGINEER
>
> Red Hat
>
> ga...@redhat.comT: 972-9-7692230



-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/G7TXY6WCW3VKTSKEQ5GWGH2CIX35NLIZ/


resources.ovirt.org's cert expired

2022-01-01 Thread Yedidyah Bar David
Hi all,

$Subject:

Not After Sun, 02 Jan 2022 05:06:16 GMT

Do we have the renewal automated?

This currently breaks some CI flows.

Thanks and best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/FLWXAMNPHKVY2XTMBAFHC6YEL25OVLOL/


[JIRA] (OVIRT-3127) ovirt-hosted-engine-ha standard-on-merge fails in _cpickle_loads

2021-12-19 Thread Yedidyah Bar David (oVirt JIRA)

[ 
https://ovirt-jira.atlassian.net/browse/OVIRT-3127?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=41081#comment-41081
 ] 

Yedidyah Bar David commented on OVIRT-3127:
---

On Thu, Dec 16, 2021 at 3:25 PM Yedidyah Bar David (oVirt JIRA)
 wrote:
>
> Yedidyah Bar David created OVIRT-3127:
> -
>
>  Summary: ovirt-hosted-engine-ha standard-on-merge fails in 
> _cpickle_loads
>  Key: OVIRT-3127
>  URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3127
>  Project: oVirt - virtualization made easy
>   Issue Type: By-EMAIL
>     Reporter: Yedidyah Bar David
> Assignee: infra
>
>
> Hi,
>
> Please see e.g. [1][2] (console log shows nothing):
> [2021-12-16T13:11:30.371Z]   File
> "/home/jenkins/workspace/ovirt-4.4_change-queue/jenkins/stdci_libs/jenkins_objects.py",
> line 223, in _cpickle_loads
> [2021-12-16T13:11:30.371Z] return pk.load()
> [2021-12-16T13:11:30.371Z] ValueError: unsupported pickle protocol: 3
> script returned exit code 1

The build has no GERRIT_CHANGE_OWNER nor GERRIT_CHANGE_OWNER_NAME, for
some reason:

https://jenkins.ovirt.org/job/ovirt-hosted-engine-ha_standard-on-merge/84/parameters/

I suppose the issue can be ignored for now, let's see if the next merge passes.

Best regards,
-- 
Didi



> ovirt-hosted-engine-ha standard-on-merge fails in _cpickle_loads
> 
>
> Key: OVIRT-3127
> URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3127
> Project: oVirt - virtualization made easy
>  Issue Type: By-EMAIL
>Reporter: Yedidyah Bar David
>Assignee: infra
>
> Hi,
> Please see e.g. [1][2] (console log shows nothing):
> [2021-12-16T13:11:30.371Z]   File
> "/home/jenkins/workspace/ovirt-4.4_change-queue/jenkins/stdci_libs/jenkins_objects.py",
> line 223, in _cpickle_loads
> [2021-12-16T13:11:30.371Z] return pk.load()
> [2021-12-16T13:11:30.371Z] ValueError: unsupported pickle protocol: 3
> script returned exit code 1
> Thanks,
> [1] https://jenkins.ovirt.org/job/ovirt-hosted-engine-ha_standard-on-merge/85/
> [2] 
> https://jenkins.ovirt.org/blue/organizations/jenkins/ovirt-4.4_change-queue/detail/ovirt-4.4_change-queue/2045/pipeline
> -- 
> Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100187)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/4LWLG56TXY77CDL2MBZSP5RJTQHDU2YP/


Re: [JIRA] (OVIRT-3127) ovirt-hosted-engine-ha standard-on-merge fails in _cpickle_loads

2021-12-19 Thread Yedidyah Bar David
On Thu, Dec 16, 2021 at 3:25 PM Yedidyah Bar David (oVirt JIRA)
 wrote:
>
> Yedidyah Bar David created OVIRT-3127:
> -
>
>  Summary: ovirt-hosted-engine-ha standard-on-merge fails in 
> _cpickle_loads
>  Key: OVIRT-3127
>  URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3127
>  Project: oVirt - virtualization made easy
>   Issue Type: By-EMAIL
>     Reporter: Yedidyah Bar David
> Assignee: infra
>
>
> Hi,
>
> Please see e.g. [1][2] (console log shows nothing):
> [2021-12-16T13:11:30.371Z]   File
> "/home/jenkins/workspace/ovirt-4.4_change-queue/jenkins/stdci_libs/jenkins_objects.py",
> line 223, in _cpickle_loads
> [2021-12-16T13:11:30.371Z] return pk.load()
> [2021-12-16T13:11:30.371Z] ValueError: unsupported pickle protocol: 3
> script returned exit code 1

The build has no GERRIT_CHANGE_OWNER nor GERRIT_CHANGE_OWNER_NAME, for
some reason:

https://jenkins.ovirt.org/job/ovirt-hosted-engine-ha_standard-on-merge/84/parameters/

I suppose the issue can be ignored for now, let's see if the next merge passes.

Best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/KG2J2VSSASUKTLCET2W3ML4F4S4VHJVG/


[JIRA] (OVIRT-3127) ovirt-hosted-engine-ha standard-on-merge fails in _cpickle_loads

2021-12-16 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3127:
-

 Summary: ovirt-hosted-engine-ha standard-on-merge fails in 
_cpickle_loads
 Key: OVIRT-3127
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3127
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


Hi,

Please see e.g. [1][2] (console log shows nothing):
[2021-12-16T13:11:30.371Z]   File
"/home/jenkins/workspace/ovirt-4.4_change-queue/jenkins/stdci_libs/jenkins_objects.py",
line 223, in _cpickle_loads
[2021-12-16T13:11:30.371Z] return pk.load()
[2021-12-16T13:11:30.371Z] ValueError: unsupported pickle protocol: 3
script returned exit code 1

Thanks,

[1] https://jenkins.ovirt.org/job/ovirt-hosted-engine-ha_standard-on-merge/85/

[2] 
https://jenkins.ovirt.org/blue/organizations/jenkins/ovirt-4.4_change-queue/detail/ovirt-4.4_change-queue/2045/pipeline
-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100187)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/RMA5LKG62AENL5TYPVKYIEVZ2WBOKUBJ/


ovirt-hosted-engine-ha standard-on-merge fails in _cpickle_loads

2021-12-16 Thread Yedidyah Bar David
Hi,

Please see e.g. [1][2] (console log shows nothing):
[2021-12-16T13:11:30.371Z]   File
"/home/jenkins/workspace/ovirt-4.4_change-queue/jenkins/stdci_libs/jenkins_objects.py",
line 223, in _cpickle_loads
[2021-12-16T13:11:30.371Z] return pk.load()
[2021-12-16T13:11:30.371Z] ValueError: unsupported pickle protocol: 3
script returned exit code 1

Thanks,

[1] https://jenkins.ovirt.org/job/ovirt-hosted-engine-ha_standard-on-merge/85/

[2] 
https://jenkins.ovirt.org/blue/organizations/jenkins/ovirt-4.4_change-queue/detail/ovirt-4.4_change-queue/2045/pipeline
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/AGL6I43EORJM7UUGHX3KY6UWEN3RJRB5/


[JIRA] (OVIRT-3119) Re: [ovirt-users] Public Glance repository for oVirt

2021-10-13 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3119:
-

 Summary: Re: [ovirt-users] Public Glance repository for oVirt
 Key: OVIRT-3119
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3119
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


On Tue, Oct 12, 2021 at 6:45 PM  wrote:
>
> Hello,
>   can anyone get to the public openstack image repository  
> ovirt-image-repository?
>
> Also when I try  browse the site I get
>
> This site can’t be reached
> https://glance.ovirt.org/ is unreachable.

Thanks for the report.

Adding infra/infra-support. Any clue? This is affecting also CI,
breaking at least
test_undo_preview_snapshot_when_mac_used_reassigns_a_new_mac in
ovirt-system-tests.

Thanks and best regards,
-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100181)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/B7GUJAGU7CWILGCLMZ64C2MXSOUJNNLF/


Re: [ovirt-users] Public Glance repository for oVirt

2021-10-13 Thread Yedidyah Bar David
On Tue, Oct 12, 2021 at 6:45 PM  wrote:
>
> Hello,
>   can anyone get to the public openstack image repository  
> ovirt-image-repository?
>
> Also when I try  browse the site I get
>
> This site can’t be reached
> https://glance.ovirt.org/ is unreachable.

Thanks for the report.

Adding infra/infra-support. Any clue? This is affecting also CI,
breaking at least
test_undo_preview_snapshot_when_mac_used_reassigns_a_new_mac in
ovirt-system-tests.

Thanks and best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/IOC25K7CDTOK2L7OLXQJHZKZOQMIPWAO/


[JIRA] (OVIRT-3118) Missing "Keep this build forever" button

2021-10-06 Thread Yedidyah Bar David (oVirt JIRA)

[ 
https://ovirt-jira.atlassian.net/browse/OVIRT-3118?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=41062#comment-41062
 ] 

Yedidyah Bar David commented on OVIRT-3118:
---

An example build: 
[https://jenkins.ovirt.org/job/otopi_standard-on-merge/143/|https://jenkins.ovirt.org/job/otopi_standard-on-merge/143/|smart-link]
 

> Missing "Keep this build forever" button
> 
>
> Key: OVIRT-3118
> URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3118
> Project: oVirt - virtualization made easy
>  Issue Type: By-EMAIL
>    Reporter: Yedidyah Bar David
>Assignee: infra
>
> Hi,
> I do not have anymore "Keep this build forever". I suspect it's
> something to do with my account - I guess my previous account was with
> an alias as email address, whereas with the SSO login it's my main
> address. Please help :-).
> Best regards,
> -- 
> Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100180)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/J6CLRMNO2HN5M66VW4I7R5SZN2XKHGLU/


[JIRA] (OVIRT-3118) Missing "Keep this build forever" button

2021-10-05 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3118:
-

 Summary: Missing "Keep this build forever" button
 Key: OVIRT-3118
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3118
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


Hi,

I do not have anymore "Keep this build forever". I suspect it's
something to do with my account - I guess my previous account was with
an alias as email address, whereas with the SSO login it's my main
address. Please help :-).

Best regards,
-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100179)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/T4PD7TY5QIRM2DW2MFRTIFJCZGCYDMT7/


Missing "Keep this build forever" button

2021-10-05 Thread Yedidyah Bar David
Hi,

I do not have anymore "Keep this build forever". I suspect it's
something to do with my account - I guess my previous account was with
an alias as email address, whereas with the SSO login it's my main
address. Please help :-).

Best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/IYFPAJKFBSCPUTJSOC7RPVL6BQOGBH46/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2171 - Still Failing!

2021-10-04 Thread Yedidyah Bar David
On Thu, Sep 9, 2021 at 6:03 PM Marcin Sobczyk  wrote:
>
>
>
> On 9/9/21 4:22 PM, Yedidyah Bar David wrote:
> > On Thu, Sep 9, 2021 at 4:40 PM Marcin Sobczyk  wrote:
> >>
> >>
> >> On 9/9/21 1:39 PM, Yedidyah Bar David wrote:
> >>> On Thu, Sep 9, 2021 at 12:07 PM Marcin Sobczyk  >>> <mailto:msobc...@redhat.com>> wrote:
> >>>
> >>>  Hi,
> >>>
> >>>  On 9/9/21 10:45 AM, Yedidyah Bar David wrote:
> >>>  > On Thu, Sep 9, 2021 at 4:59 AM  >>>  <mailto:jenk...@jenkins.phx.ovirt.org>
> >>>  > <mailto:jenk...@jenkins.phx.ovirt.org
> >>>  <mailto:jenk...@jenkins.phx.ovirt.org>>> wrote:
> >>>  >
> >>>  > Project:
> >>>  >
> >>>  
> >>> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> >>>  
> >>> <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/>
> >>>  >
> >>>   
> >>> <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> >>>  
> >>> <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/>>
> >>>  >
> >>>  > Build:
> >>>  >
> >>>  
> >>> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2171/
> >>>  
> >>> <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2171/>
> >>>  >
> >>>   
> >>> <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2171/
> >>>  
> >>> <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2171/>>
> >>>  >
> >>>  >
> >>>  >   # /usr/bin/dnf --installroot
> >>>  
> >>> /var/lib/mock/epel-8-x86_64-31c5ec78d5c6c0fb3c72e2b5ad18fba9-bootstrap-2910831/root/
> >>>  --releasever 8 --setopt=deltarpm=False --allowerasing
> >>>  --disableplugin=local --disableplugin=spacewalk
> >>>  --disableplugin=versionlock install dnf dnf-plugins-core
> >>>  distribution-gpg-keys --setopt=tsflags=nocontexts
> >>>  > No matches found for the following disable plugin patterns:
> >>>  local, spacewalk, versionlock
> >>>  > Errors during downloading metadata for repository 'ost-images':
> >>>  >- Status code: 503
> >>>  forhttp://nike15.rhev.lab.eng.brq.redhat.com/repodata/repomd.xml
> >>>  <http://nike15.rhev.lab.eng.brq.redhat.com/repodata/repomd.xml>
> >>>  <http://nike15.rhev.lab.eng.brq.redhat.com/repodata/repomd.xml
> >>>  <http://nike15.rhev.lab.eng.brq.redhat.com/repodata/repomd.xml>>
> >>>  (IP: 66.187.230.40)
> >>>  > Error: Failed to download metadata for repo 'ost-images': Cannot
> >>>  download repomd.xml: Cannot download repodata/repomd.xml: All
> >>>  mirrors were tried
> >>>  >
> >>>  > Is anyone tracking this? Thanks.
> >>>  Running OST on jenkins.ovirt.org <http://jenkins.ovirt.org> is not
> >>>  supported anymore.
> >>>  Please see: https://gerrit.ovirt.org/c/ovirt-system-tests/+/116458
> >>>  <https://gerrit.ovirt.org/c/ovirt-system-tests/+/116458>
> >>>
> >>>
> >>> OK, thanks for the clarification. Do we, then, remove the
> >>> scheduled/nightly jobs, among which above is one?
> >> Fine with me. I'm not even sure - are they defined in jenkins project?
> > They are, AFAIU - in jobs/confs/projects/ovirt/system-tests.yaml .
> @Galit, could you please handle this?

Any update? I still get daily reports and consciously ignore them...

Now found "remove ovirt-system-tests jobs and zuul config":

https://gerrit.ovirt.org/c/jenkins/+/116728

>
> >
> > I admit this is a bit sad for me - I happened to check and point at these
> > jobs as a demonstration/reference of how a good run should look like (in
> > terms of output, logs etc.). Was pretty convenient...
> I understand, but this is probably not a good enough
> reason to keep maintaining lago and living with mock.
>
> >
> >>> I see in your patch that not all suites are removed from automation,
> >>> not sure about the status of each.
> >> PSI CI environment still uses the stuff from 'automation/'. We simply
> >> limited the runs to that environment.
> > OK.
> >
> > Do we intend to publish artifacts from there?
> Yes, Galit is working on it.

Would love to get an update about this one as well.

Thanks and best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/TINWC22EEN26Z3CHKMVAIJVETJJ4GJRL/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2171 - Still Failing!

2021-09-09 Thread Yedidyah Bar David
On Thu, Sep 9, 2021 at 4:40 PM Marcin Sobczyk  wrote:
>
>
>
> On 9/9/21 1:39 PM, Yedidyah Bar David wrote:
> > On Thu, Sep 9, 2021 at 12:07 PM Marcin Sobczyk  > <mailto:msobc...@redhat.com>> wrote:
> >
> > Hi,
> >
> > On 9/9/21 10:45 AM, Yedidyah Bar David wrote:
> > > On Thu, Sep 9, 2021 at 4:59 AM  > <mailto:jenk...@jenkins.phx.ovirt.org>
> > > <mailto:jenk...@jenkins.phx.ovirt.org
> > <mailto:jenk...@jenkins.phx.ovirt.org>>> wrote:
> > >
> > > Project:
> > >
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> > 
> > <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/>
> > >
> >  
> > <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> > 
> > <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/>>
> > >
> > > Build:
> > >
> > 
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2171/
> > 
> > <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2171/>
> > >
> >  
> > <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2171/
> > 
> > <https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2171/>>
> > >
> > >
> > >   # /usr/bin/dnf --installroot
> > 
> > /var/lib/mock/epel-8-x86_64-31c5ec78d5c6c0fb3c72e2b5ad18fba9-bootstrap-2910831/root/
> > --releasever 8 --setopt=deltarpm=False --allowerasing
> > --disableplugin=local --disableplugin=spacewalk
> > --disableplugin=versionlock install dnf dnf-plugins-core
> > distribution-gpg-keys --setopt=tsflags=nocontexts
> > > No matches found for the following disable plugin patterns:
> > local, spacewalk, versionlock
> > > Errors during downloading metadata for repository 'ost-images':
> > >- Status code: 503
> > forhttp://nike15.rhev.lab.eng.brq.redhat.com/repodata/repomd.xml
> > <http://nike15.rhev.lab.eng.brq.redhat.com/repodata/repomd.xml>
> > <http://nike15.rhev.lab.eng.brq.redhat.com/repodata/repomd.xml
> > <http://nike15.rhev.lab.eng.brq.redhat.com/repodata/repomd.xml>>
> > (IP: 66.187.230.40)
> > > Error: Failed to download metadata for repo 'ost-images': Cannot
> > download repomd.xml: Cannot download repodata/repomd.xml: All
> > mirrors were tried
> > >
> > > Is anyone tracking this? Thanks.
> > Running OST on jenkins.ovirt.org <http://jenkins.ovirt.org> is not
> > supported anymore.
> > Please see: https://gerrit.ovirt.org/c/ovirt-system-tests/+/116458
> > <https://gerrit.ovirt.org/c/ovirt-system-tests/+/116458>
> >
> >
> > OK, thanks for the clarification. Do we, then, remove the
> > scheduled/nightly jobs, among which above is one?
> Fine with me. I'm not even sure - are they defined in jenkins project?

They are, AFAIU - in jobs/confs/projects/ovirt/system-tests.yaml .

I admit this is a bit sad for me - I happened to check and point at these
jobs as a demonstration/reference of how a good run should look like (in
terms of output, logs etc.). Was pretty convenient...

>
> >
> > I see in your patch that not all suites are removed from automation,
> > not sure about the status of each.
> PSI CI environment still uses the stuff from 'automation/'. We simply
> limited the runs to that environment.

OK.

Do we intend to publish artifacts from there?

Thanks,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/ABLGIMDYKR7WTT7QKTDC56EAD3QKD6RS/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2171 - Still Failing!

2021-09-09 Thread Yedidyah Bar David
On Thu, Sep 9, 2021 at 12:07 PM Marcin Sobczyk  wrote:

> Hi,
>
> On 9/9/21 10:45 AM, Yedidyah Bar David wrote:
> > On Thu, Sep 9, 2021 at 4:59 AM  > <mailto:jenk...@jenkins.phx.ovirt.org>> wrote:
> >
> > Project:
> >
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> > <
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/>
> >
> > Build:
> >
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2171/
> > <
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2171/
> >
> >
> >
> >   # /usr/bin/dnf --installroot
> /var/lib/mock/epel-8-x86_64-31c5ec78d5c6c0fb3c72e2b5ad18fba9-bootstrap-2910831/root/
> --releasever 8 --setopt=deltarpm=False --allowerasing --disableplugin=local
> --disableplugin=spacewalk --disableplugin=versionlock install dnf
> dnf-plugins-core distribution-gpg-keys --setopt=tsflags=nocontexts
> > No matches found for the following disable plugin patterns: local,
> spacewalk, versionlock
> > Errors during downloading metadata for repository 'ost-images':
> >- Status code: 503 forhttp://
> nike15.rhev.lab.eng.brq.redhat.com/repodata/repomd.xml  <
> http://nike15.rhev.lab.eng.brq.redhat.com/repodata/repomd.xml>  (IP:
> 66.187.230.40)
> > Error: Failed to download metadata for repo 'ost-images': Cannot
> download repomd.xml: Cannot download repodata/repomd.xml: All mirrors were
> tried
> >
> > Is anyone tracking this? Thanks.
> Running OST on jenkins.ovirt.org is not supported anymore.
> Please see: https://gerrit.ovirt.org/c/ovirt-system-tests/+/116458
>
>
OK, thanks for the clarification. Do we, then, remove the scheduled/nightly
jobs, among which above is one?

I see in your patch that not all suites are removed from automation, not
sure about the status of each.

Thanks,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/WRLJ44FZQ22HQHLVNPUTXHFWLE6TWYW6/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2171 - Still Failing!

2021-09-09 Thread Yedidyah Bar David
On Thu, Sep 9, 2021 at 4:59 AM  wrote:

> Project:
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> Build:
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2171/


 # /usr/bin/dnf --installroot
/var/lib/mock/epel-8-x86_64-31c5ec78d5c6c0fb3c72e2b5ad18fba9-bootstrap-2910831/root/
--releasever 8 --setopt=deltarpm=False --allowerasing
--disableplugin=local --disableplugin=spacewalk
--disableplugin=versionlock install dnf dnf-plugins-core
distribution-gpg-keys --setopt=tsflags=nocontexts
No matches found for the following disable plugin patterns: local,
spacewalk, versionlock
Errors during downloading metadata for repository 'ost-images':
  - Status code: 503 for
http://nike15.rhev.lab.eng.brq.redhat.com/repodata/repomd.xml (IP:
66.187.230.40)
Error: Failed to download metadata for repo 'ost-images': Cannot
download repomd.xml: Cannot download repodata/repomd.xml: All mirrors
were tried

Is anyone tracking this? Thanks.

-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/P7EHAANFZNDXJ4N2TSZDTW2QIZIFGTVB/


Re: [oVirt Jenkins] ovirt-system-tests_basic-suite-4.3_el7-nightly - Build # 694 - Still Failing!

2021-07-27 Thread Yedidyah Bar David
On Wed, Jul 28, 2021 at 8:01 AM  wrote:
>
> Project: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-4.3_el7-nightly/
> Build: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-4.3_el7-nightly/694/

This fails for a few weeks now. If we do not intend to fix, perhaps
disable/remove.

It seems like an infra issue, before starting OST code. I didn't dive
too deeply.

Best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/RC4MW2HVMTUYF7TUNF2N3AO4HI72CSVE/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2046 - Failure!

2021-07-21 Thread Yedidyah Bar David
On Wed, Jul 7, 2021 at 4:42 PM Eitan Raviv  wrote:
>
> Adding Ales as well.
> AFAIK vdsm does not actively poll engine for liveness, nor does any retries. 
> But retries might be at a deeper infra level where Marcin is the person to 
> ask IIUC.

I now filed a bug to track this [1], and (think I) managed to verify
my patch, to use +tcp [2]. So please review/merge the patch and
flag/target/ack the bug. Thanks.

[1] https://bugzilla.redhat.com/1984356

[2] https://gerrit.ovirt.org/c/ovirt-hosted-engine-ha/+/115596

>
> On Wed, Jul 7, 2021 at 4:40 PM Yedidyah Bar David  wrote:
>>
>> On Wed, Jun 23, 2021 at 12:30 PM Yedidyah Bar David  wrote:
>> >
>> > On Wed, Jun 23, 2021 at 12:02 PM Sandro Bonazzola  
>> > wrote:
>> >>
>> >>
>> >>
>> >> Il giorno mer 23 giu 2021 alle ore 07:48 Yedidyah Bar David 
>> >>  ha scritto:
>> >>>
>> >>> On Wed, Jun 9, 2021 at 12:13 PM Yedidyah Bar David  
>> >>> wrote:
>> >>> >
>> >>> > On Tue, Jun 8, 2021 at 9:01 AM Yedidyah Bar David  
>> >>> > wrote:
>> >>> > >
>> >>> > > On Tue, Jun 8, 2021 at 6:08 AM  wrote:
>> >>> > > >
>> >>> > > > Project: 
>> >>> > > > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
>> >>> > > > Build: 
>> >>> > > > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/
>> >>> > > > Build Number: 2046
>> >>> > > > Build Status:  Failure
>> >>> > > > Triggered By: Started by timer
>> >>> > > >
>> >>> > > > -
>> >>> > > > Changes Since Last Success:
>> >>> > > > -
>> >>> > > > Changes for Build #2046
>> >>> > > > [Eitan Raviv] network: force select spm - wait for dc status
>> >>> > > >
>> >>> > > >
>> >>> > > >
>> >>> > > >
>> >>> > > > -
>> >>> > > > Failed Tests:
>> >>> > > > -
>> >>> > > > 1 tests failed.
>> >>> > > > FAILED:  
>> >>> > > > he-basic-suite-master.test-scenarios.test_004_basic_sanity.test_template_export
>> >>> > > >
>> >>> > > > Error Message:
>> >>> > > > ovirtsdk4.Error: Failed to read response: [(> >>> > > > 0x5624fe64d108>, 7, 'Failed to connect to 192.168.200.99 port 443: 
>> >>> > > > Connection refused')]
>> >>> > >
>> >>> > > - The engine VM went down:
>> >>> > >
>> >>> > > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/agent.log
>> >>> > >
>> >>> > > MainThread::INFO::2021-06-08
>> >>> > > 05:07:34,414::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
>> >>> > > Current state EngineUp (score: 3400)
>> >>> > > MainThread::INFO::2021-06-08
>> >>> > > 05:07:44,575::states::135::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
>> >>> > > Penalizing score by 960 due to network status
>> >>> > >
>> >>> > > - Because HA monitoring failed to get a reply from the dns server:
>> >>> > >
>> >>> > > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/broker.log
>> >>> > >
>> >>> > > Thread-1::WARNING::2021-06-08
>> >>> > > 05:07:25,486::network::120::network.Network::(_dns) DNS query failed:
>> >>> > > ; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
>> >>> > > ;; global options: +cmd
>> >>> > > ;; connection timed out; no servers could be reached
>> >>> > >
>> >>> > > Thre

Re: package in automation/*.packages not installed

2021-07-07 Thread Yedidyah Bar David
On Wed, Jul 7, 2021 at 11:26 PM Yedidyah Bar David  wrote:
>
> On Wed, Jul 7, 2021 at 10:54 PM Code Review  wrote:
> >
> > From Jenkins CI :
> >
> > Jenkins CI has posted comments on this change. ( 
> > https://gerrit.ovirt.org/c/ovirt-system-tests/+/115392 )
> >
> > Change subject: test, do not merge: HE: Use node image
> > ..
> >
> >
> > Patch Set 3: Continuous-Integration-1
> >
> > Build Failed
> >
> > https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/17737/
> >  : FAILURE
>
> Hi all,
>
> Above failed with:
>
>  [36m  # Create disks for VM lago-he-basic-suite-master-host-0:
>  [31mERROR [0m (in 0:00:00) [0m
>  [36m  # Missing current link, setting it to default [0m
>  [36m@ Initialize and populate prefix:  [31mERROR [0m (in 0:03:23) [0m
>  [31mError occured, aborting
> ...
> RuntimeError: Failed to get info for /var/lib/lago/ost-images-2h8yHA/
> b"qemu-img: Could not open '/var/lib/lago/ost-images-2h8yHA/': A
> regular file was expected by the 'file' driver, but something else was
> given\n"
>
> This job was for [1]. [1] also includes a patch to
> he-basic_suite_master.packages.el8 to install ost-images-node-base. It
> also removes ost-images-el8stream-he-installed. This should still
> work, because the latest published node image is a result of manually
> running the job in jenkins with  [2]).
>
> Anyway, the console log [3] shows that ost-images-node-base was not
> installed. Any idea why, or what I should do?

Sorry, please ignore. Above run was for patchset 3, which didn't include
the .packages change. Sorry for the noise.

Best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/ZBH7PPGJQAMDJAFESYGBRTLQQ4OKVGEC/


package in automation/*.packages not installed

2021-07-07 Thread Yedidyah Bar David
On Wed, Jul 7, 2021 at 10:54 PM Code Review  wrote:
>
> From Jenkins CI :
>
> Jenkins CI has posted comments on this change. ( 
> https://gerrit.ovirt.org/c/ovirt-system-tests/+/115392 )
>
> Change subject: test, do not merge: HE: Use node image
> ..
>
>
> Patch Set 3: Continuous-Integration-1
>
> Build Failed
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/17737/ 
> : FAILURE

Hi all,

Above failed with:

 [36m  # Create disks for VM lago-he-basic-suite-master-host-0:
 [31mERROR [0m (in 0:00:00) [0m
 [36m  # Missing current link, setting it to default [0m
 [36m@ Initialize and populate prefix:  [31mERROR [0m (in 0:03:23) [0m
 [31mError occured, aborting
...
RuntimeError: Failed to get info for /var/lib/lago/ost-images-2h8yHA/
b"qemu-img: Could not open '/var/lib/lago/ost-images-2h8yHA/': A
regular file was expected by the 'file' driver, but something else was
given\n"

This job was for [1]. [1] also includes a patch to
he-basic_suite_master.packages.el8 to install ost-images-node-base. It
also removes ost-images-el8stream-he-installed. This should still
work, because the latest published node image is a result of manually
running the job in jenkins with  [2]).

Anyway, the console log [3] shows that ost-images-node-base was not
installed. Any idea why, or what I should do?

Thanks and best regards,

[1] https://gerrit.ovirt.org/c/ovirt-system-tests/+/115392

[2] https://gerrit.ovirt.org/c/ost-images/+/115453

>
>
> --
> To view, visit https://gerrit.ovirt.org/c/ovirt-system-tests/+/115392
> To unsubscribe, or for help writing mail filters, visit 
> https://gerrit.ovirt.org/settings
>
> Gerrit-Project: ovirt-system-tests
> Gerrit-Branch: master
> Gerrit-Change-Id: I0fd1313100997edde22fc6b5869ab98abb217f42
> Gerrit-Change-Number: 115392
> Gerrit-PatchSet: 3
> Gerrit-Owner: Yedidyah Bar David 
> Gerrit-Reviewer: Anton Marchukov 
> Gerrit-Reviewer: Dafna Ron 
> Gerrit-Reviewer: Dusan Fodor 
> Gerrit-Reviewer: Gal Ben Haim 
> Gerrit-Reviewer: Galit Rosenthal 
> Gerrit-Reviewer: Jenkins CI 
> Gerrit-Reviewer: Name of user not set #1001916
> Gerrit-Reviewer: Yedidyah Bar David 
> Gerrit-Comment-Date: Wed, 07 Jul 2021 19:54:31 +
> Gerrit-HasComments: No
> Gerrit-Has-Labels: Yes
> Gerrit-MessageType: comment
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/OM4YBJKVGRWUXD2RSKKCVQCQHOJUOPO7/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2046 - Failure!

2021-07-07 Thread Yedidyah Bar David
On Wed, Jun 23, 2021 at 12:30 PM Yedidyah Bar David  wrote:
>
> On Wed, Jun 23, 2021 at 12:02 PM Sandro Bonazzola  wrote:
>>
>>
>>
>> Il giorno mer 23 giu 2021 alle ore 07:48 Yedidyah Bar David 
>>  ha scritto:
>>>
>>> On Wed, Jun 9, 2021 at 12:13 PM Yedidyah Bar David  wrote:
>>> >
>>> > On Tue, Jun 8, 2021 at 9:01 AM Yedidyah Bar David  wrote:
>>> > >
>>> > > On Tue, Jun 8, 2021 at 6:08 AM  wrote:
>>> > > >
>>> > > > Project: 
>>> > > > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
>>> > > > Build: 
>>> > > > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/
>>> > > > Build Number: 2046
>>> > > > Build Status:  Failure
>>> > > > Triggered By: Started by timer
>>> > > >
>>> > > > -
>>> > > > Changes Since Last Success:
>>> > > > -
>>> > > > Changes for Build #2046
>>> > > > [Eitan Raviv] network: force select spm - wait for dc status
>>> > > >
>>> > > >
>>> > > >
>>> > > >
>>> > > > -
>>> > > > Failed Tests:
>>> > > > -
>>> > > > 1 tests failed.
>>> > > > FAILED:  
>>> > > > he-basic-suite-master.test-scenarios.test_004_basic_sanity.test_template_export
>>> > > >
>>> > > > Error Message:
>>> > > > ovirtsdk4.Error: Failed to read response: [(>> > > > 0x5624fe64d108>, 7, 'Failed to connect to 192.168.200.99 port 443: 
>>> > > > Connection refused')]
>>> > >
>>> > > - The engine VM went down:
>>> > >
>>> > > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/agent.log
>>> > >
>>> > > MainThread::INFO::2021-06-08
>>> > > 05:07:34,414::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
>>> > > Current state EngineUp (score: 3400)
>>> > > MainThread::INFO::2021-06-08
>>> > > 05:07:44,575::states::135::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
>>> > > Penalizing score by 960 due to network status
>>> > >
>>> > > - Because HA monitoring failed to get a reply from the dns server:
>>> > >
>>> > > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/broker.log
>>> > >
>>> > > Thread-1::WARNING::2021-06-08
>>> > > 05:07:25,486::network::120::network.Network::(_dns) DNS query failed:
>>> > > ; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
>>> > > ;; global options: +cmd
>>> > > ;; connection timed out; no servers could be reached
>>> > >
>>> > > Thread-3::INFO::2021-06-08
>>> > > 05:07:28,543::mem_free::51::mem_free.MemFree::(action) memFree: 1801
>>> > > Thread-5::INFO::2021-06-08
>>> > > 05:07:28,972::engine_health::246::engine_health.EngineHealth::(_result_from_stats)
>>> > > VM is up on this host with healthy engine
>>> > > Thread-2::INFO::2021-06-08
>>> > > 05:07:31,532::mgmt_bridge::65::mgmt_bridge.MgmtBridge::(action) Found
>>> > > bridge ovirtmgmt in up state
>>> > > Thread-1::WARNING::2021-06-08
>>> > > 05:07:33,011::network::120::network.Network::(_dns) DNS query failed:
>>> > > ; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
>>> > > ;; global options: +cmd
>>> > > ;; connection timed out; no servers could be reached
>>> > >
>>> > > Thread-4::INFO::2021-06-08
>>> > > 05:07:37,433::cpu_load_no_engine::126::cpu_load_no_engine.CpuLoadNoEngine::(calculate_load)
>>> > > System load total=0.3196, engine=0.1724, non-engine=0.1472
>>> > > Thread-3::INFO::2021-06-08
>>> > > 05:07:37,839::mem_free::5

Re: OST fails in deploy: Failed to connect to the host via ssh

2021-07-06 Thread Yedidyah Bar David
On Tue, Jul 6, 2021 at 12:44 PM Marcin Sobczyk  wrote:
>
>
>
> On 7/6/21 11:32 AM, Yedidyah Bar David wrote:
> > Hi all,
> >
> > Recently we had several cases of OST runs that seemed to have
> > successfully started their VMs, but ansible later failed to connect to
> > them in the "wait_for_connection" done at start of deployment. This
> > happened to both he-basic and basic, e.g. [1][2][3]. Some do pass,
> > e.g. [4]. This looks like e.g. (from [1]):
> >
> > 20:29:58  [36m  # Start vms:  [32mSuccess [0m (in 0:00:02) [0m
> > ...
> > 20:30:08 
> > he-basic-suite-master/test-scenarios/test_001_he_deploy.py::test_he_deploy
> > started at 2021-07-05 17:30:08
> > 20:30:08  live log setup
> > 
> > 20:30:09 2021-07-05 17:30:09+,180 INFO
> > [ost_utils.pytest.fixtures.deployment] Waiting for SSH on the VMs
> > (deployment:60)
> > 20:30:09 ERROR (122s)Traceback (most recent call last):
> > ...
> > 20:32:15 "msg": "Failed to connect to the host via ssh: ssh:
> > connect to host 192.168.200.3 port 22: No route to host",
> >
> > Any idea?
> Yes, please see [5]. We have working images since Sunday I think,
> but we need to wait until mock cache invalidates on all CI agents and
> picks up the new images.

Ok, thanks.

I now pushed this trivial long-overdue patch, to allow us know which
ost-images are used:

https://gerrit.ovirt.org/c/ovirt-system-tests/+/115577

Best regards,

>
> Regards, Marcin
>
> [5] https://gerrit.ovirt.org/#/c/ost-images/+/115535/
>
> >
> > Thanks and best regards,
> >
> > [1] 
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2077/
> >
> > [2] 
> > https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/1287/
> >
> > [3] 
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2076/
> >
> > [4] 
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2078/
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/HWKLTEILC2JDOUPE4S6PMAMA7X7LHIOX/


OST fails in deploy: Failed to connect to the host via ssh

2021-07-06 Thread Yedidyah Bar David
Hi all,

Recently we had several cases of OST runs that seemed to have
successfully started their VMs, but ansible later failed to connect to
them in the "wait_for_connection" done at start of deployment. This
happened to both he-basic and basic, e.g. [1][2][3]. Some do pass,
e.g. [4]. This looks like e.g. (from [1]):

20:29:58  [36m  # Start vms:  [32mSuccess [0m (in 0:00:02) [0m
...
20:30:08 
he-basic-suite-master/test-scenarios/test_001_he_deploy.py::test_he_deploy
started at 2021-07-05 17:30:08
20:30:08  live log setup

20:30:09 2021-07-05 17:30:09+,180 INFO
[ost_utils.pytest.fixtures.deployment] Waiting for SSH on the VMs
(deployment:60)
20:30:09 ERROR (122s)Traceback (most recent call last):
...
20:32:15 "msg": "Failed to connect to the host via ssh: ssh:
connect to host 192.168.200.3 port 22: No route to host",

Any idea?

Thanks and best regards,

[1] https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2077/

[2] 
https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/1287/

[3] https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2076/

[4] https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2078/
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/LZX2XXV2A65NUQB6KJTTECJWC2OUWSWI/


nothing provides rpm = 4.14.3-14.el8_4 needed by rpm-build-4.14.3-14.el8_4.x86_64

2021-06-30 Thread Yedidyah Bar David
On Wed, Jun 30, 2021 at 3:20 PM Code Review  wrote:
>
> From Jenkins CI :
>
> Jenkins CI has posted comments on this change. ( 
> https://gerrit.ovirt.org/c/ovirt-system-tests/+/115276 )
>
> Change subject: Get rid of default_backend
> ..
>
>
> Patch Set 21: Continuous-Integration-1
>
> Build Failed
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/17555/ 
> : FAILURE

Failed with:
nothing provides rpm = 4.14.3-14.el8_4 needed by
rpm-build-4.14.3-14.el8_4.x86_64

Any clue?

Thanks,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/3FOQY5GLYHJQ3JYOCZWRUO26J7BPS3ES/


[JIRA] (OVIRT-3101) Add "ci system-test" command

2021-06-24 Thread Yedidyah Bar David (oVirt JIRA)

[ 
https://ovirt-jira.atlassian.net/browse/OVIRT-3101?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=41031#comment-41031
 ] 

Yedidyah Bar David commented on OVIRT-3101:
---

On Thu, Jun 24, 2021 at 1:51 PM Marcin Sobczyk  wrote:
>
>
>
> On 6/23/21 5:44 PM, Nir Soffer wrote:
> > Similar to "ci build", "ci test", "ci merge" add a new command that
> > triggers OST run.
> >
> > Running OST is tied now in vdsm (and engine?) to Code-Review: +2.
> > This causes trouble and does not allow non-maintainers to use the 
> > convenient OST
> > infrastructure.
> >
> > Expected flow:
> >
> > 1. User add a comment with "ci system-test"
> "ci system-test" is sooo long, I vote for "ci ost".

+1.

Perhaps we can add an optional suite name? E.g. 'ci ost ansible-suite-master'

Best regards,
-- 
Didi

> Add "ci system-test" command
> 
>
> Key: OVIRT-3101
> URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3101
> Project: oVirt - virtualization made easy
>  Issue Type: By-EMAIL
>Reporter: Nir Soffer
>Assignee: infra
>
> Similar to "ci build", "ci test", "ci merge" add a new command that
> triggers OST run.
> Running OST is tied now in vdsm (and engine?) to Code-Review: +2.
> This causes trouble and does not allow non-maintainers to use the convenient 
> OST
> infrastructure.
> Expected flow:
> 1. User add a comment with "ci system-test"
> 2. OST flow building and running OST triggered



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100166)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/5MYXL2F362W6B7NW7TDYGNVIBSIDT2UP/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2046 - Failure!

2021-06-23 Thread Yedidyah Bar David
On Wed, Jun 23, 2021 at 12:02 PM Sandro Bonazzola 
wrote:

>
>
> Il giorno mer 23 giu 2021 alle ore 07:48 Yedidyah Bar David <
> d...@redhat.com> ha scritto:
>
>> On Wed, Jun 9, 2021 at 12:13 PM Yedidyah Bar David 
>> wrote:
>> >
>> > On Tue, Jun 8, 2021 at 9:01 AM Yedidyah Bar David 
>> wrote:
>> > >
>> > > On Tue, Jun 8, 2021 at 6:08 AM  wrote:
>> > > >
>> > > > Project:
>> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
>> > > > Build:
>> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/
>> > > > Build Number: 2046
>> > > > Build Status:  Failure
>> > > > Triggered By: Started by timer
>> > > >
>> > > > -
>> > > > Changes Since Last Success:
>> > > > -
>> > > > Changes for Build #2046
>> > > > [Eitan Raviv] network: force select spm - wait for dc status
>> > > >
>> > > >
>> > > >
>> > > >
>> > > > -
>> > > > Failed Tests:
>> > > > -
>> > > > 1 tests failed.
>> > > > FAILED:
>> he-basic-suite-master.test-scenarios.test_004_basic_sanity.test_template_export
>> > > >
>> > > > Error Message:
>> > > > ovirtsdk4.Error: Failed to read response: [(> 0x5624fe64d108>, 7, 'Failed to connect to 192.168.200.99 port 443:
>> Connection refused')]
>> > >
>> > > - The engine VM went down:
>> > >
>> > >
>> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/agent.log
>> > >
>> > > MainThread::INFO::2021-06-08
>> > >
>> 05:07:34,414::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
>> > > Current state EngineUp (score: 3400)
>> > > MainThread::INFO::2021-06-08
>> > >
>> 05:07:44,575::states::135::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
>> > > Penalizing score by 960 due to network status
>> > >
>> > > - Because HA monitoring failed to get a reply from the dns server:
>> > >
>> > >
>> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/broker.log
>> > >
>> > > Thread-1::WARNING::2021-06-08
>> > > 05:07:25,486::network::120::network.Network::(_dns) DNS query failed:
>> > > ; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
>> > > ;; global options: +cmd
>> > > ;; connection timed out; no servers could be reached
>> > >
>> > > Thread-3::INFO::2021-06-08
>> > > 05:07:28,543::mem_free::51::mem_free.MemFree::(action) memFree: 1801
>> > > Thread-5::INFO::2021-06-08
>> > >
>> 05:07:28,972::engine_health::246::engine_health.EngineHealth::(_result_from_stats)
>> > > VM is up on this host with healthy engine
>> > > Thread-2::INFO::2021-06-08
>> > > 05:07:31,532::mgmt_bridge::65::mgmt_bridge.MgmtBridge::(action) Found
>> > > bridge ovirtmgmt in up state
>> > > Thread-1::WARNING::2021-06-08
>> > > 05:07:33,011::network::120::network.Network::(_dns) DNS query failed:
>> > > ; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
>> > > ;; global options: +cmd
>> > > ;; connection timed out; no servers could be reached
>> > >
>> > > Thread-4::INFO::2021-06-08
>> > >
>> 05:07:37,433::cpu_load_no_engine::126::cpu_load_no_engine.CpuLoadNoEngine::(calculate_load)
>> > > System load total=0.3196, engine=0.1724, non-engine=0.1472
>> > > Thread-3::INFO::2021-06-08
>> > > 05:07:37,839::mem_free::51::mem_free.MemFree::(action) memFree: 1735
>> > > Thread-5::INFO::2021-06-08
>> > >
>> 05:07:39,146::engine_health::246::engine_health.EngineHealth::(_result_from_stats)
>> > > VM is up on this host with healthy engine
>> > > Thread-1::WARNING::2021-06-08
>> > > 05:07:40,535::network::120::network.Network::(_dns) 

Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2046 - Failure!

2021-06-22 Thread Yedidyah Bar David
On Wed, Jun 9, 2021 at 12:13 PM Yedidyah Bar David  wrote:
>
> On Tue, Jun 8, 2021 at 9:01 AM Yedidyah Bar David  wrote:
> >
> > On Tue, Jun 8, 2021 at 6:08 AM  wrote:
> > >
> > > Project: 
> > > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> > > Build: 
> > > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/
> > > Build Number: 2046
> > > Build Status:  Failure
> > > Triggered By: Started by timer
> > >
> > > -
> > > Changes Since Last Success:
> > > -
> > > Changes for Build #2046
> > > [Eitan Raviv] network: force select spm - wait for dc status
> > >
> > >
> > >
> > >
> > > -
> > > Failed Tests:
> > > -
> > > 1 tests failed.
> > > FAILED:  
> > > he-basic-suite-master.test-scenarios.test_004_basic_sanity.test_template_export
> > >
> > > Error Message:
> > > ovirtsdk4.Error: Failed to read response: [( > > 0x5624fe64d108>, 7, 'Failed to connect to 192.168.200.99 port 443: 
> > > Connection refused')]
> >
> > - The engine VM went down:
> >
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/agent.log
> >
> > MainThread::INFO::2021-06-08
> > 05:07:34,414::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> > Current state EngineUp (score: 3400)
> > MainThread::INFO::2021-06-08
> > 05:07:44,575::states::135::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
> > Penalizing score by 960 due to network status
> >
> > - Because HA monitoring failed to get a reply from the dns server:
> >
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/broker.log
> >
> > Thread-1::WARNING::2021-06-08
> > 05:07:25,486::network::120::network.Network::(_dns) DNS query failed:
> > ; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
> > ;; global options: +cmd
> > ;; connection timed out; no servers could be reached
> >
> > Thread-3::INFO::2021-06-08
> > 05:07:28,543::mem_free::51::mem_free.MemFree::(action) memFree: 1801
> > Thread-5::INFO::2021-06-08
> > 05:07:28,972::engine_health::246::engine_health.EngineHealth::(_result_from_stats)
> > VM is up on this host with healthy engine
> > Thread-2::INFO::2021-06-08
> > 05:07:31,532::mgmt_bridge::65::mgmt_bridge.MgmtBridge::(action) Found
> > bridge ovirtmgmt in up state
> > Thread-1::WARNING::2021-06-08
> > 05:07:33,011::network::120::network.Network::(_dns) DNS query failed:
> > ; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
> > ;; global options: +cmd
> > ;; connection timed out; no servers could be reached
> >
> > Thread-4::INFO::2021-06-08
> > 05:07:37,433::cpu_load_no_engine::126::cpu_load_no_engine.CpuLoadNoEngine::(calculate_load)
> > System load total=0.3196, engine=0.1724, non-engine=0.1472
> > Thread-3::INFO::2021-06-08
> > 05:07:37,839::mem_free::51::mem_free.MemFree::(action) memFree: 1735
> > Thread-5::INFO::2021-06-08
> > 05:07:39,146::engine_health::246::engine_health.EngineHealth::(_result_from_stats)
> > VM is up on this host with healthy engine
> > Thread-1::WARNING::2021-06-08
> > 05:07:40,535::network::120::network.Network::(_dns) DNS query failed:
> > ; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
> > ;; global options: +cmd
> > ;; connection timed out; no servers could be reached
> >
> > Thread-1::WARNING::2021-06-08
> > 05:07:40,535::network::92::network.Network::(action) Failed to verify
> > network status, (2 out of 5)
> >
> > - Not sure why. DNS servers:
> >
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_etc_resolv.conf
> >
> > # Generated by NetworkManager
> > search lago.local
> > nameserver 192.168.200.1
> > nameserver fe80::5054:ff:fe0c:9ad0%ovirtmgmt
> > nameserver fd8f:1391:3a82:200::1

Now happened again:

https://jenkins.ovirt.org

[JIRA] (OVIRT-3098) Gerrit hook adds unrelated patches to bugs

2021-06-22 Thread Yedidyah Bar David (oVirt JIRA)

[ 
https://ovirt-jira.atlassian.net/browse/OVIRT-3098?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=41029#comment-41029
 ] 

Yedidyah Bar David commented on OVIRT-3098:
---

On Mon, Jun 21, 2021 at 9:58 PM Eyal Shenitzky  wrote:
>
>
> +Dusan Fodor
>
> On Mon, 21 Jun 2021 at 13:32, Nir Soffer  wrote:
>>
>> Gerrit hook is wrongly looking for https://bugzilla.redhat.com/ URLs
>> in the commit message, and adding the patch to the bug.
>>
>> Example patch:
>> https://gerrit.ovirt.org/c/vdsm/+/115339
>>
>> I had to clean up the bug after the broken hook (see screenshot).
>>
>> The hook should really look only in the single URL in (one or more)
>> Bug-Url headers:
>>
>> Bug-Url: https://bugzilla.redhat.com/
>>
>> I reported this years ago (I think for Related-To:), and I remember we had
>> a patch fixing this issue, but for some reason it was lost.
>>
>> Nir

See also: https://ovirt-jira.atlassian.net/browse/OVIRT-3075

Best regards,
-- 
Didi

> Gerrit hook adds unrelated patches to bugs
> --
>
> Key: OVIRT-3098
> URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3098
> Project: oVirt - virtualization made easy
>  Issue Type: By-EMAIL
>Reporter: Nir Soffer
>Assignee: infra
> Attachments: Screenshot from 2021-06-21 13-16-19.png
>
>
> Gerrit hook is wrongly looking for https://bugzilla.redhat.com/ URLs
> in the commit message, and adding the patch to the bug.
> Example patch:
> https://gerrit.ovirt.org/c/vdsm/+/115339
> I had to clean up the bug after the broken hook (see screenshot).
> The hook should really look only in the single URL in (one or more)
> Bug-Url headers:
> Bug-Url: https://bugzilla.redhat.com/
> I reported this years ago (I think for Related-To:), and I remember we had
> a patch fixing this issue, but for some reason it was lost.
> Nir



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100166)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/CI77F45EA4JBRD26GJJWUYX3P4HUJXAL/


Re: he-basic-suite-master failed collecting artifacts

2021-06-21 Thread Yedidyah Bar David
On Wed, Jun 16, 2021 at 2:57 PM Marcin Sobczyk  wrote:
>
> Hi,
>
> On 6/16/21 7:28 AM, Yedidyah Bar David wrote:
> > On Sun, Jun 13, 2021 at 10:30 AM Yedidyah Bar David  wrote:
> >> Hi all,
> >>
> >> Last two nightly runs of [1], 2052 and 2053, failed both after
> >> finishing all tests, during collecting artifacts, with:
> >>
> >> 06:31:39 @ Collect artifacts:
> >> 06:31:40   # [Thread-1] lago-he-basic-suite-master-engine:
> >> 06:31:40   # [Thread-2] lago-he-basic-suite-master-storage:
> >> 06:31:40   # [Thread-3] lago-he-basic-suite-master-host-0:
> >> 06:31:40   # [Thread-4] lago-he-basic-suite-master-host-1:
> >> 06:31:40 Unknown exception: q must be exactly 160, 224, or 256 bits long
> > The nightly job keeps failing like this, and also the node-upgrade suite
> > fails the same way.
> >
> > Any idea?
> >
> > Might be related to changes around artifacts collection?
> It's caused by the fact that we've switched to using FIPS in OST.
> We already had a discussion about it - it was about paramiko
> not properly supporting the RSA keys with SHA256/SHA512 checksums.
> There's no point on working on this - we should focus on switching
> to ansible-based artifact collection in these suites.
>
> If that causes CI-1 votes, then let's disable the old artifact collection
> in respective 'control.sh' files.

OK, rebased your patch for HE, let's see if it passes.

https://gerrit.ovirt.org/c/ovirt-system-tests/+/115115

Thanks!

>
> Regards, Marcin
>
> >
> >> Known issue?
> >>
> >> [1] https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> >>
> >> Thanks and best regards,
> >> --
> >> Didi
> >
> >
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/S7BGTXHLEEZEDC7ZUYSZKE3URXV5QCFI/


he-basic-suite-master failed collecting artifacts

2021-06-21 Thread Yedidyah Bar David
Hi all,

Last two nightly runs of [1], 2052 and 2053, failed both after
finishing all tests, during collecting artifacts, with:

06:31:39 @ Collect artifacts:
06:31:40   # [Thread-1] lago-he-basic-suite-master-engine:
06:31:40   # [Thread-2] lago-he-basic-suite-master-storage:
06:31:40   # [Thread-3] lago-he-basic-suite-master-host-0:
06:31:40   # [Thread-4] lago-he-basic-suite-master-host-1:
06:31:40 Unknown exception: q must be exactly 160, 224, or 256 bits long

Known issue?

[1] https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/

Thanks and best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/XYGTJJPYWQNJIJT63IFYZHIZLSCKRARA/


Re: he-basic-suite-master failed collecting artifacts

2021-06-21 Thread Yedidyah Bar David
On Sun, Jun 13, 2021 at 10:30 AM Yedidyah Bar David  wrote:
>
> Hi all,
>
> Last two nightly runs of [1], 2052 and 2053, failed both after
> finishing all tests, during collecting artifacts, with:
>
> 06:31:39 @ Collect artifacts:
> 06:31:40   # [Thread-1] lago-he-basic-suite-master-engine:
> 06:31:40   # [Thread-2] lago-he-basic-suite-master-storage:
> 06:31:40   # [Thread-3] lago-he-basic-suite-master-host-0:
> 06:31:40   # [Thread-4] lago-he-basic-suite-master-host-1:
> 06:31:40 Unknown exception: q must be exactly 160, 224, or 256 bits long

The nightly job keeps failing like this, and also the node-upgrade suite
fails the same way.

Any idea?

Might be related to changes around artifacts collection?

>
> Known issue?
>
> [1] https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
>
> Thanks and best regards,
> --
> Didi



-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/XSIHIRMHQ6ABLMEP5UAUM5SXXPROC7YN/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2046 - Failure!

2021-06-09 Thread Yedidyah Bar David
On Tue, Jun 8, 2021 at 9:01 AM Yedidyah Bar David  wrote:
>
> On Tue, Jun 8, 2021 at 6:08 AM  wrote:
> >
> > Project: 
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> > Build: 
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/
> > Build Number: 2046
> > Build Status:  Failure
> > Triggered By: Started by timer
> >
> > -
> > Changes Since Last Success:
> > -
> > Changes for Build #2046
> > [Eitan Raviv] network: force select spm - wait for dc status
> >
> >
> >
> >
> > -
> > Failed Tests:
> > -
> > 1 tests failed.
> > FAILED:  
> > he-basic-suite-master.test-scenarios.test_004_basic_sanity.test_template_export
> >
> > Error Message:
> > ovirtsdk4.Error: Failed to read response: [( > 0x5624fe64d108>, 7, 'Failed to connect to 192.168.200.99 port 443: 
> > Connection refused')]
>
> - The engine VM went down:
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/agent.log
>
> MainThread::INFO::2021-06-08
> 05:07:34,414::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state EngineUp (score: 3400)
> MainThread::INFO::2021-06-08
> 05:07:44,575::states::135::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
> Penalizing score by 960 due to network status
>
> - Because HA monitoring failed to get a reply from the dns server:
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/broker.log
>
> Thread-1::WARNING::2021-06-08
> 05:07:25,486::network::120::network.Network::(_dns) DNS query failed:
> ; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
> ;; global options: +cmd
> ;; connection timed out; no servers could be reached
>
> Thread-3::INFO::2021-06-08
> 05:07:28,543::mem_free::51::mem_free.MemFree::(action) memFree: 1801
> Thread-5::INFO::2021-06-08
> 05:07:28,972::engine_health::246::engine_health.EngineHealth::(_result_from_stats)
> VM is up on this host with healthy engine
> Thread-2::INFO::2021-06-08
> 05:07:31,532::mgmt_bridge::65::mgmt_bridge.MgmtBridge::(action) Found
> bridge ovirtmgmt in up state
> Thread-1::WARNING::2021-06-08
> 05:07:33,011::network::120::network.Network::(_dns) DNS query failed:
> ; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
> ;; global options: +cmd
> ;; connection timed out; no servers could be reached
>
> Thread-4::INFO::2021-06-08
> 05:07:37,433::cpu_load_no_engine::126::cpu_load_no_engine.CpuLoadNoEngine::(calculate_load)
> System load total=0.3196, engine=0.1724, non-engine=0.1472
> Thread-3::INFO::2021-06-08
> 05:07:37,839::mem_free::51::mem_free.MemFree::(action) memFree: 1735
> Thread-5::INFO::2021-06-08
> 05:07:39,146::engine_health::246::engine_health.EngineHealth::(_result_from_stats)
> VM is up on this host with healthy engine
> Thread-1::WARNING::2021-06-08
> 05:07:40,535::network::120::network.Network::(_dns) DNS query failed:
> ; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
> ;; global options: +cmd
> ;; connection timed out; no servers could be reached
>
> Thread-1::WARNING::2021-06-08
> 05:07:40,535::network::92::network.Network::(action) Failed to verify
> network status, (2 out of 5)
>
> - Not sure why. DNS servers:
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_etc_resolv.conf
>
> # Generated by NetworkManager
> search lago.local
> nameserver 192.168.200.1
> nameserver fe80::5054:ff:fe0c:9ad0%ovirtmgmt
> nameserver fd8f:1391:3a82:200::1
>
> - The command we run is 'dig +tries=1 +time=5', which defaults to
> querying for '.' (the dns root). This is normally cached locally, but
> has a TTL of 86400, meaning it can be cached for up to one day. So if
> we ran this query right after it expired, _and_ then the local dns
> server had some issues forwarding our request (due to external issues,
> perhaps), then it would fail like this. I am going to ignore this
> failure for now, assuming it was temporary, but it might be worth
> opening an RFE on ovirt-hosted-engine-ha asking for some more
> flexibilit

Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2046 - Failure!

2021-06-08 Thread Yedidyah Bar David
On Tue, Jun 8, 2021 at 6:08 AM  wrote:
>
> Project: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> Build: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/
> Build Number: 2046
> Build Status:  Failure
> Triggered By: Started by timer
>
> -
> Changes Since Last Success:
> -
> Changes for Build #2046
> [Eitan Raviv] network: force select spm - wait for dc status
>
>
>
>
> -
> Failed Tests:
> -
> 1 tests failed.
> FAILED:  
> he-basic-suite-master.test-scenarios.test_004_basic_sanity.test_template_export
>
> Error Message:
> ovirtsdk4.Error: Failed to read response: [( 0x5624fe64d108>, 7, 'Failed to connect to 192.168.200.99 port 443: Connection 
> refused')]

- The engine VM went down:

https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/agent.log

MainThread::INFO::2021-06-08
05:07:34,414::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
Current state EngineUp (score: 3400)
MainThread::INFO::2021-06-08
05:07:44,575::states::135::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
Penalizing score by 960 due to network status

- Because HA monitoring failed to get a reply from the dns server:

https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/broker.log

Thread-1::WARNING::2021-06-08
05:07:25,486::network::120::network.Network::(_dns) DNS query failed:
; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
;; global options: +cmd
;; connection timed out; no servers could be reached

Thread-3::INFO::2021-06-08
05:07:28,543::mem_free::51::mem_free.MemFree::(action) memFree: 1801
Thread-5::INFO::2021-06-08
05:07:28,972::engine_health::246::engine_health.EngineHealth::(_result_from_stats)
VM is up on this host with healthy engine
Thread-2::INFO::2021-06-08
05:07:31,532::mgmt_bridge::65::mgmt_bridge.MgmtBridge::(action) Found
bridge ovirtmgmt in up state
Thread-1::WARNING::2021-06-08
05:07:33,011::network::120::network.Network::(_dns) DNS query failed:
; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
;; global options: +cmd
;; connection timed out; no servers could be reached

Thread-4::INFO::2021-06-08
05:07:37,433::cpu_load_no_engine::126::cpu_load_no_engine.CpuLoadNoEngine::(calculate_load)
System load total=0.3196, engine=0.1724, non-engine=0.1472
Thread-3::INFO::2021-06-08
05:07:37,839::mem_free::51::mem_free.MemFree::(action) memFree: 1735
Thread-5::INFO::2021-06-08
05:07:39,146::engine_health::246::engine_health.EngineHealth::(_result_from_stats)
VM is up on this host with healthy engine
Thread-1::WARNING::2021-06-08
05:07:40,535::network::120::network.Network::(_dns) DNS query failed:
; <<>> DiG 9.11.26-RedHat-9.11.26-4.el8_4 <<>> +tries=1 +time=5
;; global options: +cmd
;; connection timed out; no servers could be reached

Thread-1::WARNING::2021-06-08
05:07:40,535::network::92::network.Network::(action) Failed to verify
network status, (2 out of 5)

- Not sure why. DNS servers:

https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_etc_resolv.conf

# Generated by NetworkManager
search lago.local
nameserver 192.168.200.1
nameserver fe80::5054:ff:fe0c:9ad0%ovirtmgmt
nameserver fd8f:1391:3a82:200::1

- The command we run is 'dig +tries=1 +time=5', which defaults to
querying for '.' (the dns root). This is normally cached locally, but
has a TTL of 86400, meaning it can be cached for up to one day. So if
we ran this query right after it expired, _and_ then the local dns
server had some issues forwarding our request (due to external issues,
perhaps), then it would fail like this. I am going to ignore this
failure for now, assuming it was temporary, but it might be worth
opening an RFE on ovirt-hosted-engine-ha asking for some more
flexibility - setting the query string or something similar. I think
that this bug is probably quite hard to reproduce, because normally,
all hosts will use the same dns server, and problems with it will
affect all of them similarly.

- Anyway, it seems like there were temporary connectivity issues on
the network there. A minute later:

https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2046/artifact/exported-artifacts/test_logs/he-basic-suite-master/lago-he-basic-suite-master-host-0/_var_log/ovirt-hosted-engine-ha/broker.log

Thread-1::INFO::2021-06-08
05:08:08,143::network::88::network.Network::(action) Successfully
verified network status

But that was too late and the engine VM was already on its way 

Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2027 - Failure!

2021-05-25 Thread Yedidyah Bar David
On Tue, May 25, 2021 at 5:07 AM  wrote:
>
> Project: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> Build: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2027/

STDERR
b'Error: Transaction test error:\n  installing package
ovirt-engine-appliance-4.4-20210524174442.1.el8.x86_64 needs 320MB on
the / filesystem\n  installing package unzip-6.0-45.el8.x86_64 needs
320MB on the / filesystem\n  installing package
libgcrypt-1.8.5-5.el8.x86_64 needs 321MB on the / filesystem\n
installing package cups-libs-1:2.2.6-40.el8.x86_64 needs 321MB on the
/ filesystem\n  installing package libfastjson-0.99.9-1.el8.x86_64
needs 320MB on the / filesystem\n\nError Summary\n-\nDisk
Requirements:\n   At least 321MB more space needed on the /
filesystem.\n\n'

Known issue?

Best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/5QI2GQJGNRMUDP3OS6MS3HYWNL5BUSTE/


Whitelisting CI

2021-05-19 Thread Yedidyah Bar David
Hi all,

1. Please whitelist Dmitry Voronetskiy [1]. Do we have any policy re
what should be done in such cases? Perhaps sign some contributor
agreement or whatever?

2. I think we should make CI allow existing whitelisted users to issue
'ci test'/'ci build' commands even if the patch author (or gerrit
owner or whatever) is not whitelisted (yet).

[1] https://gerrit.ovirt.org/c/ovirt-dwh/+/114804

Thanks and best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/MNFL2ZUKGQ67CWMCX2LOCHQTZRHFBR6P/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-master - Build # 2012 - Still Failing!

2021-05-09 Thread Yedidyah Bar David
On Mon, May 10, 2021 at 8:01 AM  wrote:
>
> Project: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> Build: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2012/

08:01:23 stderr: fatal: unable to access
'https://gerrit.ovirt.org/ovirt-system-tests.git/': Could not resolve
host: gerrit.ovirt.org

Known issue?
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/4PJUZXVZWC6GOQPTLOW2LKDH67NRO7PT/


Re: ovirt-appliance CI

2021-03-18 Thread Yedidyah Bar David
On Wed, Mar 17, 2021 at 4:44 PM Ehud Yonasi  wrote:
>
> I've sent a patch for example [1],
> You need to define the poll stage in order for the poll job to run.
>
>
> [1]: https://gerrit.ovirt.org/c/ovirt-appliance/+/113923

Lev merged it, I ran the job, seems to have worked well. Thanks!

>
> On Wed, Mar 17, 2021 at 3:01 PM Yedidyah Bar David  wrote:
>>
>> On Wed, Mar 17, 2021 at 2:52 PM Yedidyah Bar David  wrote:
>> >
>> > On Wed, Mar 17, 2021 at 2:01 PM Ehud Yonasi  wrote:
>> > >
>> > > Hey Didi,
>> > >
>> > > Note that the `ci build` triggers stdci v2 jobs such as 
>> > > `ovirt-appliance_standard-check-patch` and inside it will build the 
>> > > artifacts.
>> > > The other job you mentioned is a v1 job that is less flexible and is 
>> > > looking for the `build-artifacts.sh` script. to fix it we can just 
>> > > remove that job because
>> > > you are already using v2.
>> >
>> > OK. I see you merged my patch, thanks. I now ran the job, let's see if it 
>> > works.
>>
>> It didn't do anything [1]. It says:
>>
>> https://jenkins.ovirt.org/job/ovirt-appliance_master_standard-poll-upstream-sources/1/artifact/ci_build_summary.html
>> :
>>
>> Note:For post-merge runs, only check-merged threads get invoked.
>> To get build-artifacts threads to run, you need to configure the
>> branch as a release branch.
>>
>> Although I do have this:
>>
>> release_branches:
>> master: ovirt-master
>>
>> Any idea? Thanks.
>>
>> [1] 
>> https://jenkins.ovirt.org/job/ovirt-appliance_master_standard-poll-upstream-sources/1/
>>
>> >
>> > Would you now remove the old job definition?
>> >
>> > Would this affect anything else? Publishing, etc.?
>> >
>> > Thanks and best regards,
>> >
>> > >
>> > > Thanks,
>> > > Ehud.
>> > >
>> > > On Wed, Mar 17, 2021 at 1:50 PM Anton Marchukov  
>> > > wrote:
>> > >>
>> > >> Ehud, can you have a look, please?
>> > >>
>> > >> Thanks.
>> > >>
>> > >> > On 17. 3. 2021, at 12:22, Yedidyah Bar David  wrote:
>> > >> >
>> > >> > Hi all,
>> > >> >
>> > >> > We now merged a patch [1] that changes the CI configuration of
>> > >> > ovirt-appliance. In particular, it replaces check-patch and
>> > >> > build-artifacts files under automation/ with new sub-stages. It worked
>> > >> > well when I did 'ci build', before merging, but now, after merging, I
>> > >> > ran [2], and it failed, saying:
>> > >> >
>> > >> > 12:07:07 WARN: Unable to find sh file automation/build-artifacts.sh or
>> > >> > automation/build-artifacts.sh.el8, skipping sh
>> > >> >
>> > >> > I suppose this is due to its definition in jenkins [3], but not sure
>> > >> > how to fix. Perhaps best replace this with an entry in
>> > >> > standard-poll-stage-pipelines.yaml - e.g. [4] (and probably remove
>> > >> > [3])? Something else? If [4], not sure what the implications are
>> > >> > elsewhere.
>> > >> >
>> > >> > Thanks and best regards,
>> > >> >
>> > >> > [1] https://gerrit.ovirt.org/c/ovirt-appliance/+/113555
>> > >> >
>> > >> > [2] 
>> > >> > https://jenkins.ovirt.org/job/ovirt-appliance_master_build-artifacts-el8-x86_64/605/
>> > >> >
>> > >> > [3] 
>> > >> > https://github.com/oVirt/jenkins/blob/master/jobs/confs/projects/ovirt-appliance/ovirt-appliance.yaml
>> > >> >
>> > >> > [4] https://gerrit.ovirt.org/c/jenkins/+/113912
>> > >> > --
>> > >> > Didi
>> > >> > ___
>> > >> > Infra mailing list -- infra@ovirt.org
>> > >> > To unsubscribe send an email to infra-le...@ovirt.org
>> > >> > Privacy Statement: https://www.ovirt.org/privacy-policy.html
>> > >> > oVirt Code of Conduct: 
>> > >> > https://www.ovirt.org/community/about/community-guidelines/
>> > >> > List Archives: 
>> > >> > https://lists.ovirt.org/archives/list/infra@ovirt.org/message/BWXH2VXIPUNLGVSKEGTNNPYN57B36577/
>> > >>
>> > >> --
>> > >> Anton Marchukov
>> > >> Associate Manager - RHV DevOps - Red Hat
>> > >>
>> > >>
>> > >>
>> > >>
>> > >>
>> > >>
>> >
>> >
>> > --
>> > Didi
>>
>>
>>
>> --
>> Didi
>>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/PSQNH7IDKN7MPEPEX7VKUSH5VOARXTPR/


Re: [ovirt-devel] Re: basic suite fails on test_metrics_and_log_collector

2021-03-18 Thread Yedidyah Bar David
On Wed, Mar 17, 2021 at 5:38 PM Yedidyah Bar David  wrote:
>
> On Wed, Mar 17, 2021 at 4:48 PM Martin Perina  wrote:
> >
> >
> >
> > On Wed, Mar 17, 2021 at 3:24 PM Michal Skrivanek 
> >  wrote:
> >>
> >>
> >>
> >> On 17. 3. 2021, at 13:53, Dana Elfassy  wrote:
> >>
> >> Adding +Marcin Sobczyk
> >>
> >> On Mon, Mar 15, 2021 at 9:59 AM Yedidyah Bar David  wrote:
> >>>
> >>> On Mon, Mar 15, 2021 at 7:55 AM Yedidyah Bar David  
> >>> wrote:
> >>> >
> >>> > Hi all,
> >>> >
> >>> > This started a few days ago [1] and randomly happens since then:
> >>> >
> >>> > E   DEBUG: Configuration:
> >>> > E   DEBUG: command: collect
> >>> > E   DEBUG: Traceback (most recent call last):
> >>> > E   DEBUG:   File
> >>> > "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
> >>> > line 2067, in 
> >>> > E   DEBUG: '%s directory is not empty.' % 
> >>> > (conf["local_tmp_dir"])
> >>> > E   DEBUG: Exception: /dev/shm/log directory is not
> >>> > empty.ERROR: /dev/shm/log directory is not empty.non-zero return code
> >>> >
> >>> > Michal tried to fix this by using a random directory but it still fails 
> >>> > [2]:
> >>> >
> >>> > DEBUG: command: collect
> >>> > DEBUG: Traceback (most recent call last):
> >>> > DEBUG:   File 
> >>> > "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
> >>> > line 2067, in 
> >>> > DEBUG: '%s directory is not empty.' % (conf["local_tmp_dir"])
> >>> > DEBUG: Exception: /dev/shm/kaN7uY directory is not empty.ERROR:
> >>> > /dev/shm/kaN7uY directory is not empty.non-zero return code
> >>> >
> >>> > Since I suppose that the randomness of mktemp is good enough, it must
> >>> > be something else. Also, the last successful run before [1] used the
> >>> > same OST git commit (same code), so I do not think it's something in
> >>> > OST's code.
> >>> >
> >>> > Any idea?
> >>> >
> >>> > I think I'll push a patch to create and use the directory right before
> >>> > calling ovirt-log-collector, which is probably better in other ways.
> >>>
> >>> My patch [1] still fails, with a somewhat different error message, but
> >>> this made me check further, and while I still do not understand, I have
> >>> this to add:
> >>>
> >>> In the failing runs, ovirt-log-collector is called *twice* in parallel. 
> >>> E.g.
> >>> in [2] (the check-patch of [1]):
> >>>
> >>> Mar 15 07:38:59 lago-basic-suite-master-engine platform-python[59099]:
> >>> ansible-command Invoked with _raw_params=lctmp=$(mktemp -d -p
> >>> /dev/shm); ovirt-log-collector --verbose --batch --no-hypervisors
> >>> --local-tmp="${lctmp}" --conf-file=/root/ovirt-log-collector.conf
> >>> _uses_shell=True warn=True stdin_add_newline=True
> >>> strip_empty_ends=True argv=None chdir=None executable=None
> >>> creates=None removes=None stdin=None
> >>> Mar 15 07:38:59 lago-basic-suite-master-engine platform-python[59124]:
> >>> ansible-command Invoked with _raw_params=lctmp=$(mktemp -d -p
> >>> /dev/shm); ovirt-log-collector --verbose --batch --no-hypervisors
> >>> --local-tmp="${lctmp}" --conf-file=/root/ovirt-log-collector.conf
> >>> _uses_shell=True warn=True stdin_add_newline=True
> >>> strip_empty_ends=True argv=None chdir=None executable=None
> >>> creates=None removes=None stdin=None
> >>>
> >>> It also generates two logs, which you can check/compare.
> >>>
> >>> It's the same for previous ones, e.g. latest nightly [3][4]:
> >>>
> >>> Mar 15 06:23:30 lago-basic-suite-master-engine platform-python[59343]:
> >>> ansible-command Invoked with _raw_params=ovirt-log-collector --verbose
> >>> --batch --no-hypervisors --conf-file=/root/ovirt-log-collector.conf
> >>> _uses_shell=True warn=True stdin_add_newline=True
> >>> strip_empty_ends=True argv=None chdir=None executable=None
> >>> creates=None removes=None stdin=None
> &g

Re: [ovirt-devel] Re: basic suite fails on test_metrics_and_log_collector

2021-03-17 Thread Yedidyah Bar David
On Wed, Mar 17, 2021 at 4:48 PM Martin Perina  wrote:
>
>
>
> On Wed, Mar 17, 2021 at 3:24 PM Michal Skrivanek 
>  wrote:
>>
>>
>>
>> On 17. 3. 2021, at 13:53, Dana Elfassy  wrote:
>>
>> Adding +Marcin Sobczyk
>>
>> On Mon, Mar 15, 2021 at 9:59 AM Yedidyah Bar David  wrote:
>>>
>>> On Mon, Mar 15, 2021 at 7:55 AM Yedidyah Bar David  wrote:
>>> >
>>> > Hi all,
>>> >
>>> > This started a few days ago [1] and randomly happens since then:
>>> >
>>> > E   DEBUG: Configuration:
>>> > E   DEBUG: command: collect
>>> > E   DEBUG: Traceback (most recent call last):
>>> > E   DEBUG:   File
>>> > "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
>>> > line 2067, in 
>>> > E   DEBUG: '%s directory is not empty.' % 
>>> > (conf["local_tmp_dir"])
>>> > E   DEBUG: Exception: /dev/shm/log directory is not
>>> > empty.ERROR: /dev/shm/log directory is not empty.non-zero return code
>>> >
>>> > Michal tried to fix this by using a random directory but it still fails 
>>> > [2]:
>>> >
>>> > DEBUG: command: collect
>>> > DEBUG: Traceback (most recent call last):
>>> > DEBUG:   File 
>>> > "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
>>> > line 2067, in 
>>> > DEBUG: '%s directory is not empty.' % (conf["local_tmp_dir"])
>>> > DEBUG: Exception: /dev/shm/kaN7uY directory is not empty.ERROR:
>>> > /dev/shm/kaN7uY directory is not empty.non-zero return code
>>> >
>>> > Since I suppose that the randomness of mktemp is good enough, it must
>>> > be something else. Also, the last successful run before [1] used the
>>> > same OST git commit (same code), so I do not think it's something in
>>> > OST's code.
>>> >
>>> > Any idea?
>>> >
>>> > I think I'll push a patch to create and use the directory right before
>>> > calling ovirt-log-collector, which is probably better in other ways.
>>>
>>> My patch [1] still fails, with a somewhat different error message, but
>>> this made me check further, and while I still do not understand, I have
>>> this to add:
>>>
>>> In the failing runs, ovirt-log-collector is called *twice* in parallel. E.g.
>>> in [2] (the check-patch of [1]):
>>>
>>> Mar 15 07:38:59 lago-basic-suite-master-engine platform-python[59099]:
>>> ansible-command Invoked with _raw_params=lctmp=$(mktemp -d -p
>>> /dev/shm); ovirt-log-collector --verbose --batch --no-hypervisors
>>> --local-tmp="${lctmp}" --conf-file=/root/ovirt-log-collector.conf
>>> _uses_shell=True warn=True stdin_add_newline=True
>>> strip_empty_ends=True argv=None chdir=None executable=None
>>> creates=None removes=None stdin=None
>>> Mar 15 07:38:59 lago-basic-suite-master-engine platform-python[59124]:
>>> ansible-command Invoked with _raw_params=lctmp=$(mktemp -d -p
>>> /dev/shm); ovirt-log-collector --verbose --batch --no-hypervisors
>>> --local-tmp="${lctmp}" --conf-file=/root/ovirt-log-collector.conf
>>> _uses_shell=True warn=True stdin_add_newline=True
>>> strip_empty_ends=True argv=None chdir=None executable=None
>>> creates=None removes=None stdin=None
>>>
>>> It also generates two logs, which you can check/compare.
>>>
>>> It's the same for previous ones, e.g. latest nightly [3][4]:
>>>
>>> Mar 15 06:23:30 lago-basic-suite-master-engine platform-python[59343]:
>>> ansible-command Invoked with _raw_params=ovirt-log-collector --verbose
>>> --batch --no-hypervisors --conf-file=/root/ovirt-log-collector.conf
>>> _uses_shell=True warn=True stdin_add_newline=True
>>> strip_empty_ends=True argv=None chdir=None executable=None
>>> creates=None removes=None stdin=None
>>> Mar 15 06:23:30 lago-basic-suite-master-engine setroubleshoot[58889]:
>>> SELinux is preventing /usr/lib/systemd/systemd from unlink access on
>>> the sock_file ansible-ssh-lago-basic-suite-master-host-1-22-root. For
>>> complete SELinux messages run: sealert -l
>>> d03a8655-9430-4fcf-9892-3b4df1939899
>>> Mar 15 06:23:30 lago-basic-suite-master-engine setroubleshoot[58889]:
>>> SELinux is preventing /usr/lib/systemd/systemd from unlink access on
>>> the s

Re: ovirt-appliance CI

2021-03-17 Thread Yedidyah Bar David
On Wed, Mar 17, 2021 at 2:52 PM Yedidyah Bar David  wrote:
>
> On Wed, Mar 17, 2021 at 2:01 PM Ehud Yonasi  wrote:
> >
> > Hey Didi,
> >
> > Note that the `ci build` triggers stdci v2 jobs such as 
> > `ovirt-appliance_standard-check-patch` and inside it will build the 
> > artifacts.
> > The other job you mentioned is a v1 job that is less flexible and is 
> > looking for the `build-artifacts.sh` script. to fix it we can just remove 
> > that job because
> > you are already using v2.
>
> OK. I see you merged my patch, thanks. I now ran the job, let's see if it 
> works.

It didn't do anything [1]. It says:

https://jenkins.ovirt.org/job/ovirt-appliance_master_standard-poll-upstream-sources/1/artifact/ci_build_summary.html
:

Note:For post-merge runs, only check-merged threads get invoked.
To get build-artifacts threads to run, you need to configure the
branch as a release branch.

Although I do have this:

release_branches:
master: ovirt-master

Any idea? Thanks.

[1] 
https://jenkins.ovirt.org/job/ovirt-appliance_master_standard-poll-upstream-sources/1/

>
> Would you now remove the old job definition?
>
> Would this affect anything else? Publishing, etc.?
>
> Thanks and best regards,
>
> >
> > Thanks,
> > Ehud.
> >
> > On Wed, Mar 17, 2021 at 1:50 PM Anton Marchukov  wrote:
> >>
> >> Ehud, can you have a look, please?
> >>
> >> Thanks.
> >>
> >> > On 17. 3. 2021, at 12:22, Yedidyah Bar David  wrote:
> >> >
> >> > Hi all,
> >> >
> >> > We now merged a patch [1] that changes the CI configuration of
> >> > ovirt-appliance. In particular, it replaces check-patch and
> >> > build-artifacts files under automation/ with new sub-stages. It worked
> >> > well when I did 'ci build', before merging, but now, after merging, I
> >> > ran [2], and it failed, saying:
> >> >
> >> > 12:07:07 WARN: Unable to find sh file automation/build-artifacts.sh or
> >> > automation/build-artifacts.sh.el8, skipping sh
> >> >
> >> > I suppose this is due to its definition in jenkins [3], but not sure
> >> > how to fix. Perhaps best replace this with an entry in
> >> > standard-poll-stage-pipelines.yaml - e.g. [4] (and probably remove
> >> > [3])? Something else? If [4], not sure what the implications are
> >> > elsewhere.
> >> >
> >> > Thanks and best regards,
> >> >
> >> > [1] https://gerrit.ovirt.org/c/ovirt-appliance/+/113555
> >> >
> >> > [2] 
> >> > https://jenkins.ovirt.org/job/ovirt-appliance_master_build-artifacts-el8-x86_64/605/
> >> >
> >> > [3] 
> >> > https://github.com/oVirt/jenkins/blob/master/jobs/confs/projects/ovirt-appliance/ovirt-appliance.yaml
> >> >
> >> > [4] https://gerrit.ovirt.org/c/jenkins/+/113912
> >> > --
> >> > Didi
> >> > ___
> >> > Infra mailing list -- infra@ovirt.org
> >> > To unsubscribe send an email to infra-le...@ovirt.org
> >> > Privacy Statement: https://www.ovirt.org/privacy-policy.html
> >> > oVirt Code of Conduct: 
> >> > https://www.ovirt.org/community/about/community-guidelines/
> >> > List Archives: 
> >> > https://lists.ovirt.org/archives/list/infra@ovirt.org/message/BWXH2VXIPUNLGVSKEGTNNPYN57B36577/
> >>
> >> --
> >> Anton Marchukov
> >> Associate Manager - RHV DevOps - Red Hat
> >>
> >>
> >>
> >>
> >>
> >>
>
>
> --
> Didi



-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/33U7I4K7CQAXUTLGJRLAT7VPWBRLTEWT/


Re: ovirt-appliance CI

2021-03-17 Thread Yedidyah Bar David
On Wed, Mar 17, 2021 at 2:01 PM Ehud Yonasi  wrote:
>
> Hey Didi,
>
> Note that the `ci build` triggers stdci v2 jobs such as 
> `ovirt-appliance_standard-check-patch` and inside it will build the artifacts.
> The other job you mentioned is a v1 job that is less flexible and is looking 
> for the `build-artifacts.sh` script. to fix it we can just remove that job 
> because
> you are already using v2.

OK. I see you merged my patch, thanks. I now ran the job, let's see if it works.

Would you now remove the old job definition?

Would this affect anything else? Publishing, etc.?

Thanks and best regards,

>
> Thanks,
> Ehud.
>
> On Wed, Mar 17, 2021 at 1:50 PM Anton Marchukov  wrote:
>>
>> Ehud, can you have a look, please?
>>
>> Thanks.
>>
>> > On 17. 3. 2021, at 12:22, Yedidyah Bar David  wrote:
>> >
>> > Hi all,
>> >
>> > We now merged a patch [1] that changes the CI configuration of
>> > ovirt-appliance. In particular, it replaces check-patch and
>> > build-artifacts files under automation/ with new sub-stages. It worked
>> > well when I did 'ci build', before merging, but now, after merging, I
>> > ran [2], and it failed, saying:
>> >
>> > 12:07:07 WARN: Unable to find sh file automation/build-artifacts.sh or
>> > automation/build-artifacts.sh.el8, skipping sh
>> >
>> > I suppose this is due to its definition in jenkins [3], but not sure
>> > how to fix. Perhaps best replace this with an entry in
>> > standard-poll-stage-pipelines.yaml - e.g. [4] (and probably remove
>> > [3])? Something else? If [4], not sure what the implications are
>> > elsewhere.
>> >
>> > Thanks and best regards,
>> >
>> > [1] https://gerrit.ovirt.org/c/ovirt-appliance/+/113555
>> >
>> > [2] 
>> > https://jenkins.ovirt.org/job/ovirt-appliance_master_build-artifacts-el8-x86_64/605/
>> >
>> > [3] 
>> > https://github.com/oVirt/jenkins/blob/master/jobs/confs/projects/ovirt-appliance/ovirt-appliance.yaml
>> >
>> > [4] https://gerrit.ovirt.org/c/jenkins/+/113912
>> > --
>> > Didi
>> > ___
>> > Infra mailing list -- infra@ovirt.org
>> > To unsubscribe send an email to infra-le...@ovirt.org
>> > Privacy Statement: https://www.ovirt.org/privacy-policy.html
>> > oVirt Code of Conduct: 
>> > https://www.ovirt.org/community/about/community-guidelines/
>> > List Archives: 
>> > https://lists.ovirt.org/archives/list/infra@ovirt.org/message/BWXH2VXIPUNLGVSKEGTNNPYN57B36577/
>>
>> --
>> Anton Marchukov
>> Associate Manager - RHV DevOps - Red Hat
>>
>>
>>
>>
>>
>>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/F4BM4CGQHCEEVD6OBM42LVUDSASQGGL2/


ovirt-appliance CI

2021-03-17 Thread Yedidyah Bar David
Hi all,

We now merged a patch [1] that changes the CI configuration of
ovirt-appliance. In particular, it replaces check-patch and
build-artifacts files under automation/ with new sub-stages. It worked
well when I did 'ci build', before merging, but now, after merging, I
ran [2], and it failed, saying:

12:07:07 WARN: Unable to find sh file automation/build-artifacts.sh or
automation/build-artifacts.sh.el8, skipping sh

I suppose this is due to its definition in jenkins [3], but not sure
how to fix. Perhaps best replace this with an entry in
standard-poll-stage-pipelines.yaml - e.g. [4] (and probably remove
[3])? Something else? If [4], not sure what the implications are
elsewhere.

Thanks and best regards,

[1] https://gerrit.ovirt.org/c/ovirt-appliance/+/113555

[2] 
https://jenkins.ovirt.org/job/ovirt-appliance_master_build-artifacts-el8-x86_64/605/

[3] 
https://github.com/oVirt/jenkins/blob/master/jobs/confs/projects/ovirt-appliance/ovirt-appliance.yaml

[4] https://gerrit.ovirt.org/c/jenkins/+/113912
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/BWXH2VXIPUNLGVSKEGTNNPYN57B36577/


dummy-master_change-queue is broken

2021-03-17 Thread Yedidyah Bar David
Hi all,

On Wed, Mar 17, 2021 at 10:43 AM Code Review  wrote:
>
> From Jenkins CI :
>
> Jenkins CI has posted comments on this change. ( 
> https://gerrit.ovirt.org/c/jenkins/+/113907 )
>
> Change subject: Email didi on node suite
> ..
>
>
> Patch Set 1:
>
> Build Failed
>
> https://jenkins.ovirt.org/job/standard-enqueue/32675/ :
> This change was not submitted to any change queues for system testing. You 
> will need to create some 'build-artifacts' jobs if you want changes to be 
> submitted to change queues, take part in the system tests and be deployed to 
> the nightly snapshot repositories. If your project uses STDCI V2 and you have 
> release branches configured, you may disregard this message.
>
>
> https://jenkins.ovirt.org/job/jenkins_standard-on-merge/1126/ : UNSTABLE

AFAIU, this ^^^ ran:

https://jenkins.ovirt.org/job/dummy-master_change-queue/1284/

Which failed with:

10:46:24File
"/home/jenkins/workspace/dummy-master_change-queue@tmp/durable-d1ad9c18/script.sh",
line 7, in 
10:46:24  with JenkinsChangeQueue.persist_in_artifacts() as cq:
10:46:24File "/usr/lib64/python3.6/contextlib.py", line 81, in __enter__
10:46:24  return next(self.gen)
10:46:24File
"/home/jenkins/workspace/dummy-master_change-queue/jenkins/stdci_libs/jenkins_objects.py",
line 315, in persist_in_artifacts
10:46:24  obj = cls.load_from_artifact(artifact_file)
10:46:24File
"/home/jenkins/workspace/dummy-master_change-queue/jenkins/stdci_libs/jenkins_objects.py",
line 296, in load_from_artifact
10:46:24  artifact_file, cls if fallback_to_new else None
10:46:24File
"/home/jenkins/workspace/dummy-master_change-queue/jenkins/stdci_libs/jenkins_objects.py",
line 270, in object_from_artifact
10:46:24  return _cpickle_loads(fd.read())
10:46:24File
"/home/jenkins/workspace/dummy-master_change-queue/jenkins/stdci_libs/jenkins_objects.py",
line 223, in _cpickle_loads
10:46:24  return pk.load()
10:46:24File
"/home/jenkins/workspace/dummy-master_change-queue/jenkins/stdci_libs/jenkins_objects.py",
line 207, in _find_global
10:46:24  mod = import_module(mod_name)
10:46:24File "/usr/lib64/python3.6/importlib/__init__.py", line
126, in import_module
10:46:24  return _bootstrap._gcd_import(name[level:], package, level)
10:46:24File "", line 994, in _gcd_import
10:46:24File "", line 971, in _find_and_load
10:46:24File "", line 953, in
_find_and_load_unlocked
10:46:24  ModuleNotFoundError: No module named 'copy_reg'

Quickly looking at the code, it seems like we keep somewhere in
"exported-artifacts" cached picked objects, and above fails when
trying to load them.

I guess it might include too-old code and should be refreshed
(hopefully, if we did this the way I expect we did, by just finding
the file (".dat") and removing it).

Last successful run of this job was a month ago.

Anyone is taking care of this?

Thanks,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/6XXM7NONBOCUB4RFKWD3IM7WH54HTF3I/


Re: basic suite fails on test_metrics_and_log_collector

2021-03-17 Thread Yedidyah Bar David
On Tue, Mar 16, 2021 at 9:07 PM Yedidyah Bar David  wrote:

> On Tue, Mar 16, 2021 at 1:11 PM Yedidyah Bar David 
> wrote:
> >
> > On Tue, Mar 16, 2021 at 12:50 PM Dana Elfassy 
> wrote:
> > >
> > > About [5] https://bugzilla.redhat.com/show_bug.cgi?id=1917707, I have
> found that when executing the http request of the playbook, sometimes in
> the response there are some events that are missing. The way we handle the
> events is via while loop, where in each loop we continue from the
> lastEventId. Since not all events were listed, the lastEventId (which is
> returned as total number of events) is smaller than the actual eventId that
> was handled. Thus, in the next iteration the last events are handled again.
> It seems that this comes from Ansible Runner, and I'm investigating that
> area now
> >
> > Thanks
> >
> > > I'm not sure if it's the same issue as you get with the logs, though,
> as I'm not familiar with this area in the OSTs. Is the creation of the 2
> logs done by different threads?
> >
> > I have no idea either.
> >
> > >
> > > Regarding the initial error (' directory is not empty.non-zero return
> code') I'm not sure, will have to loop more into it. Martin, do you have
> any idea?
> >
> > I think it's quite obvious that the reason for this is that we run two
> > ovirt-log-collector processes in parallel. Why we (sometimes) do this
> > is an open question, but unrelated to ovirt-log-collector (or the
> > specific error message it emits).
>
> I now did this:
>
> 1. Wrote a trivial dummy module [1] to test this
>
> 2. Ran it on an already finished basic-suite env with lagofy/run_tc [2].
>
> As you can see, if you pipe [2] through grep -Ev
> 'do_something_2.*something 2|do_something_1.*something 1', in some
> cases do_something_1 actually ran the ansible code of do_something_2.
> I currently think it's some kind of race condition in ansible-runner,
> causing it to misbehave when ran in parallel, as we do in
> test_metrics_and_log_collector. I still didn't try to check its code
> and/or add debugging stuff to it to try and find the exact location.
>
> 2021-03-16 20:54:18+0200,681 INFO[test_003_01_dummy]
> do_something_1: Tue Mar 16 19:54:18 CET 2021
> lago-basic-suite-master-engine something 2 (test_003_01_dummy:39)
> 2021-03-16 20:55:41+0200,155 INFO[test_003_01_dummy]
> do_something_1: Tue Mar 16 19:55:40 CET 2021
> lago-basic-suite-master-engine something 2 (test_003_01_dummy:39)
> 2021-03-16 20:56:56+0200,196 INFO[test_003_01_dummy]
> do_something_1: Tue Mar 16 19:56:55 CET 2021
> lago-basic-suite-master-engine something 2 (test_003_01_dummy:39)
> 2021-03-16 20:58:19+0200,665 INFO[test_003_01_dummy]
> do_something_1: Tue Mar 16 19:58:19 CET 2021
> lago-basic-suite-master-engine something 2 (test_003_01_dummy:39)
>

For now, pushed:

https://gerrit.ovirt.org/c/ovirt-system-tests/+/113906
test_metrics_and_log_collector: Run serially

Best regards,


>
> [1]
> https://github.com/didib/ovirt-system-tests/commit/6a5327976eaf5bc27e352723bf5c17dab4e60fa4
>
> [2] https://gist.github.com/didib/fe1b3bc28b8364cf6976d3d93e6cd9b6
>
> >
> > >
> > > On Mon, Mar 15, 2021 at 9:59 AM Yedidyah Bar David 
> wrote:
> > >>
> > >> On Mon, Mar 15, 2021 at 7:55 AM Yedidyah Bar David 
> wrote:
> > >> >
> > >> > Hi all,
> > >> >
> > >> > This started a few days ago [1] and randomly happens since then:
> > >> >
> > >> > E   DEBUG: Configuration:
> > >> > E   DEBUG: command: collect
> > >> > E   DEBUG: Traceback (most recent call last):
> > >> > E   DEBUG:   File
> > >> > "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
> > >> > line 2067, in 
> > >> > E   DEBUG: '%s directory is not empty.' %
> (conf["local_tmp_dir"])
> > >> > E   DEBUG: Exception: /dev/shm/log directory is not
> > >> > empty.ERROR: /dev/shm/log directory is not empty.non-zero return
> code
> > >> >
> > >> > Michal tried to fix this by using a random directory but it still
> fails [2]:
> > >> >
> > >> > DEBUG: command: collect
> > >> > DEBUG: Traceback (most recent call last):
> > >> > DEBUG:   File
> "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
> > >> > line 2067, in 
> > >> > DEBUG: '%s directory is not empty.' % (conf["local_tmp_dir"])
> > 

Re: basic suite fails on test_metrics_and_log_collector

2021-03-16 Thread Yedidyah Bar David
On Tue, Mar 16, 2021 at 1:11 PM Yedidyah Bar David  wrote:
>
> On Tue, Mar 16, 2021 at 12:50 PM Dana Elfassy  wrote:
> >
> > About [5] https://bugzilla.redhat.com/show_bug.cgi?id=1917707, I have found 
> > that when executing the http request of the playbook, sometimes in the 
> > response there are some events that are missing. The way we handle the 
> > events is via while loop, where in each loop we continue from the 
> > lastEventId. Since not all events were listed, the lastEventId (which is 
> > returned as total number of events) is smaller than the actual eventId that 
> > was handled. Thus, in the next iteration the last events are handled again. 
> > It seems that this comes from Ansible Runner, and I'm investigating that 
> > area now
>
> Thanks
>
> > I'm not sure if it's the same issue as you get with the logs, though, as 
> > I'm not familiar with this area in the OSTs. Is the creation of the 2 logs 
> > done by different threads?
>
> I have no idea either.
>
> >
> > Regarding the initial error (' directory is not empty.non-zero return 
> > code') I'm not sure, will have to loop more into it. Martin, do you have 
> > any idea?
>
> I think it's quite obvious that the reason for this is that we run two
> ovirt-log-collector processes in parallel. Why we (sometimes) do this
> is an open question, but unrelated to ovirt-log-collector (or the
> specific error message it emits).

I now did this:

1. Wrote a trivial dummy module [1] to test this

2. Ran it on an already finished basic-suite env with lagofy/run_tc [2].

As you can see, if you pipe [2] through grep -Ev
'do_something_2.*something 2|do_something_1.*something 1', in some
cases do_something_1 actually ran the ansible code of do_something_2.
I currently think it's some kind of race condition in ansible-runner,
causing it to misbehave when ran in parallel, as we do in
test_metrics_and_log_collector. I still didn't try to check its code
and/or add debugging stuff to it to try and find the exact location.

2021-03-16 20:54:18+0200,681 INFO[test_003_01_dummy]
do_something_1: Tue Mar 16 19:54:18 CET 2021
lago-basic-suite-master-engine something 2 (test_003_01_dummy:39)
2021-03-16 20:55:41+0200,155 INFO[test_003_01_dummy]
do_something_1: Tue Mar 16 19:55:40 CET 2021
lago-basic-suite-master-engine something 2 (test_003_01_dummy:39)
2021-03-16 20:56:56+0200,196 INFO[test_003_01_dummy]
do_something_1: Tue Mar 16 19:56:55 CET 2021
lago-basic-suite-master-engine something 2 (test_003_01_dummy:39)
2021-03-16 20:58:19+0200,665 INFO[test_003_01_dummy]
do_something_1: Tue Mar 16 19:58:19 CET 2021
lago-basic-suite-master-engine something 2 (test_003_01_dummy:39)

[1] 
https://github.com/didib/ovirt-system-tests/commit/6a5327976eaf5bc27e352723bf5c17dab4e60fa4

[2] https://gist.github.com/didib/fe1b3bc28b8364cf6976d3d93e6cd9b6

>
> >
> > On Mon, Mar 15, 2021 at 9:59 AM Yedidyah Bar David  wrote:
> >>
> >> On Mon, Mar 15, 2021 at 7:55 AM Yedidyah Bar David  wrote:
> >> >
> >> > Hi all,
> >> >
> >> > This started a few days ago [1] and randomly happens since then:
> >> >
> >> > E   DEBUG: Configuration:
> >> > E   DEBUG: command: collect
> >> > E   DEBUG: Traceback (most recent call last):
> >> > E   DEBUG:   File
> >> > "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
> >> > line 2067, in 
> >> > E   DEBUG: '%s directory is not empty.' % 
> >> > (conf["local_tmp_dir"])
> >> > E   DEBUG: Exception: /dev/shm/log directory is not
> >> > empty.ERROR: /dev/shm/log directory is not empty.non-zero return code
> >> >
> >> > Michal tried to fix this by using a random directory but it still fails 
> >> > [2]:
> >> >
> >> > DEBUG: command: collect
> >> > DEBUG: Traceback (most recent call last):
> >> > DEBUG:   File 
> >> > "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
> >> > line 2067, in 
> >> > DEBUG: '%s directory is not empty.' % (conf["local_tmp_dir"])
> >> > DEBUG: Exception: /dev/shm/kaN7uY directory is not empty.ERROR:
> >> > /dev/shm/kaN7uY directory is not empty.non-zero return code
> >> >
> >> > Since I suppose that the randomness of mktemp is good enough, it must
> >> > be something else. Also, the last successful run before [1] used the
> >> > same OST git commit (same code), so I do not think it's something in
> >> > OST's code.
> >> >
> &g

Re: basic suite fails on test_metrics_and_log_collector

2021-03-16 Thread Yedidyah Bar David
On Tue, Mar 16, 2021 at 12:50 PM Dana Elfassy  wrote:
>
> About [5] https://bugzilla.redhat.com/show_bug.cgi?id=1917707, I have found 
> that when executing the http request of the playbook, sometimes in the 
> response there are some events that are missing. The way we handle the events 
> is via while loop, where in each loop we continue from the lastEventId. Since 
> not all events were listed, the lastEventId (which is returned as total 
> number of events) is smaller than the actual eventId that was handled. Thus, 
> in the next iteration the last events are handled again. It seems that this 
> comes from Ansible Runner, and I'm investigating that area now

Thanks

> I'm not sure if it's the same issue as you get with the logs, though, as I'm 
> not familiar with this area in the OSTs. Is the creation of the 2 logs done 
> by different threads?

I have no idea either.

>
> Regarding the initial error (' directory is not empty.non-zero return code') 
> I'm not sure, will have to loop more into it. Martin, do you have any idea?

I think it's quite obvious that the reason for this is that we run two
ovirt-log-collector processes in parallel. Why we (sometimes) do this
is an open question, but unrelated to ovirt-log-collector (or the
specific error message it emits).

>
> On Mon, Mar 15, 2021 at 9:59 AM Yedidyah Bar David  wrote:
>>
>> On Mon, Mar 15, 2021 at 7:55 AM Yedidyah Bar David  wrote:
>> >
>> > Hi all,
>> >
>> > This started a few days ago [1] and randomly happens since then:
>> >
>> > E   DEBUG: Configuration:
>> > E   DEBUG: command: collect
>> > E   DEBUG: Traceback (most recent call last):
>> > E   DEBUG:   File
>> > "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
>> > line 2067, in 
>> > E   DEBUG: '%s directory is not empty.' % 
>> > (conf["local_tmp_dir"])
>> > E   DEBUG: Exception: /dev/shm/log directory is not
>> > empty.ERROR: /dev/shm/log directory is not empty.non-zero return code
>> >
>> > Michal tried to fix this by using a random directory but it still fails 
>> > [2]:
>> >
>> > DEBUG: command: collect
>> > DEBUG: Traceback (most recent call last):
>> > DEBUG:   File 
>> > "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
>> > line 2067, in 
>> > DEBUG: '%s directory is not empty.' % (conf["local_tmp_dir"])
>> > DEBUG: Exception: /dev/shm/kaN7uY directory is not empty.ERROR:
>> > /dev/shm/kaN7uY directory is not empty.non-zero return code
>> >
>> > Since I suppose that the randomness of mktemp is good enough, it must
>> > be something else. Also, the last successful run before [1] used the
>> > same OST git commit (same code), so I do not think it's something in
>> > OST's code.
>> >
>> > Any idea?
>> >
>> > I think I'll push a patch to create and use the directory right before
>> > calling ovirt-log-collector, which is probably better in other ways.
>>
>> My patch [1] still fails, with a somewhat different error message, but
>> this made me check further, and while I still do not understand, I have
>> this to add:
>>
>> In the failing runs, ovirt-log-collector is called *twice* in parallel. E.g.
>> in [2] (the check-patch of [1]):
>>
>> Mar 15 07:38:59 lago-basic-suite-master-engine platform-python[59099]:
>> ansible-command Invoked with _raw_params=lctmp=$(mktemp -d -p
>> /dev/shm); ovirt-log-collector --verbose --batch --no-hypervisors
>> --local-tmp="${lctmp}" --conf-file=/root/ovirt-log-collector.conf
>> _uses_shell=True warn=True stdin_add_newline=True
>> strip_empty_ends=True argv=None chdir=None executable=None
>> creates=None removes=None stdin=None
>> Mar 15 07:38:59 lago-basic-suite-master-engine platform-python[59124]:
>> ansible-command Invoked with _raw_params=lctmp=$(mktemp -d -p
>> /dev/shm); ovirt-log-collector --verbose --batch --no-hypervisors
>> --local-tmp="${lctmp}" --conf-file=/root/ovirt-log-collector.conf
>> _uses_shell=True warn=True stdin_add_newline=True
>> strip_empty_ends=True argv=None chdir=None executable=None
>> creates=None removes=None stdin=None
>>
>> It also generates two logs, which you can check/compare.
>>
>> It's the same for previous ones, e.g. latest nightly [3][4]:
>>
>> Mar 15 06:23:30 lago-basic-suite-master-engine platform-python[59343]:
>> ansible-command Invoked with _raw_params=ovirt-log-collector --verbose
>> --ba

Re: basic suite fails on test_metrics_and_log_collector

2021-03-15 Thread Yedidyah Bar David
On Mon, Mar 15, 2021 at 7:55 AM Yedidyah Bar David  wrote:
>
> Hi all,
>
> This started a few days ago [1] and randomly happens since then:
>
> E   DEBUG: Configuration:
> E   DEBUG: command: collect
> E   DEBUG: Traceback (most recent call last):
> E   DEBUG:   File
> "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
> line 2067, in 
> E   DEBUG: '%s directory is not empty.' % (conf["local_tmp_dir"])
> E   DEBUG: Exception: /dev/shm/log directory is not
> empty.ERROR: /dev/shm/log directory is not empty.non-zero return code
>
> Michal tried to fix this by using a random directory but it still fails [2]:
>
> DEBUG: command: collect
> DEBUG: Traceback (most recent call last):
> DEBUG:   File 
> "/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
> line 2067, in 
> DEBUG: '%s directory is not empty.' % (conf["local_tmp_dir"])
> DEBUG: Exception: /dev/shm/kaN7uY directory is not empty.ERROR:
> /dev/shm/kaN7uY directory is not empty.non-zero return code
>
> Since I suppose that the randomness of mktemp is good enough, it must
> be something else. Also, the last successful run before [1] used the
> same OST git commit (same code), so I do not think it's something in
> OST's code.
>
> Any idea?
>
> I think I'll push a patch to create and use the directory right before
> calling ovirt-log-collector, which is probably better in other ways.

My patch [1] still fails, with a somewhat different error message, but
this made me check further, and while I still do not understand, I have
this to add:

In the failing runs, ovirt-log-collector is called *twice* in parallel. E.g.
in [2] (the check-patch of [1]):

Mar 15 07:38:59 lago-basic-suite-master-engine platform-python[59099]:
ansible-command Invoked with _raw_params=lctmp=$(mktemp -d -p
/dev/shm); ovirt-log-collector --verbose --batch --no-hypervisors
--local-tmp="${lctmp}" --conf-file=/root/ovirt-log-collector.conf
_uses_shell=True warn=True stdin_add_newline=True
strip_empty_ends=True argv=None chdir=None executable=None
creates=None removes=None stdin=None
Mar 15 07:38:59 lago-basic-suite-master-engine platform-python[59124]:
ansible-command Invoked with _raw_params=lctmp=$(mktemp -d -p
/dev/shm); ovirt-log-collector --verbose --batch --no-hypervisors
--local-tmp="${lctmp}" --conf-file=/root/ovirt-log-collector.conf
_uses_shell=True warn=True stdin_add_newline=True
strip_empty_ends=True argv=None chdir=None executable=None
creates=None removes=None stdin=None

It also generates two logs, which you can check/compare.

It's the same for previous ones, e.g. latest nightly [3][4]:

Mar 15 06:23:30 lago-basic-suite-master-engine platform-python[59343]:
ansible-command Invoked with _raw_params=ovirt-log-collector --verbose
--batch --no-hypervisors --conf-file=/root/ovirt-log-collector.conf
_uses_shell=True warn=True stdin_add_newline=True
strip_empty_ends=True argv=None chdir=None executable=None
creates=None removes=None stdin=None
Mar 15 06:23:30 lago-basic-suite-master-engine setroubleshoot[58889]:
SELinux is preventing /usr/lib/systemd/systemd from unlink access on
the sock_file ansible-ssh-lago-basic-suite-master-host-1-22-root. For
complete SELinux messages run: sealert -l
d03a8655-9430-4fcf-9892-3b4df1939899
Mar 15 06:23:30 lago-basic-suite-master-engine setroubleshoot[58889]:
SELinux is preventing /usr/lib/systemd/systemd from unlink access on
the sock_file ansible-ssh-lago-basic-suite-master-host-1-22-root.#012#012*
 Plugin catchall (100. confidence) suggests
**#012#012If you believe that systemd should
be allowed unlink access on the
ansible-ssh-lago-basic-suite-master-host-1-22-root sock_file by
default.#012Then you should report this as a bug.#012You can generate
a local policy module to allow this access.#012Do#012allow this access
for now by executing:#012# ausearch -c 'systemd' --raw | audit2allow
-M my-systemd#012# semodule -X 300 -i my-systemd.pp#012
Mar 15 06:23:30 lago-basic-suite-master-engine platform-python[59361]:
ansible-command Invoked with _raw_params=ovirt-log-collector --verbose
--batch --no-hypervisors --conf-file=/root/ovirt-log-collector.conf
_uses_shell=True warn=True stdin_add_newline=True
strip_empty_ends=True argv=None chdir=None executable=None
creates=None removes=None stdin=None

Any idea what might have caused this to start happening? Perhaps
a bug in ansible, or ansible-runner? It reminds me of [5].
Adding Dana and Martin.

I think [5] is quite a serious bug, btw, should be a 4.4.5 blocker.

Best regards,

[1] https://gerrit.ovirt.org/c/ovirt-system-tests/+/113875

[2] 
https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/15980/artifact/check-patch.basic_suite_master.el8.x86_64/test_logs/lago-basic-suite-master-engine/var/log/mess

basic suite fails on test_metrics_and_log_collector

2021-03-14 Thread Yedidyah Bar David
Hi all,

This started a few days ago [1] and randomly happens since then:

E   DEBUG: Configuration:
E   DEBUG: command: collect
E   DEBUG: Traceback (most recent call last):
E   DEBUG:   File
"/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
line 2067, in 
E   DEBUG: '%s directory is not empty.' % (conf["local_tmp_dir"])
E   DEBUG: Exception: /dev/shm/log directory is not
empty.ERROR: /dev/shm/log directory is not empty.non-zero return code

Michal tried to fix this by using a random directory but it still fails [2]:

DEBUG: command: collect
DEBUG: Traceback (most recent call last):
DEBUG:   File 
"/usr/lib/python3.6/site-packages/ovirt_log_collector/__main__.py",
line 2067, in 
DEBUG: '%s directory is not empty.' % (conf["local_tmp_dir"])
DEBUG: Exception: /dev/shm/kaN7uY directory is not empty.ERROR:
/dev/shm/kaN7uY directory is not empty.non-zero return code

Since I suppose that the randomness of mktemp is good enough, it must
be something else. Also, the last successful run before [1] used the
same OST git commit (same code), so I do not think it's something in
OST's code.

Any idea?

I think I'll push a patch to create and use the directory right before
calling ovirt-log-collector, which is probably better in other ways.

Best regards,

[1] 
https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/949/

[2] 
https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/959/


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/TVKEQK3IIQUZ4QVFVE2PJA6NAMOFCMJG/


Re: Change in ovirt-system-tests[master]: log_utils: close_children_tasks: Fix for python3

2021-02-16 Thread Yedidyah Bar David
On Tue, Feb 16, 2021 at 2:46 PM Code Review  wrote:
>
> From Jenkins CI :
>
> Jenkins CI has posted comments on this change. ( 
> https://gerrit.ovirt.org/c/ovirt-system-tests/+/113504 )
>
> Change subject: log_utils: close_children_tasks: Fix for python3
> ..
>
>
> Patch Set 1:
>
> Build Failed
>
> https://jenkins.ovirt.org/job/standard-enqueue/32334/ :
> This change was not submitted to any change queues for system testing. You 
> will need to create some 'build-artifacts' jobs if you want changes to be 
> submitted to change queues, take part in the system tests and be deployed to 
> the nightly snapshot repositories. If your project uses STDCI V2 and you have 
> release branches configured, you may disregard this message.
>
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-on-merge/1467/ : 
> FAILURE

This is failing for some time now. Anyone looking at this? Thanks.

>
>
> --
> To view, visit https://gerrit.ovirt.org/c/ovirt-system-tests/+/113504
> To unsubscribe, or for help writing mail filters, visit 
> https://gerrit.ovirt.org/settings
>
> Gerrit-Project: ovirt-system-tests
> Gerrit-Branch: master
> Gerrit-Change-Id: Ia7a069089e4e6d128747a8ced6d157802b878608
> Gerrit-Change-Number: 113504
> Gerrit-PatchSet: 1
> Gerrit-Owner: Yedidyah Bar David 
> Gerrit-Reviewer: Andrej Cernek 
> Gerrit-Reviewer: Anton Marchukov 
> Gerrit-Reviewer: Dafna Ron 
> Gerrit-Reviewer: Dusan Fodor 
> Gerrit-Reviewer: Gal Ben Haim 
> Gerrit-Reviewer: Galit Rosenthal 
> Gerrit-Reviewer: Jenkins CI 
> Gerrit-Reviewer: Marcin Sobczyk 
> Gerrit-Reviewer: Name of user not set #1001916
> Gerrit-Reviewer: Yedidyah Bar David 
> Gerrit-Reviewer: harel braha 
> Gerrit-Comment-Date: Tue, 16 Feb 2021 12:46:14 +
> Gerrit-HasComments: No
> Gerrit-Has-Labels: No
> Gerrit-MessageType: comment
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/CX2KLTVWZJ3QL2XA6G25PLWB66N4CPWZ/


Re: manual job does not use custom repo

2021-02-01 Thread Yedidyah Bar David
On Mon, Feb 1, 2021 at 12:46 PM Marcin Sobczyk  wrote:

>
>
> On 2/1/21 11:41 AM, Marcin Sobczyk wrote:
> > Hi,
> >
> > On 1/31/21 4:22 PM, Yedidyah Bar David wrote:
> >> On Sun, Jan 17, 2021 at 2:42 PM Yedidyah Bar David  >> <mailto:d...@redhat.com>> wrote:
> >>> Hi all,
> >>>
> >>> Ran [1] for verifying [2]. Passed [3] as CUSTOM_REPOS.
> >>>
> >>> lago.log [4] shows that it did add it to internal_repo, but it was not
> >>> actually installed later. Perhaps something changed in local repos
> >>> handling or something like that?
> >> Now happened again, so I spent some more time looking. I tried this
> >> locally (on buri03) and it did work - passing '-s URL' to run_suite.sh
> >> did install/update packages from there. In CI it does not, e.g. [1].
> >>
> >> I think that the problem is around here, inside global_setup.sh:
> >>
> >> *11:04:29* + sudo -nl firewall-cmd
> >>
> >> *11:04:29* ++ sudo -n firewall-cmd --query-service=ovirtlago
> >>
> >> *11:04:30* + [[ ! -n yes ]]
> >>
> >> *11:04:30* + known_hosts
> >>
> >> Meaning, there were no sudo permissions to run firewall-cmd,
> >> so it didn't add 8585, so later (in lago.log):
> >>
> >> + ADDR=192.168.200.1
> >> + curl_res=0
> >> + curl -f http://192.168.200.1:8585/default/el8/repodata/repomd.xml
> >> <http://192.168.200.1:8585/default/el8/repodata/repomd.xml>
> >>% Total% Received % Xferd  Average Speed   TimeTime Time
> >>   Current
> >>   Dload  Upload   Total   SpentLeft
> >>   Speed
> >> \r  0 00 00 0  0  0 --:--:-- --:--:--
> >> --:--:-- 0curl: (7) Failed to connect to 192.168.200.1 port 8585:
> >> Connection refused
> >> + curl_res=7
>

Are you sure this failed because the proxy wasn't running, and not because
it was blocked? How did this work for me on buri03, then?


> >>
> >> Since this seems like an infra issue, moving to infra list.
> >>
> >> Can someone have a look there? Thanks!
> > The topic of adding and using extra repos in OST is complicated...
> BTW we have an RFE for this that Harel is working on:
>
> https://issues.redhat.com/browse/RHV-40844


Thanks :-). Commented there.


>
>
> > Since HE suite is half-migrated to pytest and still depends on lago OST
> > plugin, we need
> > to add some legacy compatibility layer to 'conftest.py' like the one
> > that still exists
> > in basic suite:
> >
> >
> https://github.com/oVirt/ovirt-system-tests/blob/ebab2fc2c1c10a36bd835f4a8cdb18bd71903cf8/basic-suite-master/test-scenarios/conftest.py#L31
> >
> > Please try using custom repos with this added ^^^.
>

I already verified manually and merged the patch I wanted to test (to
ovirt-hosted-engine-ha).

I'll try to keep that in mind for the next time.

Best regards,


> >
> > Regards, Marcin
> >
> >> Best regards,
> >>
> >> [1]
> >>
> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7726/
> >> <
> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7726/
> >
> >>
> >>
> >> On Sun, Jan 17, 2021 at 2:42 PM Yedidyah Bar David  >> <mailto:d...@redhat.com>> wrote:
> >>
> >>  Hi all,
> >>
> >>  Ran [1] for verifying [2]. Passed [3] as CUSTOM_REPOS.
> >>
> >>  lago.log [4] shows that it did add it to internal_repo, but it was
> not
> >>  actually installed later. Perhaps something changed in local repos
> >>  handling or something like that?
> >>
> >>  Thanks and best regards,
> >>
> >>  [1]
> >>
> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7681/
> >>  <
> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7681/
> >
> >>
> >>  [2] https://github.com/oVirt/ovirt-ansible-collection/pull/202
> >>  <https://github.com/oVirt/ovirt-ansible-collection/pull/202>
> >>
> >>  [3]
> >>
> https://jenkins.ovirt.org/job/oVirt_ovirt-ansible-collection_standard-check-pr/468/artifact/check-patch.el8.x86_64/
> >>  <
> https://jenkins.ovirt.org/job/oVirt_ovirt-ansible-collection_standard-check-pr/468/artifact/check-patch.el8.x86_64/
> >
> >&g

Re: manual job does not use custom repo

2021-01-31 Thread Yedidyah Bar David
On Sun, Jan 17, 2021 at 2:42 PM Yedidyah Bar David  wrote:
>
> Hi all,
>
> Ran [1] for verifying [2]. Passed [3] as CUSTOM_REPOS.
>
> lago.log [4] shows that it did add it to internal_repo, but it was not
> actually installed later. Perhaps something changed in local repos
> handling or something like that?

Now happened again, so I spent some more time looking. I tried this
locally (on buri03) and it did work - passing '-s URL' to run_suite.sh
did install/update packages from there. In CI it does not, e.g. [1].

I think that the problem is around here, inside global_setup.sh:

*11:04:29* + sudo -nl firewall-cmd
*11:04:29* ++ sudo -n firewall-cmd --query-service=ovirtlago

*11:04:30* + [[ ! -n yes ]]
*11:04:30* + known_hosts


Meaning, there were no sudo permissions to run firewall-cmd,
so it didn't add 8585, so later (in lago.log):

+ ADDR=192.168.200.1
+ curl_res=0
+ curl -f http://192.168.200.1:8585/default/el8/repodata/repomd.xml
  % Total% Received % Xferd  Average Speed   TimeTime Time
 Current
 Dload  Upload   Total   SpentLeft
 Speed
\r  0 00 00 0  0  0 --:--:-- --:--:-- --:--:--
0curl: (7) Failed to connect to 192.168.200.1 port 8585: Connection
refused
+ curl_res=7

Since this seems like an infra issue, moving to infra list.

Can someone have a look there? Thanks!

Best regards,

[1]
https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7726/


On Sun, Jan 17, 2021 at 2:42 PM Yedidyah Bar David  wrote:

> Hi all,
>
> Ran [1] for verifying [2]. Passed [3] as CUSTOM_REPOS.
>
> lago.log [4] shows that it did add it to internal_repo, but it was not
> actually installed later. Perhaps something changed in local repos
> handling or something like that?
>
> Thanks and best regards,
>
> [1]
> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7681/
>
> [2] https://github.com/oVirt/ovirt-ansible-collection/pull/202
>
> [3]
> https://jenkins.ovirt.org/job/oVirt_ovirt-ansible-collection_standard-check-pr/468/artifact/check-patch.el8.x86_64/
>
> [4]
> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7681/artifact/exported-artifacts/test_logs/he-basic-suite-master/post-012_local_maintenance_sdk_pytest.py/lago_logs/lago.log
> --
> Didi
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/UJUBSSYXBFHAZUPDDJZQVOAG7CCOQX6L/


Running basic-suite-master more often

2021-01-20 Thread Yedidyah Bar David
Hi all,

We currently run [1], despite its name, 3 times at night - at 01:00,
03:00, 05:00 (UTC+00, I think). What do you think about running it
also during the day? This should help identify breakages caused by
merging stuff during the day (most developers, I think, are in EMEA).

I know we worked on gating in the past and gave up on this. I do not
suggest to start that discussion again, nor to block anything any
other way. Just run a bit more. Even once an hour will not add too
much load IMO and can be useful.

I also do not suggest to run per-patch (say, for the engine git repo)
- this sometimes causes too much load, and sometimes is not enough -
if a patch is not always failing a run, running more than once does
add value.

Best regards,

[1] https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/CF5Y67DYCCHMOS7FDEIWY2JJGJFIIRHG/


[JIRA] (OVIRT-3076) mock_runner fails on automation/basic_suite_master.sh

2020-12-23 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3076:
-

 Summary: mock_runner fails on automation/basic_suite_master.sh
 Key: OVIRT-3076
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3076
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


Hi all,

See e.g. [1][2][3]:

22:29:59 ERROR: Command failed:
22:29:59  # /bin/mount -n --bind /var/run/docker.sock
/var/lib/mock/epel-7-x86_64-243a67eeccf40e93705fec68eb3d7680-1846418/root/var/run/docker.sock
22:29:59
22:29:59 mv: missing destination file operand after
'/tmp/mock_logs.Z5WlIgD3/script'
22:29:59 Try 'mv --help' for more information.
22:29:59 rmdir: failed to remove './mock_logs.Wo5BsLbN': No such file
or directory
22:29:59 @@
22:29:59 @@ Wed Dec 23 20:29:59 UTC 2020
automation/basic_suite_master.sh chroot finished

I am not sure it's a bug in mock_runner, or perhaps in above
(basic_suite_master.sh), or elsewhere - it's a bit hard to understand
from the logs.

Thanks and best regards,

[1] 
https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7621/

[2] 
https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7622/

[3] 
https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7623/

-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100153)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/LGK2YKQJFTGUONZPDAFUA355WOP7JAIO/


mock_runner fails on automation/basic_suite_master.sh

2020-12-23 Thread Yedidyah Bar David
Hi all,

See e.g. [1][2][3]:

22:29:59 ERROR: Command failed:
22:29:59  # /bin/mount -n --bind /var/run/docker.sock
/var/lib/mock/epel-7-x86_64-243a67eeccf40e93705fec68eb3d7680-1846418/root/var/run/docker.sock
22:29:59
22:29:59 mv: missing destination file operand after
'/tmp/mock_logs.Z5WlIgD3/script'
22:29:59 Try 'mv --help' for more information.
22:29:59 rmdir: failed to remove './mock_logs.Wo5BsLbN': No such file
or directory
22:29:59 @@
22:29:59 @@ Wed Dec 23 20:29:59 UTC 2020
automation/basic_suite_master.sh chroot finished

I am not sure it's a bug in mock_runner, or perhaps in above
(basic_suite_master.sh), or elsewhere - it's a bit hard to understand
from the logs.

Thanks and best regards,

[1] 
https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7621/

[2] 
https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7622/

[3] 
https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7623/

-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/JPRJBJKH5S4AS4QF4PMGPRL5J5KZ4HS3/


[JIRA] (OVIRT-3075) link added to bugzilla from gerrit without bug-url

2020-12-20 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3075:
-

 Summary: link added to bugzilla from gerrit without bug-url
 Key: OVIRT-3075
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3075
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


Hi all,

Please see [1][2]. Something added in [1] a link to [2], presumably
because [1] was mentioned in [2], but without a "Bug-Url:".

This is basically a duplicate of [3], fixed at the time by [4]. [4] is
almost two years old, and is still the branch HEAD - nothing was
changed there since. No idea what else might have caused this.

If the code handling this is now maintained elsewhere, perhaps also
add a README and/or some commit pointing at the new place.

Thanks and best regards,

[1] https://bugzilla.redhat.com/show_bug.cgi?id=1846294
[2] https://gerrit.ovirt.org/c/ovirt-dwh/+/112745
[3] https://ovirt-jira.atlassian.net/browse/OVIRT-2669
[4] https://gerrit.ovirt.org/c/gerrit-admin/+/97605
-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100153)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/GW2OXWJYQ6KZKRGLM4IGFSOHSDIGTEOT/


link added to bugzilla from gerrit without bug-url

2020-12-20 Thread Yedidyah Bar David
Hi all,

Please see [1][2]. Something added in [1] a link to [2], presumably
because [1] was mentioned in [2], but without a "Bug-Url:".

This is basically a duplicate of [3], fixed at the time by [4]. [4] is
almost two years old, and is still the branch HEAD - nothing was
changed there since. No idea what else might have caused this.

If the code handling this is now maintained elsewhere, perhaps also
add a README and/or some commit pointing at the new place.

Thanks and best regards,

[1] https://bugzilla.redhat.com/show_bug.cgi?id=1846294
[2] https://gerrit.ovirt.org/c/ovirt-dwh/+/112745
[3] https://ovirt-jira.atlassian.net/browse/OVIRT-2669
[4] https://gerrit.ovirt.org/c/gerrit-admin/+/97605
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/XUZBZXSY3EUKHT6DLYFU3R7I4RL537V4/


[JIRA] (OVIRT-3074) Long filenames chopped in directory index

2020-12-16 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3074:
-

 Summary: Long filenames chopped in directory index
 Key: OVIRT-3074
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3074
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


Hi all,

Can you please configure stuff so that [1] will show full file names?
If that's apache httpd, should be doable by adding to conf (e.g.
.htaccess or somewhere in /etc/httpd):

IndexOptions NameWidth=*

Thanks and best regards,

[1] https://templates.ovirt.org/yum/
-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100153)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/UWHKXYZQRGXJWODVSZBL7OQZ6D7QXUYP/


Long filenames chopped in directory index

2020-12-16 Thread Yedidyah Bar David
Hi all,

Can you please configure stuff so that [1] will show full file names?
If that's apache httpd, should be doable by adding to conf (e.g.
.htaccess or somewhere in /etc/httpd):

IndexOptions NameWidth=*

Thanks and best regards,

[1] https://templates.ovirt.org/yum/
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/EWKDXM7GHGOP2WNTLNYGHQBFHSS6SWNM/


[JIRA] (OVIRT-3073) Re: Change in ovirt-engine[master]: packaging: setup: Move engine-specific pg conf to common

2020-12-14 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3073:
-

 Summary: Re: Change in ovirt-engine[master]: packaging: setup: 
Move engine-specific pg conf to common
 Key: OVIRT-3073
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3073
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


On Mon, Dec 14, 2020 at 12:21 PM Code Review  wrote:
>
> From Jenkins CI :
>
> Jenkins CI has posted comments on this change. ( 
> https://gerrit.ovirt.org/c/ovirt-engine/+/112414 )
>
> Change subject: packaging: setup: Move engine-specific pg conf to common
> ..
>
>
> Patch Set 4: Continuous-Integration-1
>
> Build Failed
>
> https://jenkins.ovirt.org/job/ovirt-engine_standard-check-patch/9236/ : 
> FAILURE

12:21:32  + 
usrc=/home/jenkins/workspace/ovirt-engine_standard-check-patch/jenkins/stdci_tools/usrc.py
12:21:32  + [[ -x
/home/jenkins/workspace/ovirt-engine_standard-check-patch/jenkins/stdci_tools/usrc.py
]]
12:21:32  + 
/home/jenkins/workspace/ovirt-engine_standard-check-patch/jenkins/stdci_tools/usrc.py
--log -d get
12:21:32  /usr/bin/env: ‘python’: No such file or directory

Any idea?

This is a result of manual 'ci test'. Automatic check-patch on the
same patch did succeed.

Thanks,

>
>
> --
> To view, visit https://gerrit.ovirt.org/c/ovirt-engine/+/112414
> To unsubscribe, or for help writing mail filters, visit 
> https://gerrit.ovirt.org/settings
>
> Gerrit-Project: ovirt-engine
> Gerrit-Branch: master
> Gerrit-Change-Id: I8af2e259389fef59251a69c6443e3e440b93a843
> Gerrit-Change-Number: 112414
> Gerrit-PatchSet: 4
> Gerrit-Owner: Yedidyah Bar David 
> Gerrit-Reviewer: Asaf Rachmani 
> Gerrit-Reviewer: Evgeny Slutsky 
> Gerrit-Reviewer: Gal Zaidman 
> Gerrit-Reviewer: Jenkins CI 
> Gerrit-Reviewer: Lev Veyde 
> Gerrit-Reviewer: Sandro Bonazzola 
> Gerrit-Reviewer: Yedidyah Bar David 
> Gerrit-Reviewer: gerrit-hooks 
> Gerrit-Comment-Date: Mon, 14 Dec 2020 10:21:34 +
> Gerrit-HasComments: No
> Gerrit-Has-Labels: Yes
> Gerrit-MessageType: comment
>


-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100152)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/YD7YCTBUIAT65ZE62CTV5764WXM7JBJ4/


Re: Change in ovirt-engine[master]: packaging: setup: Move engine-specific pg conf to common

2020-12-14 Thread Yedidyah Bar David
On Mon, Dec 14, 2020 at 12:21 PM Code Review  wrote:
>
> From Jenkins CI :
>
> Jenkins CI has posted comments on this change. ( 
> https://gerrit.ovirt.org/c/ovirt-engine/+/112414 )
>
> Change subject: packaging: setup: Move engine-specific pg conf to common
> ..
>
>
> Patch Set 4: Continuous-Integration-1
>
> Build Failed
>
> https://jenkins.ovirt.org/job/ovirt-engine_standard-check-patch/9236/ : 
> FAILURE

12:21:32  + 
usrc=/home/jenkins/workspace/ovirt-engine_standard-check-patch/jenkins/stdci_tools/usrc.py
12:21:32  + [[ -x
/home/jenkins/workspace/ovirt-engine_standard-check-patch/jenkins/stdci_tools/usrc.py
]]
12:21:32  + 
/home/jenkins/workspace/ovirt-engine_standard-check-patch/jenkins/stdci_tools/usrc.py
--log -d get
12:21:32  /usr/bin/env: ‘python’: No such file or directory

Any idea?

This is a result of manual 'ci test'. Automatic check-patch on the
same patch did succeed.

Thanks,

>
>
> --
> To view, visit https://gerrit.ovirt.org/c/ovirt-engine/+/112414
> To unsubscribe, or for help writing mail filters, visit 
> https://gerrit.ovirt.org/settings
>
> Gerrit-Project: ovirt-engine
> Gerrit-Branch: master
> Gerrit-Change-Id: I8af2e259389fef59251a69c6443e3e440b93a843
> Gerrit-Change-Number: 112414
> Gerrit-PatchSet: 4
> Gerrit-Owner: Yedidyah Bar David 
> Gerrit-Reviewer: Asaf Rachmani 
> Gerrit-Reviewer: Evgeny Slutsky 
> Gerrit-Reviewer: Gal Zaidman 
> Gerrit-Reviewer: Jenkins CI 
> Gerrit-Reviewer: Lev Veyde 
> Gerrit-Reviewer: Sandro Bonazzola 
> Gerrit-Reviewer: Yedidyah Bar David 
> Gerrit-Reviewer: gerrit-hooks 
> Gerrit-Comment-Date: Mon, 14 Dec 2020 10:21:34 +
> Gerrit-HasComments: No
> Gerrit-Has-Labels: Yes
> Gerrit-MessageType: comment
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/MZOBMYE4X7FDB6SQIFAXBO3ONX2JPPEU/


[JIRA] (OVIRT-3071) CI OST: hardware acceleration not available, kvm executable not found

2020-12-08 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3071:
-

 Summary: CI OST: hardware acceleration not available, kvm 
executable not found
 Key: OVIRT-3071
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3071
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


Hi all,

Recent runs fail with $subject, e.g.:

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14269/console

08:40:07 Package
qemu-kvm-15:4.2.0-34.module_el8.3.0+555+a55c8938.x86_64 is already
installed.

08:45:03  hardware acceleration not available
08:45:03  emulator not found b'\n\n  \n
4c4c4544-0038-4410-8039-b7c04f333832\n\n
x86_64\n  Haswell-noTSX-IBRS\n
Intel\n  \n
\n
\n  \n  \n  \n  \n  \n  \n  \n
 \n  \n
\n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n
 \n\n  \n
\n  \n\n
\n\n  \n
\ntcp\n
rdma\n  \n
\n\n  \n
\n  115148964\n
   28787241\n  0\n  0\n  \n\n  \n  \n\n\n\n
\n
 \n\n\n\n
\n
\n\n\n\n
\n
  \n\n  \n\n
  \n\n\n  \n\n
\n  selinux\n  0\n
system_u:system_r:svirt_t:s0\n
system_u:system_r:svirt_tcg_t:s0\n
  \n\n  dac\n
0\n  +107:+107\n
+107:+107\n\n
\n\n'

As you can see, above does not include any '' sections.
Searching for relevant keywords I can find:

https://bugzilla.redhat.com/show_bug.cgi?id=1066630

So perhaps some dependency of qemu is not installed, or not updated?

Thanks and best regards,
-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100152)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/WOYSG3T5N3NKZO4GJ26XM5N2BFYG3HPK/


CI OST: hardware acceleration not available, kvm executable not found

2020-12-08 Thread Yedidyah Bar David
Hi all,

Recent runs fail with $subject, e.g.:

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14269/console

08:40:07 Package
qemu-kvm-15:4.2.0-34.module_el8.3.0+555+a55c8938.x86_64 is already
installed.

08:45:03  hardware acceleration not available
08:45:03  emulator not found b'\n\n  \n
4c4c4544-0038-4410-8039-b7c04f333832\n\n
x86_64\n  Haswell-noTSX-IBRS\n
Intel\n  \n
\n
\n  \n  \n  \n  \n  \n  \n  \n
 \n  \n
\n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n  \n
 \n\n  \n
\n  \n\n
\n\n  \n
\ntcp\n
rdma\n  \n
\n\n  \n
\n  115148964\n
   28787241\n  0\n  0\n  \n\n  \n  \n\n\n\n
\n
 \n\n\n\n
\n
\n\n\n\n
\n
  \n\n  \n\n
  \n\n\n  \n\n
\n  selinux\n  0\n
system_u:system_r:svirt_t:s0\n
system_u:system_r:svirt_tcg_t:s0\n
  \n\n  dac\n
0\n  +107:+107\n
+107:+107\n\n
\n\n'

As you can see, above does not include any '' sections.
Searching for relevant keywords I can find:

https://bugzilla.redhat.com/show_bug.cgi?id=1066630

So perhaps some dependency of qemu is not installed, or not updated?

Thanks and best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/6YVGXLKJJEICRQYM4NB7XNL4O7X3PK2O/


[JIRA] (OVIRT-3063) gtar Cannot write: Input/output error (was: Change in ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)

2020-11-30 Thread Yedidyah Bar David (oVirt JIRA)

[ 
https://ovirt-jira.atlassian.net/browse/OVIRT-3063?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=40957#comment-40957
 ] 

Yedidyah Bar David commented on OVIRT-3063:
---

On Sun, Nov 29, 2020 at 2:39 PM Shlomi Zidmi (oVirt JIRA) <
j...@ovirt-jira.atlassian.net> wrote:

>
> [
> https://ovirt-jira.atlassian.net/browse/OVIRT-3063?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=40955#comment-40955
> ]
>
> Shlomi Zidmi commented on OVIRT-3063:
> -
>
> Looks like recent builds ran without this error. Taking ovirt-srv22 back
> online to see how it behaves
>

Now it failed again, on ovirt-srv22:

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14124/

I am pretty certain that it didn't always fail, also after it started
failing - it was sporadic.


>
> > gtar Cannot write: Input/output error (was: Change in
> ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)
> >
> --
> >
> > Key: OVIRT-3063
> > URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3063
> > Project: oVirt - virtualization made easy
> >  Issue Type: By-EMAIL
> >Reporter: Yedidyah Bar David
> >Assignee: Shlomi Zidmi
> >
> > Hi all,
> > On Tue, Nov 24, 2020 at 8:54 PM Code Review  wrote:
> > >
> > > From Jenkins CI :
> > >
> > > Jenkins CI has posted comments on this change.
> > >
> > > Change subject: pytest: he: Port 008_restart_he_vm.py to pytest
> > > ..
> > >
> > >
> > > Patch Set 12: Continuous-Integration-1
> > >
> > > Build Failed
> > >
> > >
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14018/
> : FAILURE
> >
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14018//artifact/check-patch.he-basic_suite_master.el8.x86_64/mock_logs/script/stdout_stderr.log
> > [ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Extract appliance to
> > local VM directory]
> > [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "dest":
> > "/var/tmp/localvm3i6cjs21", "extract_results": {"cmd":
> > ["/usr/bin/gtar", "--extract", "-C", "/var/tmp/localvm3i6cjs21", "-z",
> > "--show-transformed-names", "--sparse", "-f",
> >
> "/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova"],
> > "err": "/usr/bin/gtar:
> >
> images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
> > Cannot write: Input/output error\n/usr/bin/gtar:
> >
> images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
> > Cannot utime: Input/output error\n/usr/bin/gtar:
> >
> images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
> > Cannot change ownership to uid 0, gid 0: Input/output
> > error\n/usr/bin/gtar:
> >
> images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
> > Cannot change mode to rwxr-xr-x: Input/output error\n/usr/bin/gtar:
> >
> images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d.meta:
> > Cannot open: Input/output error\n/usr/bin/gtar: Exiting with failure
> > status due to previous errors\n", "out": "", "rc": 2}, "handler":
> > "TgzArchive", "msg": "failed to unpack
> >
> /usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova
> > to /var/tmp/localvm3i6cjs21", "src":
> >
> "/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova"}
> > Other such failures already happened several times recently. Perhaps
> > some disk-space issue? Or something similar?
> > Thanks,
> > >
> > >
> > > --
> > > To view, visit https://gerrit.ovirt.org/112273
> > > To unsubscribe, visit https://gerrit.ovirt.org/settings
> > >
> > > Gerrit-Project: ovirt-system-tests
> > > Gerrit-Branch: master
> > > Gerrit-MessageType: comment
> > > Gerrit-Change-Id: Ib510a1624ac5baad0f637a96919c5fd6040e

Re: [JIRA] (OVIRT-3063) gtar Cannot write: Input/output error (was: Change in ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)

2020-11-30 Thread Yedidyah Bar David
On Sun, Nov 29, 2020 at 2:39 PM Shlomi Zidmi (oVirt JIRA) <
j...@ovirt-jira.atlassian.net> wrote:

>
> [
> https://ovirt-jira.atlassian.net/browse/OVIRT-3063?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=40955#comment-40955
> ]
>
> Shlomi Zidmi commented on OVIRT-3063:
> -
>
> Looks like recent builds ran without this error. Taking ovirt-srv22 back
> online to see how it behaves
>

Now it failed again, on ovirt-srv22:

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14124/

I am pretty certain that it didn't always fail, also after it started
failing - it was sporadic.


>
> > gtar Cannot write: Input/output error (was: Change in
> ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)
> >
> --
> >
> > Key: OVIRT-3063
> > URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3063
> > Project: oVirt - virtualization made easy
> >  Issue Type: By-EMAIL
> >Reporter: Yedidyah Bar David
> >Assignee: Shlomi Zidmi
> >
> > Hi all,
> > On Tue, Nov 24, 2020 at 8:54 PM Code Review  wrote:
> > >
> > > From Jenkins CI :
> > >
> > > Jenkins CI has posted comments on this change.
> > >
> > > Change subject: pytest: he: Port 008_restart_he_vm.py to pytest
> > > ..
> > >
> > >
> > > Patch Set 12: Continuous-Integration-1
> > >
> > > Build Failed
> > >
> > >
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14018/
> : FAILURE
> >
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14018//artifact/check-patch.he-basic_suite_master.el8.x86_64/mock_logs/script/stdout_stderr.log
> > [ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Extract appliance to
> > local VM directory]
> > [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "dest":
> > "/var/tmp/localvm3i6cjs21", "extract_results": {"cmd":
> > ["/usr/bin/gtar", "--extract", "-C", "/var/tmp/localvm3i6cjs21", "-z",
> > "--show-transformed-names", "--sparse", "-f",
> >
> "/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova"],
> > "err": "/usr/bin/gtar:
> >
> images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
> > Cannot write: Input/output error\n/usr/bin/gtar:
> >
> images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
> > Cannot utime: Input/output error\n/usr/bin/gtar:
> >
> images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
> > Cannot change ownership to uid 0, gid 0: Input/output
> > error\n/usr/bin/gtar:
> >
> images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
> > Cannot change mode to rwxr-xr-x: Input/output error\n/usr/bin/gtar:
> >
> images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d.meta:
> > Cannot open: Input/output error\n/usr/bin/gtar: Exiting with failure
> > status due to previous errors\n", "out": "", "rc": 2}, "handler":
> > "TgzArchive", "msg": "failed to unpack
> >
> /usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova
> > to /var/tmp/localvm3i6cjs21", "src":
> >
> "/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova"}
> > Other such failures already happened several times recently. Perhaps
> > some disk-space issue? Or something similar?
> > Thanks,
> > >
> > >
> > > --
> > > To view, visit https://gerrit.ovirt.org/112273
> > > To unsubscribe, visit https://gerrit.ovirt.org/settings
> > >
> > > Gerrit-Project: ovirt-system-tests
> > > Gerrit-Branch: master
> > > Gerrit-MessageType: comment
> > > Gerrit-Change-Id: Ib510a1624ac5baad0f637a96919c5fd6040e89aa
> > > Gerrit-Change-Number: 112273
> > > Gerrit-PatchSet: 12
> > > Gerrit-Owner: Yedidyah Bar David 
> > > Gerrit-Reviewer: Anonymous Coward (1001916)
> > > Gerrit-Reviewer: Anton Marchukov

[JIRA] (OVIRT-3063) gtar Cannot write: Input/output error (was: Change in ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)

2020-11-25 Thread Yedidyah Bar David (oVirt JIRA)

[ 
https://ovirt-jira.atlassian.net/browse/OVIRT-3063?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=40944#comment-40944
 ] 

Yedidyah Bar David commented on OVIRT-3063:
---

On Wed, Nov 25, 2020 at 3:00 PM Shlomi Zidmi (oVirt JIRA) <
j...@ovirt-jira.atlassian.net> wrote:

> Shlomi Zidmi
> <https://ovirt-jira.atlassian.net/secure/ViewProfile.jspa?accountId=5c977456c430371a3c67dbf6>
> *commented* on [image: By-EMAIL] OVIRT-3063
> <https://ovirt-jira.atlassian.net/browse/OVIRT-3063?atlOrigin=eyJpIjoiMmNlZjRiNTAzMTdjNDM3MTkzZTQwOGE3N2QzOTU1NWQiLCJwIjoiaiJ9>
>
> Re: gtar Cannot write: Input/output error (was: Change in
> ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)
> <https://ovirt-jira.atlassian.net/browse/OVIRT-3063?atlOrigin=eyJpIjoiMmNlZjRiNTAzMTdjNDM3MTkzZTQwOGE3N2QzOTU1NWQiLCJwIjoiaiJ9>
>
> I’m still reviewing this issue. I don’t think that’s a disk space issue
> since only 15% of the disk is being used. Also no info is returned from
> dmesg regarding any errrors/failures.
>
> For now I’ve disabled the node (ovirt-srv22) on Jenkins until we figure
> this out
>

This happens also on other nodes, e.g.:

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14022/

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14022/consoleText

[2020-11-24T12:41:02.482Z] Running on node:
openshift-integ-tests-container-sp0sc (integ-tests-container el7)

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14022//artifact/check-patch.he-basic_suite_master.el8.x86_64/mock_logs/script/stdout_stderr.log

[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Extract appliance to
local VM directory]
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "dest":
"/var/tmp/localvmwdymnd_d", "extract_results": {"cmd":
["/usr/bin/gtar", "--extract", "-C", "/var/tmp/localvmwdymnd_d", "-z",
"--show-transformed-names", "--sparse", "-f",
"/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201123175824.1.el8.ova"],
"err": "/usr/bin/gtar:
images/81ceaf89-2550-4724-9bdf-56c520fa14c0/e3728613-7016-4a4e-b95f-68145cd3c028:
Cannot write: Input/output error\n/usr/bin/gtar:
images/81ceaf89-2550-4724-9bdf-56c520fa14c0/e3728613-7016-4a4e-b95f-68145cd3c028:
Cannot utime: Input/output error\n/usr/bin/gtar:
images/81ceaf89-2550-4724-9bdf-56c520fa14c0/e3728613-7016-4a4e-b95f-68145cd3c028:
Cannot change ownership to uid 0, gid 0: Input/output
error\n/usr/bin/gtar:
images/81ceaf89-2550-4724-9bdf-56c520fa14c0/e3728613-7016-4a4e-b95f-68145cd3c028:
Cannot change mode to rwxr-xr-x: Input/output error\n/usr/bin/gtar:
images/81ceaf89-2550-4724-9bdf-56c520fa14c0/e3728613-7016-4a4e-b95f-68145cd3c028.meta:
Cannot open: Input/output error\n/usr/bin/gtar: Exiting with failure
status due to previous errors\n", "out": "", "rc": 2}, "handler":
"TgzArchive", "msg": "failed to unpack
/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201123175824.1.el8.ova
to /var/tmp/localvmwdymnd_d", "src":
"/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201123175824.1.el8.ova"}



> [image: Add Comment]
> <https://ovirt-jira.atlassian.net/browse/OVIRT-3063#add-comment?atlOrigin=eyJpIjoiMmNlZjRiNTAzMTdjNDM3MTkzZTQwOGE3N2QzOTU1NWQiLCJwIjoiaiJ9>
>  Add
> Comment
> <https://ovirt-jira.atlassian.net/browse/OVIRT-3063#add-comment?atlOrigin=eyJpIjoiMmNlZjRiNTAzMTdjNDM3MTkzZTQwOGE3N2QzOTU1NWQiLCJwIjoiaiJ9>
>
> Get Jira notifications on your phone! Download the Jira Cloud app for
> Android
> <https://play.google.com/store/apps/details?id=com.atlassian.android.jira.core=utm_source%3DNotificationLink%26utm_medium%3DEmail>
> or iOS
> <https://itunes.apple.com/app/apple-store/id1006972087?pt=696495=EmailNotificationLink=8>
> --
> This message was sent by Atlassian Jira (v1001.0.0-SNAPSHOT#100151-
> sha1:c8fcc1e)
> [image: Atlassian logo]
>


-- 
Didi


> gtar Cannot write: Input/output error (was: Change in 
> ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)
> --
>
> Key: OVIRT-3063
> URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3063
> Project: oVirt - virtualization made easy
>  Issue Type: By-EMAIL
>Reporter: Yedidyah Bar David
>Assignee: infra
>
> Hi all,
> On Tue, Nov 24, 2020 at 8:54 PM Code Review  wr

Re: [JIRA] (OVIRT-3063) gtar Cannot write: Input/output error (was: Change in ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)

2020-11-25 Thread Yedidyah Bar David
On Wed, Nov 25, 2020 at 3:00 PM Shlomi Zidmi (oVirt JIRA) <
j...@ovirt-jira.atlassian.net> wrote:

> Shlomi Zidmi
> 
> *commented* on [image: By-EMAIL] OVIRT-3063
> 
>
> Re: gtar Cannot write: Input/output error (was: Change in
> ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)
> 
>
> I’m still reviewing this issue. I don’t think that’s a disk space issue
> since only 15% of the disk is being used. Also no info is returned from
> dmesg regarding any errrors/failures.
>
> For now I’ve disabled the node (ovirt-srv22) on Jenkins until we figure
> this out
>

This happens also on other nodes, e.g.:

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14022/

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14022/consoleText

[2020-11-24T12:41:02.482Z] Running on node:
openshift-integ-tests-container-sp0sc (integ-tests-container el7)

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14022//artifact/check-patch.he-basic_suite_master.el8.x86_64/mock_logs/script/stdout_stderr.log

[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Extract appliance to
local VM directory]
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "dest":
"/var/tmp/localvmwdymnd_d", "extract_results": {"cmd":
["/usr/bin/gtar", "--extract", "-C", "/var/tmp/localvmwdymnd_d", "-z",
"--show-transformed-names", "--sparse", "-f",
"/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201123175824.1.el8.ova"],
"err": "/usr/bin/gtar:
images/81ceaf89-2550-4724-9bdf-56c520fa14c0/e3728613-7016-4a4e-b95f-68145cd3c028:
Cannot write: Input/output error\n/usr/bin/gtar:
images/81ceaf89-2550-4724-9bdf-56c520fa14c0/e3728613-7016-4a4e-b95f-68145cd3c028:
Cannot utime: Input/output error\n/usr/bin/gtar:
images/81ceaf89-2550-4724-9bdf-56c520fa14c0/e3728613-7016-4a4e-b95f-68145cd3c028:
Cannot change ownership to uid 0, gid 0: Input/output
error\n/usr/bin/gtar:
images/81ceaf89-2550-4724-9bdf-56c520fa14c0/e3728613-7016-4a4e-b95f-68145cd3c028:
Cannot change mode to rwxr-xr-x: Input/output error\n/usr/bin/gtar:
images/81ceaf89-2550-4724-9bdf-56c520fa14c0/e3728613-7016-4a4e-b95f-68145cd3c028.meta:
Cannot open: Input/output error\n/usr/bin/gtar: Exiting with failure
status due to previous errors\n", "out": "", "rc": 2}, "handler":
"TgzArchive", "msg": "failed to unpack
/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201123175824.1.el8.ova
to /var/tmp/localvmwdymnd_d", "src":
"/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201123175824.1.el8.ova"}



> [image: Add Comment]
> 
>  Add
> Comment
> 
>
> Get Jira notifications on your phone! Download the Jira Cloud app for
> Android
> 
> or iOS
> 
> --
> This message was sent by Atlassian Jira (v1001.0.0-SNAPSHOT#100151-
> sha1:c8fcc1e)
> [image: Atlassian logo]
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/7IL74SEK6X77NCKLLCPMTAMIJ5YUMIHG/


[JIRA] (OVIRT-3063) gtar Cannot write: Input/output error (was: Change in ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)

2020-11-24 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3063:
-

 Summary: gtar Cannot write: Input/output error (was: Change in 
ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)
 Key: OVIRT-3063
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3063
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


Hi all,

On Tue, Nov 24, 2020 at 8:54 PM Code Review  wrote:
>
> From Jenkins CI :
>
> Jenkins CI has posted comments on this change.
>
> Change subject: pytest: he: Port 008_restart_he_vm.py to pytest
> ..
>
>
> Patch Set 12: Continuous-Integration-1
>
> Build Failed
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14018/ 
> : FAILURE

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14018//artifact/check-patch.he-basic_suite_master.el8.x86_64/mock_logs/script/stdout_stderr.log

[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Extract appliance to
local VM directory]
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "dest":
"/var/tmp/localvm3i6cjs21", "extract_results": {"cmd":
["/usr/bin/gtar", "--extract", "-C", "/var/tmp/localvm3i6cjs21", "-z",
"--show-transformed-names", "--sparse", "-f",
"/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova"],
"err": "/usr/bin/gtar:
images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
Cannot write: Input/output error\n/usr/bin/gtar:
images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
Cannot utime: Input/output error\n/usr/bin/gtar:
images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
Cannot change ownership to uid 0, gid 0: Input/output
error\n/usr/bin/gtar:
images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
Cannot change mode to rwxr-xr-x: Input/output error\n/usr/bin/gtar:
images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d.meta:
Cannot open: Input/output error\n/usr/bin/gtar: Exiting with failure
status due to previous errors\n", "out": "", "rc": 2}, "handler":
"TgzArchive", "msg": "failed to unpack
/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova
to /var/tmp/localvm3i6cjs21", "src":
"/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova"}

Other such failures already happened several times recently. Perhaps
some disk-space issue? Or something similar?

Thanks,

>
>
> --
> To view, visit https://gerrit.ovirt.org/112273
> To unsubscribe, visit https://gerrit.ovirt.org/settings
>
> Gerrit-Project: ovirt-system-tests
> Gerrit-Branch: master
> Gerrit-MessageType: comment
> Gerrit-Change-Id: Ib510a1624ac5baad0f637a96919c5fd6040e89aa
> Gerrit-Change-Number: 112273
> Gerrit-PatchSet: 12
> Gerrit-Owner: Yedidyah Bar David 
> Gerrit-Reviewer: Anonymous Coward (1001916)
> Gerrit-Reviewer: Anton Marchukov 
> Gerrit-Reviewer: Dafna Ron 
> Gerrit-Reviewer: Dusan Fodor 
> Gerrit-Reviewer: Gal Ben Haim 
> Gerrit-Reviewer: Galit Rosenthal 
> Gerrit-Reviewer: Jenkins CI 
> Gerrit-Reviewer: Marcin Sobczyk 
> Gerrit-Reviewer: Yedidyah Bar David 
> Gerrit-Comment-Date: Tue, 24 Nov 2020 18:54:46 +
> Gerrit-HasComments: No
>


-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100151)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/5VIOL5K2BQYH4FUCBYNVGLIMFAX5OE7A/


gtar Cannot write: Input/output error (was: Change in ovirt-system-tests[master]: pytest: he: Port 008_restart_he_vm.py to pytest)

2020-11-24 Thread Yedidyah Bar David
Hi all,

On Tue, Nov 24, 2020 at 8:54 PM Code Review  wrote:
>
> From Jenkins CI :
>
> Jenkins CI has posted comments on this change.
>
> Change subject: pytest: he: Port 008_restart_he_vm.py to pytest
> ..
>
>
> Patch Set 12: Continuous-Integration-1
>
> Build Failed
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14018/ 
> : FAILURE

https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/14018//artifact/check-patch.he-basic_suite_master.el8.x86_64/mock_logs/script/stdout_stderr.log

[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Extract appliance to
local VM directory]
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "dest":
"/var/tmp/localvm3i6cjs21", "extract_results": {"cmd":
["/usr/bin/gtar", "--extract", "-C", "/var/tmp/localvm3i6cjs21", "-z",
"--show-transformed-names", "--sparse", "-f",
"/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova"],
"err": "/usr/bin/gtar:
images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
Cannot write: Input/output error\n/usr/bin/gtar:
images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
Cannot utime: Input/output error\n/usr/bin/gtar:
images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
Cannot change ownership to uid 0, gid 0: Input/output
error\n/usr/bin/gtar:
images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d:
Cannot change mode to rwxr-xr-x: Input/output error\n/usr/bin/gtar:
images/1519176a-e693-4425-b47b-7acfdd7f180b/4904ac49-1535-47d9-9d52-b803df4f869d.meta:
Cannot open: Input/output error\n/usr/bin/gtar: Exiting with failure
status due to previous errors\n", "out": "", "rc": 2}, "handler":
"TgzArchive", "msg": "failed to unpack
/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova
to /var/tmp/localvm3i6cjs21", "src":
"/usr/share/ovirt-engine-appliance/ovirt-engine-appliance-4.4-20201124175128.1.el8.ova"}

Other such failures already happened several times recently. Perhaps
some disk-space issue? Or something similar?

Thanks,

>
>
> --
> To view, visit https://gerrit.ovirt.org/112273
> To unsubscribe, visit https://gerrit.ovirt.org/settings
>
> Gerrit-Project: ovirt-system-tests
> Gerrit-Branch: master
> Gerrit-MessageType: comment
> Gerrit-Change-Id: Ib510a1624ac5baad0f637a96919c5fd6040e89aa
> Gerrit-Change-Number: 112273
> Gerrit-PatchSet: 12
> Gerrit-Owner: Yedidyah Bar David 
> Gerrit-Reviewer: Anonymous Coward (1001916)
> Gerrit-Reviewer: Anton Marchukov 
> Gerrit-Reviewer: Dafna Ron 
> Gerrit-Reviewer: Dusan Fodor 
> Gerrit-Reviewer: Gal Ben Haim 
> Gerrit-Reviewer: Galit Rosenthal 
> Gerrit-Reviewer: Jenkins CI 
> Gerrit-Reviewer: Marcin Sobczyk 
> Gerrit-Reviewer: Yedidyah Bar David 
> Gerrit-Comment-Date: Tue, 24 Nov 2020 18:54:46 +
> Gerrit-HasComments: No
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/QGOSWABQS34JNIXI55UWDTPZXVB3HQQZ/


github: check-patch not triggered on push

2020-11-23 Thread Yedidyah Bar David
Hi,

I recently opened [1], and a few hours ago (force-) pushed to it, and
I do not see [2] triggered. It did work for me in the past, e.g. [3].

Any idea why?

Is it possible to manually run a build of [2]?

I tried pushing "Build Now" a few times, and only after several,
noticed that it triggers a build, but without prompting for
parameters, sorry...

E.g. is it safe to "Rebuild" [3] and guess which params need changing?
Trying  that now anyway [4].

[1] https://github.com/oVirt/ovirt-ansible-collection/pull/181

[2] 
https://jenkins.ovirt.org/job/oVirt_ovirt-ansible-collection_standard-check-pr/

[3] 
https://jenkins.ovirt.org/job/oVirt_ovirt-ansible-collection_standard-check-pr/439/

[4] 
https://jenkins.ovirt.org/job/oVirt_ovirt-ansible-collection_standard-check-pr/446/
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/JJNTE3NG3NMW5ULSCG6XBUK67O3ELIR5/


Re: upgrade-from- master or 4.4 suite

2020-11-17 Thread Yedidyah Bar David
On Tue, Nov 17, 2020 at 9:50 AM Galit Rosenthal  wrote:

>
>
> On Tue, Nov 17, 2020 at 9:33 AM Yedidyah Bar David 
> wrote:
>
>> On Tue, Nov 17, 2020 at 6:55 AM Galit Rosenthal 
>> wrote:
>>
>>> Hi Didi,
>>>
>>> You can revert the patch, but not merge till you fix the suite.
>>>
>>
>> Pushed a revert and changed it a bit, it's still broken:
>>
>> https://gerrit.ovirt.org/112249
>>
>>
>>> This suite was removed because it failed.
>>>
>>
>> Fine, but why do we still have an option to use it from the manual job,
>> if it's broken? NVM... We should better fix it soon rather than remove from
>> the manual job.
>>
>
>
> You still have upgrade from release suite 4.3.
>

Right. Sorry. Thanks. :-)


>
>> Thanks,
>> --
>> Didi
>>
>
>
> --
>
> Galit Rosenthal
>
> SOFTWARE ENGINEER
>
> Red Hat <https://www.redhat.com/>
>
> ga...@redhat.comT: 972-9-7692230
> <https://www.redhat.com/>
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/AL2LS22TFNYKX2KUBVQ6CUWK73K4AZZD/


Re: upgrade-from- master or 4.4 suite

2020-11-16 Thread Yedidyah Bar David
On Tue, Nov 17, 2020 at 6:55 AM Galit Rosenthal  wrote:

> Hi Didi,
>
> You can revert the patch, but not merge till you fix the suite.
>

Pushed a revert and changed it a bit, it's still broken:

https://gerrit.ovirt.org/112249


> This suite was removed because it failed.
>

Fine, but why do we still have an option to use it from the manual job, if
it's broken? NVM... We should better fix it soon rather than remove from
the manual job.

Thanks,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/NPE2GTYWE6EDIDPLYLRKQTFLCWLDG63M/


Re: upgrade-from- master or 4.4 suite

2020-11-16 Thread Yedidyah Bar David
On Mon, Nov 16, 2020 at 2:19 PM Galit Rosenthal  wrote:

> Hi Didi
>
> We have this suite upgrade-from-release-suite-master
>

We do, but it's broken, at least for me - see link from previous mail.


>
> It might require updates.
>
We had to remove automation/upgrade-from-release_suite_master.sh this file
> because of the CQ.
> https://gerrit.ovirt.org/#/c/104988/
>

Is it enough to revert ^^? Something else?

Thanks,


>
>
> Regards,
> Galit
>
> On Mon, Nov 16, 2020 at 1:21 PM Yedidyah Bar David 
> wrote:
>
>> Hi all,
>>
>> I'd like to have a suite that tests upgrade from current master to a
>> build I want to test, and/or from latest released 4.4 to a build I
>> want to test. Do we have anything like that?
>>
>> At some point we removed [1] upgrade from 4.3 to 4.4, because this
>> requires reinstalling the hosts from el7 to el8. Perhaps it's bad that
>> we didn't invest in automating this. But now I want to test 4.4->4.4
>> upgrades. Tried choosing it from manual job [2] and it failed with:
>>
>> 11:51:28 WARN: Unable to find sh file
>> automation/upgrade-from-release_suite_master.sh or
>> automation/upgrade-from-release_suite_master.sh.el7, skipping sh
>>
>> Thanks and best regards,
>>
>> [1] https://gerrit.ovirt.org/104988
>>
>> [2]
>> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7579/
>> --
>> Didi
>>
>>
>
> --
>
> Galit Rosenthal
>
> SOFTWARE ENGINEER
>
> Red Hat <https://www.redhat.com/>
>
> ga...@redhat.comT: 972-9-7692230
> <https://www.redhat.com/>
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/C53IMKWLG63JJMOT5F4VHIN2GM4TTZZS/


upgrade-from- master or 4.4 suite

2020-11-16 Thread Yedidyah Bar David
Hi all,

I'd like to have a suite that tests upgrade from current master to a
build I want to test, and/or from latest released 4.4 to a build I
want to test. Do we have anything like that?

At some point we removed [1] upgrade from 4.3 to 4.4, because this
requires reinstalling the hosts from el7 to el8. Perhaps it's bad that
we didn't invest in automating this. But now I want to test 4.4->4.4
upgrades. Tried choosing it from manual job [2] and it failed with:

11:51:28 WARN: Unable to find sh file
automation/upgrade-from-release_suite_master.sh or
automation/upgrade-from-release_suite_master.sh.el7, skipping sh

Thanks and best regards,

[1] https://gerrit.ovirt.org/104988

[2] 
https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7579/
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/ZOR7UXX5HA4ILZFJ364ZSJSARECPE2WN/


Re: [oVirt Jenkins] ovirt-system-tests_basic-suite-master_nightly - Build # 561 - Failure!

2020-11-08 Thread Yedidyah Bar David
On Sun, Nov 8, 2020 at 6:08 AM  wrote:
>
> Project: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/
> Build: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/561/
> Build Number: 561
> Build Status:  Failure
> Triggered By: Started by timer
>
> -
> Changes Since Last Success:
> -
> Changes for Build #561
> [Marcin Sobczyk] basic: Report start times and runnings time of tests
>
>
>
>
> -
> Failed Tests:
> -
> 1 tests failed.
> FAILED:  basic-suite-master.test-scenarios.004_basic_sanity.test_run_vms
>
> Error Message:
> Error: Fault reason is "Operation Failed". Fault detail is "[Network error 
> during communication with the Host.]". HTTP response code is 400.

06:05:28 ../basic-suite-master/test-scenarios/004_basic_sanity.py::test_run_vms
started at 04:05:28 FAILED (30s)
06:06:00
06:06:00 === FAILURES
===

https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/561/artifact/exported-artifacts/test_logs/basic-suite-master/lago-basic-suite-master-engine/_var_log/ovirt-engine/engine.log

2020-11-07 23:05:57,825-05 ERROR
[org.ovirt.engine.core.bll.RunVmOnceCommand] (default task-2)
[06bc4dc1-f61e-4004-8623-8e91e2e2db24] Command
'org.ovirt.engine.core.bll.RunVmOnceCommand' failed: EngineException:
org.ovirt.engine.core.vdsbroker.vdsbroker.VDSNetworkException:
VDSGenericException: VDSNetworkException: Connection timeout for host
'lago-basic-suite-master-host-1', last response arrived 23158 ms ago.
(Failed with error VDS_NETWORK_ERROR and code 5022)

Logs on host-1 seem ok to me, but finish a bit before this, so perhaps
indeed we lost communication with it (temporarily?):

https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/561/artifact/exported-artifacts/test_logs/basic-suite-master/lago-basic-suite-master-host-1/_var_log/log/vdsm/vdsm.log

Last line is:

2020-11-07 23:05:04,434-0500 DEBUG (jsonrpc/6) [jsonrpc.JsonRpcServer]
Return 'Host.getStats' in bridge with (suppressed) (__init__:360)

https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/561/artifact/exported-artifacts/test_logs/basic-suite-master/lago-basic-suite-master-host-1/_var_log/log/messages/*view*

Last line is:

Nov  7 23:04:10 lago-basic-suite-master-host-1 systemd[1]: Started
Session c177 of user root.

This run (561) isn't the last one - there were a few successful runs
after it, and the following (and last) failed one, 565, is due to
virt-sparsify (see other thread).

But this also happened to me now:

https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7566/

Not copying/pasting again - it looks very similar. engine.log has same
error, vdsm.log looks ok but finishes a bit earlier.

https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7566/artifact/exported-artifacts/test_logs/basic-suite-master/lago_logs/lago.log

2020-11-08 20:51:57,693::log_utils.py::__exit__::611::lago.prefix::INFO::
 # Deploy environment:  [32mSuccess [0m (in 0:14:33)
2020-11-08 20:51:57,957::log_utils.py::__exit__::611::ovirtlago.prefix::INFO::@
Deploy oVirt environment:  [32mSuccess [0m (in 0:14:33)
2020-11-08 21:32:41,201::log_utils.py::__enter__::600::lago.prefix::INFO::@
Collect artifacts:  [0m [0m
2020-11-08 21:32:41,295::log_utils.py::__enter__::600::lago.prefix::INFO::
 # [Thread-1] lago-basic-suite-master-host-0:  [0m [0m
2020-11-08 21:32:41,296::log_utils.py::__enter__::600::lago.prefix::INFO::
 # [Thread-2] lago-basic-suite-master-host-1:  [0m [0m
2020-11-08 21:32:41,297::log_utils.py::__enter__::600::lago.prefix::INFO::
 # [Thread-3] lago-basic-suite-master-engine:  [0m [0m
2020-11-08 
21:32:41,298::vm.py::extract_paths::402::lago.providers.libvirt.vm::DEBUG::lago-basic-suite-master-host-0:
failed extracting files: Unable to extract paths from
lago-basic-suite-master-host-0: unreachable with SSH
2020-11-08 
21:32:41,299::vm.py::extract_paths_dead::437::lago.providers.libvirt.vm::DEBUG::lago-basic-suite-master-host-0:
attempting to extract files with libguestfs
2020-11-08 
21:32:41,507::vm.py::extract_paths::402::lago.providers.libvirt.vm::DEBUG::lago-basic-suite-master-host-1:
failed extracting files: Unable to extract paths from
lago-basic-suite-master-host-1: unreachable with SSH
2020-11-08 
21:32:41,508::vm.py::extract_paths_dead::437::lago.providers.libvirt.vm::DEBUG::lago-basic-suite-master-host-1:
attempting to extract files with libguestfs

So it seems to me like an infra issue - also lago failed to ssh there
(and also to host-0). Can anyone have a look please? Thanks.

Best regards,

>
> Stack Trace:
> assert_vm_is_alive = 
> engine_api = 
> management_gw_ip = '192.168.201.1'
>
> @order_by(_TEST_LIST)
> def test_run_vms(assert_vm_is_alive, engine_api, management_gw_ip):
> engine = 

[JIRA] (OVIRT-3049) ovirt-ansible-collection: Automatic bugzilla linking

2020-10-20 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3049:
-

 Summary: ovirt-ansible-collection: Automatic bugzilla linking
 Key: OVIRT-3049
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3049
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


Hi all,

I now pushed my first PR to ovirt-ansible-collection [1].

It has a Bug-Url link to [2].

The bug wasn't updated with a link to the PR. Should it? Please make
it so that it does.

Also, perhaps, when creating new projects, do this automatically for
them (and ask what bugzilla product should be affected).

Thanks and best regards,

[1] https://github.com/oVirt/ovirt-ansible-collection/pull/151
[2] https://bugzilla.redhat.com/show_bug.cgi?id=1844965
-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100149)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/D6WPX2XP3TDHLFKJNENUWALHIZX6MY65/


ovirt-ansible-collection: Automatic bugzilla linking

2020-10-20 Thread Yedidyah Bar David
Hi all,

I now pushed my first PR to ovirt-ansible-collection [1].

It has a Bug-Url link to [2].

The bug wasn't updated with a link to the PR. Should it? Please make
it so that it does.

Also, perhaps, when creating new projects, do this automatically for
them (and ask what bugzilla product should be affected).

Thanks and best regards,

[1] https://github.com/oVirt/ovirt-ansible-collection/pull/151
[2] https://bugzilla.redhat.com/show_bug.cgi?id=1844965
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/7AC6QLZCH3D5E23L54YKZWMT456LL6BB/


[JIRA] (OVIRT-3042) git fetch fails - Network is unreachable (was: [CQ]: 111471,2 (ovirt-engine) failed "ovirt-master" system tests)

2020-10-15 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3042:
-

 Summary: git fetch fails - Network is unreachable (was: [CQ]: 
111471,2 (ovirt-engine) failed "ovirt-master" system tests)
 Key: OVIRT-3042
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3042
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


On Thu, Oct 15, 2020 at 10:47 AM Radoslaw Szwajkowski
 wrote:
>
> Fixed in https://gerrit.ovirt.org/#/c/111741/

It failed due to infra issue, adding infra-support and changing subject:

https://jenkins.ovirt.org/job/ovirt-engine_standard-check-patch/8460/console

10:42:37  stderr: fatal: unable to access
'https://gerrit.ovirt.org/ovirt-engine/': Failed to connect to
2600:1f18:26e2:::: Network is unreachable

I think we already saw several other similar cases recently. Is anyone
checking this? Is it a network issue? Routine maintenance? Something
else? Perhaps we should patch our scripts to try a few times before
failing.

Thanks,
-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100149)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/776D6WUS2UCW6LWJXBDDCZUCRP7ZV3QP/


[JIRA] (OVIRT-3009) errors on infra

2020-09-10 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-3009:
-

 Summary: errors on infra
 Key: OVIRT-3009
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-3009
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


Hi,

1. Please see thread "[ovirt-devel] gerrit pull/clone timeout".

2. Please see [1] - it failed check-merged, I tried again, and it
failed again. In both cases, e.g. [2]:

Error: Error downloading packages:
  Cannot download x86_64/ovirt-engine-wildfly-19.1.0-2.el8.x86_64.rpm:
All mirrors were tried

Of course, I can't be sure these are due to a single root cause.

Please fix :-)

Thanks,

[1] https://gerrit.ovirt.org/111098

[2] 
https://jenkins.ovirt.org/job/ovirt-dwh_standard-on-merge/124/artifact/build-artifacts.el8.x86_64/mock_logs/script/stdout_stderr.log
-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100145)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/RAEHUSOMCJ777YH73LSHSFNN2ON3VPTW/


Re: [ovirt-users] invalid spf record for ovirt.org

2020-07-22 Thread Yedidyah Bar David
On Wed, Jul 22, 2020 at 3:57 PM Jorick Astrego  wrote:

> Hi,
>
> During routine maintenance on our mailserver I noticed the following in
> the log:
>
> [22/Jul/2020 14:33:33] Error when parsing SPF TXT record for domain:
> ovirt.org, envelope-from=users-boun...@ovirt.org, message: Invalid
> character found near "" in "66.187.233.88"
>
> A check on MXtoolbox also gives an invalid systax error:
>
> v=spf1 a:mail.ovirt.org a:gerrit.ovirt.org 66.187.233.88 ~all
>
> Prefix Type Value PrefixDesc Description Error
>
> v spf1
> The SPF record version
> + a mail.ovirt.org Pass Match if IP has a DNS 'A' record in given domain.
> + a gerrit.ovirt.org Pass Match if IP has a DNS 'A' record in given
> domain.
> + 66.187.233.88
> Pass Unknown Unknown mechanisms are not allowed
> ~ all
> SoftFail Always matches. It goes at the end of your record.
>
>
> Test Result
>
> 
>  SPF
> Syntax Check Invalid syntax found
>
>
Thanks for the report. Adding some more relevant people.

Best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/WQ3VL3CFYC7GIE5VWW37GX77WYLQES2R/


Re: ovirt-engine-wildfly 19.1.0-2 publishing

2020-07-16 Thread Yedidyah Bar David
On Thu, Jul 16, 2020 at 1:40 PM Yedidyah Bar David  wrote:
>
> On Thu, Jul 16, 2020 at 1:14 PM Ehud Yonasi  wrote:
> >
> > Hi Didi,
> > You recall correctly.
> >
> > Right now the process of getting rpms publishing to nightly is as follows:
> >
> > 1) Merge the patch in the project
> > 2) It will end up in CQ, and if the build-artifacts job (check-merged) 
> > finishes successfully, it will be published to tested.
> > 3) There is a nightly run job called ovirt_${RELEASE}_publish-rpms_nightly 
> > that copy tested ${RELEASE} rpms into it.
>
> Thanks for the explanation.
>
> >
> > If the rpm is not in nightly that might be because the check merged job 
> > failed (which it didn't because you clearly see it in tested), then it 
> > probably happened after the nightly job ran, and it will appear tomorrow.
>
> But, if I got it right, the last "publish" run [1] started 5 hours
> after the check-merged [2] (one tells me 10 hours ago, other 15). Does
> it take time (e.g. waiting for resources) between finish of
> check-merged and copying to tested? We saw rather long queues lately
> (you already talked about this in other threads).
>
> Anyway, I'd not call it "urgent", but I'd rather make he-basic-master
> Green again before the weekend... (and this requires also building the
> appliance after you publish).

I ran it myself, and it finished, saying successfully, but has an error -
and it didn't update wildfly (at least):

https://jenkins.ovirt.org/job/ovirt_master_publish-rpms_nightly/1801/consoleText

2020-07-16 11:44:54,374::INFO::repoman.common.stores.RPM::Adding
package 
/srv/resources/repos/ovirt/tested/master/rpm/fc30/noarch/python2-ovirt-engine-lib-4.4.0-0.0.master.20200116192242.git2d9dd76235a.fc30.noarch.rpm
to repo Non persistent RPMStore
2020-07-16 11:44:54,376::ERROR::repoman.cmd::
2020-07-16 11:44:54,376::ERROR::repoman.cmd::Error while adding
/srv/resources/repos/ovirt/tested/master/rpm:latest:only-missing

Can you please check? Thanks.

>
> Thanks,
>
> [1] https://jenkins.ovirt.org/job/ovirt_master_publish-rpms_nightly/1800/
> [2] https://jenkins.ovirt.org/job/ovirt-engine-wildfly_standard-on-merge/33/
>
> >
> > If it's urgent, I can rerun the job. Please let me know.
> >
> > Regards,
> > Ehud.
> >
> > On Thu, Jul 16, 2020 at 12:57 PM Yedidyah Bar David  wrote:
> >>
> >> Hi all,
> >>
> >> 1. Yesterday mperina pushed and merged a patch [1] to hopefully solve
> >> a rather-long-time breakage of, among others, he-basic-suite. I still
> >> do not see it in master-snapshot [2] - only 19.1.0-1. I do see it in
> >> tested [3]. Can someone please publish it to snapshot?
> >>
> >> 2. I tried to debug this myself a bit, but failed to find
> >> documentation for the publishing part. A few months ago it was said
> >> (not very clearly/officially, IIRC) that we do not anymore pass
> >> built-artifacts results through OST, because it's too fragile. Did we
> >> update docs? Where can I find a clear overview of the process from
> >> pressing Merge to having an rpm in master snapshot?
> >>
> >> Thanks and best regards,
> >>
> >> [1] https://gerrit.ovirt.org/110324
> >> [2] https://resources.ovirt.org/pub/ovirt-master-snapshot/rpm/el8/
> >> [3] https://resources.ovirt.org/repos/ovirt/tested/master/rpm/el8/
> >> --
> >> Didi
> >> ___
> >> Infra mailing list -- infra@ovirt.org
> >> To unsubscribe send an email to infra-le...@ovirt.org
> >> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> >> oVirt Code of Conduct: 
> >> https://www.ovirt.org/community/about/community-guidelines/
> >> List Archives: 
> >> https://lists.ovirt.org/archives/list/infra@ovirt.org/message/Q7QAW24L5AYRJVBFYWCTE7UW4DSFTSFN/
>
>
>
> --
> Didi



-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/4GIMHC3XTUGC3DIWI75RQXSUQBO3ANUB/


Missing build-artifacts

2020-07-16 Thread Yedidyah Bar David
Hi all,

Any idea why [1] has no collected artifacts?

I ran it to try to reproduce something similar to [2], using the same
OST patch [3] - both as-is [1], for reproduction, and with a patched
[4] engine [5], for verification. [5] passed, so I am happy with the
verification. Former [1] failed, as expected, and at a reasonable
point to be considered a reproduction (based on console output), but I
can't be sure without some more logs. Going to mark the patch verified
anyway, just wondering - mainly for the next time I'll want to use CI
like that (which I do often).

Thanks and best regards,

[1] https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/10297/
[2] https://bugzilla.redhat.com/show_bug.cgi?id=1847963
[3] https://gerrit.ovirt.org/110302
[4] https://gerrit.ovirt.org/110284
[5] 
https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/7097/
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/WAUWQG3KHOP4URSWVTT265CP6WBVLK4C/


Re: ovirt-engine-wildfly 19.1.0-2 publishing

2020-07-16 Thread Yedidyah Bar David
On Thu, Jul 16, 2020 at 1:14 PM Ehud Yonasi  wrote:
>
> Hi Didi,
> You recall correctly.
>
> Right now the process of getting rpms publishing to nightly is as follows:
>
> 1) Merge the patch in the project
> 2) It will end up in CQ, and if the build-artifacts job (check-merged) 
> finishes successfully, it will be published to tested.
> 3) There is a nightly run job called ovirt_${RELEASE}_publish-rpms_nightly 
> that copy tested ${RELEASE} rpms into it.

Thanks for the explanation.

>
> If the rpm is not in nightly that might be because the check merged job 
> failed (which it didn't because you clearly see it in tested), then it 
> probably happened after the nightly job ran, and it will appear tomorrow.

But, if I got it right, the last "publish" run [1] started 5 hours
after the check-merged [2] (one tells me 10 hours ago, other 15). Does
it take time (e.g. waiting for resources) between finish of
check-merged and copying to tested? We saw rather long queues lately
(you already talked about this in other threads).

Anyway, I'd not call it "urgent", but I'd rather make he-basic-master
Green again before the weekend... (and this requires also building the
appliance after you publish).

Thanks,

[1] https://jenkins.ovirt.org/job/ovirt_master_publish-rpms_nightly/1800/
[2] https://jenkins.ovirt.org/job/ovirt-engine-wildfly_standard-on-merge/33/

>
> If it's urgent, I can rerun the job. Please let me know.
>
> Regards,
> Ehud.
>
> On Thu, Jul 16, 2020 at 12:57 PM Yedidyah Bar David  wrote:
>>
>> Hi all,
>>
>> 1. Yesterday mperina pushed and merged a patch [1] to hopefully solve
>> a rather-long-time breakage of, among others, he-basic-suite. I still
>> do not see it in master-snapshot [2] - only 19.1.0-1. I do see it in
>> tested [3]. Can someone please publish it to snapshot?
>>
>> 2. I tried to debug this myself a bit, but failed to find
>> documentation for the publishing part. A few months ago it was said
>> (not very clearly/officially, IIRC) that we do not anymore pass
>> built-artifacts results through OST, because it's too fragile. Did we
>> update docs? Where can I find a clear overview of the process from
>> pressing Merge to having an rpm in master snapshot?
>>
>> Thanks and best regards,
>>
>> [1] https://gerrit.ovirt.org/110324
>> [2] https://resources.ovirt.org/pub/ovirt-master-snapshot/rpm/el8/
>> [3] https://resources.ovirt.org/repos/ovirt/tested/master/rpm/el8/
>> --
>> Didi
>> ___
>> Infra mailing list -- infra@ovirt.org
>> To unsubscribe send an email to infra-le...@ovirt.org
>> Privacy Statement: https://www.ovirt.org/privacy-policy.html
>> oVirt Code of Conduct: 
>> https://www.ovirt.org/community/about/community-guidelines/
>> List Archives: 
>> https://lists.ovirt.org/archives/list/infra@ovirt.org/message/Q7QAW24L5AYRJVBFYWCTE7UW4DSFTSFN/



-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/WC6YJCCJ7N2DTUFQQ3IJGMDUW5ZTHPBL/


ovirt-engine-wildfly 19.1.0-2 publishing

2020-07-16 Thread Yedidyah Bar David
Hi all,

1. Yesterday mperina pushed and merged a patch [1] to hopefully solve
a rather-long-time breakage of, among others, he-basic-suite. I still
do not see it in master-snapshot [2] - only 19.1.0-1. I do see it in
tested [3]. Can someone please publish it to snapshot?

2. I tried to debug this myself a bit, but failed to find
documentation for the publishing part. A few months ago it was said
(not very clearly/officially, IIRC) that we do not anymore pass
built-artifacts results through OST, because it's too fragile. Did we
update docs? Where can I find a clear overview of the process from
pressing Merge to having an rpm in master snapshot?

Thanks and best regards,

[1] https://gerrit.ovirt.org/110324
[2] https://resources.ovirt.org/pub/ovirt-master-snapshot/rpm/el8/
[3] https://resources.ovirt.org/repos/ovirt/tested/master/rpm/el8/
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/Q7QAW24L5AYRJVBFYWCTE7UW4DSFTSFN/


Re: No module named ost_utils.memoized (was: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-4.3 - Build # 480 - Still Failing!)

2020-06-21 Thread Yedidyah Bar David
On Sun, Jun 21, 2020 at 11:23 AM Galit Rosenthal  wrote:
>
> Hi,
>
> This is the fix: https://gerrit.ovirt.org/#/c/109785/
>
> I've already saw the issue.

Thanks!

Can you please check why it was not caught on check-patch? Thanks.

>
> Galit
>
> On Sun, Jun 21, 2020 at 10:54 AM Yedidyah Bar David  wrote:
>>
>> On Sun, Jun 21, 2020 at 6:18 AM  wrote:
>> >
>> > Project: 
>> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-4.3/
>> > Build: 
>> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-4.3/480/
>>
>> This, and other HE suites in recent days, failed with:
>>
>> Traceback (most recent call last):
>>   File "/usr/lib64/python2.7/unittest/case.py", line 369, in run
>> testMethod()
>>   File "/usr/lib/python2.7/site-packages/nose/loader.py", line 418, in
>> loadTestsFromName
>> addr.filename, addr.module)
>>   File "/usr/lib/python2.7/site-packages/nose/importer.py", line 47,
>> in importFromPath
>> return self.importFromDir(dir_path, fqname)
>>   File "/usr/lib/python2.7/site-packages/nose/importer.py", line 94,
>> in importFromDir
>> mod = load_module(part_fqname, fh, filename, desc)
>>   File 
>> "/home/jenkins/agent/workspace/ovirt-system-tests_he-basic-suite-4.3/ovirt-system-tests/he-basic-suite-4.3/test-scenarios/002_bootstrap.py",
>> line 35, in 
>> import test_utils
>>   File 
>> "/home/jenkins/agent/workspace/ovirt-system-tests_he-basic-suite-4.3/ovirt-system-tests/he-basic-suite-4.3/test-scenarios/test_utils/__init__.py",
>> line 30, in 
>> from ost_utils.memoized import memoized
>> ImportError: No module named ost_utils.memoized
>>
>> I assume that this is a result of:
>>
>> https://gerrit.ovirt.org/109337
>>
>> and of the fact that some suites (HE, not sure about others) are still
>> using nose.
>>
>> Marcin, can you please have a look? Thanks.
>>
>> Galit, can you please check why check-patch on above patch didn't
>> fail? If it's because we removed HE suites from check-patch, because
>> they were unstable, then IMO we should add them back. They are rather
>> stable now, and if they fail, we should fix them. Thanks.
>>
>> Best regards,
>> --
>> Didi
>>
>
>
> --
>
> GALIT ROSENTHAL
>
> SOFTWARE ENGINEER
>
> Red Hat
>
> ga...@redhat.comT: 972-9-7692230



-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/FDD4F5ERCZTHXSMEFTY4SF56BK6O3KPY/


No module named ost_utils.memoized (was: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-4.3 - Build # 480 - Still Failing!)

2020-06-21 Thread Yedidyah Bar David
On Sun, Jun 21, 2020 at 6:18 AM  wrote:
>
> Project: https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-4.3/
> Build: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-4.3/480/

This, and other HE suites in recent days, failed with:

Traceback (most recent call last):
  File "/usr/lib64/python2.7/unittest/case.py", line 369, in run
testMethod()
  File "/usr/lib/python2.7/site-packages/nose/loader.py", line 418, in
loadTestsFromName
addr.filename, addr.module)
  File "/usr/lib/python2.7/site-packages/nose/importer.py", line 47,
in importFromPath
return self.importFromDir(dir_path, fqname)
  File "/usr/lib/python2.7/site-packages/nose/importer.py", line 94,
in importFromDir
mod = load_module(part_fqname, fh, filename, desc)
  File 
"/home/jenkins/agent/workspace/ovirt-system-tests_he-basic-suite-4.3/ovirt-system-tests/he-basic-suite-4.3/test-scenarios/002_bootstrap.py",
line 35, in 
import test_utils
  File 
"/home/jenkins/agent/workspace/ovirt-system-tests_he-basic-suite-4.3/ovirt-system-tests/he-basic-suite-4.3/test-scenarios/test_utils/__init__.py",
line 30, in 
from ost_utils.memoized import memoized
ImportError: No module named ost_utils.memoized

I assume that this is a result of:

https://gerrit.ovirt.org/109337

and of the fact that some suites (HE, not sure about others) are still
using nose.

Marcin, can you please have a look? Thanks.

Galit, can you please check why check-patch on above patch didn't
fail? If it's because we removed HE suites from check-patch, because
they were unstable, then IMO we should add them back. They are rather
stable now, and if they fail, we should fix them. Thanks.

Best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/JCIPZS7XG7UG76XXZW2JFXFTPHOCX2JA/


Missing collectd-write_syslog (was: Change in ovirt-system-tests[master]: WIP: basic-suite: Run rename right after engine-setup)

2020-06-18 Thread Yedidyah Bar David
On Thu, Jun 18, 2020 at 4:06 PM Code Review  wrote:
>
> From Jenkins CI :
>
> Jenkins CI has posted comments on this change.
>
> Change subject: WIP: basic-suite: Run rename right after engine-setup
> ..
>
>
> Patch Set 1: Continuous-Integration-1
>
> Build Failed
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/9807/ : 
> FAILURE

+ yum -y install ovirt-host
Error:
 Problem: package
ovirt-host-4.4.2-0.0.20200615143243.git79d37bd.el8.x86_64 requires
ovirt-host-dependencies = 4.4.2-0.0.20200615143243.git79d37bd.el8, but
none of the providers can be installed
  - cannot install the best candidate for the job
  - nothing provides collectd-write_syslog needed by
ovirt-host-dependencies-4.4.2-0.0.20200615143243.git79d37bd.el8.x86_64

Known issue?

>
> --
> To view, visit https://gerrit.ovirt.org/109734
> To unsubscribe, visit https://gerrit.ovirt.org/settings
>
> Gerrit-Project: ovirt-system-tests
> Gerrit-Branch: master
> Gerrit-MessageType: comment
> Gerrit-Change-Id: Iedc6b9f317fa9942a0c0a56f8920b5e8b2afc97d
> Gerrit-Change-Number: 109734
> Gerrit-PatchSet: 1
> Gerrit-Owner: Yedidyah Bar David 
> Gerrit-Reviewer: Anton Marchukov 
> Gerrit-Reviewer: Dafna Ron 
> Gerrit-Reviewer: Dusan Fodor 
> Gerrit-Reviewer: Ehud Yonasi 
> Gerrit-Reviewer: Gal Ben Haim 
> Gerrit-Reviewer: Galit Rosenthal 
> Gerrit-Reviewer: Jenkins CI 
> Gerrit-Reviewer: Yedidyah Bar David 
> Gerrit-Comment-Date: Thu, 18 Jun 2020 13:05:58 +
> Gerrit-HasComments: No
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/VDEDEPWHWMQBKA5MRL4RVSM25JRMLKL7/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-4.3 - Build # 450 - Failure!

2020-05-25 Thread Yedidyah Bar David via Infra
On Sun, May 24, 2020 at 9:11 AM Yedidyah Bar David  wrote:
>
> On Fri, May 22, 2020 at 6:42 AM  wrote:
> >
> > Project: 
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-4.3/
> > Build: 
> > https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-4.3/450/
> > Build Number: 450
> > Build Status:  Failure
> > Triggered By: Started by timer
> >
> > -
> > Changes Since Last Success:
> > -
> > Changes for Build #450
> > [hbraha] network: explicitly set the filter to None
> >
> >
> >
> >
> > -
> > Failed Tests:
> > -
> > 1 tests failed.
> > FAILED:  008_restart_he_vm.restart_he_vm
> >
> > Error Message:
> > could not parse JSON: {"1": {"conf_on_shared_storage": true, "live-data": 
> > true, "extra": 
> > "metadata_parse_version=1\nmetadata_feature_version=1\ntimestamp=5768 (Thu 
> > May 21 23:39:31 2020)\nhost-id=1\nscore=3400\nvm_conf_refresh_time=5768 
> > (Thu May 21 23:39:32 
> > 2020)\nconf_on_shared_storage=True\nmaintenance=False\nstate=GlobalMaintenance\nstopped=False\n",
> >  "hostname": "lago-he-basic-suite-4-3-host-0.lago.local", "host-id": 1, 
> > "engine-status": {"reason": "failed liveliness check", "health": "bad", 
> > "vm": "up", "detail": "Up"}, "score": 3400, "stopped": false, 
> > "maintenance": false, "crc32": "4f6a7bea", "local_conf_timestamp": 5768, 
> > "host-ts": 5768}, "2": {"conf_on_shared_storage": true, "live-data": true, 
> > "extra": 
> > "metadata_parse_version=1\nmetadata_feature_version=1\ntimestamp=5767 (Thu 
> > May 21 23:39:31 2020)\nhost-id=2\nscore=3400\nvm_conf_refresh_time=5768 
> > (Thu May 21 23:39:32 
> > 2020)\nconf_on_shared_storage=True\nmaintenance=False\nstate=GlobalMaintenance\nstopped=False\n",
> >  "hostname": "lago-he-basic-suite-4-3-host-1", "host-id": 2, "en
>
> Again output cut after 1024 bytes. Still not sure why, pushed this to
> try a few times:
>
> https://gerrit.ovirt.org/109243

Marked verified.

Galit - please review/merge. Thanks.

CI check-patch didn't trigger a case where it had to run more than
once (per loop. It does run several tens of times), but hopefully
that's also ok.

Best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: ${hyperkitty_url}


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-suite-4.3 - Build # 450 - Failure!

2020-05-24 Thread Yedidyah Bar David
On Fri, May 22, 2020 at 6:42 AM  wrote:
>
> Project: https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-4.3/
> Build: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-4.3/450/
> Build Number: 450
> Build Status:  Failure
> Triggered By: Started by timer
>
> -
> Changes Since Last Success:
> -
> Changes for Build #450
> [hbraha] network: explicitly set the filter to None
>
>
>
>
> -
> Failed Tests:
> -
> 1 tests failed.
> FAILED:  008_restart_he_vm.restart_he_vm
>
> Error Message:
> could not parse JSON: {"1": {"conf_on_shared_storage": true, "live-data": 
> true, "extra": 
> "metadata_parse_version=1\nmetadata_feature_version=1\ntimestamp=5768 (Thu 
> May 21 23:39:31 2020)\nhost-id=1\nscore=3400\nvm_conf_refresh_time=5768 (Thu 
> May 21 23:39:32 
> 2020)\nconf_on_shared_storage=True\nmaintenance=False\nstate=GlobalMaintenance\nstopped=False\n",
>  "hostname": "lago-he-basic-suite-4-3-host-0.lago.local", "host-id": 1, 
> "engine-status": {"reason": "failed liveliness check", "health": "bad", "vm": 
> "up", "detail": "Up"}, "score": 3400, "stopped": false, "maintenance": false, 
> "crc32": "4f6a7bea", "local_conf_timestamp": 5768, "host-ts": 5768}, "2": 
> {"conf_on_shared_storage": true, "live-data": true, "extra": 
> "metadata_parse_version=1\nmetadata_feature_version=1\ntimestamp=5767 (Thu 
> May 21 23:39:31 2020)\nhost-id=2\nscore=3400\nvm_conf_refresh_time=5768 (Thu 
> May 21 23:39:32 
> 2020)\nconf_on_shared_storage=True\nmaintenance=False\nstate=GlobalMaintenance\nstopped=False\n",
>  "hostname": "lago-he-basic-suite-4-3-host-1", "host-id": 2, "en

Again output cut after 1024 bytes. Still not sure why, pushed this to
try a few times:

https://gerrit.ovirt.org/109243

Best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/EYRWIDZJIJL4YGDXPYT5EY4XWIZN52R4/


otopi and change-queue

2020-05-17 Thread Yedidyah Bar David
Hi all,

I recently merged a patch to otopi:

https://gerrit.ovirt.org/#/c/108590/

It failed in CQ:

https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/23512/

Now commented there "ci re-merge", and it passed:

https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/23736/

But didn't run basic-suite-master, and I do not understand if it will
be published.

Was anything changed in how change-queue works?

Should I do anything else to make sure it's published?

Thanks,
--
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/5QJCEAVOEDOYLWRWUL5IG5U3OYRG6ZRP/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-role-remote-suite-master - Build # 406 - Still Failing!

2020-05-10 Thread Yedidyah Bar David
On Sun, May 10, 2020 at 4:35 AM  wrote:
>
> Project: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-role-remote-suite-master/
> Build: 
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-role-remote-suite-master/406/

Fails for a long time now, but recently, since he-basic-suite-master
is fixed, fails with:

https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-role-remote-suite-master/406/console

03:32:03 TASK [ovirt.hosted_engine_setup : Wait for the local VM]
***
04:32:09 fatal: [lago-he-basic-role-remote-suite-master-host-0 ->
lago-he-basic-role-remote-suite-master-engine.lago.local]: FAILED! =>
{"changed": false, "elapsed": 3605, "msg": "timed out waiting for ping
module test success: [Errno 24] Too many open files"}

Looked around and failed to find other relevant information.
I guess we should somehow (in the playbook or in OST, not sure which
is better) collect 'ulimit -a' and open files.

Searching google for the exact error message (meaning, both that the
ping module failed, and that the reason was too many open files),
finds only one place, does not seem related, so I do not think it's a
regression in ansible.

>
>
> Build Number: 406
> Build Status:  Still Failing
> Triggered By: Started by timer
>
> -
> Changes Since Last Success:
> -
> Changes for Build #397
> [Sandro Bonazzola] python3: use print function
>
>
> Changes for Build #398
> [Sandro Bonazzola] python3: use print function
>
>
> Changes for Build #399
> [Sandro Bonazzola] python3: use print function
>
>
> Changes for Build #400
> [Sandro Bonazzola] python3: use print function
>
>
> Changes for Build #401
> [Michal Skrivanek] fix list of packages for storage setup
>
>
> Changes for Build #402
> [Lucia Jelinkova] UI test refactoring
>
>
> Changes for Build #403
> [Galit Rosenthal] Remove reposync from upgrade-from-prevrelease-suite-4.3
>
>
> Changes for Build #404
> [Marcin Sobczyk] selenium: podman: Add grid setup retries
>
>
> Changes for Build #405
> [Marcin Sobczyk] selenium: podman: Add grid setup retries
>
>
> Changes for Build #406
> [Marcin Sobczyk] selenium: podman: Add grid setup retries
>
>
>
>
> -
> Failed Tests:
> -
> No tests ran.



-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/7GQMNURKXJKKD3GE5RDYCMU7S36VKXMP/


[JIRA] (OVIRT-2936) lists.ovirt.org is missing in google searches

2020-05-06 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-2936:
-

 Summary: lists.ovirt.org is missing in google searches
 Key: OVIRT-2936
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-2936
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


Hi all,

I now searched google for "engine certificate problems in MacOS",
expecting to find [1], but didn't get it.

I don't think it's due to our robots.txt [2], because google does find
it in [3]. Perhaps it's due to configuration for our site in google or
whatever, no idea.

Can you please handle? Thanks.

[1] 
https://lists.ovirt.org/archives/list/us...@ovirt.org/message/YNL6NSW6GP3IR7GECYE6DNPJA6H2X3RB/
[2] https://lists.ovirt.org/robots.txt
[3] https://www.mail-archive.com/users@ovirt.org/msg60558.html
-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100126)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/EBMWGTYZFLO2JKKY5JBFGARZRW2VHNID/


fc31/fc32 support in CI

2020-05-03 Thread Yedidyah Bar David
Hi all,

Do we plan $subject?

Tried pushing [1] to otopi, fails with:

Unable to find mock env

[1] https://gerrit.ovirt.org/108768

Thanks,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/5N4KNZXLACAFE33WDWXBW4BDKMZH3ZNP/


Re: [CQ]: 107489, 11 (ovirt-engine) failed "ovirt-master" system tests, but isn't the failure root cause

2020-04-19 Thread Yedidyah Bar David
On Sun, Apr 19, 2020 at 5:28 AM oVirt Jenkins  wrote:
>
> A system test invoked by the "ovirt-master" change queue including change
> 107489,11 (ovirt-engine) failed. However, this change seems not to be the root
> cause for this failure. Change 108233,6 (ovirt-engine) that this change 
> depends
> on or is based on, was detected as the cause of the testing failures.
>
> This change had been removed from the testing queue. Artifacts built from this
> change will not be released until either change 108233,6 (ovirt-engine) is
> fixed and this change is updated to refer to or rebased on the fixed version,
> or this change is modified to no longer depend on it.
>
> For further details about the change see:
> https://gerrit.ovirt.org/#/c/107489/11

This change was merged 12 days ago

>
> For further details about the change that seems to be the root cause behind 
> the
> testing failures see:
> https://gerrit.ovirt.org/#/c/108233/6

Also this

>
> For failed test results see:
> https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/23260/

This ran yesterday.

Where can I see the queue of jobs/builds/whatever that should test all
the patches since this? Its status? Etc.? Thanks.

Best regards,
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/2N24ESFHWF4S63KVWEWYJXNVU4W6S5PT/


Re: [JIRA] (OVIRT-2898) CQ Changes is empty

2020-04-05 Thread Yedidyah Bar David
On Sun, Apr 5, 2020 at 12:23 PM Barak Korren (oVirt JIRA)
 wrote:
>
>
> [ 
> https://ovirt-jira.atlassian.net/browse/OVIRT-2898?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=40276#comment-40276
>  ]
>
> Barak Korren commented on OVIRT-2898:
> -
>
> The changes screen in Jenkins only shows a log of Git changes that were 
> cloned via the Git (or other SCM plugin). Its only useful for a 
> single-project, single-branch job. Having it show the CQ changes would 
> require writing a jenkins plugin that would make Jenkins think the CQ is some 
> kind of an SCM (This is not trivial, SCM plugins in jenkns are strange...)
>
> WRT the linked path, it works as expected:
> https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/22202/execution/node/99/log/

That's helpful, thanks.

>
> This show up under the "loading change data" stage both in blue ocean and in 
> the pipeline view.
>
> Closing NOT A BUG.
>
> > CQ Changes is empty
> > ---
> >
> > Key: OVIRT-2898
> > URL: https://ovirt-jira.atlassian.net/browse/OVIRT-2898
> >         Project: oVirt - virtualization made easy
> >  Issue Type: By-EMAIL
> >Reporter: Yedidyah Bar David
> >Assignee: infra
> >
> > Hi all,
> > If I look e.g. at:
> > https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/22202/
> > It says:
> > Testing 58 changes:
> > But then no list of changes.
> > If I then press Changes:
> > https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/22202/changes
> > I get an empty page (other than red circle and "Changes" title).
> > I think this is a bug, which was already supposed to fixed by:
> > https://gerrit.ovirt.org/79036
> > But for some reason this does not work as expected.
> > Please handle.
> > Thanks,
> > --
> > Didi
>
>
>
> --
> This message was sent by Atlassian Jira
> (v1001.0.0-SNAPSHOT#100124)
> ___
> Infra mailing list -- infra@ovirt.org
> To unsubscribe send an email to infra-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives: 
> https://lists.ovirt.org/archives/list/infra@ovirt.org/message/AIU7RVA4Q7KOJBY3A7E5XBKFGGHFIUCM/



-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/4HMYC4TW3PUVFQCW4IASL4SVXLYF65LX/


[JIRA] (OVIRT-2898) CQ Changes is empty

2020-04-05 Thread Yedidyah Bar David (oVirt JIRA)
Yedidyah Bar David created OVIRT-2898:
-

 Summary: CQ Changes is empty
 Key: OVIRT-2898
 URL: https://ovirt-jira.atlassian.net/browse/OVIRT-2898
 Project: oVirt - virtualization made easy
  Issue Type: By-EMAIL
Reporter: Yedidyah Bar David
Assignee: infra


Hi all,

If I look e.g. at:

https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/22202/

It says:

Testing 58 changes:

But then no list of changes.

If I then press Changes:

https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/22202/changes

I get an empty page (other than red circle and "Changes" title).

I think this is a bug, which was already supposed to fixed by:

https://gerrit.ovirt.org/79036

But for some reason this does not work as expected.

Please handle.

Thanks,
-- 
Didi



--
This message was sent by Atlassian Jira
(v1001.0.0-SNAPSHOT#100124)
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/MD7GMKQ4RRKK4JRTH4GWLP6HWAT5DOKT/


Re: mirrors.phx.ovirt.org is down

2020-04-04 Thread Yedidyah Bar David
On Thu, Apr 2, 2020 at 7:41 PM Barak Korren  wrote:
>
> Not sure why you're seeing any access to the mirrors from your laptop, they 
> should only be used in CI.

I just checked it manually, copy/pasted the address from the job below.

>
> As for the CI code, of the mirrors are down, it ignores then and goes to the 
> upstream repos.

OK, so:

>
> Is there some script that is hardwired to get something from a specific 
> mirror url?
>
>
> בתאריך יום ה׳, 2 באפר׳ 2020, 17:32, מאת Yedidyah Bar David ‏:
>>
>> Both from my laptop and e.g.:
>>
>> https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/6729/

Why did above run fail, then?

>>
>> 17:23:21 + yum install '--disablerepo=*'
>> --enablerepo=ovirt-master-tested-el8,centos-base-el8,centos-appstream-el8,centos-powertools-el8,epel-el8,ovirt-master-glusterfs-7-el8,ovirt-master-virtio-win-latest-el8,ovirt-master-copr-sbonazzo-collection-el8,ovirt-master-copr:copr.fedorainfracloud.org:sac:gluster-ansible-el8,ovirt-master-copr:copr.fedorainfracloud.org:mdbarroso:ovsdbapp-el8,ovirt-master-copr-nmstate-0.2-el8,ovirt-master-copr-NetworkManager-1.22-el8,ovirt-master-centos-advanced-virtualization-el8,ovirt-master-centos-ovirt44-el8
>> -y yum-utils
>>
>> 17:23:21 Error: Error downloading packages:
>>
>> 17:23:21   Status code: 404 for
>> http://mirrors.phx.ovirt.org/repos/yum/centos-base-el8/base/Packages/yum-utils-4.0.8-3.el8.noarch.rpm
>>
>> Known problem?
>> --
>> Didi
>> ___
>> Infra mailing list -- infra@ovirt.org
>> To unsubscribe send an email to infra-le...@ovirt.org
>> Privacy Statement: https://www.ovirt.org/privacy-policy.html
>> oVirt Code of Conduct: 
>> https://www.ovirt.org/community/about/community-guidelines/
>> List Archives: 
>> https://lists.ovirt.org/archives/list/infra@ovirt.org/message/BPER52OO73BBJW22W4UCEYKWCEQRI77M/



-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/QM6TZC5UUB63PKY7Q64XWSNVIYB4YILV/


mirrors.phx.ovirt.org is down

2020-04-02 Thread Yedidyah Bar David
Both from my laptop and e.g.:

https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/6729/

17:23:21 + yum install '--disablerepo=*'
--enablerepo=ovirt-master-tested-el8,centos-base-el8,centos-appstream-el8,centos-powertools-el8,epel-el8,ovirt-master-glusterfs-7-el8,ovirt-master-virtio-win-latest-el8,ovirt-master-copr-sbonazzo-collection-el8,ovirt-master-copr:copr.fedorainfracloud.org:sac:gluster-ansible-el8,ovirt-master-copr:copr.fedorainfracloud.org:mdbarroso:ovsdbapp-el8,ovirt-master-copr-nmstate-0.2-el8,ovirt-master-copr-NetworkManager-1.22-el8,ovirt-master-centos-advanced-virtualization-el8,ovirt-master-centos-ovirt44-el8
-y yum-utils

17:23:21 Error: Error downloading packages:

17:23:21   Status code: 404 for
http://mirrors.phx.ovirt.org/repos/yum/centos-base-el8/base/Packages/yum-utils-4.0.8-3.el8.noarch.rpm

Known problem?
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/BPER52OO73BBJW22W4UCEYKWCEQRI77M/


basic suite: Error: Failed to download metadata for repo 'alocalsync'

2020-03-30 Thread Yedidyah Bar David
Hi all,

I get $subject in [1][2]. Any idea?

Thanks,

[1] 
https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/6707/
[2] 
https://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt-system-tests_manual/6707/artifact/exported-artifacts/mock_logs/script/stdout_stderr.log
-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/POALKSROLFLRITLLUSZMX4HF2UDXIFEE/


Re: [oVirt Jenkins] ovirt-system-tests_he-basic-role-remote-suite-4.3 - Build # 304 - Still Failing!

2020-03-15 Thread Yedidyah Bar David
On Wed, Mar 11, 2020 at 1:22 PM Yedidyah Bar David  wrote:
>
> On Wed, Mar 11, 2020 at 10:41 AM Galit Rosenthal  wrote:
> >
> > Hi Didi,
> >
> > Check the error of the suite he-basic-role-remote=suite-4.3
> > https://jenkins.ovirt.org/blue/organizations/jenkins/ovirt-system-tests_standard-check-patch/detail/ovirt-system-tests_standard-check-patch/8305/pipeline/173
> >
> >
> >
> >
> > DEPRECATION WARNING]: The 'ovirt_host_facts' module has been renamed to
> >
> >
> > 'ovirt_host_info', and the renamed one no longer returns ansible_facts. This
> >
> >
> > feature will be removed in version 2.13. Deprecation warnings can be 
> > disabled
> >
> >
> > by setting deprecation_warnings=False in ansible.cfg.
>
>
> That's just a warning, can be ignored for now. We do have an open bug
> to fix it (perhaps only in 4.4, we'll see).
>
> >
> >
> >
> > fatal: [lago-he-basic-role-remote-suite-4-3-host-0]: FAILED! => 
> > {"ansible_facts": {"ovirt_hosts": [{"address": 
> > "lago-he-basic-role-remote-suite-4-3-host-0.lago.local", "affinity_labels": 
> > [], "auto_numa_status": "unknown", "certificate": {"organization": 
> > "lago.local", "subject": 
> > "O=lago.local,CN=lago-he-basic-role-remote-suite-4-3-host-0.lago.local"}, 
> > "cluster": {"href": 
> > "/ovirt-engine/api/clusters/03967a8e-6368-11ea-a651-5452c0a8c863", "id": 
> > "03967a8e-6368-11ea-a651-5452c0a8c863"}, "comment": "", "cpu": {"speed": 
> > 0.0, "topology": {}}, "device_passthrough": {"enabled": false}, "devices": 
> > [], "external_network_provider_configurations": [], "external_status": 
> > "ok", "hardware_information": {"supported_rng_sources": []}, "hooks": [], 
> > "href": "/ovirt-engine/api/hosts/8d06089a-d562-458f-8720-0ba41ba082ef", 
> > "id": "8d06089a-d562-458f-8720-0ba41ba082ef", "katello_errata": [], 
> > "kdump_status": "unknown", "ksm": {"enabled": false}, 
> > "max_scheduling_memory": 0, "memory": 0, "name": 
> > "lago-he-basic-role-remote-suite-4-3-host-0", "network_attachments": [], 
> > "nics": [], "numa_nodes": [], "numa_supported": false, "os": 
> > {"custom_kernel_cmdline": ""}, "permissions": [], "port": 54321, 
> > "power_management": {"automatic_pm_enabled": true, "enabled": false, 
> > "kdump_detection": true, "pm_proxies": []}, "protocol": "stomp", 
> > "se_linux": {}, "spm": {"priority": 5, "status": "none"}, "ssh": 
> > {"fingerprint": "SHA256:tiWL/3QtoZ//DFHnGT+B3b80EBBcbpftGhcEgYpB7+8", 
> > "port": 22}, "statistics": [], "status": "install_failed", 
> > "storage_connection_extensions": [], "summary": {"total": 0}, "tags": [], 
> > "transparent_huge_pages": {"enabled": false}, "type": "rhel", 
> > "unmanaged_networks": [], "update_available": false, "vgpu_placement": 
> > "consolidated"}]}, "attempts": 120, "changed": false}
>
> This failed:
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/8305/artifact/check-patch.he-basic-role-remote_suite_4.3.el7.x86_64/mock_logs/script/stdout_stderr.log
>
> TASK [ovirt.hosted_engine_setup : Wait for the host to be up] 
> **
> FAILED - RETRYING: Wait for the host to be up (120 retries left).
> ...
> FAILED - RETRYING: Wait for the host to be up (1 retries left).
>
> Why the host didn't come up?
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_standard-check-patch/8305/artifact/check-patch.he-basic-role-remote_suite_4.3.el7.x86_64/test_logs/he-basic-role-remote-suite-4.3/post-he_deploy/lago-he-basic-role-remote-suite-4-3-host-0/_var_log/ovirt-hosted-engine-setup/engine-logs-2020-03-11T07%3A08%3A40Z/ovirt-engine/host-deploy/ovirt-host-deploy-20200311032223-lago-he-basic-role-remote-suite-4-3-host-0.lago.local-54583b6f.log
>
> 2020-03-11 03:22:22,561-0400 DEBUG
> otopi.plugins.otopi.packagers.yumpackager yumpackager.verbose:

No space left on device (was: Change in ovirt-appliance[master]: Add serial console)

2020-03-15 Thread Yedidyah Bar David
On Sun, Mar 15, 2020 at 12:54 PM Code Review  wrote:
>
> From Jenkins CI :
>
> Jenkins CI has posted comments on this change.
>
> Change subject: Add serial console
> ..
>
>
> Patch Set 3: Continuous-Integration-1
>
> Build Failed
>
> https://jenkins.ovirt.org/job/ovirt-appliance_standard-check-patch/242/ : 
> FAILURE

12:53:55  Archiving artifacts
[Pipeline] echo
12:53:55  STDCI report generation skipped because not on a node
[Pipeline] }
[Pipeline] // node
[Pipeline] // node
[Pipeline] }
[Pipeline] // parallel
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // timestamps
[Pipeline] End of Pipeline
java.io.IOException: No space left on device

Known issue?

>
> --
> To view, visit https://gerrit.ovirt.org/106943
> To unsubscribe, visit https://gerrit.ovirt.org/settings
>
> Gerrit-Project: ovirt-appliance
> Gerrit-Branch: master
> Gerrit-MessageType: comment
> Gerrit-Change-Id: I70a5faaf039d0debef730b08be87e524c2c65c98
> Gerrit-Change-Number: 106943
> Gerrit-PatchSet: 3
> Gerrit-Owner: Yedidyah Bar David 
> Gerrit-Reviewer: Asaf Rachmani 
> Gerrit-Reviewer: Douglas Schilling Landgraf 
> Gerrit-Reviewer: Evgeny Slutsky 
> Gerrit-Reviewer: Jenkins CI 
> Gerrit-Reviewer: Nir Levy 
> Gerrit-Reviewer: Ryan Barry 
> Gerrit-Reviewer: Sandro Bonazzola 
> Gerrit-Reviewer: Yedidyah Bar David 
> Gerrit-Reviewer: Yuval Turgeman 
> Gerrit-Reviewer: gerrit-hooks 
> Gerrit-Comment-Date: Sun, 15 Mar 2020 10:53:56 +
> Gerrit-HasComments: No
>


-- 
Didi
___
Infra mailing list -- infra@ovirt.org
To unsubscribe send an email to infra-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/infra@ovirt.org/message/AT5NNWFJZNHBDTLZBZIMXAAITPXTTE4N/


  1   2   3   4   5   >