On Fri, 21 Feb 2020 at 14:52, Nathanaël Blanchet <blanc...@abes.fr> wrote:

> Hello,
>
> It seems that the work for including ovirt as a provider in the master
> branch of openshift installer has been done. I compiled the master code and
> ovirt does appear in the survey.
> I don't have much time to test it for now but is it operationnal? If yes,
> I will prior to have a look to it.
>

It is operational, yes - which OS are you going to use?

> Thanks.
> Le 06/01/2020 à 21:30, Roy Golan a écrit :
>
>
>
> The merge window is now open for the masters branches of the various
> origin components.
> Post merge there should be an OKD release - this is not under my control,
> but when it will be available I'll let you know.
>
> On Mon, 6 Jan 2020 at 20:54, Nathanaël Blanchet <blanc...@abes.fr> wrote:
>
>> Hello Roy
>> Le 21/11/2019 à 13:57, Roy Golan a écrit :
>>
>>
>>
>> On Thu, 21 Nov 2019 at 08:48, Roy Golan <rgo...@redhat.com> wrote:
>>
>>>
>>>
>>> On Wed, 20 Nov 2019 at 09:49, Nathanaël Blanchet <blanc...@abes.fr>
>>> wrote:
>>>
>>>>
>>>> Le 19/11/2019 à 19:23, Nathanaël Blanchet a écrit :
>>>>
>>>>
>>>> Le 19/11/2019 à 13:43, Roy Golan a écrit :
>>>>
>>>>
>>>>
>>>> On Tue, 19 Nov 2019 at 14:34, Nathanaël Blanchet <blanc...@abes.fr>
>>>> wrote:
>>>>
>>>>> Le 19/11/2019 à 08:55, Roy Golan a écrit :
>>>>>
>>>>> oc get -o json clusterversion
>>>>>
>>>>> This is the output of the previous failed deployment, I'll give a try
>>>>> to a newer one when I'll have a minute to test
>>>>>
>>>> Without changing nothing with template,  I gave a new try and...
>>>> nothing works anymore now, none of provided IPs can be pingued : dial tcp
>>>> 10.34.212.51:6443: connect: no route to host", so none of masters can
>>>> be provisonned by bootstrap.
>>>>
>>>> I tried with the latest rhcos and latest ovirt 4.3.7, it is the same.
>>>> Obviously something changed since my first attempt 12 days ago... is your
>>>> docker image for openshift-installer up to date?
>>>>
>>>> Are you still able to your side to deploy a valid cluster ?
>>>>
>>>> I investigated looking at bootstrap logs (attached) and it seems that
>>>> every containers die immediately after been started.
>>>>
>>>> Nov 20 07:02:33 localhost podman[2024]: 2019-11-20 07:02:33.60107571
>>>> +0000 UTC m=+0.794838407 container init
>>>> 446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603 (image=
>>>> registry.svc.ci.openshift.org/origin/release:4.3, name=eager_cannon)
>>>> Nov 20 07:02:33 localhost podman[2024]: 2019-11-20 07:02:33.623197173
>>>> +0000 UTC m=+0.816959853 container start
>>>> 446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603 (image=
>>>> registry.svc.ci.openshift.org/origin/release:4.3, name=eager_cannon)
>>>> Nov 20 07:02:33 localhost podman[2024]: 2019-11-20 07:02:33.623814258
>>>> +0000 UTC m=+0.817576965 container attach
>>>> 446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603 (image=
>>>> registry.svc.ci.openshift.org/origin/release:4.3, name=eager_cannon)
>>>> Nov 20 07:02:34 localhost systemd[1]:
>>>> libpod-446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603.scope:
>>>> Consumed 814ms CPU time
>>>> Nov 20 07:02:34 localhost podman[2024]: 2019-11-20 07:02:34.100569998
>>>> +0000 UTC m=+1.294332779 container died
>>>> 446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603 (image=
>>>> registry.svc.ci.openshift.org/origin/release:4.3, name=eager_cannon)
>>>> Nov 20 07:02:35 localhost podman[2024]: 2019-11-20 07:02:35.138523102
>>>> +0000 UTC m=+2.332285844 container remove
>>>> 446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603 (image=
>>>> registry.svc.ci.openshift.org/origin/release:4.3, name=eager_cannon)
>>>>
>>>> and this:
>>>>
>>>> Nov 20 07:04:16 localhost hyperkube[1909]: E1120 07:04:16.489527
>>>> 1909 remote_runtime.go:200] CreateContainer in sandbox
>>>> "58f2062aa7b6a5b2bdd6b9cf7b41a9f94ca2b30ad5a20e4fa4dec8a9b82f05e5" from
>>>> runtime service failed: rpc error: code = Unknown desc = container create
>>>> failed: container_linux.go:345: starting container process caused "exec:
>>>> \"runtimecfg\": executable file not found in $PATH"
>>>> Nov 20 07:04:16 localhost hyperkube[1909]: E1120 07:04:16.489714
>>>> 1909 kuberuntime_manager.go:783] init container start failed:
>>>> CreateContainerError: container create failed: container_linux.go:345:
>>>> starting container process caused "exec: \"runtimecfg\": executable file
>>>> not found in $PATH"
>>>>
>>>> What do you think about this?
>>>>
>>>
>>> I'm seeing the same now, checking...
>>>
>>>
>> Because of the move upstream to release OKD the release-image that comes
>> with the installer I gave you are no longer valid.
>>
>> I need to prepare an installer version with the preview of OKD, you can
>> find the details here
>> https://mobile.twitter.com/smarterclayton/status/1196477646885965824
>>
>> I tested your last openshift-installer container on quay.io, but the
>> ovirt provider is not available anymore. Will ovirt be supported as an OKD
>> 4.2 iaas provider ?
>>
>>
>>
>>
>> (do I need to use the terraform-workers tag instead of latest?)
>>>>>
>>>>> docker pull quay.io/rgolangh/openshift-installer:terraform-workers
>>>>>
>>>>>
>>>>> [root@openshift-installer
>>>>> openshift-origin-client-tools-v3.11.0-0cbc58b-linux-64bit]# ./oc get -o
>>>>> json clusterversion
>>>>> {
>>>>>     "apiVersion": "v1",
>>>>>     "items": [
>>>>>         {
>>>>>             "apiVersion": "config.openshift.io/v1",
>>>>>             "kind": "ClusterVersion",
>>>>>             "metadata": {
>>>>>                 "creationTimestamp": "2019-11-07T12:23:06Z",
>>>>>                 "generation": 1,
>>>>>                 "name": "version",
>>>>>                 "namespace": "",
>>>>>                 "resourceVersion": "3770202",
>>>>>                 "selfLink": "/apis/
>>>>> config.openshift.io/v1/clusterversions/version",
>>>>>                 "uid": "77600bba-6e71-4b35-a60b-d8ee6e0f545c"
>>>>>             },
>>>>>             "spec": {
>>>>>                 "channel": "stable-4.3",
>>>>>                 "clusterID": "6f87b719-e563-4c0b-ab5a-1144172bc983",
>>>>>                 "upstream":
>>>>> "https://api.openshift.com/api/upgrades_info/v1/graph";
>>>>> <https://api.openshift.com/api/upgrades_info/v1/graph>
>>>>>             },
>>>>>             "status": {
>>>>>                 "availableUpdates": null,
>>>>>                 "conditions": [
>>>>>                     {
>>>>>                         "lastTransitionTime": "2019-11-07T12:23:12Z",
>>>>>                         "status": "False",
>>>>>                         "type": "Available"
>>>>>                     },
>>>>>                     {script
>>>>>                         "lastTransitionTime": "2019-11-07T12:56:15Z",
>>>>>                         "message": "Cluster operator image-registry is
>>>>> still updating",
>>>>>                         "reason": "ClusterOperatorNotAvailable",
>>>>>                         "status": "True",
>>>>>                         "type": "Failing"
>>>>>                     },
>>>>>                     {
>>>>>                         "lastTransitionTime": "2019-11-07T12:23:12Z",
>>>>>                         "message": "Unable to apply
>>>>> 4.3.0-0.okd-2019-10-29-180250: the cluster operator image-registry has not
>>>>> yet successfully rolled out",
>>>>>                         "reason": "ClusterOperatorNotAvailable",
>>>>>                         "status": "True",
>>>>>                         "type": "Progressing"
>>>>>                     },
>>>>>                     {
>>>>>                         "lastTransitionTime": "2019-11-07T12:23:12Z",
>>>>>                         "message": "Unable to retrieve available
>>>>> updates: currently installed version 4.3.0-0.okd-2019-10-29-180250 not
>>>>> found in the \"stable-4.3\" channel",
>>>>>                         "reason": "RemoteFailed",
>>>>>                         "status": "False",
>>>>>                         "type": "RetrievedUpdates"
>>>>>                     }
>>>>>                 ],
>>>>>                 "desired": {
>>>>>                     "force": false,
>>>>>                     "image": "
>>>>> registry.svc.ci.openshift.org/origin/release@sha256:68286e07f7d68ebc8a067389aabf38dee9f9b810c5520d6ee4593c38eb48ddc9
>>>>> ",
>>>>>                     "version": "4.3.0-0.okd-2019-10-29-180250"
>>>>>
>>>>
>>>> Indeed this version is not the latest and is missing the aforementioned
>>>> fix for the registry.
>>>>
>>>>                 },
>>>>>                 "history": [
>>>>>                     {
>>>>>                         "completionTime": null,
>>>>>                         "image": "
>>>>> registry.svc.ci.openshift.org/origin/release@sha256:68286e07f7d68ebc8a067389aabf38dee9f9b810c5520d6ee4593c38eb48ddc9
>>>>> ",
>>>>>                         "startedTime": "2019-11-07T12:23:12Z",
>>>>>                         "state": "Partial",
>>>>>                         "verified": false,
>>>>>                         "version": "4.3.0-0.okd-2019-10-29-180250"
>>>>>                     }
>>>>>                 ],
>>>>>                 "observedGeneration": 1,
>>>>>                 "versionHash": "-3onP9QpPTg="
>>>>>             }
>>>>>         }
>>>>>     ],
>>>>>     "kind": "List",
>>>>>     "metadata": {
>>>>>         "resourceVersion": "",
>>>>>         "selfLink": ""
>>>>>     }
>>>>>
>>>>> }
>>>>>
>>>>>
>>>>> Can you answer to these few questions please?
>>>>>
>>>>>    - The latest stable OKD version is 4.2.4. Is it possible to chose
>>>>>    the version of okd when deploying (seems to use 4.3) or does the 
>>>>> installer
>>>>>    always download the latest OKD?
>>>>>
>>>>>
>>>>
>>>>>
>>>>>    - Can we use FCOS instead of RHCOS?
>>>>>
>>>>>
>>>>
>>>>>    - About the pull secret, do we absolutely need a redhat login to
>>>>>    get this file to deploy an upstream OKD cluster and not downstream
>>>>>    openshift?
>>>>>
>>>>>
>>>>> To answer the 3 of those, this specific build is not really OKD, and
>>>> will use 4.3 and Red Hat artifact and must use RHCOs, hence the pull secret
>>>> thing.
>>>> I frankly don't know when OKD 4.3 is going to be released, I guess it
>>>> will be on top FCOS.
>>>> I'll update the list once we have the oVirt installer for OKD ready for
>>>> testing (on FCOS)
>>>>
>>>>
>>>>
>>>> --
>>>>> Nathanaël Blanchet
>>>>>
>>>>> Supervision réseau
>>>>> Pôle Infrastrutures Informatiques
>>>>> 227 avenue Professeur-Jean-Louis-Viala
>>>>> 34193 MONTPELLIER CEDEX 5         
>>>>> Tél. 33 (0)4 67 54 84 55
>>>>> Fax  33 (0)4 67 54 84 14blanc...@abes.fr
>>>>>
>>>>> --
>>>> Nathanaël Blanchet
>>>>
>>>> Supervision réseau
>>>> Pôle Infrastrutures Informatiques
>>>> 227 avenue Professeur-Jean-Louis-Viala
>>>> 34193 MONTPELLIER CEDEX 5  
>>>> Tél. 33 (0)4 67 54 84 55
>>>> Fax  33 (0)4 67 54 84 14blanc...@abes.fr
>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list -- users@ovirt.org
>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>> oVirt Code of Conduct: 
>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>> List Archives: 
>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/MLO4NW7NWR2TXKOJ4MJLW4N3DFU2PJ7V/
>>>>
>>>> --
>>>> Nathanaël Blanchet
>>>>
>>>> Supervision réseau
>>>> Pôle Infrastrutures Informatiques
>>>> 227 avenue Professeur-Jean-Louis-Viala
>>>> 34193 MONTPELLIER CEDEX 5  
>>>> Tél. 33 (0)4 67 54 84 55
>>>> Fax  33 (0)4 67 54 84 14blanc...@abes.fr
>>>>
>>>> --
>> Nathanaël Blanchet
>>
>> Supervision réseau
>> SIRE
>> 227 avenue Professeur-Jean-Louis-Viala
>> 34193 MONTPELLIER CEDEX 5    
>> Tél. 33 (0)4 67 54 84 55
>> Fax  33 (0)4 67 54 84 14blanc...@abes.fr
>>
>> --
> Nathanaël Blanchet
>
> Supervision réseau
> SIRE
> 227 avenue Professeur-Jean-Louis-Viala
> 34193 MONTPELLIER CEDEX 5     
> Tél. 33 (0)4 67 54 84 55
> Fax  33 (0)4 67 54 84 14blanc...@abes.fr
>
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BVJ5ZVXPLYE2ADG54OXWZYQYHH3D24A3/

Reply via email to