On Thu, Nov 7, 2019 at 12:57 PM Roy Golan <rgo...@redhat.com> wrote:

>
>
> On Thu, 7 Nov 2019 at 12:28, Nathanaël Blanchet <blanc...@abes.fr> wrote:
>
>>
>> Le 07/11/2019 à 11:16, Roy Golan a écrit :
>>
>>
>>
>> On Thu, 7 Nov 2019 at 11:23, Nathanaël Blanchet <blanc...@abes.fr> wrote:
>>
>>>
>>> Le 07/11/2019 à 07:18, Roy Golan a écrit :
>>>
>>>
>>>
>>> On Thu, 7 Nov 2019 at 00:10, Nathanaël Blanchet <blanc...@abes.fr>
>>> wrote:
>>>
>>>>
>>>> Le 05/11/2019 à 21:50, Roy Golan a écrit :
>>>>
>>>>
>>>>
>>>> On Tue, 5 Nov 2019 at 22:46, Roy Golan <rgo...@redhat.com> wrote:
>>>>
>>>>>
>>>>>
>>>>> On Tue, 5 Nov 2019 at 20:28, Nathanaël Blanchet <blanc...@abes.fr>
>>>>> wrote:
>>>>>
>>>>>>
>>>>>> Le 05/11/2019 à 18:22, Roy Golan a écrit :
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, 5 Nov 2019 at 19:12, Nathanaël Blanchet <blanc...@abes.fr>
>>>>>> wrote:
>>>>>>
>>>>>>>
>>>>>>> Le 05/11/2019 à 13:54, Roy Golan a écrit :
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Tue, 5 Nov 2019 at 14:52, Nathanaël Blanchet <blanc...@abes.fr>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I tried openshift-install after compiling but no ovirt provider is
>>>>>>>> available... So waht do you mean when you say "give a try"? Maybe only
>>>>>>>> provisionning ovirt with the terraform module?
>>>>>>>>
>>>>>>>> [root@vm5 installer]# bin/openshift-install create cluster
>>>>>>>> ? Platform  [Use arrows to move, space to select, type to filter, ?
>>>>>>>> for more help]
>>>>>>>> > aws
>>>>>>>>   azure
>>>>>>>>   gcp
>>>>>>>>   openstack
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>> Its not merged yet. Please pull this image and work with it as a
>>>>>>> container
>>>>>>> quay.io/rgolangh/openshift-installer
>>>>>>>
>>>>>>> A little feedback as you asked:
>>>>>>>
>>>>>>> [root@openshift-installer ~]# docker run -it 56e5b667100f create
>>>>>>> cluster
>>>>>>> ? Platform ovirt
>>>>>>> ? Enter oVirt's api endpoint URL
>>>>>>> https://air-dev.v100.abes.fr/ovirt-engine/api
>>>>>>> ? Enter ovirt-engine username admin@internal
>>>>>>> ? Enter password **********
>>>>>>> ? Pick the oVirt cluster Default
>>>>>>> ? Pick a VM template centos7.x
>>>>>>> ? Enter the internal API Virtual IP 10.34.212.200
>>>>>>> ? Enter the internal DNS Virtual IP 10.34.212.100
>>>>>>> ? Enter the ingress IP  10.34.212.50
>>>>>>> ? Base Domain oc4.localdomain
>>>>>>> ? Cluster Name test
>>>>>>> ? Pull Secret [? for help] *************************************
>>>>>>> INFO Creating infrastructure resources...
>>>>>>> INFO Waiting up to 30m0s for the Kubernetes API at
>>>>>>> https://api.test.oc4.localdomain:6443...
>>>>>>> ERROR Attempted to gather ClusterOperator status after installation
>>>>>>> failure: listing ClusterOperator objects: Get
>>>>>>> https://api.test.oc4.localdomain:6443/apis/config.openshift.io/v1/clusteroperators:
>>>>>>> dial tcp: lookup api.test.oc4.localdomain on 10.34.212.100:53: no
>>>>>>> such host
>>>>>>> INFO Pulling debug logs from the bootstrap machine
>>>>>>> ERROR Attempted to gather debug logs after installation failure:
>>>>>>> failed to create SSH client, ensure the proper ssh key is in your 
>>>>>>> keyring
>>>>>>> or specify with --key: failed to initialize the SSH agent: failed to 
>>>>>>> read
>>>>>>> directory "/output/.ssh": open /output/.ssh: no such file or directory
>>>>>>> FATAL Bootstrap failed to complete: waiting for Kubernetes API:
>>>>>>> context deadline exceeded
>>>>>>>
>>>>>>>    - 6 vms are successfully created thin dependent from the template
>>>>>>>
>>>>>>>
>>>>>>>    - each vm is provisionned by cloud-init
>>>>>>>    - the step "INFO Waiting up to 30m0s for the Kubernetes API at
>>>>>>>    https://api.test.oc4.localdomain:6443..."; fails. It seems that
>>>>>>>    the DNS pod is not up at this time.
>>>>>>>    - Right this moment, there is no more visibility on what is
>>>>>>>    done, what goes wrong... what's happening there? supposing a kind of
>>>>>>>    playbook downloading a kind of images...
>>>>>>>    - The" pull secret step" is not clear: we must have a redhat
>>>>>>>    account to https://cloud.redhat.com/openshift/install/ to get a
>>>>>>>    key like
>>>>>>>    -
>>>>>>>    {"auths":{"cloud.openshift.com
>>>>>>>    
>>>>>>> ":{"auth":"b3BlbnNoaWZ0LXJlbGVhc2UtZGV2K2V4cGxvaXRhYmVzZnIxdGN0ZnR0dmFnMHpuazMxd2IwMnIwenV1MDg6TE9XVzFQODM1NzNJWlI4MlZDSUEyTFdEVlJJS0U5VTVWM0NTSUdOWjJH********************==","email":
>>>>>>>    "expl...@abes.fr" <expl...@abes.fr>},"quay.io
>>>>>>>    
>>>>>>> ":{"auth":"b3BlbnNoaWZ0LXJlbGVhc2UtZGV2K2V4cGxvaXRhYmVzZnIxdGN0ZnR0dmFnMHpuazMxd2IwMnIwenV1MDg6TE9XVzFQODM1NzNJWlI4MlZDSUEyTFdEVlJJS0U5VTVWM0NTSUdOWjJH********************==","email":
>>>>>>>    "expl...@abes.fr" <expl...@abes.fr>},"registry.connect.redhat.com
>>>>>>>    
>>>>>>> ":{"auth":"NTI0MjkwMnx1aGMtMVRDVEZUVFZBRzBaTkszMXdCMDJSMFp1VTA4OmV5SmhiR2NpT2lKU1V6VXhNaUo5LmV5SnpkV0lpT2lJMk4ySTJNREV3WXpObE1HSTBNbVE0T1RGbVpUZGxa**********************","email":
>>>>>>>    "expl...@abes.fr" <expl...@abes.fr>},"registry.redhat.io
>>>>>>>    
>>>>>>> ":{"auth":"NTI0MjkwMnx1aGMtMVRDVEZUVFZBRzBaTkszMXdCMDJSMFp1VTA4OmV5SmhiR2NpT2lKU1V6VXhNaUo5LmV5SnpkV0lpT2lJMk4ySTJNREV3WXpObE1HSTBNbVE0T1RGbVpUZGxa**********************","email":
>>>>>>>    "expl...@abes.fr" <expl...@abes.fr>}}}
>>>>>>>
>>>>>>>
>>>>>>> Can you tell me if I'm doing wrong?
>>>>>>>
>>>>>>
>>>>>> What is the template you are using? I don't think its RHCOS(Red Hat
>>>>>> CoreOs) template, it looks like Centos?
>>>>>>
>>>>>> Use this gist to import the template
>>>>>> https://gist.github.com/rgolangh/adccf6d6b5eaecaebe0b0aeba9d3331b
>>>>>>
>>>>>> Unfortunately, the result is the same with the RHCOS template...
>>>>>>
>>>>>
>>>>> Make sure that:
>>>>> - the IPs supplied are taken, and belong to the VM network of those
>>>>> master VMs
>>>>> - localdomain or local domain suffix shouldn't be used
>>>>> - your ovirt-engine is version 4.3.7 or master
>>>>>
>>>>> I didn't mention that you can provide any domain name, even
>>>> non-existing.
>>>> When the bootstrap phase will be done, the instllation will teardown
>>>> the bootsrap mahchine.
>>>> At this stage if you are using a non-existing domain you would need to
>>>> add the DNS Virtual IP
>>>> you provided to your resolv.conf so the installation could resolve
>>>> api.$CLUSTER_NAME.$CLUSTER_DOMAIN.
>>>>
>>>> Also, you have a log under your $INSTALL_DIR/.openshift_install.log
>>>>
>>>> I tried several things with your advices, but I'm still stuck  at the
>>>> https://api.test.oc4.localdomain:6443/version?timeout=32s test
>>>>
>>>> with logs:
>>>>
>>>> time="2019-11-06T20:21:15Z" level=debug msg="Still waiting for the
>>>> Kubernetes API: the server could not find the requested resource"
>>>>
>>>> So it means DNS resolution and network are now good and ignition
>>>> provisionning is is OK but something goes wrong with the bootstrap vm.
>>>>
>>>> Now if I log into the bootstrap vm, I can see a selinux message, but it
>>>> may be not relevant...
>>>>
>>>> SELinux: mount invalid. Same Superblock, different security settings
>>>> for (dev nqueue, type nqueue).
>>>>
>>>> Some other cluewWith journalctl:
>>>>
>>>> journalctl -b -f -u bootkube
>>>>
>>>> Nov 06 21:55:40 localhost bootkube.sh[2101]:
>>>> {"level":"warn","ts":"2019-11-06T21:55:40.661Z","caller":"clientv3/retry_interceptor.go:61","msg":"retrying
>>>> of unary invoker
>>>> failed","target":"endpoint://client-7beef51d-daad-4b46-9497-8e135e528f7c/etcd-1.test.oc4.localdomain:2379","attempt":0,"error":"rpc
>>>> error: code = DeadlineExceeded desc = latest connection error: connection
>>>> error: desc = \"transport: Error while dialing dial tcp: lookup
>>>> etcd-1.test.oc4.localdomain on 10.34.212.101:53: no such host\""}
>>>> Nov 06 21:55:40 localhost bootkube.sh[2101]:
>>>> {"level":"warn","ts":"2019-11-06T21:55:40.662Z","caller":"clientv3/retry_interceptor.go:61","msg":"retrying
>>>> of unary invoker
>>>> failed","target":"endpoint://client-03992fc6-5a87-4160-9b87-44ec6e82f7cd/etcd-2.test.oc4.localdomain:2379","attempt":0,"error":"rpc
>>>> error: code = DeadlineExceeded desc = latest connection error: connection
>>>> error: desc = \"transport: Error while dialing dial tcp: lookup
>>>> etcd-2.test.oc4.localdomain on 10.34.212.101:53: no such host\""}
>>>> Nov 06 21:55:40 localhost bootkube.sh[2101]:
>>>> {"level":"warn","ts":"2019-11-06T21:55:40.662Z","caller":"clientv3/retry_interceptor.go:61","msg":"retrying
>>>> of unary invoker
>>>> failed","target":"endpoint://client-00db28a7-5188-4666-896b-e37c88ad3ae9/etcd-0.test.oc4.localdomain:2379","attempt":0,"error":"rpc
>>>> error: code = DeadlineExceeded desc = latest connection error: connection
>>>> error: desc = \"transport: Error while dialing dial tcp: lookup
>>>> etcd-0.test.oc4.localdomain on 10.34.212.101:53: no such host\""}
>>>> Nov 06 21:55:40 localhost bootkube.sh[2101]:
>>>> https://etcd-1.test.oc4.localdomain:2379 is unhealthy: failed to
>>>> commit proposal: context deadline exceeded
>>>> Nov 06 21:55:40 localhost bootkube.sh[2101]:
>>>> https://etcd-2.test.oc4.localdomain:2379 is unhealthy: failed to
>>>> commit proposal: context deadline exceeded
>>>> Nov 06 21:55:40 localhost bootkube.sh[2101]:
>>>> https://etcd-0.test.oc4.localdomain:2379 is unhealthy: failed to
>>>> commit proposal: context deadline exceeded
>>>> Nov 06 21:55:40 localhost bootkube.sh[2101]: Error: unhealthy cluster
>>>> Nov 06 21:55:40 localhost podman[61210]: 2019-11-06 21:55:40.720514151
>>>> +0000 UTC m=+5.813853296 container died
>>>> 7db3014e3f19c61775bac2a7a155eeb8521a6b78fea0d512384dd965cb0b8b01 (image=
>>>> registry.svc.ci.openshift.org/origin/4.3-2019-10-29-180250@sha256:696a0ba7d344e625ec71d915d74d387fc2a951b879d4d54bdc69d460724c01ae,
>>>> name=etcdctl)
>>>> Nov 06 21:55:40 localhost podman[61210]: 2019-11-06 21:55:40.817475095
>>>> +0000 UTC m=+5.910814273 container remove
>>>> 7db3014e3f19c61775bac2a7a155eeb8521a6b78fea0d512384dd965cb0b8b01 (image=
>>>> registry.svc.ci.openshift.org/origin/4.3-2019-10-29-180250@sha256:696a0ba7d344e625ec71d915d74d387fc2a951b879d4d54bdc69d460724c01ae,
>>>> name=etcdctl)
>>>> Nov 06 21:55:40 localhost bootkube.sh[2101]: etcdctl failed. Retrying
>>>> in 5 seconds...
>>>>
>>>> It seems to be again a dns resolution issue.
>>>>
>>>> [user1@localhost ~]$ dig api.test.oc4.localdomain +short
>>>> 10.34.212.201
>>>>
>>>> [user1@localhost ~]$ dig etcd-2.test.oc4.localdomain +short
>>>> nothing
>>>>
>>>>
>>>> So what do you think about that?
>>>>
>>>>
>>>> Key here is the masters - they need to boot, get ignition from the
>>> bootstrap machine and start publishing their IPs and hostnames.
>>>
>>> Connect to a master, check its hostname, check its running or failing
>>> containers `crictl ps -a` by root user.
>>>
>>> You were right:
>>> # crictl ps -a
>>> CONTAINER ID
>>> IMAGE
>>> CREATED             STATE               NAME
>>> ATTEMPT             POD ID
>>> 744cb8e654705
>>> e77034cf36baff5e625acbba15331db68e1d84571f977d254fd833341158daa8
>>> 4 minutes ago       Running             discovery
>>> 75                  9462e9a8ca478
>>> 912ba9db736c3
>>> e77034cf36baff5e625acbba15331db68e1d84571f977d254fd833341158daa8
>>> 14 minutes ago      Exited              discovery
>>> 74                  9462e9a8ca478
>>>
>>> # crictl logs 744cb8e654705
>>> E1107 08:10:04.262330       1 run.go:67] error looking up self for
>>> candidate IP 10.34.212.227: lookup
>>> _etcd-server-ssl._tcp.test.oc4.localdomain on 10.34.212.51:53: no such
>>> host
>>>
>>> # hostname
>>> localhost
>>>
>>> Conclusion: discovery didn't publish IPs and hostname to coreDNS because
>>> the master didn't get its name master-0.test.oc4.localdomain during
>>> provisionning phase.
>>>
>>> I changed the master-0 hostname and reinitiates ignition to verify:
>>>
>>> # hostnamectl set-hostname master-0.test.oc4.localdomain
>>>
>>> # touch /boot/ignition.firstboot && rm -rf /etc/machine-id && reboot
>>>
>>> After reboot is completed, no more exited discovery container:
>>>
>>> CONTAINER ID
>>> IMAGE
>>> CREATED             STATE               NAME
>>> ATTEMPT             POD ID
>>> e701efa8bc583
>>> 77ec5e26cc676ef2bf5c42dd40e55394a11fb45a3e2d7e95cbaf233a1eef472f
>>> 20 seconds ago      Running             coredns
>>> 1                   cbabc53322ac8
>>> 2c7bc6abb5b65
>>> d73eca122bd567a3a1f70fa5021683bc17dd87003d05d88b1cdd0215c55049f6
>>> 20 seconds ago      Running             mdns-publisher
>>> 1                   6f8914ff9db35
>>> b3f619d5afa2c
>>> 7aa184de043265814f9a775968234ac3280a285056da773f1aba0917e9615370
>>> 21 seconds ago      Running             haproxy-monitor
>>> 1                   0e5c209496787
>>> 07769ce79b032
>>> 7aa184de043265814f9a775968234ac3280a285056da773f1aba0917e9615370
>>> 21 seconds ago      Running             keepalived-monitor
>>> 1                   02cf141d01a29
>>> fb20d66b81254
>>> e77034cf36baff5e625acbba15331db68e1d84571f977d254fd833341158daa8
>>> 21 seconds ago      Running             discovery
>>> 77                  562f32067e0a7
>>> 476b07599260e
>>> 86a34bc5edd3e70073313f97bfd51ed8937658b341dc52334fb98ea6896ebdc2
>>> 22 seconds ago      Running             haproxy
>>> 1                   0e5c209496787
>>> 26b53050a412b
>>> 9f94e500f85a735ec212ffb7305e0b63f7151a5346e41c2d5d293c8456f6fa42
>>> 22 seconds ago      Running             keepalived
>>> 1                   02cf141d01a29
>>> 30ce48453854b
>>> 7aa184de043265814f9a775968234ac3280a285056da773f1aba0917e9615370
>>> 22 seconds ago      Exited              render-config
>>> 1                   cbabc53322ac8
>>> ad3ab0ae52077
>>> 7aa184de043265814f9a775968234ac3280a285056da773f1aba0917e9615370
>>> 22 seconds ago      Exited              render-config
>>> 1                   6f8914ff9db35
>>> 650d62765e9e1
>>> registry.svc.ci.openshift.org/origin/4.3-2019-10-29-180250@sha256:9a7e8297f7b70ee3a11fcbe4a78c59a5861e1afda5657a7437de6934bdc2458e
>>> 13 hours ago        Exited              coredns
>>> 0                   2ae0512b3b6ac
>>> 481969ce49bb9
>>> registry.svc.ci.openshift.org/origin/4.3-2019-10-29-180250@sha256:768194132b4dbde077e32de8801c952265643da00ae161f1ee560fabf6ed1f8e
>>> 13 hours ago        Exited              mdns-publisher
>>> 0                   d49754042b792
>>> 3594d9d261ca7
>>> registry.svc.ci.openshift.org/origin/4.3-2019-10-29-180250@sha256:8c3b0223699801a9cb93246276da4746fa4d6fa66649929b2d9b702c17dac75d
>>> 13 hours ago        Exited              haproxy-monitor
>>> 0                   3476219058ba8
>>> 88b13ec02a5c1
>>> 7aa184de043265814f9a775968234ac3280a285056da773f1aba0917e9615370
>>> 13 hours ago        Exited              keepalived-monitor
>>> 0                   a3e13cf07c04f
>>> 1ab721b5599ed
>>> registry.svc.ci.openshift.org/origin/4.3-2019-10-29-180250@sha256:629d73fdd28beafda788d2248608f90c7ed048357e250f3e855b9462b92cfe60
>>> 13 hours ago
>>>
>>> because DNS registration is OK:
>>>
>>> [user1@master-0 ~]$ dig etcd-0.test.oc4.localdomain +short
>>> 10.34.212.227
>>>
>>> CONCLUSION:
>>>
>>>    - none of rhcos vm is correctly provisionned to their targeted
>>>    hostname, so they all stay with localhost.
>>>
>>>
>> What is your engine version? the hostname support for ignition is merged
>> into 4.3.7 and master
>>
>> 4.3.7.1-1.el7
>>
> https://gerrit.ovirt.org/c/100397/ merged 2 days ago, so it will apear in
> 4.3.7.2.
>
> Sandro when is 4.7.3.2 is due?
>

You can also use the nightly 4.3 snapshot - it's not really nightly anymore
- it's updated per every run of CI Change-Queue, IIUC:

https://www.ovirt.org/develop/dev-process/install-nightly-snapshot.html


>
> I only upgraded engine and not vdsm on hosts, but I suppose hosts are not
>> important for ignition
>>
> Correct.
>
>>
>>
>>>    - Cloud-init syntax for the hostname is ok, but it is not
>>>    provisioned by ignition:
>>>
>>> Why not provisionning these hostnames with a json snippet or else?
>>>
>>> {
>>>   "ignition": { "version": "2.2.0" },
>>>   "storage": {
>>>     "files": [{
>>>       "filesystem": "root",
>>>       "path": "/etc/hostname",
>>>       "mode": 420,
>>>       "contents": { "source": "data:,master-0.test.oc4.localdomain" }
>>>     }]
>>>   }}
>>>
>>>
>>>
>>>
>>>
>>>>
>>>>>
>>>>>>
>>>>>>
>>>>>> Le 05/11/2019 à 12:24, Roy Golan a écrit :
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, 5 Nov 2019 at 13:22, Nathanaël Blanchet <blanc...@abes.fr>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hello,
>>>>>>>>>
>>>>>>>>> I'm interested by installing okd on ovirt with the official
>>>>>>>>> openshift
>>>>>>>>> installer (https://github.com/openshift/installer), but ovirt is
>>>>>>>>> not yet
>>>>>>>>> supported.
>>>>>>>>>
>>>>>>>>>
>>>>>>>> If you want to give a try and supply feedback I'll be glad.
>>>>>>>>
>>>>>>>>
>>>>>>>>> Regarding https://bugzilla.redhat.com/show_bug.cgi?id=1578255 and
>>>>>>>>>
>>>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/thread/EF7OQUVTY53GV3A7NVQVUT7UCUYKK5CH/
>>>>>>>>> , how ovirt 4.3.7 should integrate openshift installer integration
>>>>>>>>> with
>>>>>>>>> terraform?
>>>>>>>>>
>>>>>>>>>
>>>>>>>> Terraform is part of it, yes, It is what we use to spin the first 3
>>>>>>>> masters, plus a bootstraping machine.
>>>>>>>>
>>>>>>>> --
>>>>>>>>> Nathanaël Blanchet
>>>>>>>>>
>>>>>>>>> Supervision réseau
>>>>>>>>> Pôle Infrastrutures Informatiques
>>>>>>>>> 227 avenue Professeur-Jean-Louis-Viala
>>>>>>>>> 34193 MONTPELLIER CEDEX 5
>>>>>>>>> Tél. 33 (0)4 67 54 84 55
>>>>>>>>> Fax  33 (0)4 67 54 84 14
>>>>>>>>> blanc...@abes.fr
>>>>>>>>>
>>>>>>>>> --
>>>>>>>> Nathanaël Blanchet
>>>>>>>>
>>>>>>>> Supervision réseau
>>>>>>>> Pôle Infrastrutures Informatiques
>>>>>>>> 227 avenue Professeur-Jean-Louis-Viala
>>>>>>>> 34193 MONTPELLIER CEDEX 5      
>>>>>>>> Tél. 33 (0)4 67 54 84 55
>>>>>>>> Fax  33 (0)4 67 54 84 14blanc...@abes.fr
>>>>>>>>
>>>>>>>> --
>>>>>>> Nathanaël Blanchet
>>>>>>>
>>>>>>> Supervision réseau
>>>>>>> Pôle Infrastrutures Informatiques
>>>>>>> 227 avenue Professeur-Jean-Louis-Viala
>>>>>>> 34193 MONTPELLIER CEDEX 5       
>>>>>>> Tél. 33 (0)4 67 54 84 55
>>>>>>> Fax  33 (0)4 67 54 84 14blanc...@abes.fr
>>>>>>>
>>>>>>> --
>>>>>> Nathanaël Blanchet
>>>>>>
>>>>>> Supervision réseau
>>>>>> Pôle Infrastrutures Informatiques
>>>>>> 227 avenue Professeur-Jean-Louis-Viala
>>>>>> 34193 MONTPELLIER CEDEX 5        
>>>>>> Tél. 33 (0)4 67 54 84 55
>>>>>> Fax  33 (0)4 67 54 84 14blanc...@abes.fr
>>>>>>
>>>>>> --
>>>> Nathanaël Blanchet
>>>>
>>>> Supervision réseau
>>>> Pôle Infrastrutures Informatiques
>>>> 227 avenue Professeur-Jean-Louis-Viala
>>>> 34193 MONTPELLIER CEDEX 5  
>>>> Tél. 33 (0)4 67 54 84 55
>>>> Fax  33 (0)4 67 54 84 14blanc...@abes.fr
>>>>
>>>> --
>>> Nathanaël Blanchet
>>>
>>> Supervision réseau
>>> Pôle Infrastrutures Informatiques
>>> 227 avenue Professeur-Jean-Louis-Viala
>>> 34193 MONTPELLIER CEDEX 5   
>>> Tél. 33 (0)4 67 54 84 55
>>> Fax  33 (0)4 67 54 84 14blanc...@abes.fr
>>>
>>> --
>> Nathanaël Blanchet
>>
>> Supervision réseau
>> Pôle Infrastrutures Informatiques
>> 227 avenue Professeur-Jean-Louis-Viala
>> 34193 MONTPELLIER CEDEX 5    
>> Tél. 33 (0)4 67 54 84 55
>> Fax  33 (0)4 67 54 84 14blanc...@abes.fr
>>
>> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/GZ64UU7KYDYJMZW2BSPWVKLX3EXMYMYO/
>


-- 
Didi
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/IIHU2AIL4IIPTM6YYR5NFQIBLIOBHQDX/

Reply via email to