Something was definitely wrong ; as indicated, qemu process
for guest=HostedEngineLocal was running but the disk file did not exist
anymore...
No surprise I could not connect

I am retrying


Guillaume Pavese
Ingénieur Système et Réseau
Interactiv-Group


On Mon, Feb 25, 2019 at 11:15 PM Guillaume Pavese <
[email protected]> wrote:

> It fails too :
> I made sure PermitTunnel=yes in sshd config but when I try to connect to
> the forwarded port I get the following error on the openened host ssh
> session :
>
> [gpavese@sheepora-X230 ~]$ ssh -v -L 5900:
> vs-inf-int-kvm-fr-301-210.hostics.fr:5900
> [email protected]
> ...
> [root@vs-inf-int-kvm-fr-301-210 ~]#
> debug1: channel 3: free: direct-tcpip: listening port 5900 for
> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900, connect from ::1 port
> 42144 to ::1 port 5900, nchannels 4
> debug1: Connection to port 5900 forwarding to
> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900 requested.
> debug1: channel 3: new [direct-tcpip]
> channel 3: open failed: connect failed: Connection refused
> debug1: channel 3: free: direct-tcpip: listening port 5900 for
> vs-inf-int-kvm-fr-301-210.hostics.fr port 5900, connect from 127.0.0.1
> port 32778 to 127.0.0.1 port 5900, nchannels 4
>
>
> and in journalctl :
>
> févr. 25 14:55:38 vs-inf-int-kvm-fr-301-210.hostics.fr sshd[19595]:
> error: connect_to vs-inf-int-kvm-fr-301-210.hostics.fr port 5900: failed.
>
>
> Guillaume Pavese
> Ingénieur Système et Réseau
> Interactiv-Group
>
>
> On Mon, Feb 25, 2019 at 10:44 PM Simone Tiraboschi <[email protected]>
> wrote:
>
>>
>>
>>
>> On Mon, Feb 25, 2019 at 2:35 PM Guillaume Pavese <
>> [email protected]> wrote:
>>
>>> I made sure of everything and even stopped firewalld but still can't
>>> connect :
>>>
>>> [root@vs-inf-int-kvm-fr-301-210 ~]# cat
>>> /var/run/libvirt/qemu/HostedEngineLocal.xml
>>>      <graphics type='vnc' port='*5900*' autoport='yes'
>>> *listen='127.0.0.1*'>
>>>         <listen type='address' address='*127.0.0.1*' fromConfig='1'
>>> autoGenerated='no'/>
>>>
>>> [root@vs-inf-int-kvm-fr-301-210 ~]# netstat -pan | grep 59
>>> tcp        0      0 127.0.0.1:5900          0.0.0.0:*
>>>  LISTEN      13376/qemu-kvm
>>>
>>
>>
>> I suggest to try ssh tunneling, run
>> ssh -L 5900:vs-inf-int-kvm-fr-301-210.hostics.fr:5900
>> [email protected]
>>
>> and then
>> remote-viewer vnc://localhost:5900
>>
>>
>>
>>>
>>> [root@vs-inf-int-kvm-fr-301-210 ~]# systemctl status firewalld.service
>>> ● firewalld.service - firewalld - dynamic firewall daemon
>>>    Loaded: loaded (/usr/lib/systemd/system/firewalld.service; enabled;
>>> vendor preset: enabled)
>>>    *Active: inactive (dead)*
>>> *févr. 25 14:24:03 vs-inf-int-kvm-fr-301-210.hostics.fr
>>> <http://vs-inf-int-kvm-fr-301-210.hostics.fr> systemd[1]: Stopped firewalld
>>> - dynamic firewall daemon.*
>>>
>>> From my laptop :
>>> [gpavese@sheepora-X230 ~]$ telnet vs-inf-int-kvm-fr-301-210.hostics.fr
>>> *5900*
>>> Trying 10.199.210.11...
>>> [*nothing gets through...*]
>>> ^C
>>>
>>> For making sure :
>>> [gpavese@sheepora-X230 ~]$ telnet vs-inf-int-kvm-fr-301-210.hostics.fr
>>> *9090*
>>> Trying 10.199.210.11...
>>> *Connected* to vs-inf-int-kvm-fr-301-210.hostics.fr.
>>> Escape character is '^]'.
>>>
>>>
>>>
>>>
>>>
>>> Guillaume Pavese
>>> Ingénieur Système et Réseau
>>> Interactiv-Group
>>>
>>>
>>> On Mon, Feb 25, 2019 at 10:24 PM Parth Dhanjal <[email protected]>
>>> wrote:
>>>
>>>> Hey!
>>>>
>>>> You can check under /var/run/libvirt/qemu/HostedEngine.xml
>>>> Search for 'vnc'
>>>> From there you can look up the port on which the HE VM is available and
>>>> connect to the same.
>>>>
>>>>
>>>> On Mon, Feb 25, 2019 at 6:47 PM Guillaume Pavese <
>>>> [email protected]> wrote:
>>>>
>>>>> 1) I am running in a Nested env, but under libvirt/kvm on remote
>>>>> Centos 7.4 Hosts
>>>>>
>>>>> Please advise how to connect with VNC to the local HE vm. I see it's
>>>>> running, but this is on a remote host, not my local machine :
>>>>> qemu     13376  100  3.7 17679424 845216 ?     Sl   12:46  85:08
>>>>> /usr/libexec/qemu-kvm -name guest=HostedEngineLocal,debug-threads=on -S
>>>>> -object
>>>>> secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-HostedEngineLocal/master-key.aes
>>>>> -machine pc-i440fx-rhel7.6.0,accel=kvm,usb=off,dump-guest-core=off -cpu
>>>>> Haswell-noTSX,+kvmclock -m 16384 -realtime mlock=off -smp
>>>>> 4,sockets=4,cores=1,threads=1 -uuid 6fe7c1c3-ea93-4343-a385-0d9e14bb563a
>>>>> -no-user-config -nodefaults -chardev
>>>>> socket,id=charmonitor,fd=27,server,nowait -mon
>>>>> chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown
>>>>> -global PIIX4_PM.disable_s3=1 -global PIIX4_PM.disable_s4=1 -boot
>>>>> menu=off,strict=on -device
>>>>> virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive
>>>>> file=/var/tmp/localvmgmyYik/images/15023c8a-e3a7-4851-a97d-3b90996b423b/07fdcff3-11ce-4f7c-af05-0a878593e78e,format=qcow2,if=none,id=drive-virtio-disk0
>>>>> -device
>>>>> virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
>>>>> -drive
>>>>> file=/var/tmp/localvmgmyYik/seed.iso,format=raw,if=none,id=drive-ide0-0-0,readonly=on
>>>>> -device ide-cd,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0 -netdev
>>>>> tap,fd=29,id=hostnet0,vhost=on,vhostfd=30 -device
>>>>> virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:3e:fe:28,bus=pci.0,addr=0x3
>>>>> -chardev pty,id=charserial0 -device
>>>>> isa-serial,chardev=charserial0,id=serial0 -chardev
>>>>> socket,id=charchannel0,fd=31,server,nowait -device
>>>>> virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=org.qemu.guest_agent.0
>>>>> *-vnc 127.0.0.1:0 <http://127.0.0.1:0> -device 
>>>>> VGA*,id=video0,vgamem_mb=16,bus=pci.0,addr=0x2
>>>>> -object rng-random,id=objrng0,filename=/dev/random -device
>>>>> virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x6 -sandbox
>>>>> on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny
>>>>> -msg timestamp=on
>>>>>
>>>>>
>>>>> 2) [root@vs-inf-int-kvm-fr-301-210 ~]# cat
>>>>> /etc/libvirt/qemu/networks/default.xml
>>>>> <!--
>>>>> WARNING: THIS IS AN AUTO-GENERATED FILE. CHANGES TO IT ARE LIKELY TO BE
>>>>> OVERWRITTEN AND LOST. Changes to this xml configuration should be made
>>>>> using:
>>>>>   virsh net-edit default
>>>>> or other application using the libvirt API.
>>>>> -->
>>>>>
>>>>> <network>
>>>>>   <name>default</name>
>>>>>   <uuid>ba7bbfc8-28b8-459e-a42d-c2d6218e2cb6</uuid>
>>>>>   <forward mode='nat'/>
>>>>>   <bridge name='virbr0' stp='on' delay='0'/>
>>>>>   <mac address='52:54:00:e5:fe:3b'/>
>>>>>   <ip address='192.168.122.1' netmask='255.255.255.0'>
>>>>>     <dhcp>
>>>>>       <range start='192.168.122.2' end='192.168.122.254'/>
>>>>>     </dhcp>
>>>>>   </ip>
>>>>> </network>
>>>>> You have new mail in /var/spool/mail/root
>>>>> [root@vs-inf-int-kvm-fr-301-210 ~]
>>>>>
>>>>>
>>>>>
>>>>> Guillaume Pavese
>>>>> Ingénieur Système et Réseau
>>>>> Interactiv-Group
>>>>>
>>>>>
>>>>> On Mon, Feb 25, 2019 at 9:57 PM Simone Tiraboschi <[email protected]>
>>>>> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Mon, Feb 25, 2019 at 1:14 PM Guillaume Pavese <
>>>>>> [email protected]> wrote:
>>>>>>
>>>>>>> He deployment with "hosted-engine --deploy" fails at TASK
>>>>>>> [ovirt.hosted_engine_setup : Get local VM IP]
>>>>>>>
>>>>>>> See following Error :
>>>>>>>
>>>>>>> 2019-02-25 12:46:50,154+0100 INFO
>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils
>>>>>>> ansible_utils._process_output:109 TASK [ovirt.hosted_engine_setup : Get
>>>>>>> local VM IP]
>>>>>>> 2019-02-25 12:55:26,823+0100 DEBUG
>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils
>>>>>>> ansible_utils._process_output:103 {u'_ansible_parsed': True,
>>>>>>> u'stderr_lines': [], u'cmd': u"virsh -r net-dhcp-leases default | grep 
>>>>>>> -i 00
>>>>>>> :16:3e:3e:fe:28 | awk '{ print $5 }' | cut -f1 -d'/'", u'end':
>>>>>>> u'2019-02-25 12:55:26.666925', u'_ansible_no_log': False, u'stdout': 
>>>>>>> u'',
>>>>>>> u'changed': True, u'invocation': {u'module_args': {u'warn': True,
>>>>>>> u'executable':
>>>>>>> None, u'_uses_shell': True, u'_raw_params': u"virsh -r
>>>>>>> net-dhcp-leases default | grep -i 00:16:3e:3e:fe:28 | awk '{ print $5 
>>>>>>> }' |
>>>>>>> cut -f1 -d'/'", u'removes': None, u'argv': None, u'creates': None,
>>>>>>> u'chdir': None, u'std
>>>>>>> in': None}}, u'start': u'2019-02-25 12:55:26.584686', u'attempts':
>>>>>>> 50, u'stderr': u'', u'rc': 0, u'delta': u'0:00:00.082239', 
>>>>>>> u'stdout_lines':
>>>>>>> []}
>>>>>>> 2019-02-25 12:55:26,924+0100 ERROR
>>>>>>> otopi.ovirt_hosted_engine_setup.ansible_utils
>>>>>>> ansible_utils._process_output:107 fatal: [localhost]: FAILED! =>
>>>>>>> {"attempts": 50, "changed": true, "cmd": "virsh -r net-dhcp-leases 
>>>>>>> default
>>>>>>> | grep -i 00:16:3e:3e:fe:28 | awk '{ print $5 }' | cut -f1 -d'/'", 
>>>>>>> "delta":
>>>>>>> "0:00:00.082239", "end": "2019-02-25 12:55:26.666925", "rc": 0, "start":
>>>>>>> "2019-02-25 12:55:26.584686", "stderr": "", "stderr_lines": [], 
>>>>>>> "stdout":
>>>>>>> "", "stdout_lines": []}
>>>>>>>
>>>>>>
>>>>>> Here we are just waiting for the bootstrap engine VM to fetch an IP
>>>>>> address from default libvirt network over DHCP but it your case it never
>>>>>> happened.
>>>>>> Possible issues: something went wrong in the bootstrap process for
>>>>>> the engine VM or the default libvirt network is not correctly configured.
>>>>>>
>>>>>> 1. can you try to reach the engine VM via VNC and check what's
>>>>>> happening there? (another question, are you running it nested? AFAIK it
>>>>>> will not work if nested over ESXi)
>>>>>> 2. can you please share the output of
>>>>>> cat /etc/libvirt/qemu/networks/default.xml
>>>>>>
>>>>>>
>>>>>>>
>>>>>>> Guillaume Pavese
>>>>>>> Ingénieur Système et Réseau
>>>>>>> Interactiv-Group
>>>>>>> _______________________________________________
>>>>>>> Users mailing list -- [email protected]
>>>>>>> To unsubscribe send an email to [email protected]
>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>>> oVirt Code of Conduct:
>>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>>> List Archives:
>>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/VXRMU3SQWTMB2YYNMOMD7I5NX7RZQ2IW/
>>>>>>>
>>>>>> _______________________________________________
>>>>> Users mailing list -- [email protected]
>>>>> To unsubscribe send an email to [email protected]
>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>> oVirt Code of Conduct:
>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>> List Archives:
>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/45UR44ITQTV7YVE6GLJWEV2AV22TIGXK/
>>>>>
>>>>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/TKHBH7AZH57BQHL2LDKNWYMZTBONRK7W/

Reply via email to