[ovirt-users] Can not connect to gluster storage
Hello, I ve a host that can not connet to gluster storage. It has worked since I ve set up the environment, and today it stoped working this are the error messages in the webui The error message for connection kvm380.durchhalten.intern:/data returned by VDSM was: Failed to fetch Gluster Volume List Failed to connect Host kvm380.durchhalten.intern to the Storage Domains data. Failed to connect Host kvm380.durchhalten.intern to the Storage Domains hosted_storage. and here the vdsm.log StorageDomainDoesNotExist: Storage domain does not exist: (u'36663740-576a-4498-b28e-0a402628c6a7',) 2020-11-27 12:59:07,665+ INFO (jsonrpc/2) [storage.TaskManager.Task] (Task='8bed48b8-0696-4d3f-966a-119219f3b013') aborting: Task is aborted: "Storage domain does not exist: (u'36663740-576a-4498-b28e-0a402628c6a7',)" - code 358 (task:1181) 2020-11-27 12:59:07,665+ ERROR (jsonrpc/2) [storage.Dispatcher] FINISH getStorageDomainInfo error=Storage domain does not exist: (u'36663740-576a-4498-b28e-0a402628c6a7',) (dispatcher:83) 2020-11-27 12:59:07,666+ INFO (jsonrpc/2) [jsonrpc.JsonRpcServer] RPC call StorageDomain.getInfo failed (error 358) in 0.38 seconds (__init__:312) 2020-11-27 12:59:07,698+ INFO (jsonrpc/7) [vdsm.api] START connectStorageServer(domType=7, spUUID=u'----', conList=[{u'id': u'e29cf818-5ee5-46e1-85c1-8aeefa33e95d', u'vfs_type': u'glusterfs', u'connection': u'kvm380.durchhalten.intern:/engine', u'user': u'kvm'}], options=None) from=::1,40964, task_id=3a3eeb80-50ef-4710-a4f4-9d35da2ff281 (api:48) 2020-11-27 12:59:07,871+ ERROR (jsonrpc/7) [storage.HSM] Could not connect to storageServer (hsm:2420) Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 2417, in connectStorageServer conObj.connect() File "/usr/lib/python2.7/site-packages/vdsm/storage/storageServer.py", line 167, in connect self.validate() File "/usr/lib/python2.7/site-packages/vdsm/storage/storageServer.py", line 297, in validate if not self.volinfo: File "/usr/lib/python2.7/site-packages/vdsm/storage/storageServer.py", line 284, in volinfo self._volinfo = self._get_gluster_volinfo() File "/usr/lib/python2.7/site-packages/vdsm/storage/storageServer.py", line 329, in _get_gluster_volinfo self._volfileserver) File "/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 56, in __call__ return callMethod() File "/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 54, in **kwargs) File "", line 2, in glusterVolumeInfo File "/usr/lib64/python2.7/multiprocessing/managers.py", line 773, in _callmethod raise convert_to_error(kind, result) GlusterVolumesListFailedException: Volume list failed: rc=30806 out=() err=['Volume does not exist'] 2020-11-27 12:59:07,871+ INFO (jsonrpc/7) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'status': 4149, 'id': u'e29cf818-5ee5-46e1-85c1-8aeefa33e95d'}]} from=::1,40964, task_id=3a3eeb80-50ef-4710-a4f4-9d35da2ff281 (api:54) 2020-11-27 12:59:07,871+ INFO (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC call StoragePool.connectStorageServer succeeded in 0.18 seconds (__init__:312) 2020-11-27 12:59:08,474+ INFO (Reactor thread) [ProtocolDetector.AcceptorImpl] Accepted connection from ::1:40966 (protocoldetector:61) 2020-11-27 12:59:08,484+ INFO (Reactor thread) [ProtocolDetector.Detector] Detected protocol stomp from ::1:40966 (protocoldetector:125) 2020-11-27 12:59:08,484+ INFO (Reactor thread) [Broker.StompAdapter] Processing CONNECT request (stompserver:95) 2020-11-27 12:59:08,485+ INFO (JsonRpc (StompReactor)) [Broker.StompAdapter] Subscribe command received (stompserver:124) 2020-11-27 12:59:08,525+ INFO (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call Host.ping2 succeeded in 0.00 seconds (__init__:312) 2020-11-27 12:59:08,529+ INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call Host.ping2 succeeded in 0.00 seconds (__init__:312) 2020-11-27 12:59:08,533+ INFO (jsonrpc/6) [vdsm.api] START getStorageDomainInfo(sdUUID=u'36663740-576a-4498-b28e-0a402628c6a7', options=None) from=::1,40966, task_id=ee3ac98e-6a93-4cb2-a626-5533c8fb78ad (api:48) 2020-11-27 12:59:08,909+ INFO (jsonrpc/6) [vdsm.api] FINISH getStorageDomainInfo error=Storage domain does not exist: (u'36663740-576a-4498-b28e-0a402628c6a7',) from=::1,40966, task_id=ee3ac98e-6a93-4cb2-a626-5533c8fb78ad (api:52) 2020-11-27 12:59:08,910+ ERROR (jsonrpc/6) [storage.TaskManager.Task] (Task='ee3ac98e-6a93-4cb2-a626-5533c8fb78ad') Unexpected error (task:875) Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, in _run return fn(*args, **kargs) File "", line 2, in getStorageDomainInfo File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File
[ovirt-users] Re: How to Backup a VM
I think, I found the problem. It is case sensitive. For the export it is NOT case sensitive but for the step "wait for export" it is. I ve changed it and now it seems to be working ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/RYFMBHZTJF76RT56HWUK5EV3ETB5QCSV/
[ovirt-users] Re: How to Backup a VM
OK, I ve run the backup three times . I still have two machines, where it still fails on TASK [Wait for export] I think the Problem is not the timeout, in oVirt engine the export has already finished : " Exporting VM VMName as an OVA to /home/backup/in_progress/VMName.ova on Host kvm360" But [Wait for export] still counts to 1 exit with error and move on to the next task bye shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/W65G6ZUL6C6UJAJI627WVGITGIUUJ2XZ/
[ovirt-users] Re: How to Backup a VM
yes you are right, I ve already found. But this was not realy my problem. It causes from the HostedEngine. Long time ago I ve decreased the memory. It seems that this was the problem. now it is seems to be working pretty well. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/KVTFZHUV7URK6GBEEO5OE2BDYIFPRJ5F/
[ovirt-users] Re: How to Backup a VM
Hello, >https://github.com/silverorange/ovirt_ansible_backup I am also still using 4.3. In my opinion this is by far the best and easiest solution for disaster recovery. No need to install an appliance, and if there is a need to recover, you can import the ova in every hypervisor - no databases, no dependency. Sometimes I ve issues with "TASK [Wait for export] " sometime it takes to long to export the ova. an I also had the problem, that the export already finished, but it was not realized by the script. In ovirt the export was finished and the filename was renamed from *.tmp to *.ova maybe you have an idea for me. thanks bye ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/Q7TKVK5TL6HT7DQZCY354ICK5J3JRDH4/
[ovirt-users] How to Backup a VM
Hello to all I try to backup a normal VM. But It seems that I don't really understand the concept. At first I found the possibility to backup with the api https://www.ovirt.org/documentation/administration_guide/#Setting_a_storage_domain_to_be_a_backup_domain_backup_domain. Create a snapshot of the VM, finding the ID of the snapshot and the configuration of the VM makes sense to me. But at this point, I would download the config an the snapshot and put it to my backup storage. And not create a new VM attach the disk and run a backup with backup programm. And for restoring do the sam way backwards. If i look at other project, there seems do be a way to download the snapshot and configfile, or am I wrong? Maybe someone can explain to me why I should use additional software to install in an additional machine. Or even better someone can explain to me how I don't have to use additional backup software. And to the same topic backup. There is in the documentation the possibility to set up a backup storage It is nearly the same, create a snapshot, or clone the machine and export it to backup storage > Export the new virtual machine to a backup domain. See Exporting a Virtual > Machine to a Data Domain in the Virtual Machine Management Guide. Sadly there is just writen what to do, not how, the link points to 404 page. maybe someone can explain to me how to use backup storage thank you very much shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/COR6VIV477XUFDKJAVEO2ODCESVENKLV/
[ovirt-users] Re: oVirt 4.4.1 HCI single server deployment failed nested-kvm
Hello, I ve the same problem, I ve already set up glusterfs and like to deplay a self hosted engine. I ve allready ovirt self hosted engine deployed, with no problems. But here I get the same error with hosted-engine --deploy and in the web frontend. [ INFO ] TASK [ovirt.hosted_engine_setup : Failed if only teaming devices are availible] [ ERROR ] fatal: [localhost]: FAILED! => {"msg": "The conditional check '(otopi_host_net.ansible_facts.otopi_host_net | length == 0)' failed. The error was: error while evaluating conditional ((otopi_host_net.ansible_facts.otopi_host_net | length == 0)): 'list object' has no attribute 'ansible_facts'\n\nThe error appears to be in '/usr/share/ansible/roles/ovirt.hosted_engine_setup/tasks/filter_team_devices.yml': line 29, column 13, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\nThe offending line appears to be:\n\n- debug: var=otopi_host_net\n^ here\n\nThere appears to be both 'k=v' shorthand syntax and YAML in this task. Only one syntax may be used.\n"} [ ERROR ] Failed to execute stage 'Environment customization': Failed executing ansible-playbook is there a solution or just use an older ovirt `node iso? bye shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/L2T6ENJRXFLQNJ6DMQ5H4WWV2TJLJJCE/
[ovirt-users] Host after update NonResponsive
Hello to all, I ve done a normal host updadet at the webfrontend. After the reboot the host is NonResponsive. It is a HCI setup with glusterfs But mount shows on this host that only engine is mountet. The data volume is not mounted. [root@kvm320 ~]# mount|grep _engine kvm380.durchhalten.intern:/engine on /rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine type fuse.glusterfs (rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072) [root@kvm320 ~]# mount|grep _data [root@kvm320 ~]# # on every other host it looks like this [root@kvm10 ~]# mount|grep _engine kvm380.durchhalten.intern:/engine on /rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine type fuse.glusterfs (rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072) [root@kvm10 ~]# mount|grep _data kvm380.durchhalten.intern:/data on /rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_data type fuse.glusterfs (rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072) [root@kvm10 ~]# here so more informaton [root@kvm320 ~]# systemctl status glusterd -l ● glusterd.service - GlusterFS, a clustered file-system server Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled; vendor preset: disabled) Active: active (running) since Fr 2020-03-13 14:19:03 CET; 1h 49min ago Docs: man:glusterd(8) Process: 9263 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid --log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS) Main PID: 9264 (glusterd) Tasks: 114 CGroup: /system.slice/glusterd.service ├─3839 /usr/sbin/glusterfsd -s kvm320.durchhalten.intern --volfile-id data.kvm320.durchhalten.intern.gluster_bricks-data -p /var/run/gluster/vols/data/kvm320.durchhalten.intern-gluster_bricks-data.pid -S /var/run/gluster/1fd58e7c80335308.socket --brick-name /gluster_bricks/data -l /var/log/glusterfs/bricks/gluster_bricks-data.log --xlator-option *-posix.glusterd-uuid=ce474774-436a-41d3-bfdd-ab153ac77830 --process-name brick --brick-port 49152 --xlator-option data-server.listen-port=49152 ├─3896 /usr/sbin/glusterfsd -s kvm320.durchhalten.intern --volfile-id engine.kvm320.durchhalten.intern.gluster_bricks-engine -p /var/run/gluster/vols/engine/kvm320.durchhalten.intern-gluster_bricks-engine.pid -S /var/run/gluster/5d4bcd552e3a3806.socket --brick-name /gluster_bricks/engine -l /var/log/glusterfs/bricks/gluster_bricks-engine.log --xlator-option *-posix.glusterd-uuid=ce474774-436a-41d3-bfdd-ab153ac77830 --process-name brick --brick-port 49153 --xlator-option engine-server.listen-port=49153 ├─4032 /usr/sbin/glusterfsd -s kvm320.durchhalten.intern --volfile-id home.kvm320.durchhalten.intern.gluster_bricks-home -p /var/run/gluster/vols/home/kvm320.durchhalten.intern-gluster_bricks-home.pid -S /var/run/gluster/050dbbce51bc7cb8.socket --brick-name /gluster_bricks/home -l /var/log/glusterfs/bricks/gluster_bricks-home.log --xlator-option *-posix.glusterd-uuid=ce474774-436a-41d3-bfdd-ab153ac77830 --process-name brick --brick-port 49154 --xlator-option home-server.listen-port=49154 ├─9264 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level INFO └─9381 /usr/sbin/glusterfs -s localhost --volfile-id gluster/glustershd -p /var/run/gluster/glustershd/glustershd.pid -l /var/log/glusterfs/glustershd.log -S /var/run/gluster/6b30e1d260d31419.socket --xlator-option *replicate*.node-uuid=ce474774-436a-41d3-bfdd-ab153ac77830 --process-name glustershd --client-pid=-6 Mär 13 14:19:02 kvm320.durchhalten.intern systemd[1]: Starting GlusterFS, a clustered file-system server... Mär 13 14:19:03 kvm320.durchhalten.intern systemd[1]: Started GlusterFS, a clustered file-system server. Mär 13 14:19:12 kvm320.durchhalten.intern glusterd[9264]: [2020-03-13 13:19:12.869310] C [MSGID: 106003] [glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action] 0-management: Server quorum regained for volume data. Starting local bricks. Mär 13 14:19:13 kvm320.durchhalten.intern glusterd[9264]: [2020-03-13 13:19:13.115484] C [MSGID: 106003] [glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action] 0-management: Server quorum regained for volume engine. Starting local bricks. everything seems to be connected [root@kvm380 ~]# gluster peer status Number of Peers: 3 Hostname: kvm360.durchhalten.intern Uuid: 2f6cf198-afe5-4e61-b018-e7e4c10793b4 State: Peer in Cluster (Connected) Hostname: kvm320.durchhalten.intern Uuid: ce474774-436a-41d3-bfdd-ab153ac77830 State: Peer in Cluster (Connected) Other names: 192.168.200.231 Hostname: kvm10 Uuid: 33cd77a6-3cda-4e21-bd45-a907044f410b State: Peer in Cluster (Connected) Other names: kvm10 and this is the state of the brick from engine and data [root@kvm320 ~]# gluster volume status engine Status of volume: engine Gluster process TCP Port RDMA Port Online Pid
[ovirt-users] Memory problem
Hi to all, I ve a memory problem I got this error: Used memory of host kvm380.durchhalten.intern in cluster Default [96%] exceeded defined threshold [95%]. after reviewing the server with top command, I found ovn-controller with heavy memory usage: 45055 root 10 -10 46,5g 45,4g 2400 S 51,3 72,4 627:58.60 ovn-controller after restarting ovn-controller, every thing is fine again. after some days, it uses so much memory again. I ve also tried to wait a day or two, after that it seems to restart it self. is there a solution that ovn-controller does not use so much memory or restart automatic? thx shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/CW4YPJ5PEOR3RQQ7VDBZ7BYQ6A4LKP3Q/
[ovirt-users] Re: Gluster: a lof of Number of ntries in heal pending
Hello >I hope you plan to add another brick or arbiter, as you are now prone to >split-brain and other issues. Yes I will add an other one, but I think this is not a problem. I ve set cluster.server-quorum-ratio to 51% to avid the split brain problem. of course I know I just have failure tolarance of one. I ve solved the probelem with removeing the brick gluster volume remove-brick data replica 3 kvm320.durchhalten.intern:/gluster_bricks/data/ force Remove-brick force will not migrate files from the removed bricks, so they will no longer be available on the volume. Do you want to continue? (y/n) y volume remove-brick commit force: success [root@kvm380 ~]# gluster volume heal data info summary Brick kvm10:/gluster_bricks/data Status: Connected Total Number of entries: 1 Number of entries in heal pending: 1 Number of entries in split-brain: 0 Number of entries possibly healing: 0 Brick kvm360.durchhalten.intern:/gluster_bricks/data Status: Connected Total Number of entries: 0 Number of entries in heal pending: 0 Number of entries in split-brain: 0 Number of entries possibly healing: 0 Brick kvm380.durchhalten.intern:/gluster_bricks/data Status: Connected Total Number of entries: 1 Number of entries in heal pending: 1 Number of entries in split-brain: 0 Number of entries possibly healing: 0 after this heal pending was ok for me. than I removed all files - from this node - and added it back again. now everything is find Best Regards Stefan ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/3HZTSL3YY5GBVSREKXBY2KSZXJCPGVMJ/
[ovirt-users] Re: Gluster: a lof of Number of ntries in heal pending
Hi Strahil, yes it is a replica 4 set I ve tried to stop and stop every gluster server, and Ive rebooted every server. or should I remove the brick and add it again? bye stefan ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/LLZ7NIGVYB33PXA3ZLJBK4GDNSSFWJHU/
[ovirt-users] Re: Gluster: a lof of Number of ntries in heal pending
yes, I ve already tried a full heal a week a go. how do i perform a manual heal? I only have this gfid: Status: Connected Number of entries: 868 I ve tried to heal it with: [root@kvm10 ~]# gluster volume heal data split-brain latest-mtime gfid:c2b47c5c-89b6-49ac-bf10-1733dd8f0902 Healing gfid:c2b47c5c-89b6-49ac-bf10-1733dd8f0902 failed: File not in split-brain. Volume heal failed. (the last entry) And if i understood it correct ther is no split-brain [root@kvm10 ~]# gluster volume heal data info split-brain Brick kvm10:/gluster_bricks/data Status: Connected Number of entries in split-brain: 0 Brick kvm320.durchhalten.intern:/gluster_bricks/data Status: Connected Number of entries in split-brain: 0 Brick kvm360.durchhalten.intern:/gluster_bricks/data Status: Connected Number of entries in split-brain: 0 Brick kvm380.durchhalten.intern:/gluster_bricks/data Status: Connected Number of entries in split-brain: 0 ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/AQMSKCYGKKKUZIRXSKBW2VWWPLYVEX7A/
[ovirt-users] Gluster: a lof of Number of ntries in heal pending
Hello to all, I ve a problem with gluster [root@kvm10 ~]# gluster volume heal data info summary Brick kvm10:/gluster_bricks/data Status: Connected Total Number of entries: 868 Number of entries in heal pending: 868 Number of entries in split-brain: 0 Number of entries possibly healing: 0 Brick kvm320.durchhalten.intern:/gluster_bricks/data Status: Connected Total Number of entries: 1 Number of entries in heal pending: 1 Number of entries in split-brain: 0 Number of entries possibly healing: 0 Brick kvm360.durchhalten.intern:/gluster_bricks/data Status: Connected Total Number of entries: 867 Number of entries in heal pending: 867 Number of entries in split-brain: 0 Number of entries possibly healing: 0 Brick kvm380.durchhalten.intern:/gluster_bricks/data Status: Connected Total Number of entries: 868 Number of entries in heal pending: 868 Number of entries in split-brain: 0 Number of entries possibly healing: 0 [root@kvm10 ~]# gluster volume heal data info split-brain Brick kvm10:/gluster_bricks/data Status: Connected Number of entries in split-brain: 0 Brick kvm320.durchhalten.intern:/gluster_bricks/data Status: Connected Number of entries in split-brain: 0 Brick kvm360.durchhalten.intern:/gluster_bricks/data Status: Connected Number of entries in split-brain: 0 Brick kvm380.durchhalten.intern:/gluster_bricks/data Status: Connected Number of entries in split-brain: 0 As I understand there is no split-brain but 868 files ar in state heal pending. I ve restarted every node. I ve also tried: [root@kvm10 ~]# gluster volume heal data full Launching heal operation to perform full self heal on volume data has been successful Use heal info commands to check status. but even after a week there is no really change ( I started with 912 Number of entries in heal pending) can somebody tell what exactly is the problem and how can I solve it. thank you very much ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/PN63LC3OBQOM7IQY763ZS5V6VZDUFPNP/
[ovirt-users] Node not starting | blk_cloned_rq_check_limits: over max size limit
hi all, i ve 4 nodes running with current ovirt. I ve only a problem on one host even after a fresh installation . I ve installed the latest image. Than I add the node to the cluster Everything is working good. After this I configure the network. BUT, after a restart the host does not come up again. I got this error: blk_cloned_rq_check_limits: over max size limit every 5 seconds I can continue with control-D or I can login with root password to fix the problem. but i dont know what is the problem and where does it came from I ve also changed the sas disk to nvme storage, but I ve changed this on every host. And this problem exists only on one host i found this https://lists.centos.org/pipermail/centos/2017-December/167727.html the output is [root@kvm380 ~]# ./test.sh Sys Block Node: Devicemax_sectors_kb max_hw_sectors_kb /sys/block/dm-0 : onn_kvm380-pool00_tmeta 256 4096 /sys/block/dm-1 : onn_kvm380-pool00_tdata 256 4096 /sys/block/dm-10 : onn_kvm380-var256 4096 /sys/block/dm-11 : onn_kvm380-tmp256 4096 /sys/block/dm-12 : onn_kvm380-home 256 4096 /sys/block/dm-13 : onn_kvm380-var_crash 256 4096 /sys/block/dm-2 : onn_kvm380-pool00-tpool 256 4096 /sys/block/dm-3 : onn_kvm380-ovirt--node--ng--4.3.7--0.20191121.0+1 256 4096 /sys/block/dm-4 : onn_kvm380-swap 256 4096 /sys/block/dm-5 : eui.0025385991b1e27a 512 2048 /sys/block/dm-6 : eui.0025385991b1e27a1 512 2048 /sys/block/dm-7 : onn_kvm380-pool00 256 4096 /sys/block/dm-8 : onn_kvm380-var_log_audit 256 4096 /sys/block/dm-9 : onn_kvm380-var_log256 4096 cat: /sys/block/nvme0n1/device/vendor: Datei oder Verzeichnis nicht gefunden /sys/block/nvme0n1: Samsung SSD 970 EVO 1TB 512 2048 /sys/block/sda: HP LOGICAL VOLUME 256 4096 is the nvme not starting correct [root@kvm380 ~]# systemctl status multipathd ● multipathd.service - Device-Mapper Multipath Device Controller Loaded: loaded (/usr/lib/systemd/system/multipathd.service; enabled; vendor preset: enabled) Active: active (running) since Di 2019-12-31 16:16:32 CET; 31min ago Process: 1919 ExecStart=/sbin/multipathd (code=exited, status=0/SUCCESS) Process: 1916 ExecStartPre=/sbin/multipath -A (code=exited, status=0/SUCCESS) Process: 1911 ExecStartPre=/sbin/modprobe dm-multipath (code=exited, status=0/SUCCESS) Main PID: 1921 (multipathd) Tasks: 7 CGroup: /system.slice/multipathd.service └─1921 /sbin/multipathd Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: nvme0n1: mark as failed Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: Entering recovery mode: max_retries=4 Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: remaining active paths: 0 Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: 259:0: reinstated Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: queue_if_no_path enabled Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: Recovered to normal mode Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: remaining active paths: 1 Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: nvme0n1: mark as failed Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: Entering recovery mode: max_retries=4 Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: remaining active paths: 0 why is it marked as failed? if i create a new volume with cockpit and use it for bricks for gluster, every thing is fine. until reboot maybe some one can point me the direction ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/MHHFFWAY5T5BVTD5TAMW2R6PC6SMGGA2/
[ovirt-users] Re: Cannot start VM 2 hosts
Problem found. on these 2 hosts was the firewall disabled bye shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/Q64MJ7IXIEA2VE6BFDQNWZMCOCK3JRIN/
[ovirt-users] Cannot start VM 2 hosts
I ve got 4 Hosts after changeing the harddrive on every Host and normal updates I am not able anymore to start VM's on 2 of these 4 hosts. I am also not able to migrate a running vm to these hosts. This is the error log for start up 2019-12-17 16:01:28,326+01 INFO [org.ovirt.engine.core.bll.RunVmOnceCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] Lock Acquired to object 'EngineLock:{exclusiveLocks='[40cf6c27-6464-4fa7-bc01-9158cb03328b=VM]', sharedLocks=''}' 2019-12-17 16:01:28,609+01 INFO [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] START, IsVmDuringInitiatingVDSCommand( IsVmDuringInitiatingVDSCommandParameters:{vmId='40cf6c27-6464-4fa7-bc01-9158cb03328b'}), log id: 5706d380 2019-12-17 16:01:28,611+01 INFO [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] FINISH, IsVmDuringInitiatingVDSCommand, return: false, log id: 5706d380 2019-12-17 16:01:28,854+01 INFO [org.ovirt.engine.core.bll.RunVmOnceCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] Running command: RunVmOnceCommand internal: false. Entities affected : ID: 40cf6c27-6464-4fa7-bc01-9158cb03328b Type: VMAction group RUN_VM with role type USER, ID: 40cf6c27-6464-4fa7-bc01-9158cb03328b Type: VMAction group EDIT_ADMIN_VM_PROPERTIES with role type ADMIN 2019-12-17 16:01:28,999+01 INFO [org.ovirt.engine.core.vdsbroker.UpdateVmDynamicDataVDSCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] START, UpdateVmDynamicDataVDSCommand( UpdateVmDynamicDataVDSCommandParameters:{hostId='null', vmId='40cf6c27-6464-4fa7-bc01-9158cb03328b', vmDynamic='org.ovirt.engine.core.common.businessentities.VmDynamic@147620b'}), log id: 612dabdf 2019-12-17 16:01:29,097+01 INFO [org.ovirt.engine.core.vdsbroker.UpdateVmDynamicDataVDSCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] FINISH, UpdateVmDynamicDataVDSCommand, return: , log id: 612dabdf 2019-12-17 16:01:29,106+01 INFO [org.ovirt.engine.core.vdsbroker.CreateVDSCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] START, CreateVDSCommand( CreateVDSCommandParameters:{hostId='d38bae4c-8494-4861-ae5a-38992db338e5', vmId='40cf6c27-6464-4fa7-bc01-9158cb03328b', vm='VM [z-push]'}), log id: 56abde11 2019-12-17 16:01:29,110+01 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] START, CreateBrokerVDSCommand(HostName = kvm360.durchhalten.intern, CreateVDSCommandParameters:{hostId='d38bae4c-8494-4861-ae5a-38992db338e5', vmId='40cf6c27-6464-4fa7-bc01-9158cb03328b', vm='VM [z-push]'}), log id: 36d088ab 2019-12-17 16:01:29,163+01 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] VM http://ovirt.org/vm/tune/1.0; xmlns:ovirt-vm="http://ovirt.org/vm/1.0;> z-push 40cf6c27-6464-4fa7-bc01-9158cb03328b 4194304 4194304 1 16777216 16 oVirt OS-NAME: OS-VERSION: HOST-SERIAL: 40cf6c27-6464-4fa7-bc01-9158cb03328b SandyBridge /dev/urandom 38f2e3bb-687f-4f01-b769-3a5403136181 hvm 4096 4.3 0e279f02-0c73-11e9-bc47-00163e150480 1a9627d8-181b-4790-aa47-569dd1e6e9fb 38f2e3bb-687f-4f01-b769-3a5403136181 80364cc8-8afd-4482-884d-d63d5a7988df false auto_resume 2019-12-17 16:01:29,198+01 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] FINISH, CreateBrokerVDSCommand, return: , log id: 36d088ab 2019-12-17 16:01:29,252+01 INFO [org.ovirt.engine.core.vdsbroker.CreateVDSCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] FINISH, CreateVDSCommand, return: WaitForLaunch, log id: 56abde11 2019-12-17 16:01:29,252+01 INFO [org.ovirt.engine.core.bll.RunVmOnceCommand] (default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] Lock freed to object 'EngineLock:{exclusiveLocks='[40cf6c27-6464-4fa7-bc01-9158cb03328b=VM]', sharedLocks=''}' 2019-12-17 16:01:29,319+01 INFO
[ovirt-users] Re: Increase memory of hosted engine
thank you for your replay, I am happy if it works 1. check 2. check 3.1 check -> I cant see where is use this config file later, or it is just a backup? 4.1 check i changed to 16384 16384 and 15625000 15625000 15625000 after write, quite und open these values have changed back to 4194304 4194304 BREAK ;) after stopping the hosted engine with hosted-engine --vm-shutdown and powering it up with hosted-engine vm-start everything is reverted ;( BREAK hosted engine is now up and running, maintenance ist still global, changed again with virtsh edit HostedEngine. as written above it changes back I will wait for tomorrow. Or did I missunderstand something In which file the maxmem is hidden? thx shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/DWVBZO2OE6U62FAEMIIS3TG2C6RRV4OH/
[ovirt-users] Increase memory of hosted engine
hello, I ve decrease the memory of the hosted engine. now I am not able to increase the memory permantly right now the memory has 4096 MB Max Memory is 7936 MB and guaranted memory is 7936MB I can increase the memory up to 7936 MB in the manager, it changes immediately. I can not increase to more then 7936 MB because this is max Memory If I change Max memory to 16384 the gueanted memery gets a red border an the error is max memory ist 4096 MB I ve found this one https://access.redhat.com/solutions/964693 and tried to change the memSize in /etc/ovirt-hosted-engine/vm.conf to 16384 I ve also stoppt every ovirt-ha-agent and changed every vm.conf. After starting ovirt-ha-agent every vm.conf changed to 4096 again has somebody a solution to increase the memory again? thx shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/LZHYZ5IIFVBJT2UTDSQBTMEP4TZ67CTO/
[ovirt-users] Re: ovirt-ha-agent not running
you are right, thank you now it is up and running again [root@kvm380 ~]# ls /rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern\:_engine/36663740-576a-4498-b28e-0a402628c6a7/ha_agent/ -lha insgesamt 0 drwxr-xr-x. 2 vdsm kvm 67 1. Jan 1970 . drwxr-xr-x. 6 vdsm kvm 64 7. Dez 12:08 .. lrwxrwxrwx. 1 vdsm kvm 132 8. Dez 06:49 hosted-engine.lockspace -> /var/run/vdsm/storage/36663740-576a-4498-b28e-0a402628c6a7/de9b53fb-6c86-4036-9f1c-1f08d1828e2f/d41a5061-94b8-4e0d-84ec-8af0c547955a lrwxrwxrwx. 1 vdsm kvm 132 8. Dez 06:49 hosted-engine.metadata -> /var/run/vdsm/storage/36663740-576a-4498-b28e-0a402628c6a7/aec89d8d-0c7c-42b8-8889-1713d7e8a667/39ec2ecd-57b5-4424-8666-d7b673b1e8b6 ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/3XDFLVVU67T6NMGGS6QA5ERTIQIQ6XRA/
[ovirt-users] Re: ovirt-ha-agent not running
the contet is [root@kvm380 ~]# ls /rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine/36663740-576a-4498-b28e-0a402628c6a7/ha_agent/ -lha ls: Zugriff auf /rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine/36663740-576a-4498-b28e-0a402628c6a7/ha_agent/hosted-engine.lockspace nicht möglich: Der Socket ist nicht verbunden ls: Zugriff auf /rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine/36663740-576a-4498-b28e-0a402628c6a7/ha_agent/hosted-engine.metadata nicht möglich: Der Socket ist nicht verbunden insgesamt 512 drwxr-xr-x. 2 vdsm kvm 67 20. Nov 09:20 . drwxr-xr-x. 6 vdsm kvm 64 7. Dez 12:08 .. l?? ? ??? ? hosted-engine.lockspace l?? ? ??? ? hosted-engine.metadata I ve already tryed to stop and start the broker again, ovirt-ha-agant will not start. And as wrote i ve done a complete reinstall of the ovirt host bye stefan ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/HCM6XS7FBSDQLXZWTSXHP7HWIYH23KXI/
[ovirt-users] Re: ovirt-ha-agent not running
and here is the broker.log MainThread::INFO::2019-12-07 15:20:03,563::broker::47::ovirt_hosted_engine_ha.broker.broker.Broker::(run) ovirt-hosted-engine-ha broker 2.3.6 started MainThread::INFO::2019-12-07 15:20:03,564::monitor::40::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Searching for submonitors in /usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/submonitors MainThread::INFO::2019-12-07 15:20:03,564::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor mem-free MainThread::INFO::2019-12-07 15:20:03,565::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor network MainThread::INFO::2019-12-07 15:20:03,566::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor mem-free MainThread::INFO::2019-12-07 15:20:03,566::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor engine-health MainThread::INFO::2019-12-07 15:20:03,567::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor cpu-load MainThread::INFO::2019-12-07 15:20:03,567::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor storage-domain MainThread::INFO::2019-12-07 15:20:03,568::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor mgmt-bridge MainThread::INFO::2019-12-07 15:20:03,569::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor cpu-load MainThread::INFO::2019-12-07 15:20:03,569::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor network MainThread::INFO::2019-12-07 15:20:03,574::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor cpu-load-no-engine MainThread::INFO::2019-12-07 15:20:03,575::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor cpu-load-no-engine MainThread::INFO::2019-12-07 15:20:03,576::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor mgmt-bridge MainThread::INFO::2019-12-07 15:20:03,577::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor engine-health MainThread::INFO::2019-12-07 15:20:03,577::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Loaded submonitor storage-domain MainThread::INFO::2019-12-07 15:20:03,577::monitor::50::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors) Finished loading submonitors MainThread::INFO::2019-12-07 15:20:03,651::storage_backends::373::ovirt_hosted_engine_ha.lib.storage_backends::(connect) Connecting the storage MainThread::INFO::2019-12-07 15:20:03,652::storage_server::349::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server) Connecting storage server MainThread::INFO::2019-12-07 15:20:03,716::storage_server::356::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server) Connecting storage server MainThread::INFO::2019-12-07 15:20:03,748::storage_server::413::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server) Refreshing the storage domain MainThread::WARNING::2019-12-07 15:20:06,985::storage_broker::97::ovirt_hosted_engine_ha.broker.storage_broker.StorageBroker::(__init__) Can't connect vdsm storage: [Errno 107] Transport endpoint is not connected: '/rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine/36663740-576a-4498-b28e-0a402628c6a7/ha_agent/hosted-engine.lockspace' maybe it helps bye shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/KECDUEVCHP3OSUREULALUEJ6WSALLUDQ/
[ovirt-users] ovirt-ha-agent not running
hello, since some days ovirt-ha-agent is not running anymore i ve 4 ovirt hosts and only on one host the agent is running. maybe it was from an update, because i lost one agent after an other. i ve done a complete fresh install for the host with the latest ovirt node. I ve got on tree hosts this error [root@kvm380 ~]# systemctl status ovirt-ha-agent ● ovirt-ha-agent.service - oVirt Hosted Engine High Availability Monitoring Agent Loaded: loaded (/usr/lib/systemd/system/ovirt-ha-agent.service; enabled; vendor preset: disabled) Active: activating (auto-restart) (Result: exit-code) since Sa 2019-12-07 14:56:21 UTC; 5s ago Process: 28002 ExecStart=/usr/share/ovirt-hosted-engine-ha/ovirt-ha-agent (code=exited, status=157) Main PID: 28002 (code=exited, status=157) Tasks: 0 CGroup: /system.slice/ovirt-ha-agent.service Dez 07 14:56:21 kvm380.durchhalten.intern systemd[1]: ovirt-ha-agent.service: main process exited, code=exited, status=157/n/a Dez 07 14:56:21 kvm380.durchhalten.intern systemd[1]: Unit ovirt-ha-agent.service entered failed state. Dez 07 14:56:21 kvm380.durchhalten.intern systemd[1]: ovirt-ha-agent.service failed. and this is in /var/log/ovirt-hosted-engine-ha/agent.log MainThread::INFO::2019-12-07 15:01:51,048::agent::67::ovirt_hosted_engine_ha.agent.agent.Agent::(run) ovirt-hosted-engine-ha agent 2.3.6 started MainThread::INFO::2019-12-07 15:01:51,161::hosted_engine::234::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_hostname) Found certificate common name: kvm380.durchhalten.intern MainThread::INFO::2019-12-07 15:01:51,374::hosted_engine::543::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_broker) Initializing ha-broker connection MainThread::INFO::2019-12-07 15:01:51,378::brokerlink::80::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(start_monitor) Starting monitor network, options {'tcp_t_address': None, 'network_test': None, 'tcp_t_port': None, 'addr': '192.168.200.1'} MainThread::ERROR::2019-12-07 15:01:51,379::hosted_engine::559::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_broker) Failed to start necessary monitors MainThread::ERROR::2019-12-07 15:01:51,381::agent::144::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 131, in _run_agent return action(he) File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 55, in action_proper return he.start_monitoring() File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py", line 432, in start_monitoring self._initialize_broker() File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py", line 556, in _initialize_broker m.get('options', {})) File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", line 89, in start_monitor ).format(t=type, o=options, e=e) RequestError: brokerlink - failed to start monitor via ovirt-ha-broker: [Errno 2] No such file or directory, [monitor: 'network', options: {'tcp_t_address': None, 'network_test': None, 'tcp_t_port': None, 'addr': '192.168.200.1'}] MainThread::ERROR::2019-12-07 15:01:51,381::agent::145::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) Trying to restart agent MainThread::INFO::2019-12-07 15:01:51,382::agent::89::ovirt_hosted_engine_ha.agent.agent.Agent::(run) Agent shutting down maybe someone can give me an advice thx shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/UE4TMWWT27BHFJWITYRV52CHG5AA7ZYD/
[ovirt-users] hosted engine does not start
Hello all, after a powerloss the hosted engine wont start up anymore. I ve the current ovirt installed. Storage is glusterfs und it is up and running It is trying to start up hosted engine but it does not work, but I cant see where the problem is. [root@kvm320 ~]# hosted-engine --vm-status --== Host 1 status ==-- conf_on_shared_storage : True Status up-to-date : True Hostname : kvm380.durchhalten.intern Host ID: 1 Engine status : {"reason": "bad vm status", "health": "bad", "vm": "down", "detail": "Down"} Score : 1800 stopped: False Local maintenance : False crc32 : 3ad6d0bd local_conf_timestamp : 14594 Host timestamp : 14594 Extra metadata (valid at timestamp): metadata_parse_version=1 metadata_feature_version=1 timestamp=14594 (Mon Apr 15 21:25:12 2019) host-id=1 score=1800 vm_conf_refresh_time=14594 (Mon Apr 15 21:25:12 2019) conf_on_shared_storage=True maintenance=False state=GlobalMaintenance stopped=False --== Host 2 status ==-- conf_on_shared_storage : True Status up-to-date : True Hostname : kvm320.durchhalten.intern Host ID: 2 Engine status : {"reason": "failed liveliness check", "health": "bad", "vm": "up", "detail": "Up"} Score : 0 stopped: False Local maintenance : False crc32 : e7d4840d local_conf_timestamp : 21500 Host timestamp : 21500 Extra metadata (valid at timestamp): metadata_parse_version=1 metadata_feature_version=1 timestamp=21500 (Mon Apr 15 21:25:22 2019) host-id=2 score=0 vm_conf_refresh_time=21500 (Mon Apr 15 21:25:22 2019) conf_on_shared_storage=True maintenance=False state=ReinitializeFSM stopped=False --== Host 3 status ==-- conf_on_shared_storage : True Status up-to-date : True Hostname : kvm360.durchhalten.intern Host ID: 3 Engine status : {"reason": "vm not running on this host", "health": "bad", "vm": "down", "detail": "unknown"} Score : 1800 stopped: False Local maintenance : False crc32 : cf9221cb local_conf_timestamp : 22121 Host timestamp : 22120 Extra metadata (valid at timestamp): metadata_parse_version=1 metadata_feature_version=1 timestamp=22120 (Mon Apr 15 21:25:18 2019) host-id=3 score=1800 vm_conf_refresh_time=22121 (Mon Apr 15 21:25:18 2019) conf_on_shared_storage=True maintenance=False state=GlobalMaintenance stopped=False [root@kvm320 ~]# virsh -r list IdName Status 6 HostedEngine laufend [root@kvm320 ~]# hosted-engine --console The engine VM is running on this host Verbunden mit der Domain: HostedEngine Escape-Zeichen ist ^] Fehler: Interner Fehler: Zeichengerät kann nicht gefunden warden In engish it should be this [root@mgmt~]# hosted-engine --console The engine VM is running on this host Connected to domain HostedEngine Escape character is ^] error: internal error: cannot find character device This is in the log [root@kvm320 ~]# tail -f /var/log/ovirt-hosted-engine-ha/agent.log MainThread::INFO::2019-04-15 21:28:33,032::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine .HostedEngine::(_monitoring_loop) Current state EngineStarting (score: 1800) MainThread::INFO::2019-04-15 21:28:43,050::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.Hosted Engine::(consume) VM is powering up.. MainThread::INFO::2019-04-15 21:28:43,165::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine .HostedEngine::(_monitoring_loop) Current state EngineStarting (score: 1800) MainThread::INFO::2019-04-15 21:28:53,183::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.Hosted Engine::(consume) VM is powering up.. MainThread::INFO::2019-04-15 21:28:53,300::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine .HostedEngine::(_monitoring_loop) Current state EngineStarting (score: 1800) MainThread::INFO::2019-04-15 21:29:03,317::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.Hosted Engine::(consume) VM is powering up..
[ovirt-users] Re: Hyperconvergend Setup stuck
It is gdeploy 2.0.2 rpm -qa |grep gdeploy gdeploy-2.0.8-1.el7.noarch ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/3QUCUGUHSSRHESRSPGWENRRUND2K3QLK/
[ovirt-users] Re: Active Storage Domains as Problematic
i 've mounted it during the hosted-engine --deploy process I selected glusterfs and entered server:/engine I dont enter any mount options yes it is enabled for both. I dont got errors for the second one, but may it doesn't check after the first fail ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/SYX6BQBW2MMV4YIXHG24KMXA7FTWL46X/
[ovirt-users] Re: Active Storage Domains as Problematic
yes i think this too, but as you see at the top >[root@kvm380 ~]# gluster volume info >... > performance.strict-o-direct: on ... it was already set i did a one cluster setup with ovirt and I uses this result Volume Name: engine Type: Distribute Volume ID: a40e848b-a8f1-4990-9d32-133b46db6f1d Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: kvm360.durchhalten.intern:/gluster_bricks/engine/engine Options Reconfigured: cluster.eager-lock: enable performance.io-cache: off performance.read-ahead: off performance.quick-read: off user.cifs: off network.ping-timeout: 30 network.remote-dio: off performance.strict-o-direct: on performance.low-prio-threads: 32 features.shard: on storage.owner-gid: 36 storage.owner-uid: 36 transport.address-family: inet nfs.disable: on could there be an other reason? ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/HST6ODMSZQZEP6Q2OMLCAMD27HI2CDHQ/
[ovirt-users] Re: Active Storage Domains as Problematic
here is what i found in the logs of the hosts 2018-12-20 12:34:04,824+0100 INFO (periodic/0) [vdsm.api] START repoStats(domains=()) from=internal, task_id=09235382-a5b5-48da-853d-f94cae092684 (api:46) 2018-12-20 12:34:04,825+0100 INFO (periodic/0) [vdsm.api] FINISH repoStats return={u'20651d3d-08d7-482a-ae4e-7cd0e33cc907': {'code': 399, 'actual': True, 'version': -1, 'acquired': False, 'delay': '0', 'lastCheck': '6.1', 'valid': False}, u'ae9e4cbd-3946-481d-b01a-e8a38bf00efb': {'code': 0, 'actual': True, 'version': 4, 'acquired': True, 'delay': '0.0013974', 'lastCheck': '1.1', 'valid': True}} from=internal, task_id=09235382-a5b5-48da-853d-f94cae092684 (api:52) 2018-12-20 12:34:04,826+0100 INFO (periodic/0) [vdsm.api] START multipath_health() from=internal, task_id=8f6166cb-aa41-4f46-823d-d38e4e85f02a (api:46) 2018-12-20 12:34:04,826+0100 INFO (periodic/0) [vdsm.api] FINISH multipath_health return={} from=internal, task_id=8f6166cb-aa41-4f46-823d-d38e4e85f02a (api:52) 2018-12-20 12:34:04,832+0100 INFO (jsonrpc/4) [vdsm.api] START prepareImage(sdUUID=u'20651d3d-08d7-482a-ae4e-7cd0e33cc907', spUUID=u'----', imgUUID=u'c1ebc7ad-dfb6-4cc1-8e24-40f0be3f4afe', leafUUID=u'e7e76dd8-d166-46a0-9761-fa6391aa047b', allowIllegal=False) from=::1,55348, task_id=996206fc-65eb-4056-b3b9-2ac0e1780c2c (api:46) 2018-12-20 12:34:04,836+0100 ERROR (periodic/0) [root] failed to retrieve Hosted Engine HA score '[Errno 2] No such file or directory'Is the Hosted Engine setup finished? (api:196) 2018-12-20 12:34:04,847+0100 ERROR (jsonrpc/4) [storage.fileSD] Underlying file system doesn't supportdirect IO (fileSD:108) 2018-12-20 12:34:04,847+0100 INFO (jsonrpc/4) [vdsm.api] FINISH prepareImage error=Storage Domain target is unsupported: () from=::1,55348, task_id=996206fc-65eb-4056-b3b9-2ac0e1780c2c (api:50) 2018-12-20 12:34:04,847+0100 ERROR (jsonrpc/4) [storage.TaskManager.Task] (Task='996206fc-65eb-4056-b3b9-2ac0e1780c2c') Unexpected error (task:875) Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, in _run return fn(*args, **kargs) File "", line 2, in prepareImage File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, in method ret = func(*args, **kwargs) File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 3173, in prepareImage dom = sdCache.produce(sdUUID) File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 110, in produce domain.getRealDomain() File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 51, in getRealDomain return self._cache._realProduce(self._sdUUID) File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 134, in _realProduce domain = self._findDomain(sdUUID) File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 151, in _findDomain return findMethod(sdUUID) File "/usr/lib/python2.7/site-packages/vdsm/storage/glusterSD.py", line 55, in findDomain return GlusterStorageDomain(GlusterStorageDomain.findDomainPath(sdUUID)) File "/usr/lib/python2.7/site-packages/vdsm/storage/fileSD.py", line 400, in __init__ validateFileSystemFeatures(manifest.sdUUID, manifest.mountpoint) File "/usr/lib/python2.7/site-packages/vdsm/storage/fileSD.py", line 110, in validateFileSystemFeatures raise se.StorageDomainTargetUnsupported() StorageDomainTargetUnsupported: Storage Domain target is unsupported: () 2018-12-20 12:34:04,847+0100 INFO (jsonrpc/4) [storage.TaskManager.Task] (Task='996206fc-65eb-4056-b3b9-2ac0e1780c2c') aborting: Task is aborted: 'Storage Domain target is unsupported: ()' - code 399 (task:1181) 2018-12-20 12:34:04,848+0100 ERROR (jsonrpc/4) [storage.Dispatcher] FINISH prepareImage error=Storage Domain target is unsupported: () (dispatcher:82) 2018-12-20 12:34:04,848+0100 INFO (jsonrpc/4) [jsonrpc.JsonRpcServer] RPC call Image.prepare failed (error 399) in 0.02 seconds (__init__:573) ^ especially this part 2018-12-20 12:34:04,847+0100 ERROR (jsonrpc/4) [storage.fileSD] Underlying file system doesn't supportdirect IO (fileSD:108) 2018-12-20 12:34:04,847+0100 INFO (jsonrpc/4) [vdsm.api] FINISH prepareImage error=Storage Domain target is unsupported: () from=::1,55348, task_id=996206fc-65eb-4056-b3b9-2ac0e1780c2c (api:50) and i am not sure why he is asking this 2018-12-20 12:34:04,836+0100 ERROR (periodic/0) [root] failed to retrieve Hosted Engine HA score '[Errno 2] No such file or directory'Is the Hosted Engine setup finished? (api:196) ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/WKCPU5I4Z75TMZQQCQBAD4WK3QOLVTFR/
[ovirt-users] Active Storage Domains as Problematic
Hello, I ,ve setup a test lab with 3 nodes installed with centos 7 I configured manualy gluster fs. Glusterfs is up and running [root@kvm380 ~]# gluster peer status Number of Peers: 2 Hostname: kvm320.durchhalten.intern Uuid: dac066db-55f7-4770-900d-4830c740ffbf State: Peer in Cluster (Connected) Hostname: kvm360.durchhalten.intern Uuid: 4291be40-f77f-4f41-98f6-dc48fd993842 State: Peer in Cluster (Connected) [root@kvm380 ~]# gluster volume info Volume Name: data Type: Replicate Volume ID: 3586de82-e504-4c62-972b-448abead13d3 Status: Started Snapshot Count: 0 Number of Bricks: 1 x 3 = 3 Transport-type: tcp Bricks: Brick1: kvm380.durchhalten.intern:/gluster/data Brick2: kvm360.durchhalten.intern:/gluster/data Brick3: kvm320.durchhalten.intern:/gluster/data Options Reconfigured: storage.owner-uid: 36 storage.owner-gid: 36 features.shard: on performance.low-prio-threads: 32 performance.strict-o-direct: on network.ping-timeout: 30 user.cifs: off network.remote-dio: off performance.quick-read: off performance.read-ahead: off performance.io-cache: off cluster.eager-lock: enable transport.address-family: inet nfs.disable: on performance.client-io-threads: off Volume Name: engine Type: Replicate Volume ID: dcfbd322-5dd0-4bfe-a775-99ecc79e1416 Status: Started Snapshot Count: 0 Number of Bricks: 1 x 3 = 3 Transport-type: tcp Bricks: Brick1: kvm380.durchhalten.intern:/gluster/engine Brick2: kvm360.durchhalten.intern:/gluster/engine Brick3: kvm320.durchhalten.intern:/gluster/engine Options Reconfigured: storage.owner-uid: 36 storage.owner-gid: 36 features.shard: on performance.low-prio-threads: 32 performance.strict-o-direct: on network.remote-dio: off network.ping-timeout: 30 user.cifs: off performance.quick-read: off performance.read-ahead: off performance.io-cache: off cluster.eager-lock: enable transport.address-family: inet nfs.disable: on performance.client-io-threads: off After that I deployed a selfhosted engine And add the two other hosts, at the beginning it looks good, but without changing anything I got following error by two hosts ! 20.12.2018 11:35:05 Failed to connect Host kvm320.durchhalten.intern to Storage Pool Default ! 20.12.2018 11:35:05 Host kvm320.durchhalten.intern cannot access the Storage Domain(s) hosted_storage attached to the Data Center Default. Setting Host state to Non-Operational. X 20.12.2018 11:35:05 Host kvm320.durchhalten.intern reports about one of the Active Storage Domains as Problematic. ! 20.12.2018 11:35:05 Kdump integration is enabled for host kvm320.durchhalten.intern, but kdump is not configured properly on host. ! 20.12.2018 11:35:04 Failed to connect Host kvm360.durchhalten.intern to Storage Pool Default ! 20.12.2018 11:35:04 Host kvm360.durchhalten.intern cannot access the Storage Domain(s) hosted_storage attached to the Data Center Default. Setting Host state to Non-Operational. X 20.12.2018 11:35:04 Host kvm360.durchhalten.intern reports about one of the Active Storage Domains as Problematic. Before glusterfs I had a setup with nfs on 4. Server Where is the problem? thx ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/RELXJB2LOJNAKAIJTNTSQC3AQFQNCT75/
[ovirt-users] Hyperconvergend Setup stuck
Hello I like to setup hyperconvergend I ve 3 hosts, everyone is fresh installed Kvm320 has one additional harddrive with 1TB SATA And kvm360 and kvm380 with two additional harddrives with 300gb and 600gb SAS #gdeploy configuration generated by cockpit-gluster plugin [hosts] kvm380.durchhalten.intern kvm360.durchhalten.intern kvm320.durchhalten.intern [script1:kvm380.durchhalten.intern] action=execute ignore_script_errors=no file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc -h kvm380.durchhalten.intern, kvm360.durchhalten.intern, kvm320.durchhalten.intern [script1:kvm360.durchhalten.intern] action=execute ignore_script_errors=no file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc -h kvm380.durchhalten.intern, kvm360.durchhalten.intern, kvm320.durchhalten.intern [script1:kvm320.durchhalten.intern] action=execute ignore_script_errors=no file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h kvm380.durchhalten.intern, kvm360.durchhalten.intern, kvm320.durchhalten.intern [disktype] raid6 [diskcount] 12 [stripesize] 256 [service1] action=enable service=chronyd [service2] action=restart service=chronyd [shell2] action=execute command=vdsm-tool configure --force [script3] action=execute file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh ignore_script_errors=no [pv1:kvm380.durchhalten.intern] action=create devices=sdb ignore_pv_errors=no [pv2:kvm380.durchhalten.intern] action=create devices=sdc ignore_pv_errors=no [pv1:kvm360.durchhalten.intern] action=create devices=sdb ignore_pv_errors=no [pv2:kvm360.durchhalten.intern] action=create devices=sdc ignore_pv_errors=no [pv1:kvm320.durchhalten.intern] action=create devices=sdb ignore_pv_errors=no [vg1:kvm380.durchhalten.intern] action=create vgname=gluster_vg_sdb pvname=sdb ignore_vg_errors=no [vg2:kvm380.durchhalten.intern] action=create vgname=gluster_vg_sdc pvname=sdc ignore_vg_errors=no [vg1:kvm360.durchhalten.intern] action=create vgname=gluster_vg_sdb pvname=sdb ignore_vg_errors=no [vg2:kvm360.durchhalten.intern] action=create vgname=gluster_vg_sdc pvname=sdc ignore_vg_errors=no [vg1:kvm320.durchhalten.intern] action=create vgname=gluster_vg_sdb pvname=sdb ignore_vg_errors=no [lv1:kvm380.durchhalten.intern] action=create poolname=gluster_thinpool_sdc ignore_lv_errors=no vgname=gluster_vg_sdc lvtype=thinpool size=1005GB poolmetadatasize=5GB [lv2:kvm360.durchhalten.intern] action=create poolname=gluster_thinpool_sdc ignore_lv_errors=no vgname=gluster_vg_sdc lvtype=thinpool size=1005GB poolmetadatasize=5GB [lv3:kvm320.durchhalten.intern] action=create poolname=gluster_thinpool_sdb ignore_lv_errors=no vgname=gluster_vg_sdb lvtype=thinpool size=41GB poolmetadatasize=1GB [lv4:kvm380.durchhalten.intern] action=create lvname=gluster_lv_engine ignore_lv_errors=no vgname=gluster_vg_sdb mount=/gluster_bricks/engine size=100GB lvtype=thick [lv5:kvm380.durchhalten.intern] action=create lvname=gluster_lv_data ignore_lv_errors=no vgname=gluster_vg_sdc mount=/gluster_bricks/data lvtype=thinlv poolname=gluster_thinpool_sdc virtualsize=500GB [lv6:kvm380.durchhalten.intern] action=create lvname=gluster_lv_vmstore ignore_lv_errors=no vgname=gluster_vg_sdc mount=/gluster_bricks/vmstore lvtype=thinlv poolname=gluster_thinpool_sdc virtualsize=500GB [lv7:kvm360.durchhalten.intern] action=create lvname=gluster_lv_engine ignore_lv_errors=no vgname=gluster_vg_sdb mount=/gluster_bricks/engine size=100GB lvtype=thick [lv8:kvm360.durchhalten.intern] action=create lvname=gluster_lv_data ignore_lv_errors=no vgname=gluster_vg_sdc mount=/gluster_bricks/data lvtype=thinlv poolname=gluster_thinpool_sdc virtualsize=500GB [lv9:kvm360.durchhalten.intern] action=create lvname=gluster_lv_vmstore ignore_lv_errors=no vgname=gluster_vg_sdc mount=/gluster_bricks/vmstore lvtype=thinlv poolname=gluster_thinpool_sdc virtualsize=500GB [lv10:kvm320.durchhalten.intern] action=create lvname=gluster_lv_engine ignore_lv_errors=no vgname=gluster_vg_sdb mount=/gluster_bricks/engine size=20GB lvtype=thick [lv11:kvm320.durchhalten.intern] action=create lvname=gluster_lv_data ignore_lv_errors=no vgname=gluster_vg_sdb mount=/gluster_bricks/data lvtype=thinlv poolname=gluster_thinpool_sdb virtualsize=20GB [lv12:kvm320.durchhalten.intern] action=create lvname=gluster_lv_vmstore ignore_lv_errors=no vgname=gluster_vg_sdb mount=/gluster_bricks/vmstore lvtype=thinlv poolname=gluster_thinpool_sdb virtualsize=20GB [selinux] yes [service3] action=restart service=glusterd slice_setup=yes [firewalld] action=add ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,5 4322/tcp services=glusterfs [script2] action=execute
[ovirt-users] Gluster with two ovirt nodes
Hello, i like to set up glusterfs with two ovirt nodes and on more "normal" node is this possible? i 've setup glusterfs in cli on two ovirt nodes and 3rd network storage. glusterfs is up and running. But now i like to get something like VIP with ctdb for example. is there any possibility to set this up with ovirt? Or do i *ve to setup ovirt manually in centos to install ctdb? Or are there any other ideas? thank you stefan ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/4JKFSI4XIMXCKQQEQ7W4ZPWNASYZ52TL/
[ovirt-users] Install additional software on ovirt node
Hello, if I install on an ovirt node additional software, will it removed after updating the ovirt node or run in to trouble? thx shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/JPJFPNVLUIQSHX5YPYMJQ4J5UNW64XGO/
[ovirt-users] Re: To much hosts in this cluster
No, I didn't. what are the next steps i ve allready reinstalled the host. i 've selected deploy at hosted engine, but it seems not to work. Am Mo., 10. Dez. 2018 um 09:12 Uhr schrieb Luca 'remix_tj' Lorenzetto < lorenzetto.l...@gmail.com>: > Hello, > > you did hosted-engine undeploy before removing? > > Luca > > On Sun, Dec 9, 2018 at 12:19 PM Stefan Wolf wrote: > >> Hello, >> >> >> >> i had 3 hosts with ovirt running, with one of this 3 host i had problems >> during boot up. >> >> I decied to remove the host from the cluster. Now i ve two hosts >> >> But if I take a look at Hosted Engine in cockpit i see all three hosts >> >> Why ist he kvm380 not removed? >> >> How can I remove it? >> >> >> >> Thx shb >> >> >> ___ >> Users mailing list -- users@ovirt.org >> To unsubscribe send an email to users-le...@ovirt.org >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> List Archives: >> https://lists.ovirt.org/archives/list/users@ovirt.org/message/VZAOYESLQZDD7TJ4P5COXCOR7LST4RMN/ >> > > > -- > "E' assurdo impiegare gli uomini di intelligenza eccellente per fare > calcoli che potrebbero essere affidati a chiunque se si usassero delle > macchine" > Gottfried Wilhelm von Leibnitz, Filosofo e Matematico (1646-1716) > > "Internet è la più grande biblioteca del mondo. > Ma il problema è che i libri sono tutti sparsi sul pavimento" > John Allen Paulos, Matematico (1945-vivente) > > Luca 'remix_tj' Lorenzetto, http://www.remixtj.net , < > lorenzetto.l...@gmail.com> > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/YOL24J4653E3J5H2BEXD2PDTEXE4D7NT/
[ovirt-users] Try to add Host to cluster: Command returned failure code 1 during SSH session
Hello, I try to add a new installed ovirt node and get the error message during adding to cluster Host kvm380 installation failed. Command returned failure code 1 during SSH session 'root@kvm380.durchhalten.intern'. Maybe someone can help me Thx Here is the logfile 2018-12-09 14:48:23,262+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Initializing. 2018-12-09 14:48:23,414+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Environment setup. 2018-12-09 14:48:23,479+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Environment packages setup. 2018-12-09 14:48:26,194+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Programs detection. 2018-12-09 14:48:26,373+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Environment customization. 2018-12-09 14:48:26,842+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Kdump supported. 2018-12-09 14:48:27,123+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Logs at host located at: '/tmp/ovirt-host-deploy-20181209144822-hbil3q.log'. 2018-12-09 14:48:27,188+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Host is hypervisor. 2018-12-09 14:48:27,192+01 INFO [org.ovirt.engine.core.bll.hostdeploy.VdsDeployVdsmUnit] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] Host kvm380.durchhalten.intern reports unique id 31323436-3530-5a43-3233-303430374a4e 2018-12-09 14:48:27,208+01 INFO [org.ovirt.engine.core.bll.hostdeploy.VdsDeployVdsmUnit] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] Assigning unique id 31323436-3530-5a43-3233-303430374a4e to Host kvm380.durchhalten.intern 2018-12-09 14:48:27,436+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Setup validation. 2018-12-09 14:48:27,825+01 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS_ERROR(511), An error has occurred during installation of Host kvm380: Failed to execute stage 'Setup validation': Cannot resolve kdump destination address 'ovirt.durchhalten.intern'. 2018-12-09 14:48:27,827+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Clean up. 2018-12-09 14:48:27,830+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Pre-termination. 2018-12-09 14:48:27,874+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Retrieving installation logs to: '/var/log/ovirt-engine/host-deploy/ovirt-host-deploy-20181209144827-kvm380.d urchhalten.intern-f097f8b0-ee35-4c08-a416-1f0427dd2e9e.log'. 2018-12-09 14:48:28,295+01 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Termination. 2018-12-09 14:48:28,389+01 ERROR [org.ovirt.engine.core.uutils.ssh.SSHDialog] (EE-ManagedThreadFactory-engine-Thread-1942) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] SSH error running command root@kvm380.durchhalten.intern:'umask 0077; MYTMP="$(TMPDIR="${OVIRT_TMPDIR}" mktemp -d -t ovirt-XX)"; trap "chmod -R u+rwX \"${MYTMP}\" > /dev/null 2>&1; rm -fr \"${MYTMP}\" > /dev/null 2>&1" 0; tar --warning=no-timestamp -C "${MYTMP}" -x && "${MYTMP}"/ovirt-host-deploy DIALOG/dialect=str:machine
[ovirt-users] To much hosts in this cluster
Hello, i had 3 hosts with ovirt running, with one of this 3 host i had problems during boot up. I decied to remove the host from the cluster. Now i ve two hosts But if I take a look at Hosted Engine in cockpit i see all three hosts Why ist he kvm380 not removed? How can I remove it? Thx shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/VZAOYESLQZDD7TJ4P5COXCOR7LST4RMN/
[ovirt-users] Network Setup after ovirt node Install and before add node to cluster
Hello, I ve downloaded an installed OVIRT NODE 4.2.7.1. During the installation I ‘ve setup a static ip (I ‘ve also tried dhcp, both work) The installation passes and after the reboot ovirt node starts without any problems. BUT the network is not configured. I know https://www.ovirt.org/documentation/install-guide/chap-oVirt_Nodes/ and I know https://www.ovirt.org/documentation/admin-guide/chap-Hosts/ and now I am between these two documents oVirt is installed but network is not configured to add it to datacenter of course I know the scripts under /etc/sysconfig/network-scripts do I really have do edit it manuel? if I change ip in command line ip a add … dev.. and ip route add … it fails to add node to datacenter - with one of the last version I ‘ve added the IP address in bash - added the host to datacenter, - after reboot readd the ip address in bash after reboot - an then I could change network settings in oVirt Manger What is best practices for adding a node to datacenter when the ip address is not assigned after installation Thx shb ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/TH5MRXJUU4SI62Z5XXQX2S6GKBOQIJWW/