[ovirt-users] Can not connect to gluster storage

2020-11-27 Thread Stefan Wolf
Hello,
I ve a host that can not connet to gluster storage.
It has worked since I ve set up the environment, and today it stoped working

this are the error messages in the webui
The error message for connection kvm380.durchhalten.intern:/data returned by 
VDSM was: Failed to fetch Gluster Volume List
Failed to connect Host kvm380.durchhalten.intern to the Storage Domains data.
Failed to connect Host kvm380.durchhalten.intern to the Storage Domains 
hosted_storage.


and here the vdsm.log

StorageDomainDoesNotExist: Storage domain does not exist: 
(u'36663740-576a-4498-b28e-0a402628c6a7',)
2020-11-27 12:59:07,665+ INFO  (jsonrpc/2) [storage.TaskManager.Task] 
(Task='8bed48b8-0696-4d3f-966a-119219f3b013') aborting: Task is aborted: 
"Storage domain does not exist: (u'36663740-576a-4498-b28e-0a402628c6a7',)" - 
code 358 (task:1181)
2020-11-27 12:59:07,665+ ERROR (jsonrpc/2) [storage.Dispatcher] FINISH 
getStorageDomainInfo error=Storage domain does not exist: 
(u'36663740-576a-4498-b28e-0a402628c6a7',) (dispatcher:83)
2020-11-27 12:59:07,666+ INFO  (jsonrpc/2) [jsonrpc.JsonRpcServer] RPC call 
StorageDomain.getInfo failed (error 358) in 0.38 seconds (__init__:312)
2020-11-27 12:59:07,698+ INFO  (jsonrpc/7) [vdsm.api] START 
connectStorageServer(domType=7, spUUID=u'----', 
conList=[{u'id': u'e29cf818-5ee5-46e1-85c1-8aeefa33e95d', u'vfs_type': 
u'glusterfs', u'connection': u'kvm380.durchhalten.intern:/engine', u'user': 
u'kvm'}], options=None) from=::1,40964, 
task_id=3a3eeb80-50ef-4710-a4f4-9d35da2ff281 (api:48)
2020-11-27 12:59:07,871+ ERROR (jsonrpc/7) [storage.HSM] Could not connect 
to storageServer (hsm:2420)
Traceback (most recent call last):
  File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 2417, in 
connectStorageServer
conObj.connect()
  File "/usr/lib/python2.7/site-packages/vdsm/storage/storageServer.py", line 
167, in connect
self.validate()
  File "/usr/lib/python2.7/site-packages/vdsm/storage/storageServer.py", line 
297, in validate
if not self.volinfo:
  File "/usr/lib/python2.7/site-packages/vdsm/storage/storageServer.py", line 
284, in volinfo
self._volinfo = self._get_gluster_volinfo()
  File "/usr/lib/python2.7/site-packages/vdsm/storage/storageServer.py", line 
329, in _get_gluster_volinfo
self._volfileserver)
  File "/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 56, in 
__call__
return callMethod()
  File "/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 54, in 

**kwargs)
  File "", line 2, in glusterVolumeInfo
  File "/usr/lib64/python2.7/multiprocessing/managers.py", line 773, in 
_callmethod
raise convert_to_error(kind, result)
GlusterVolumesListFailedException: Volume list failed: rc=30806 out=() 
err=['Volume does not exist']
2020-11-27 12:59:07,871+ INFO  (jsonrpc/7) [vdsm.api] FINISH 
connectStorageServer return={'statuslist': [{'status': 4149, 'id': 
u'e29cf818-5ee5-46e1-85c1-8aeefa33e95d'}]} from=::1,40964, 
task_id=3a3eeb80-50ef-4710-a4f4-9d35da2ff281 (api:54)
2020-11-27 12:59:07,871+ INFO  (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC call 
StoragePool.connectStorageServer succeeded in 0.18 seconds (__init__:312)
2020-11-27 12:59:08,474+ INFO  (Reactor thread) 
[ProtocolDetector.AcceptorImpl] Accepted connection from ::1:40966 
(protocoldetector:61)
2020-11-27 12:59:08,484+ INFO  (Reactor thread) [ProtocolDetector.Detector] 
Detected protocol stomp from ::1:40966 (protocoldetector:125)
2020-11-27 12:59:08,484+ INFO  (Reactor thread) [Broker.StompAdapter] 
Processing CONNECT request (stompserver:95)
2020-11-27 12:59:08,485+ INFO  (JsonRpc (StompReactor)) 
[Broker.StompAdapter] Subscribe command received (stompserver:124)
2020-11-27 12:59:08,525+ INFO  (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call 
Host.ping2 succeeded in 0.00 seconds (__init__:312)
2020-11-27 12:59:08,529+ INFO  (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call 
Host.ping2 succeeded in 0.00 seconds (__init__:312)
2020-11-27 12:59:08,533+ INFO  (jsonrpc/6) [vdsm.api] START 
getStorageDomainInfo(sdUUID=u'36663740-576a-4498-b28e-0a402628c6a7', 
options=None) from=::1,40966, task_id=ee3ac98e-6a93-4cb2-a626-5533c8fb78ad 
(api:48)
2020-11-27 12:59:08,909+ INFO  (jsonrpc/6) [vdsm.api] FINISH 
getStorageDomainInfo error=Storage domain does not exist: 
(u'36663740-576a-4498-b28e-0a402628c6a7',) from=::1,40966, 
task_id=ee3ac98e-6a93-4cb2-a626-5533c8fb78ad (api:52)
2020-11-27 12:59:08,910+ ERROR (jsonrpc/6) [storage.TaskManager.Task] 
(Task='ee3ac98e-6a93-4cb2-a626-5533c8fb78ad') Unexpected error (task:875)
Traceback (most recent call last):
  File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, in 
_run
return fn(*args, **kargs)
  File "", line 2, in getStorageDomainInfo
  File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 50, in method
ret = func(*args, **kwargs)
  File 

[ovirt-users] Re: How to Backup a VM

2020-08-31 Thread Stefan Wolf
I think, I found the problem.

It is case sensitive. For the export it is NOT case sensitive but for the step 
"wait for export" it is. I ve changed it and now it seems to be working
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RYFMBHZTJF76RT56HWUK5EV3ETB5QCSV/


[ovirt-users] Re: How to Backup a VM

2020-08-30 Thread Stefan Wolf
OK,

I ve run the backup three times .
I still have two machines, where it still fails on TASK [Wait for export] 
I think the Problem is not the  timeout, in oVirt engine the export has already 
finished : "
Exporting VM VMName as an OVA to /home/backup/in_progress/VMName.ova on Host 
kvm360"
But [Wait for export] still counts to 1 exit with error and move on to the next 
task

bye shb
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/W65G6ZUL6C6UJAJI627WVGITGIUUJ2XZ/


[ovirt-users] Re: How to Backup a VM

2020-08-30 Thread Stefan Wolf
yes you are right,

I ve already found. But this was not realy my problem. It causes from the 
HostedEngine. Long time ago I ve decreased the memory. It seems that this was 
the problem. now it is seems to be working pretty well. 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KVTFZHUV7URK6GBEEO5OE2BDYIFPRJ5F/


[ovirt-users] Re: How to Backup a VM

2020-08-30 Thread Stefan Wolf
Hello,

>https://github.com/silverorange/ovirt_ansible_backup
I am also still using 4.3.
In my opinion this is by far the best and easiest solution for disaster 
recovery. No need to install an appliance, and if there is a need to recover, 
you can import the ova in every hypervisor - no databases, no dependency.

Sometimes I ve issues with "TASK [Wait for export] " sometime it takes to long 
to export the ova. an I also had the problem, that the export already finished, 
but it was not realized by the script. In ovirt the export was finished and the 
filename was renamed from *.tmp to *.ova

maybe you have an idea for me.

thanks bye
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/Q7TKVK5TL6HT7DQZCY354ICK5J3JRDH4/


[ovirt-users] How to Backup a VM

2020-08-29 Thread Stefan Wolf
Hello to all

I try to backup a normal VM. But It seems that I don't really understand the 
concept. At first I found the possibility to backup with the api  
https://www.ovirt.org/documentation/administration_guide/#Setting_a_storage_domain_to_be_a_backup_domain_backup_domain.
Create a snapshot of the VM, finding the ID of the snapshot and the 
configuration of the VM makes sense to me.
But at this point, I would download the config an the snapshot and put it to my 
backup storage. And not create a new VM attach the disk and run a backup with 
backup programm. And for restoring do the sam way backwards. 

If i look at other project, there seems do be a way to download the snapshot 
and configfile, or am I wrong?
Maybe someone can explain to me why I should use additional software to install 
in an additional machine. Or even better someone can explain to me how I don't 
have to use additional backup software.

And to the same topic backup.
There is in the documentation the possibility to set up a backup storage
It is nearly the same, create a snapshot, or clone the machine and export it to 
backup storage
> Export the new virtual machine to a backup domain. See Exporting a Virtual 
> Machine to a Data Domain in the Virtual Machine Management Guide.
Sadly there is just writen what to do, not how, the link points to 404 page. 
maybe someone can explain to me how to use backup storage

thank you very much

shb
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/COR6VIV477XUFDKJAVEO2ODCESVENKLV/


[ovirt-users] Re: oVirt 4.4.1 HCI single server deployment failed nested-kvm

2020-07-30 Thread Stefan Wolf
Hello,

I ve the same problem, I ve already set up glusterfs and like to deplay a self 
hosted engine. I ve allready ovirt self hosted engine deployed, with no 
problems. But here I get the same error with hosted-engine --deploy and in the 
web frontend.

[ INFO  ] TASK [ovirt.hosted_engine_setup : Failed if only teaming devices are 
availible]
[ ERROR ] fatal: [localhost]: FAILED! => {"msg": "The conditional check 
'(otopi_host_net.ansible_facts.otopi_host_net | length == 0)' failed. The error 
was: error while evaluating conditional 
((otopi_host_net.ansible_facts.otopi_host_net | length == 0)): 'list object' 
has no attribute 'ansible_facts'\n\nThe error appears to be in 
'/usr/share/ansible/roles/ovirt.hosted_engine_setup/tasks/filter_team_devices.yml':
 line 29, column 13, but may\nbe elsewhere in the file depending on the exact 
syntax problem.\n\nThe offending line appears to be:\n\n- debug: 
var=otopi_host_net\n^ here\n\nThere appears to be both 'k=v' 
shorthand syntax and YAML in this task. Only one syntax may be used.\n"}
[ ERROR ] Failed to execute stage 'Environment customization': Failed executing 
ansible-playbook


is there a solution or just use an older ovirt `node iso?

bye shb
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/L2T6ENJRXFLQNJ6DMQ5H4WWV2TJLJJCE/


[ovirt-users] Host after update NonResponsive

2020-03-13 Thread Stefan Wolf
Hello to all,

I ve done a normal host updadet at the webfrontend.
After the reboot the host is NonResponsive.

It is a HCI setup with glusterfs
But mount shows on this host that only engine is mountet.
The data volume is not mounted.

[root@kvm320 ~]# mount|grep _engine
kvm380.durchhalten.intern:/engine on 
/rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine type 
fuse.glusterfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072)
[root@kvm320 ~]# mount|grep _data
[root@kvm320 ~]# #

on every other host it looks like this

[root@kvm10 ~]# mount|grep _engine
kvm380.durchhalten.intern:/engine on 
/rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine type 
fuse.glusterfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072)
[root@kvm10 ~]# mount|grep _data
kvm380.durchhalten.intern:/data on 
/rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_data type 
fuse.glusterfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072)
[root@kvm10 ~]#


here so more informaton
[root@kvm320 ~]# systemctl status glusterd -l
● glusterd.service - GlusterFS, a clustered file-system server
   Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled; vendor 
preset: disabled)
   Active: active (running) since Fr 2020-03-13 14:19:03 CET; 1h 49min ago
 Docs: man:glusterd(8)
  Process: 9263 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid 
--log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS)
 Main PID: 9264 (glusterd)
Tasks: 114
   CGroup: /system.slice/glusterd.service
   ├─3839 /usr/sbin/glusterfsd -s kvm320.durchhalten.intern 
--volfile-id data.kvm320.durchhalten.intern.gluster_bricks-data -p 
/var/run/gluster/vols/data/kvm320.durchhalten.intern-gluster_bricks-data.pid -S 
/var/run/gluster/1fd58e7c80335308.socket --brick-name /gluster_bricks/data -l 
/var/log/glusterfs/bricks/gluster_bricks-data.log --xlator-option 
*-posix.glusterd-uuid=ce474774-436a-41d3-bfdd-ab153ac77830 --process-name brick 
--brick-port 49152 --xlator-option data-server.listen-port=49152
   ├─3896 /usr/sbin/glusterfsd -s kvm320.durchhalten.intern 
--volfile-id engine.kvm320.durchhalten.intern.gluster_bricks-engine -p 
/var/run/gluster/vols/engine/kvm320.durchhalten.intern-gluster_bricks-engine.pid
 -S /var/run/gluster/5d4bcd552e3a3806.socket --brick-name 
/gluster_bricks/engine -l /var/log/glusterfs/bricks/gluster_bricks-engine.log 
--xlator-option *-posix.glusterd-uuid=ce474774-436a-41d3-bfdd-ab153ac77830 
--process-name brick --brick-port 49153 --xlator-option 
engine-server.listen-port=49153
   ├─4032 /usr/sbin/glusterfsd -s kvm320.durchhalten.intern 
--volfile-id home.kvm320.durchhalten.intern.gluster_bricks-home -p 
/var/run/gluster/vols/home/kvm320.durchhalten.intern-gluster_bricks-home.pid -S 
/var/run/gluster/050dbbce51bc7cb8.socket --brick-name /gluster_bricks/home -l 
/var/log/glusterfs/bricks/gluster_bricks-home.log --xlator-option 
*-posix.glusterd-uuid=ce474774-436a-41d3-bfdd-ab153ac77830 --process-name brick 
--brick-port 49154 --xlator-option home-server.listen-port=49154
   ├─9264 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level INFO
   └─9381 /usr/sbin/glusterfs -s localhost --volfile-id 
gluster/glustershd -p /var/run/gluster/glustershd/glustershd.pid -l 
/var/log/glusterfs/glustershd.log -S /var/run/gluster/6b30e1d260d31419.socket 
--xlator-option *replicate*.node-uuid=ce474774-436a-41d3-bfdd-ab153ac77830 
--process-name glustershd --client-pid=-6

Mär 13 14:19:02 kvm320.durchhalten.intern systemd[1]: Starting GlusterFS, a 
clustered file-system server...
Mär 13 14:19:03 kvm320.durchhalten.intern systemd[1]: Started GlusterFS, a 
clustered file-system server.
Mär 13 14:19:12 kvm320.durchhalten.intern glusterd[9264]: [2020-03-13 
13:19:12.869310] C [MSGID: 106003] 
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action] 0-management: 
Server quorum regained for volume data. Starting local bricks.
Mär 13 14:19:13 kvm320.durchhalten.intern glusterd[9264]: [2020-03-13 
13:19:13.115484] C [MSGID: 106003] 
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action] 0-management: 
Server quorum regained for volume engine. Starting local bricks.

everything seems to be connected

[root@kvm380 ~]# gluster peer status
Number of Peers: 3

Hostname: kvm360.durchhalten.intern
Uuid: 2f6cf198-afe5-4e61-b018-e7e4c10793b4
State: Peer in Cluster (Connected)

Hostname: kvm320.durchhalten.intern
Uuid: ce474774-436a-41d3-bfdd-ab153ac77830
State: Peer in Cluster (Connected)
Other names:
192.168.200.231

Hostname: kvm10
Uuid: 33cd77a6-3cda-4e21-bd45-a907044f410b
State: Peer in Cluster (Connected)
Other names:
kvm10

and this is the state of the brick from engine and data

[root@kvm320 ~]# gluster volume status engine
Status of volume: engine
Gluster process TCP Port  RDMA Port  Online  Pid

[ovirt-users] Memory problem

2020-01-22 Thread Stefan Wolf
Hi to all,

I ve a memory problem
I got this error:
Used memory of host kvm380.durchhalten.intern in cluster Default [96%] exceeded 
defined threshold [95%].

after reviewing the server with top command, I found ovn-controller with heavy 
memory usage:

45055 root  10 -10   46,5g  45,4g   2400 S  51,3 72,4 627:58.60 
ovn-controller

after restarting ovn-controller, every thing is fine again.
after some days, it uses so much memory again. I ve also tried to wait a day or 
two, after that it seems to restart it self.

is there a solution that ovn-controller does not use so much memory or restart 
automatic?

thx

shb
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/CW4YPJ5PEOR3RQQ7VDBZ7BYQ6A4LKP3Q/


[ovirt-users] Re: Gluster: a lof of Number of ntries in heal pending

2020-01-21 Thread Stefan Wolf
Hello

>I hope you plan to add another brick or arbiter, as you are now prone to 
>split-brain and
other issues.
Yes I will add an other one, but I think this is not a problem. I ve set 
cluster.server-quorum-ratio to 51% to avid the split brain problem. of course I 
know I just have failure tolarance of one.

I ve solved the probelem with removeing the brick

gluster volume remove-brick data replica 3 
kvm320.durchhalten.intern:/gluster_bricks/data/ force
Remove-brick force will not migrate files from the removed bricks, so they will 
no longer be available on the volume.
Do you want to continue? (y/n) y
volume remove-brick commit force: success
[root@kvm380 ~]# gluster volume heal data info summary
Brick kvm10:/gluster_bricks/data
Status: Connected
Total Number of entries: 1
Number of entries in heal pending: 1
Number of entries in split-brain: 0
Number of entries possibly healing: 0

Brick kvm360.durchhalten.intern:/gluster_bricks/data
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0

Brick kvm380.durchhalten.intern:/gluster_bricks/data
Status: Connected
Total Number of entries: 1
Number of entries in heal pending: 1
Number of entries in split-brain: 0
Number of entries possibly healing: 0

after this heal pending was ok for me.
than I removed all files - from this node - and added it back again.

now everything is find

Best Regards
Stefan
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3HZTSL3YY5GBVSREKXBY2KSZXJCPGVMJ/


[ovirt-users] Re: Gluster: a lof of Number of ntries in heal pending

2020-01-20 Thread Stefan Wolf
Hi Strahil,

yes it is a replica 4 set
I ve tried to stop and stop every gluster server,
and Ive rebooted every server.

or should I remove the brick and add it again?

bye
stefan
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LLZ7NIGVYB33PXA3ZLJBK4GDNSSFWJHU/


[ovirt-users] Re: Gluster: a lof of Number of ntries in heal pending

2020-01-20 Thread Stefan Wolf
yes, I ve already tried a full heal a week a go.

how do i perform a manual heal?

I only have this gfid:











Status: Connected
Number of entries: 868

I ve tried to heal it with:
[root@kvm10 ~]# gluster volume heal data split-brain latest-mtime 
gfid:c2b47c5c-89b6-49ac-bf10-1733dd8f0902
Healing gfid:c2b47c5c-89b6-49ac-bf10-1733dd8f0902 failed: File not in 
split-brain.
Volume heal failed.

(the last entry)

And if i understood it correct ther is no split-brain

[root@kvm10 ~]# gluster volume heal data info split-brain
Brick kvm10:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

Brick kvm320.durchhalten.intern:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

Brick kvm360.durchhalten.intern:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

Brick kvm380.durchhalten.intern:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AQMSKCYGKKKUZIRXSKBW2VWWPLYVEX7A/


[ovirt-users] Gluster: a lof of Number of ntries in heal pending

2020-01-20 Thread Stefan Wolf
Hello to all,

I ve a problem with gluster

[root@kvm10 ~]# gluster volume heal data info summary
Brick kvm10:/gluster_bricks/data
Status: Connected
Total Number of entries: 868
Number of entries in heal pending: 868
Number of entries in split-brain: 0
Number of entries possibly healing: 0

Brick kvm320.durchhalten.intern:/gluster_bricks/data
Status: Connected
Total Number of entries: 1
Number of entries in heal pending: 1
Number of entries in split-brain: 0
Number of entries possibly healing: 0

Brick kvm360.durchhalten.intern:/gluster_bricks/data
Status: Connected
Total Number of entries: 867
Number of entries in heal pending: 867
Number of entries in split-brain: 0
Number of entries possibly healing: 0

Brick kvm380.durchhalten.intern:/gluster_bricks/data
Status: Connected
Total Number of entries: 868
Number of entries in heal pending: 868
Number of entries in split-brain: 0
Number of entries possibly healing: 0

[root@kvm10 ~]# gluster volume heal data info split-brain
Brick kvm10:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

Brick kvm320.durchhalten.intern:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

Brick kvm360.durchhalten.intern:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

Brick kvm380.durchhalten.intern:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

As I understand there is no split-brain but 868 files ar in state heal pending.
I ve restarted every node.

I ve also tried:
[root@kvm10 ~]# gluster volume heal data full
Launching heal operation to perform full self heal on volume data has been 
successful
Use heal info commands to check status.

but even after a week there is no really change ( I started with 912 Number of 
entries in heal pending)

can somebody tell what exactly is the problem and how can I solve it.

thank you very much
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/PN63LC3OBQOM7IQY763ZS5V6VZDUFPNP/


[ovirt-users] Node not starting | blk_cloned_rq_check_limits: over max size limit

2019-12-31 Thread Stefan Wolf
hi all,

i ve 4 nodes running with current ovirt.
I ve only a problem on one host even after a fresh installation .
I ve installed the latest image.
Than I add the node to the cluster
Everything is working good.
After this I configure the network.
BUT, after a restart the host does not come up again.
I got this error: blk_cloned_rq_check_limits: over max size limit
every 5 seconds

I can continue with control-D
or I can login with root password to fix the problem. but i dont know what is 
the problem and where does it came from

I ve also changed the sas disk to nvme storage, but I ve changed this on every 
host. And this problem exists only on one host

i found this https://lists.centos.org/pipermail/centos/2017-December/167727.html
the output is
[root@kvm380 ~]# ./test.sh
Sys Block Node: Devicemax_sectors_kb  
max_hw_sectors_kb
/sys/block/dm-0   : onn_kvm380-pool00_tmeta   256 4096
/sys/block/dm-1   : onn_kvm380-pool00_tdata   256 4096
/sys/block/dm-10  : onn_kvm380-var256 4096
/sys/block/dm-11  : onn_kvm380-tmp256 4096
/sys/block/dm-12  : onn_kvm380-home   256 4096
/sys/block/dm-13  : onn_kvm380-var_crash  256 4096
/sys/block/dm-2   : onn_kvm380-pool00-tpool   256 4096
/sys/block/dm-3   : onn_kvm380-ovirt--node--ng--4.3.7--0.20191121.0+1 256   
  4096
/sys/block/dm-4   : onn_kvm380-swap   256 4096
/sys/block/dm-5   : eui.0025385991b1e27a  512 2048
/sys/block/dm-6   : eui.0025385991b1e27a1 512 2048
/sys/block/dm-7   : onn_kvm380-pool00 256 4096
/sys/block/dm-8   : onn_kvm380-var_log_audit  256 4096
/sys/block/dm-9   : onn_kvm380-var_log256 4096
cat: /sys/block/nvme0n1/device/vendor: Datei oder Verzeichnis nicht gefunden
/sys/block/nvme0n1:  Samsung SSD 970 EVO 1TB  512 
2048
/sys/block/sda: HP   LOGICAL VOLUME   256 4096

is the nvme not starting correct
[root@kvm380 ~]# systemctl status multipathd
● multipathd.service - Device-Mapper Multipath Device Controller
   Loaded: loaded (/usr/lib/systemd/system/multipathd.service; enabled; vendor 
preset: enabled)
   Active: active (running) since Di 2019-12-31 16:16:32 CET; 31min ago
  Process: 1919 ExecStart=/sbin/multipathd (code=exited, status=0/SUCCESS)
  Process: 1916 ExecStartPre=/sbin/multipath -A (code=exited, status=0/SUCCESS)
  Process: 1911 ExecStartPre=/sbin/modprobe dm-multipath (code=exited, 
status=0/SUCCESS)
 Main PID: 1921 (multipathd)
Tasks: 7
   CGroup: /system.slice/multipathd.service
   └─1921 /sbin/multipathd

Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: nvme0n1: mark as 
failed
Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: 
eui.0025385991b1e27a: Entering recovery mode: max_retries=4
Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: 
eui.0025385991b1e27a: remaining active paths: 0
Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: 259:0: reinstated
Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: 
eui.0025385991b1e27a: queue_if_no_path enabled
Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: 
eui.0025385991b1e27a: Recovered to normal mode
Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: 
eui.0025385991b1e27a: remaining active paths: 1
Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: nvme0n1: mark as 
failed
Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: 
eui.0025385991b1e27a: Entering recovery mode: max_retries=4
Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: 
eui.0025385991b1e27a: remaining active paths: 0

why is it marked as failed?

if i create a new volume with cockpit and use it for bricks for gluster, every 
thing is fine. until reboot


maybe some one can point me the direction
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MHHFFWAY5T5BVTD5TAMW2R6PC6SMGGA2/


[ovirt-users] Re: Cannot start VM 2 hosts

2019-12-17 Thread Stefan Wolf
Problem found.
on these 2 hosts was the firewall disabled
bye shb
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/Q64MJ7IXIEA2VE6BFDQNWZMCOCK3JRIN/


[ovirt-users] Cannot start VM 2 hosts

2019-12-17 Thread Stefan Wolf
I ve got 4 Hosts 
after changeing the harddrive on every Host and normal updates I am not able 
anymore to start VM's on 2 of  these 4 hosts.
I am also not able to migrate a running vm to these hosts.

This is the error log for start up

2019-12-17 16:01:28,326+01 INFO  [org.ovirt.engine.core.bll.RunVmOnceCommand] 
(default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] Lock Acquired to object 
'EngineLock:{exclusiveLocks='[40cf6c27-6464-4fa7-bc01-9158cb03328b=VM]', 
sharedLocks=''}'
2019-12-17 16:01:28,609+01 INFO  
[org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] (default 
task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] START, 
IsVmDuringInitiatingVDSCommand( 
IsVmDuringInitiatingVDSCommandParameters:{vmId='40cf6c27-6464-4fa7-bc01-9158cb03328b'}),
 log id: 5706d380
2019-12-17 16:01:28,611+01 INFO  
[org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] (default 
task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] FINISH, 
IsVmDuringInitiatingVDSCommand, return: false, log id: 5706d380
2019-12-17 16:01:28,854+01 INFO  [org.ovirt.engine.core.bll.RunVmOnceCommand] 
(default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] Running command: 
RunVmOnceCommand internal: false. Entities affected :  ID: 
40cf6c27-6464-4fa7-bc01-9158cb03328b Type: VMAction group RUN_VM with role type 
USER,  ID: 40cf6c27-6464-4fa7-bc01-9158cb03328b Type: VMAction group 
EDIT_ADMIN_VM_PROPERTIES with role type ADMIN
2019-12-17 16:01:28,999+01 INFO  
[org.ovirt.engine.core.vdsbroker.UpdateVmDynamicDataVDSCommand] (default 
task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] START, 
UpdateVmDynamicDataVDSCommand( 
UpdateVmDynamicDataVDSCommandParameters:{hostId='null', 
vmId='40cf6c27-6464-4fa7-bc01-9158cb03328b', 
vmDynamic='org.ovirt.engine.core.common.businessentities.VmDynamic@147620b'}), 
log id: 612dabdf
2019-12-17 16:01:29,097+01 INFO  
[org.ovirt.engine.core.vdsbroker.UpdateVmDynamicDataVDSCommand] (default 
task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] FINISH, 
UpdateVmDynamicDataVDSCommand, return: , log id: 612dabdf
2019-12-17 16:01:29,106+01 INFO  
[org.ovirt.engine.core.vdsbroker.CreateVDSCommand] (default task-5) 
[3beff980-0f1a-461b-a6da-89c6f4dea6b3] START, CreateVDSCommand( 
CreateVDSCommandParameters:{hostId='d38bae4c-8494-4861-ae5a-38992db338e5', 
vmId='40cf6c27-6464-4fa7-bc01-9158cb03328b', vm='VM [z-push]'}), log id: 
56abde11
2019-12-17 16:01:29,110+01 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] (default 
task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] START, 
CreateBrokerVDSCommand(HostName = kvm360.durchhalten.intern, 
CreateVDSCommandParameters:{hostId='d38bae4c-8494-4861-ae5a-38992db338e5', 
vmId='40cf6c27-6464-4fa7-bc01-9158cb03328b', vm='VM [z-push]'}), log id: 
36d088ab
2019-12-17 16:01:29,163+01 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] (default 
task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] VM http://ovirt.org/vm/tune/1.0; 
xmlns:ovirt-vm="http://ovirt.org/vm/1.0;>
  z-push
  40cf6c27-6464-4fa7-bc01-9158cb03328b
  4194304
  4194304
  1
  16777216
  16
  

  oVirt
  OS-NAME:
  OS-VERSION:
  HOST-SERIAL:
  40cf6c27-6464-4fa7-bc01-9158cb03328b

  
  



  
  

  
  
SandyBridge





  

  
  
  


  
  


  
  


  /dev/urandom
  


  
  


  
  
  


  
  
  


  


  
  


  
  
  
  
  
  
  
  
  


  


  
  
  
  
  
  
  
  
  
  
  


  
  

  
  
  
  
  


  
  

  
  
  
  
  
  38f2e3bb-687f-4f01-b769-3a5403136181

  
  


  
  
hvm


  
  


  4096
  4.3
  
  

  
  
0e279f02-0c73-11e9-bc47-00163e150480

1a9627d8-181b-4790-aa47-569dd1e6e9fb

38f2e3bb-687f-4f01-b769-3a5403136181

80364cc8-8afd-4482-884d-d63d5a7988df
  
  false
  auto_resume

  


2019-12-17 16:01:29,198+01 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] (default 
task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] FINISH, CreateBrokerVDSCommand, 
return: , log id: 36d088ab
2019-12-17 16:01:29,252+01 INFO  
[org.ovirt.engine.core.vdsbroker.CreateVDSCommand] (default task-5) 
[3beff980-0f1a-461b-a6da-89c6f4dea6b3] FINISH, CreateVDSCommand, return: 
WaitForLaunch, log id: 56abde11
2019-12-17 16:01:29,252+01 INFO  [org.ovirt.engine.core.bll.RunVmOnceCommand] 
(default task-5) [3beff980-0f1a-461b-a6da-89c6f4dea6b3] Lock freed to object 
'EngineLock:{exclusiveLocks='[40cf6c27-6464-4fa7-bc01-9158cb03328b=VM]', 
sharedLocks=''}'
2019-12-17 16:01:29,319+01 INFO  

[ovirt-users] Re: Increase memory of hosted engine

2019-12-08 Thread Stefan Wolf
thank you for your replay,

I am happy if it works
1. check
2. check
3.1 check -> I cant see where is use this config file later, or it is just a 
backup?
4.1 check i changed to
  16384
  16384
and

  15625000
  15625000
  15625000

after write, quite und open these values have changed back to
  4194304
  4194304

BREAK ;) 

after stopping the hosted engine with hosted-engine --vm-shutdown and powering 
it up with hosted-engine vm-start
everything is reverted ;(

BREAK

hosted engine is now up and running, maintenance ist still global,
changed again with virtsh edit HostedEngine. as written above it changes back

I will wait for tomorrow.

Or did I missunderstand something

In which file the maxmem is hidden?

thx shb
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DWVBZO2OE6U62FAEMIIS3TG2C6RRV4OH/


[ovirt-users] Increase memory of hosted engine

2019-12-08 Thread Stefan Wolf
hello,

I ve decrease the memory of the hosted engine.
now I am not able to increase the memory permantly
right now the memory has 4096 MB
Max Memory is 7936 MB and
guaranted memory is 7936MB

I can increase the memory up to 7936 MB in the manager, it changes immediately.
I can not increase to more then 7936 MB because this is max Memory

If I change Max memory to 16384
the gueanted memery gets a red border an the error is max memory ist 4096 MB

I ve found this one https://access.redhat.com/solutions/964693
and tried to change the memSize in /etc/ovirt-hosted-engine/vm.conf to 16384 
I ve also stoppt every ovirt-ha-agent and changed every vm.conf.
After starting ovirt-ha-agent every vm.conf changed to 4096 again

has somebody a solution to increase the memory again?

thx shb
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LZHYZ5IIFVBJT2UTDSQBTMEP4TZ67CTO/


[ovirt-users] Re: ovirt-ha-agent not running

2019-12-07 Thread Stefan Wolf
you are right, thank you

now it is up and running again

[root@kvm380 ~]# ls 
/rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern\:_engine/36663740-576a-4498-b28e-0a402628c6a7/ha_agent/
 -lha
insgesamt 0
drwxr-xr-x. 2 vdsm kvm  67  1. Jan 1970  .
drwxr-xr-x. 6 vdsm kvm  64  7. Dez 12:08 ..
lrwxrwxrwx. 1 vdsm kvm 132  8. Dez 06:49 hosted-engine.lockspace -> 
/var/run/vdsm/storage/36663740-576a-4498-b28e-0a402628c6a7/de9b53fb-6c86-4036-9f1c-1f08d1828e2f/d41a5061-94b8-4e0d-84ec-8af0c547955a
lrwxrwxrwx. 1 vdsm kvm 132  8. Dez 06:49 hosted-engine.metadata -> 
/var/run/vdsm/storage/36663740-576a-4498-b28e-0a402628c6a7/aec89d8d-0c7c-42b8-8889-1713d7e8a667/39ec2ecd-57b5-4424-8666-d7b673b1e8b6
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3XDFLVVU67T6NMGGS6QA5ERTIQIQ6XRA/


[ovirt-users] Re: ovirt-ha-agent not running

2019-12-07 Thread Stefan Wolf
the contet is
[root@kvm380 ~]# ls 
/rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine/36663740-576a-4498-b28e-0a402628c6a7/ha_agent/
 -lha
ls: Zugriff auf 
/rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine/36663740-576a-4498-b28e-0a402628c6a7/ha_agent/hosted-engine.lockspace
 nicht möglich: Der Socket ist nicht verbunden
ls: Zugriff auf 
/rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine/36663740-576a-4498-b28e-0a402628c6a7/ha_agent/hosted-engine.metadata
 nicht möglich: Der Socket ist nicht verbunden
insgesamt 512
drwxr-xr-x. 2 vdsm kvm 67 20. Nov 09:20 .
drwxr-xr-x. 6 vdsm kvm 64  7. Dez 12:08 ..
l?? ? ??? ? hosted-engine.lockspace
l?? ? ??? ? hosted-engine.metadata


I ve already tryed to stop and start  the broker again, ovirt-ha-agant will not 
start. And as wrote i ve done a complete reinstall of the ovirt host

bye stefan
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HCM6XS7FBSDQLXZWTSXHP7HWIYH23KXI/


[ovirt-users] Re: ovirt-ha-agent not running

2019-12-07 Thread Stefan Wolf
and here is the broker.log

MainThread::INFO::2019-12-07 
15:20:03,563::broker::47::ovirt_hosted_engine_ha.broker.broker.Broker::(run) 
ovirt-hosted-engine-ha broker 2.3.6 started
MainThread::INFO::2019-12-07 
15:20:03,564::monitor::40::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Searching for submonitors in 
/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/submonitors
MainThread::INFO::2019-12-07 
15:20:03,564::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor mem-free
MainThread::INFO::2019-12-07 
15:20:03,565::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor network
MainThread::INFO::2019-12-07 
15:20:03,566::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor mem-free
MainThread::INFO::2019-12-07 
15:20:03,566::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor engine-health
MainThread::INFO::2019-12-07 
15:20:03,567::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor cpu-load
MainThread::INFO::2019-12-07 
15:20:03,567::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor storage-domain
MainThread::INFO::2019-12-07 
15:20:03,568::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor mgmt-bridge
MainThread::INFO::2019-12-07 
15:20:03,569::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor cpu-load
MainThread::INFO::2019-12-07 
15:20:03,569::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor network
MainThread::INFO::2019-12-07 
15:20:03,574::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor cpu-load-no-engine
MainThread::INFO::2019-12-07 
15:20:03,575::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor cpu-load-no-engine
MainThread::INFO::2019-12-07 
15:20:03,576::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor mgmt-bridge
MainThread::INFO::2019-12-07 
15:20:03,577::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor engine-health
MainThread::INFO::2019-12-07 
15:20:03,577::monitor::49::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Loaded submonitor storage-domain
MainThread::INFO::2019-12-07 
15:20:03,577::monitor::50::ovirt_hosted_engine_ha.broker.monitor.Monitor::(_discover_submonitors)
 Finished loading submonitors
MainThread::INFO::2019-12-07 
15:20:03,651::storage_backends::373::ovirt_hosted_engine_ha.lib.storage_backends::(connect)
 Connecting the storage
MainThread::INFO::2019-12-07 
15:20:03,652::storage_server::349::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
 Connecting storage server
MainThread::INFO::2019-12-07 
15:20:03,716::storage_server::356::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
 Connecting storage server
MainThread::INFO::2019-12-07 
15:20:03,748::storage_server::413::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
 Refreshing the storage domain
MainThread::WARNING::2019-12-07 
15:20:06,985::storage_broker::97::ovirt_hosted_engine_ha.broker.storage_broker.StorageBroker::(__init__)
 Can't connect vdsm storage: [Errno 107] Transport endpoint is not connected: 
'/rhev/data-center/mnt/glusterSD/kvm380.durchhalten.intern:_engine/36663740-576a-4498-b28e-0a402628c6a7/ha_agent/hosted-engine.lockspace'


maybe it helps

bye shb
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KECDUEVCHP3OSUREULALUEJ6WSALLUDQ/


[ovirt-users] ovirt-ha-agent not running

2019-12-07 Thread Stefan Wolf
hello,

since some days ovirt-ha-agent is not running anymore 
i ve 4 ovirt hosts and only on one host the agent is running.
maybe it was from an update, because i lost one agent after an other.

i ve done a complete fresh install for the host with the latest ovirt node.
I ve got on tree hosts this error

[root@kvm380 ~]# systemctl status ovirt-ha-agent
● ovirt-ha-agent.service - oVirt Hosted Engine High Availability Monitoring 
Agent
   Loaded: loaded (/usr/lib/systemd/system/ovirt-ha-agent.service; enabled; 
vendor preset: disabled)
   Active: activating (auto-restart) (Result: exit-code) since Sa 2019-12-07 
14:56:21 UTC; 5s ago
  Process: 28002 ExecStart=/usr/share/ovirt-hosted-engine-ha/ovirt-ha-agent 
(code=exited, status=157)
 Main PID: 28002 (code=exited, status=157)
Tasks: 0
   CGroup: /system.slice/ovirt-ha-agent.service

Dez 07 14:56:21 kvm380.durchhalten.intern systemd[1]: ovirt-ha-agent.service: 
main process exited, code=exited, status=157/n/a
Dez 07 14:56:21 kvm380.durchhalten.intern systemd[1]: Unit 
ovirt-ha-agent.service entered failed state.
Dez 07 14:56:21 kvm380.durchhalten.intern systemd[1]: ovirt-ha-agent.service 
failed.

and this is in /var/log/ovirt-hosted-engine-ha/agent.log

MainThread::INFO::2019-12-07 
15:01:51,048::agent::67::ovirt_hosted_engine_ha.agent.agent.Agent::(run) 
ovirt-hosted-engine-ha agent 2.3.6 started
MainThread::INFO::2019-12-07 
15:01:51,161::hosted_engine::234::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_hostname)
 Found certificate common name: kvm380.durchhalten.intern
MainThread::INFO::2019-12-07 
15:01:51,374::hosted_engine::543::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_broker)
 Initializing ha-broker connection
MainThread::INFO::2019-12-07 
15:01:51,378::brokerlink::80::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(start_monitor)
 Starting monitor network, options {'tcp_t_address': None, 'network_test': 
None, 'tcp_t_port': None, 'addr': '192.168.200.1'}
MainThread::ERROR::2019-12-07 
15:01:51,379::hosted_engine::559::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_broker)
 Failed to start necessary monitors
MainThread::ERROR::2019-12-07 
15:01:51,381::agent::144::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
 Traceback (most recent call last):
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 
131, in _run_agent
return action(he)
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 
55, in action_proper
return he.start_monitoring()
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
 line 432, in start_monitoring
self._initialize_broker()
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
 line 556, in _initialize_broker
m.get('options', {}))
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", 
line 89, in start_monitor
).format(t=type, o=options, e=e)
RequestError: brokerlink - failed to start monitor via ovirt-ha-broker: [Errno 
2] No such file or directory, [monitor: 'network', options: {'tcp_t_address': 
None, 'network_test': None, 'tcp_t_port': None, 'addr': '192.168.200.1'}]

MainThread::ERROR::2019-12-07 
15:01:51,381::agent::145::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
 Trying to restart agent
MainThread::INFO::2019-12-07 
15:01:51,382::agent::89::ovirt_hosted_engine_ha.agent.agent.Agent::(run) Agent 
shutting down

maybe someone can give me an advice 

thx shb

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UE4TMWWT27BHFJWITYRV52CHG5AA7ZYD/


[ovirt-users] hosted engine does not start

2019-04-15 Thread Stefan Wolf
Hello all,

 

after a powerloss the hosted engine won’t start up anymore.

I ‘ve the current ovirt installed.

Storage is glusterfs und it is up and running

 

It is trying to start up hosted engine but it does not work, but I can’t see
where the problem is.

 

[root@kvm320 ~]# hosted-engine --vm-status

 

 

--== Host 1 status ==--

 

conf_on_shared_storage : True

Status up-to-date  : True

Hostname   : kvm380.durchhalten.intern

Host ID: 1

Engine status  : {"reason": "bad vm status", "health":
"bad", "vm": "down", "detail": "Down"}

Score  : 1800

stopped: False

Local maintenance  : False

crc32  : 3ad6d0bd

local_conf_timestamp   : 14594

Host timestamp : 14594

Extra metadata (valid at timestamp):

   metadata_parse_version=1

metadata_feature_version=1

timestamp=14594 (Mon Apr 15 21:25:12 2019)

host-id=1

score=1800

vm_conf_refresh_time=14594 (Mon Apr 15 21:25:12 2019)

conf_on_shared_storage=True

maintenance=False

state=GlobalMaintenance

stopped=False

 

 

--== Host 2 status ==--

 

conf_on_shared_storage : True

Status up-to-date  : True

Hostname   : kvm320.durchhalten.intern

Host ID: 2

Engine status  : {"reason": "failed liveliness check",
"health": "bad", "vm": "up", "detail": "Up"}

Score  : 0

stopped: False

Local maintenance  : False

crc32  : e7d4840d

local_conf_timestamp   : 21500

Host timestamp : 21500

Extra metadata (valid at timestamp):

metadata_parse_version=1

metadata_feature_version=1

timestamp=21500 (Mon Apr 15 21:25:22 2019)

host-id=2

score=0

vm_conf_refresh_time=21500 (Mon Apr 15 21:25:22 2019)

conf_on_shared_storage=True

maintenance=False

state=ReinitializeFSM

stopped=False

 

 

--== Host 3 status ==--

 

conf_on_shared_storage : True

Status up-to-date  : True

Hostname   : kvm360.durchhalten.intern

Host ID: 3

Engine status  : {"reason": "vm not running on this
host", "health": "bad", "vm": "down", "detail": "unknown"}

Score  : 1800

stopped: False

Local maintenance  : False

crc32  : cf9221cb

local_conf_timestamp   : 22121

Host timestamp : 22120

Extra metadata (valid at timestamp):

metadata_parse_version=1

metadata_feature_version=1

timestamp=22120 (Mon Apr 15 21:25:18 2019)

host-id=3

score=1800

vm_conf_refresh_time=22121 (Mon Apr 15 21:25:18 2019)

conf_on_shared_storage=True

maintenance=False

state=GlobalMaintenance

stopped=False

 

[root@kvm320 ~]# virsh -r list

IdName   Status



6 HostedEngine   laufend

 

[root@kvm320 ~]# hosted-engine --console

The engine VM is running on this host

Verbunden mit der Domain: HostedEngine

Escape-Zeichen ist ^]

Fehler: Interner Fehler: Zeichengerät  kann nicht gefunden warden

 

In engish it should be this

 

[root@mgmt~]# hosted-engine --console
The engine VM is running on this host
Connected to domain HostedEngine
Escape character is ^]
error: internal error: cannot find character device

 

This is in the log

 

[root@kvm320 ~]# tail -f /var/log/ovirt-hosted-engine-ha/agent.log

MainThread::INFO::2019-04-15
21:28:33,032::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine
.HostedEngine::(_monitoring_loop) Current state EngineStarting (score: 1800)

MainThread::INFO::2019-04-15
21:28:43,050::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.Hosted
Engine::(consume) VM is powering up..

MainThread::INFO::2019-04-15
21:28:43,165::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine
.HostedEngine::(_monitoring_loop) Current state EngineStarting (score: 1800)

MainThread::INFO::2019-04-15
21:28:53,183::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.Hosted
Engine::(consume) VM is powering up..

MainThread::INFO::2019-04-15
21:28:53,300::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine
.HostedEngine::(_monitoring_loop) Current state EngineStarting (score: 1800)

MainThread::INFO::2019-04-15
21:29:03,317::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.Hosted
Engine::(consume) VM is powering up..


[ovirt-users] Re: Hyperconvergend Setup stuck

2018-12-20 Thread Stefan Wolf
It is gdeploy 2.0.2

rpm -qa |grep gdeploy
gdeploy-2.0.8-1.el7.noarch
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3QUCUGUHSSRHESRSPGWENRRUND2K3QLK/


[ovirt-users] Re: Active Storage Domains as Problematic

2018-12-20 Thread Stefan Wolf
i 've mounted it during the hosted-engine --deploy process
I selected glusterfs
and entered  server:/engine
I dont enter any mount options
yes it is enabled for both. I dont got errors for the second one, but may it 
doesn't check after the first fail
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SYX6BQBW2MMV4YIXHG24KMXA7FTWL46X/


[ovirt-users] Re: Active Storage Domains as Problematic

2018-12-20 Thread Stefan Wolf
yes i think this too, but as you see at the top
>[root@kvm380 ~]# gluster volume info
>...
> performance.strict-o-direct: on
...
it was already set

i did a one cluster setup with ovirt and I uses this result

Volume Name: engine
Type: Distribute
Volume ID: a40e848b-a8f1-4990-9d32-133b46db6f1d
Status: Started
Snapshot Count: 0
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: kvm360.durchhalten.intern:/gluster_bricks/engine/engine
Options Reconfigured:
cluster.eager-lock: enable
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
user.cifs: off
network.ping-timeout: 30
network.remote-dio: off
performance.strict-o-direct: on
performance.low-prio-threads: 32
features.shard: on
storage.owner-gid: 36
storage.owner-uid: 36
transport.address-family: inet
nfs.disable: on

could there be an other reason?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HST6ODMSZQZEP6Q2OMLCAMD27HI2CDHQ/


[ovirt-users] Re: Active Storage Domains as Problematic

2018-12-20 Thread Stefan Wolf
here is what i found in the logs of the hosts

2018-12-20 12:34:04,824+0100 INFO  (periodic/0) [vdsm.api] START 
repoStats(domains=()) from=internal, 
task_id=09235382-a5b5-48da-853d-f94cae092684 (api:46)
2018-12-20 12:34:04,825+0100 INFO  (periodic/0) [vdsm.api] FINISH repoStats 
return={u'20651d3d-08d7-482a-ae4e-7cd0e33cc907': {'code': 399, 'actual': True, 
'version': -1, 'acquired': False, 'delay': '0', 'lastCheck': '6.1', 'valid': 
False}, u'ae9e4cbd-3946-481d-b01a-e8a38bf00efb': {'code': 0, 'actual': True, 
'version': 4, 'acquired': True, 'delay': '0.0013974', 'lastCheck': '1.1', 
'valid': True}} from=internal, task_id=09235382-a5b5-48da-853d-f94cae092684 
(api:52)
2018-12-20 12:34:04,826+0100 INFO  (periodic/0) [vdsm.api] START 
multipath_health() from=internal, task_id=8f6166cb-aa41-4f46-823d-d38e4e85f02a 
(api:46)
2018-12-20 12:34:04,826+0100 INFO  (periodic/0) [vdsm.api] FINISH 
multipath_health return={} from=internal, 
task_id=8f6166cb-aa41-4f46-823d-d38e4e85f02a (api:52)
2018-12-20 12:34:04,832+0100 INFO  (jsonrpc/4) [vdsm.api] START 
prepareImage(sdUUID=u'20651d3d-08d7-482a-ae4e-7cd0e33cc907', 
spUUID=u'----', 
imgUUID=u'c1ebc7ad-dfb6-4cc1-8e24-40f0be3f4afe', 
leafUUID=u'e7e76dd8-d166-46a0-9761-fa6391aa047b', allowIllegal=False) 
from=::1,55348, task_id=996206fc-65eb-4056-b3b9-2ac0e1780c2c (api:46)
2018-12-20 12:34:04,836+0100 ERROR (periodic/0) [root] failed to retrieve 
Hosted Engine HA score '[Errno 2] No such file or directory'Is the Hosted 
Engine setup finished? (api:196)
2018-12-20 12:34:04,847+0100 ERROR (jsonrpc/4) [storage.fileSD] Underlying file 
system doesn't supportdirect IO (fileSD:108)
2018-12-20 12:34:04,847+0100 INFO  (jsonrpc/4) [vdsm.api] FINISH prepareImage 
error=Storage Domain target is unsupported: () from=::1,55348, 
task_id=996206fc-65eb-4056-b3b9-2ac0e1780c2c (api:50)
2018-12-20 12:34:04,847+0100 ERROR (jsonrpc/4) [storage.TaskManager.Task] 
(Task='996206fc-65eb-4056-b3b9-2ac0e1780c2c') Unexpected error (task:875)
Traceback (most recent call last):
  File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, in 
_run
return fn(*args, **kargs)
  File "", line 2, in prepareImage
  File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, in method
ret = func(*args, **kwargs)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 3173, in 
prepareImage
dom = sdCache.produce(sdUUID)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 110, in 
produce
domain.getRealDomain()
  File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 51, in 
getRealDomain
return self._cache._realProduce(self._sdUUID)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 134, in 
_realProduce
domain = self._findDomain(sdUUID)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 151, in 
_findDomain
return findMethod(sdUUID)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/glusterSD.py", line 55, 
in findDomain
return GlusterStorageDomain(GlusterStorageDomain.findDomainPath(sdUUID))
  File "/usr/lib/python2.7/site-packages/vdsm/storage/fileSD.py", line 400, in 
__init__
validateFileSystemFeatures(manifest.sdUUID, manifest.mountpoint)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/fileSD.py", line 110, in 
validateFileSystemFeatures
raise se.StorageDomainTargetUnsupported()
StorageDomainTargetUnsupported: Storage Domain target is unsupported: ()
2018-12-20 12:34:04,847+0100 INFO  (jsonrpc/4) [storage.TaskManager.Task] 
(Task='996206fc-65eb-4056-b3b9-2ac0e1780c2c') aborting: Task is aborted: 
'Storage Domain target is unsupported: ()' - code 399 (task:1181)
2018-12-20 12:34:04,848+0100 ERROR (jsonrpc/4) [storage.Dispatcher] FINISH 
prepareImage error=Storage Domain target is unsupported: () (dispatcher:82)
2018-12-20 12:34:04,848+0100 INFO  (jsonrpc/4) [jsonrpc.JsonRpcServer] RPC call 
Image.prepare failed (error 399) in 0.02 seconds (__init__:573)
^

especially this part 

2018-12-20 12:34:04,847+0100 ERROR (jsonrpc/4) [storage.fileSD] Underlying file 
system doesn't supportdirect IO (fileSD:108)
2018-12-20 12:34:04,847+0100 INFO  (jsonrpc/4) [vdsm.api] FINISH prepareImage 
error=Storage Domain target is unsupported: () from=::1,55348, 
task_id=996206fc-65eb-4056-b3b9-2ac0e1780c2c (api:50)

and i am not sure why he is asking this
2018-12-20 12:34:04,836+0100 ERROR (periodic/0) [root] failed to retrieve 
Hosted Engine HA score '[Errno 2] No such file or directory'Is the Hosted 
Engine setup finished? (api:196)
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WKCPU5I4Z75TMZQQCQBAD4WK3QOLVTFR/


[ovirt-users] Active Storage Domains as Problematic

2018-12-20 Thread Stefan Wolf
Hello,

 

I ,ve setup a test lab with 3 nodes installed with centos 7

I configured manualy gluster fs. Glusterfs is up and running

 

[root@kvm380 ~]# gluster peer status

Number of Peers: 2

 

Hostname: kvm320.durchhalten.intern

Uuid: dac066db-55f7-4770-900d-4830c740ffbf

State: Peer in Cluster (Connected)

 

Hostname: kvm360.durchhalten.intern

Uuid: 4291be40-f77f-4f41-98f6-dc48fd993842

State: Peer in Cluster (Connected)

[root@kvm380 ~]# gluster volume info

 

Volume Name: data

Type: Replicate

Volume ID: 3586de82-e504-4c62-972b-448abead13d3

Status: Started

Snapshot Count: 0

Number of Bricks: 1 x 3 = 3

Transport-type: tcp

Bricks:

Brick1: kvm380.durchhalten.intern:/gluster/data

Brick2: kvm360.durchhalten.intern:/gluster/data

Brick3: kvm320.durchhalten.intern:/gluster/data

Options Reconfigured:

storage.owner-uid: 36

storage.owner-gid: 36

features.shard: on

performance.low-prio-threads: 32

performance.strict-o-direct: on

network.ping-timeout: 30

user.cifs: off

network.remote-dio: off

performance.quick-read: off

performance.read-ahead: off

performance.io-cache: off

cluster.eager-lock: enable

transport.address-family: inet

nfs.disable: on

performance.client-io-threads: off

 

Volume Name: engine

Type: Replicate

Volume ID: dcfbd322-5dd0-4bfe-a775-99ecc79e1416

Status: Started

Snapshot Count: 0

Number of Bricks: 1 x 3 = 3

Transport-type: tcp

Bricks:

Brick1: kvm380.durchhalten.intern:/gluster/engine

Brick2: kvm360.durchhalten.intern:/gluster/engine

Brick3: kvm320.durchhalten.intern:/gluster/engine

Options Reconfigured:

storage.owner-uid: 36

storage.owner-gid: 36

features.shard: on

performance.low-prio-threads: 32

performance.strict-o-direct: on

network.remote-dio: off

network.ping-timeout: 30

user.cifs: off

performance.quick-read: off

performance.read-ahead: off

performance.io-cache: off

cluster.eager-lock: enable

transport.address-family: inet

nfs.disable: on

performance.client-io-threads: off

 

 

After that I deployed a selfhosted engine

And add the two other hosts, at the beginning it looks good, but without
changing anything I got following error by two hosts

 


!

20.12.2018 11:35:05

Failed to connect Host kvm320.durchhalten.intern to Storage Pool Default



!

20.12.2018 11:35:05

Host kvm320.durchhalten.intern cannot access the Storage Domain(s)
hosted_storage attached to the Data Center Default. Setting Host state to
Non-Operational.



X

20.12.2018 11:35:05

Host kvm320.durchhalten.intern reports about one of the Active Storage
Domains as Problematic.



!

20.12.2018 11:35:05

Kdump integration is enabled for host kvm320.durchhalten.intern, but kdump
is not configured properly on host.



!

20.12.2018 11:35:04

Failed to connect Host kvm360.durchhalten.intern to Storage Pool Default



!

20.12.2018 11:35:04

Host kvm360.durchhalten.intern cannot access the Storage Domain(s)
hosted_storage attached to the Data Center Default. Setting Host state to
Non-Operational.



X

20.12.2018 11:35:04

Host kvm360.durchhalten.intern reports about one of the Active Storage
Domains as Problematic.



 

Before glusterfs I had a setup with nfs on 4. Server

 

Where is the problem?

 

thx

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RELXJB2LOJNAKAIJTNTSQC3AQFQNCT75/


[ovirt-users] Hyperconvergend Setup stuck

2018-12-18 Thread Stefan Wolf
Hello

 

I like to setup hyperconvergend

 

I ve 3 hosts, everyone is fresh installed

Kvm320 has one additional harddrive with 1TB SATA

And kvm360 and kvm380 with two additional harddrives with 300gb and 600gb
SAS

 

 

#gdeploy configuration generated by cockpit-gluster plugin

[hosts]

kvm380.durchhalten.intern

kvm360.durchhalten.intern

kvm320.durchhalten.intern

 

[script1:kvm380.durchhalten.intern]

action=execute

ignore_script_errors=no

file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc -h
kvm380.durchhalten.intern, kvm360.durchhalten.intern,
kvm320.durchhalten.intern

 

[script1:kvm360.durchhalten.intern]

action=execute

ignore_script_errors=no

file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc -h
kvm380.durchhalten.intern, kvm360.durchhalten.intern,
kvm320.durchhalten.intern

 

[script1:kvm320.durchhalten.intern]

action=execute

ignore_script_errors=no

file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h
kvm380.durchhalten.intern, kvm360.durchhalten.intern,
kvm320.durchhalten.intern

 

[disktype]

raid6

 

[diskcount]

12

 

[stripesize]

256

 

[service1]

action=enable

service=chronyd

 

[service2]

action=restart

service=chronyd

 

[shell2]

action=execute

command=vdsm-tool configure --force

 

[script3]

action=execute

file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh

ignore_script_errors=no

 

[pv1:kvm380.durchhalten.intern]

action=create

devices=sdb

ignore_pv_errors=no

 

[pv2:kvm380.durchhalten.intern]

action=create

devices=sdc

ignore_pv_errors=no

 

[pv1:kvm360.durchhalten.intern]

action=create

devices=sdb

ignore_pv_errors=no

 

[pv2:kvm360.durchhalten.intern]

action=create

devices=sdc

ignore_pv_errors=no

 

[pv1:kvm320.durchhalten.intern]

action=create

devices=sdb

ignore_pv_errors=no

 

[vg1:kvm380.durchhalten.intern]

action=create

vgname=gluster_vg_sdb

pvname=sdb

ignore_vg_errors=no

 

[vg2:kvm380.durchhalten.intern]

action=create

vgname=gluster_vg_sdc

pvname=sdc

ignore_vg_errors=no

 

[vg1:kvm360.durchhalten.intern]

action=create

vgname=gluster_vg_sdb

pvname=sdb

ignore_vg_errors=no

 

[vg2:kvm360.durchhalten.intern]

action=create

vgname=gluster_vg_sdc

pvname=sdc

ignore_vg_errors=no

 

[vg1:kvm320.durchhalten.intern]

action=create

vgname=gluster_vg_sdb

pvname=sdb

ignore_vg_errors=no

 

[lv1:kvm380.durchhalten.intern]

action=create

poolname=gluster_thinpool_sdc

ignore_lv_errors=no

vgname=gluster_vg_sdc

lvtype=thinpool

size=1005GB

poolmetadatasize=5GB

 

[lv2:kvm360.durchhalten.intern]

action=create

poolname=gluster_thinpool_sdc

ignore_lv_errors=no

vgname=gluster_vg_sdc

lvtype=thinpool

size=1005GB

poolmetadatasize=5GB

 

[lv3:kvm320.durchhalten.intern]

action=create

poolname=gluster_thinpool_sdb

ignore_lv_errors=no

vgname=gluster_vg_sdb

lvtype=thinpool

size=41GB

poolmetadatasize=1GB

 

[lv4:kvm380.durchhalten.intern]

action=create

lvname=gluster_lv_engine

ignore_lv_errors=no

vgname=gluster_vg_sdb

mount=/gluster_bricks/engine

size=100GB

lvtype=thick

 

[lv5:kvm380.durchhalten.intern]

action=create

lvname=gluster_lv_data

ignore_lv_errors=no

vgname=gluster_vg_sdc

mount=/gluster_bricks/data

lvtype=thinlv

poolname=gluster_thinpool_sdc

virtualsize=500GB

 

[lv6:kvm380.durchhalten.intern]

action=create

lvname=gluster_lv_vmstore

ignore_lv_errors=no

vgname=gluster_vg_sdc

mount=/gluster_bricks/vmstore

lvtype=thinlv

poolname=gluster_thinpool_sdc

virtualsize=500GB

 

[lv7:kvm360.durchhalten.intern]

action=create

lvname=gluster_lv_engine

ignore_lv_errors=no

vgname=gluster_vg_sdb

mount=/gluster_bricks/engine

size=100GB

lvtype=thick

 

[lv8:kvm360.durchhalten.intern]

action=create

lvname=gluster_lv_data

ignore_lv_errors=no

vgname=gluster_vg_sdc

mount=/gluster_bricks/data

lvtype=thinlv

poolname=gluster_thinpool_sdc

virtualsize=500GB

 

[lv9:kvm360.durchhalten.intern]

action=create

lvname=gluster_lv_vmstore

ignore_lv_errors=no

vgname=gluster_vg_sdc

mount=/gluster_bricks/vmstore

lvtype=thinlv

poolname=gluster_thinpool_sdc

virtualsize=500GB

 

[lv10:kvm320.durchhalten.intern]

action=create

lvname=gluster_lv_engine

ignore_lv_errors=no

vgname=gluster_vg_sdb

mount=/gluster_bricks/engine

size=20GB

lvtype=thick

 

[lv11:kvm320.durchhalten.intern]

action=create

lvname=gluster_lv_data

ignore_lv_errors=no

vgname=gluster_vg_sdb

mount=/gluster_bricks/data

lvtype=thinlv

poolname=gluster_thinpool_sdb

virtualsize=20GB

 

[lv12:kvm320.durchhalten.intern]

action=create

lvname=gluster_lv_vmstore

ignore_lv_errors=no

vgname=gluster_vg_sdb

mount=/gluster_bricks/vmstore

lvtype=thinlv

poolname=gluster_thinpool_sdb

virtualsize=20GB

 

[selinux]

yes

 

[service3]

action=restart

service=glusterd

slice_setup=yes

 

[firewalld]

action=add

ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,5
4322/tcp

services=glusterfs

 

[script2]

action=execute


[ovirt-users] Gluster with two ovirt nodes

2018-12-12 Thread Stefan Wolf
Hello,

i like to set up glusterfs with two ovirt nodes and on more "normal" node
is this possible?
i 've setup glusterfs in cli on two ovirt nodes and 3rd network storage.
glusterfs is up and running.
But now i like to get something like VIP with ctdb for example. is there
any possibility to set this up with ovirt?
Or do i *ve to setup ovirt manually in centos to install ctdb?
Or are there any other ideas?

thank you stefan
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/4JKFSI4XIMXCKQQEQ7W4ZPWNASYZ52TL/


[ovirt-users] Install additional software on ovirt node

2018-12-11 Thread Stefan Wolf
Hello,

if I install on an ovirt node additional software, will it removed after
updating the ovirt node or run in to trouble?

thx shb
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JPJFPNVLUIQSHX5YPYMJQ4J5UNW64XGO/


[ovirt-users] Re: To much hosts in this cluster

2018-12-10 Thread Stefan Wolf
No, I didn't.
what are the next steps
i ve allready reinstalled the host. i 've selected deploy at hosted engine,
but it seems not to work.


Am Mo., 10. Dez. 2018 um 09:12 Uhr schrieb Luca 'remix_tj' Lorenzetto <
lorenzetto.l...@gmail.com>:

> Hello,
>
> you did hosted-engine undeploy before removing?
>
> Luca
>
> On Sun, Dec 9, 2018 at 12:19 PM Stefan Wolf  wrote:
>
>> Hello,
>>
>>
>>
>> i had 3 hosts with ovirt running, with one of this 3 host i had problems
>> during boot up.
>>
>> I decied to remove the host from the cluster. Now i ve two hosts
>>
>> But if I take a look at Hosted Engine in cockpit i see all three hosts
>>
>> Why ist he kvm380 not removed?
>>
>> How can I remove it?
>>
>>
>>
>> Thx shb
>>
>>
>> ___
>> Users mailing list -- users@ovirt.org
>> To unsubscribe send an email to users-le...@ovirt.org
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> oVirt Code of Conduct:
>> https://www.ovirt.org/community/about/community-guidelines/
>> List Archives:
>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/VZAOYESLQZDD7TJ4P5COXCOR7LST4RMN/
>>
>
>
> --
> "E' assurdo impiegare gli uomini di intelligenza eccellente per fare
> calcoli che potrebbero essere affidati a chiunque se si usassero delle
> macchine"
> Gottfried Wilhelm von Leibnitz, Filosofo e Matematico (1646-1716)
>
> "Internet è la più grande biblioteca del mondo.
> Ma il problema è che i libri sono tutti sparsi sul pavimento"
> John Allen Paulos, Matematico (1945-vivente)
>
> Luca 'remix_tj' Lorenzetto, http://www.remixtj.net , <
> lorenzetto.l...@gmail.com>
>
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YOL24J4653E3J5H2BEXD2PDTEXE4D7NT/


[ovirt-users] Try to add Host to cluster: Command returned failure code 1 during SSH session

2018-12-09 Thread Stefan Wolf
Hello,

 

I try to add a new installed ovirt node and get the error message during
adding to cluster

Host kvm380 installation failed. Command returned failure code 1 during SSH
session 'root@kvm380.durchhalten.intern'.

 

 

Maybe someone can help me

 

Thx 

Here is the logfile

 

2018-12-09 14:48:23,262+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Initializing.

2018-12-09 14:48:23,414+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Environment
setup.

2018-12-09 14:48:23,479+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Environment
packages setup.

2018-12-09 14:48:26,194+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Programs
detection.

2018-12-09 14:48:26,373+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Environment
customization.

2018-12-09 14:48:26,842+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Kdump supported.

2018-12-09 14:48:27,123+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Logs at host located
at: '/tmp/ovirt-host-deploy-20181209144822-hbil3q.log'.

2018-12-09 14:48:27,188+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Host is hypervisor.

2018-12-09 14:48:27,192+01 INFO
[org.ovirt.engine.core.bll.hostdeploy.VdsDeployVdsmUnit] (VdsDeploy)
[f097f8b0-ee35-4c08-a416-1f0427dd2e9e] Host kvm380.durchhalten.intern
reports unique id 31323436-3530-5a43-3233-303430374a4e

2018-12-09 14:48:27,208+01 INFO
[org.ovirt.engine.core.bll.hostdeploy.VdsDeployVdsmUnit] (VdsDeploy)
[f097f8b0-ee35-4c08-a416-1f0427dd2e9e] Assigning unique id
31323436-3530-5a43-3233-303430374a4e to Host kvm380.durchhalten.intern

2018-12-09 14:48:27,436+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Setup
validation.

2018-12-09 14:48:27,825+01 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS_ERROR(511), An error has occurred during
installation of Host kvm380: Failed to execute stage 'Setup validation':
Cannot resolve kdump destination address 'ovirt.durchhalten.intern'.

2018-12-09 14:48:27,827+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Clean up.

2018-12-09 14:48:27,830+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage:
Pre-termination.

2018-12-09 14:48:27,874+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Retrieving
installation logs to:
'/var/log/ovirt-engine/host-deploy/ovirt-host-deploy-20181209144827-kvm380.d
urchhalten.intern-f097f8b0-ee35-4c08-a416-1f0427dd2e9e.log'.

2018-12-09 14:48:28,295+01 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [f097f8b0-ee35-4c08-a416-1f0427dd2e9e] EVENT_ID:
VDS_INSTALL_IN_PROGRESS(509), Installing Host kvm380. Stage: Termination.

 

2018-12-09 14:48:28,389+01 ERROR
[org.ovirt.engine.core.uutils.ssh.SSHDialog]
(EE-ManagedThreadFactory-engine-Thread-1942)
[f097f8b0-ee35-4c08-a416-1f0427dd2e9e] SSH error running command
root@kvm380.durchhalten.intern:'umask 0077;
MYTMP="$(TMPDIR="${OVIRT_TMPDIR}" mktemp -d -t ovirt-XX)"; trap
"chmod -R u+rwX \"${MYTMP}\" > /dev/null 2>&1; rm -fr \"${MYTMP}\" >
/dev/null 2>&1" 0; tar --warning=no-timestamp -C "${MYTMP}" -x &&
"${MYTMP}"/ovirt-host-deploy DIALOG/dialect=str:machine

[ovirt-users] To much hosts in this cluster

2018-12-09 Thread Stefan Wolf
Hello,

 

i had 3 hosts with ovirt running, with one of this 3 host i had problems
during boot up. 

I decied to remove the host from the cluster. Now i ve two hosts



But if I take a look at Hosted Engine in cockpit i see all three hosts



Why ist he kvm380 not removed?

How can I remove it?

 

Thx shb

 

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VZAOYESLQZDD7TJ4P5COXCOR7LST4RMN/


[ovirt-users] Network Setup after ovirt node Install and before add node to cluster

2018-12-06 Thread Stefan Wolf
Hello,

 

I ve downloaded an installed OVIRT NODE 4.2.7.1.

During the installation I ‘ve setup a static ip (I ‘ve also tried dhcp, both 
work)

The installation passes and after the reboot ovirt node starts without any 
problems.

BUT the network is not configured.

I know https://www.ovirt.org/documentation/install-guide/chap-oVirt_Nodes/ and  
I know

https://www.ovirt.org/documentation/admin-guide/chap-Hosts/

 

and now I am between these two documents

oVirt is installed but network is not configured to add it to datacenter

 

of course I know the scripts under /etc/sysconfig/network-scripts

do I really have do edit it manuel?

 

if I change ip in command line

ip a add … dev..

and ip route add …

it fails to add node to datacenter

 

- with one of the last version I ‘ve added the IP address in bash

- added the host to datacenter,

- after reboot readd the ip address in bash after reboot

- an then I could change network settings in oVirt Manger

 

What is best practices for adding a node to datacenter when the ip address is 
not assigned after installation

 

Thx shb

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/TH5MRXJUU4SI62Z5XXQX2S6GKBOQIJWW/