On Wed, Apr 27, 2022 at 3:54 PM José Ferradeira via Users <users@ovirt.org> wrote: > > After upgrade to 4.5 host cannot be activated because cannot connect to data > domain. > > I have a data domain in NFS (master) and a GlusterFS. It complains about the > Gluster domain: > The error message for connection node1-teste.acloud.pt:/data1 returned by > VDSM was: XML error > > # rpm -qa|grep glusterfs* > glusterfs-10.1-1.el8s.x86_64 > glusterfs-selinux-2.0.1-1.el8s.noarch > glusterfs-client-xlators-10.1-1.el8s.x86_64 > glusterfs-events-10.1-1.el8s.x86_64 > libglusterfs0-10.1-1.el8s.x86_64 > glusterfs-fuse-10.1-1.el8s.x86_64 > glusterfs-server-10.1-1.el8s.x86_64 > glusterfs-cli-10.1-1.el8s.x86_64 > glusterfs-geo-replication-10.1-1.el8s.x86_64 > > > engine log: > > 2022-04-27 13:35:16,118+01 ERROR > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-66) [e > be79c6] EVENT_ID: VDS_STORAGES_CONNECTION_FAILED(188), Failed to connect Host > NODE1 to the Storage Domains DATA1. > 2022-04-27 13:35:16,169+01 ERROR > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-66) [e > be79c6] EVENT_ID: STORAGE_DOMAIN_ERROR(996), The error message for connection > node1-teste.acloud.pt:/data1 returned by VDSM was: XML error > 2022-04-27 13:35:16,170+01 ERROR > [org.ovirt.engine.core.bll.storage.connection.FileStorageHelper] > (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-66) > [ebe79c6 > ] The connection with details 'node1-teste.acloud.pt:/data1' failed because > of error code '4106' and error message is: xml error > > > > vdsm log: > 2022-04-27 13:40:07,125+0100 ERROR (jsonrpc/4) [storage.storageServer] Could > not connect to storage server (storageServer:92) > Traceback (most recent call last): > File "/usr/lib/python3.6/site-packages/vdsm/storage/storageServer.py", line > 90, in connect_all > con.connect() > File "/usr/lib/python3.6/site-packages/vdsm/storage/storageServer.py", line > 233, in connect > self.validate() > File "/usr/lib/python3.6/site-packages/vdsm/storage/storageServer.py", line > 365, in validate > if not self.volinfo: > File "/usr/lib/python3.6/site-packages/vdsm/storage/storageServer.py", line > 352, in volinfo > self._volinfo = self._get_gluster_volinfo() > File "/usr/lib/python3.6/site-packages/vdsm/storage/storageServer.py", line > 405, in _get_gluster_volinfo > self._volfileserver) > File "/usr/lib/python3.6/site-packages/vdsm/common/supervdsm.py", line 56, > in __call__ > return callMethod() > File "/usr/lib/python3.6/site-packages/vdsm/common/supervdsm.py", line 54, > in <lambda> > **kwargs) > File "<string>", line 2, in glusterVolumeInfo > File "/usr/lib64/python3.6/multiprocessing/managers.py", line 772, in > _callmethod > raise convert_to_error(kind, result) > vdsm.gluster.exception.GlusterXmlErrorException: XML error: rc=0 out=() > err=[b'<cliOutput>\n <opRet>0</opRet>\n <opErrno>0</opErrno>\n <opErrstr > />\n <volInfo>\n <volumes>\ > n <volume>\n <name>data1</name>\n > <id>d7eb2c38-2707-4774-9873-a7303d024669</id>\n <status>1</status>\n > <statusStr>Started</statusStr>\n <sn > apshotCount>0</snapshotCount>\n <brickCount>2</brickCount>\n > <distCount>2</distCount>\n <replicaCount>1</replicaCount>\n > <arbiterCount>0</arbiterCount> > \n <disperseCount>0</disperseCount>\n > <redundancyCount>0</redundancyCount>\n <type>0</type>\n > <typeStr>Distribute</typeStr>\n <transport>0</tran > sport>\n <bricks>\n <brick > uuid="08c7ba5f-9aca-49c5-abfd-8a3e42dd8c0b">node1-teste.acloud.pt:/home/brick1<name>node1-teste.acloud.pt:/home/brick1</name><hostUuid>0 > 8c7ba5f-9aca-49c5-abfd-8a3e42dd8c0b</hostUuid><isArbiter>0</isArbiter></brick>\n > <brick > uuid="08c7ba5f-9aca-49c5-abfd-8a3e42dd8c0b">node1-teste.acloud.pt:/brick2<name>nod > e1-teste.acloud.pt:/brick2</name><hostUuid>08c7ba5f-9aca-49c5-abfd-8a3e42dd8c0b</hostUuid><isArbiter>0</isArbiter></brick>\n > </bricks>\n <optCount>23</optCount>\n > <options>\n <option>\n <name>nfs.disable</name>\n > <value>on</value>\n </option>\n <option>\n > <name>transport.addre > ss-family</name>\n <value>inet</value>\n </option>\n > <option>\n <name>storage.fips-mode-rchecksum</name>\n > <value>on</value>\n > </option>\n <option>\n > <name>storage.owner-uid</name>\n <value>36</value>\n > </option>\n <option>\n <name>storag > e.owner-gid</name>\n <value>36</value>\n </option>\n > <option>\n <name>cluster.min-free-disk</name>\n > <value>5%</value>\n > </option>\n <option>\n > <name>performance.quick-read</name>\n <value>off</value>\n > </option>\n <option>\n <name>perfor > mance.read-ahead</name>\n <value>off</value>\n > </option>\n <option>\n > <name>performance.io-cache</name>\n <value>off</value>\n > </option>\n <option>\n > <name>performance.low-prio-threads</name>\n <value>32</value>\n > </option>\n <option>\n < > name>network.remote-dio</name>\n <value>enable</value>\n > </option>\n <option>\n <name>cluster.eager-lock</name>\n > <value>enable< > /value>\n </option>\n <option>\n > <name>cluster.quorum-type</name>\n <value>auto</value>\n > </option>\n <option>\n > <name>cluster.server-quorum-type</name>\n > <value>server</value>\n </option>\n <option>\n > <name>cluster.data-self-heal-algorithm</name>\n > <value>full</value>\n </option>\n <option>\n > <name>cluster.locking-scheme</name>\n > <value>granular</value>\n </option> > \n <option>\n <name>cluster.shd-wait-qlength</name>\n > <value>10000</value>\n </option>\n <option>\n > <name>features.shar > d</name>\n <value>off</value>\n </option>\n > <option>\n <name>user.cifs</name>\n > <value>off</value>\n </option>\n > <option>\n <name>cluster.choose-local</name>\n > <value>off</value>\n </option>\n <option>\n > <name>client.event-threads</name>\ > n <value>4</value>\n </option>\n <option>\n > <name>server.event-threads</name>\n <value>4</value>\n > </option>\n > <option>\n <name>performance.client-io-threads</name>\n > <value>on</value>\n </option>\n </options>\n </volume>\n > <count>1</count>\ > n </volumes>\n </volInfo>\n</cliOutput>'] > 2022-04-27 13:40:07,125+0100 INFO (jsonrpc/4) [storage.storagedomaincache] > Invalidating storage domain cache (sdc:74) > 2022-04-27 13:40:07,125+0100 INFO (jsonrpc/4) [vdsm.api] FINISH > connectStorageServer return={'statuslist': [{'id': > 'dede3145-651a-4b01-b8d2-82bff8670696', 'status': 4106}]} from= > ::ffff:192.168.5.165,42132, flow_id=4c170005, > task_id=cec6f36f-46a4-462c-9d0a-feb8d814b465 (api:54) > 2022-04-27 13:40:07,410+0100 INFO (jsonrpc/5) [api.host] START > getAllVmStats() from=::ffff:192.168.5.165,42132 (api:48) > 2022-04-27 13:40:07,411+0100 INFO (jsonrpc/5) [api.host] FINISH > getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': > (suppressed)} from=::ffff:192.168.5.1 > 65,42132 (api:54) > 2022-04-27 13:40:07,785+0100 INFO (jsonrpc/7) [api.host] START getStats() > from=::ffff:192.168.5.165,42132 (api:48) > 2022-04-27 13:40:07,797+0100 INFO (jsonrpc/7) [vdsm.api] START > repoStats(domains=()) from=::ffff:192.168.5.165,42132, > task_id=4fa4e8c4-7c65-499a-827e-8ae153aa875e (api:48) > 2022-04-27 13:40:07,797+0100 INFO (jsonrpc/7) [vdsm.api] FINISH repoStats > return={} from=::ffff:192.168.5.165,42132, > task_id=4fa4e8c4-7c65-499a-827e-8ae153aa875e (api:54) > 2022-04-27 13:40:07,797+0100 INFO (jsonrpc/7) [vdsm.api] START > multipath_health() from=::ffff:192.168.5.165,42132, > task_id=c6390f2a-845b-420b-a833-475605a24078 (api:48) > 2022-04-27 13:40:07,797+0100 INFO (jsonrpc/7) [vdsm.api] FINISH > multipath_health return={} from=::ffff:192.168.5.165,42132, > task_id=c6390f2a-845b-420b-a833-475605a24078 (api:54) > 2022-04-27 13:40:07,802+0100 INFO (jsonrpc/7) [api.host] FINISH getStats > return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} > from=::ffff:192.168.5.165,42132 ( > api:54) > 2022-04-27 13:40:11,980+0100 INFO (jsonrpc/6) [api.host] START > getAllVmStats() from=::1,37040 (api:48) > 2022-04-27 13:40:11,980+0100 INFO (jsonrpc/6) [api.host] FINISH > getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': > (suppressed)} from=::1,37040 (api:54) > 2022-04-27 13:40:12,365+0100 INFO (periodic/2) [vdsm.api] START > repoStats(domains=()) from=internal, > task_id=f5084096-e5c5-4ca8-9c47-a92fa5790484 (api:48) > 2022-04-27 13:40:12,365+0100 INFO (periodic/2) [vdsm.api] FINISH repoStats > return={} from=internal, task_id=f5084096-e5c5-4ca8-9c47-a92fa5790484 (api:54) > 2022-04-27 13:40:22,417+0100 INFO (jsonrpc/0) [api.host] START > getAllVmStats() from=::ffff:192.168.5.165,42132 (api:48) > 2022-04-27 13:40:22,417+0100 INFO (jsonrpc/0) [api.host] FINISH > getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': > (suppressed)} from=::ffff:192.168.5.1 > 65,42132 (api:54) > 2022-04-27 13:40:22,805+0100 INFO (jsonrpc/1) [api.host] START getStats() > from=::ffff:192.168.5.165,42132 (api:48) > 2022-04-27 13:40:22,816+0100 INFO (jsonrpc/1) [vdsm.api] START > repoStats(domains=()) from=::ffff:192.168.5.165,42132, > task_id=a9fb939c-ea1a-4116-a22f-d14a99e6eada (api:48) > 2022-04-27 13:40:22,816+0100 INFO (jsonrpc/1) [vdsm.api] FINISH repoStats > return={} from=::ffff:192.168.5.165,42132, > task_id=a9fb939c-ea1a-4116-a22f-d14a99e6eada (api:54) > 2022-04-27 13:40:22,816+0100 INFO (jsonrpc/1) [vdsm.api] START > multipath_health() from=::ffff:192.168.5.165,42132, > task_id=5eee2f63-2631-446a-98dd-4947f9499f8f (api:48) > 2022-04-27 13:40:22,816+0100 INFO (jsonrpc/1) [vdsm.api] FINISH > multipath_health return={} from=::ffff:192.168.5.165,42132, > task_id=5eee2f63-2631-446a-98dd-4947f9499f8f (api:54) > 2022-04-27 13:40:22,822+0100 INFO (jsonrpc/1) [api.host] FINISH getStats > return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} > from=::ffff:192.168.5.165,42132 ( > api:54)
Please file upstream issue: https://github.com/oVirt/vdsm/issues And include info about your gluster server rpm packages. I hope that Ritesh can help with this. Nir _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/73M6MASMFOUFH3JSS3J3KRMB5FPUCR6K/