I compared each file across my nodes and synced them.  It seems to have
resolved my issue.

I wonder if there is a problem with 6.5 to 6.6 upgrade that is causing the
problem?  It’s strange that it seems to have happened to more than one
person. I was also following proper upgrade procedure.



On Sun, Dec 15, 2019 at 3:09 AM <[email protected]> wrote:

> I don't know. I had the same issues when I migrated my gluster from v6.5
> to 6.6 (currently running v7.0).
> Just get the newest file and rsync it to the rest of the bricks. It will
> solve the '?????? ??????' problem.
>
> Best Regards,
> Strahil Nikolov
> В неделя, 15 декември 2019 г., 3:49:27 ч. Гринуич+2, Jayme <
> [email protected]> написа:
>
>
> on that page it says to check open bugs and the migration bug you mention
> does not appear to be on the list.  Has it been resolved or is it just
> missing from this page?
>
> On Sat, Dec 14, 2019 at 7:53 PM Strahil Nikolov <[email protected]>
> wrote:
>
> Nah... this is not gonna fix your issue and is unnecessary.
> Just compare the data from all bricks ... most probably the 'Last Updated'
> is different and the gfid of the file is different.
> Find the brick that has the most fresh data, and replace (move away as a
> backup and rsync) the file from last good copy to the other bricks.
> You can also run a 'full heal'.
>
> Best Regards,
> Strahil Nikolov
>
> В събота, 14 декември 2019 г., 21:18:44 ч. Гринуич+2, Jayme <
> [email protected]> написа:
>
>
> *Update*
>
> Situation has improved.  All VMs and engine are running.  I'm left right
> now with about 2 heal entries in each glusterfs storage volume that will
> not heal.
>
> In all cases each heal entry is related to an OVF_STORE image and the
> problem appears to be an issue with the gluster metadata for those
> ovf_store images.  When I look at the files shown in gluster volume heal
> info output I'm seeing question marks on the meta files which indicates an
> attribute/gluster problem (even though there is no split-brain).  And I get
> input/output error when attempting to do anything with the files.
>
> If I look at the files on each host in /gluster_bricks they all look
> fine.  I only see question marks on the meta files when look at the file in
> /rhev mounts
>
> Does anyone know how I can correct the attributes on these OVF_STORE
> files?  I've tried putting each host in maintenance and re-activating to
> re-mount gluster volumes.  I've also stopped and started all gluster
> volumes.
>
> I'm thinking I might be able to solve this by shutting down all VMs and
> placing all hosts in maintenance and safely restarting the entire cluster..
> but that may not be necessary?
>
> On Fri, Dec 13, 2019 at 12:59 AM Jayme <[email protected]> wrote:
>
> I believe I was able to get past this by stopping the engine volume then
> unmounting the glusterfs engine mount on all hosts and re-starting the
> volume. I was able to start hostedengine on host0.
>
> I'm still facing a few problems:
>
> 1. I'm still seeing this issue in each host's logs:
>
> Dec 13 00:57:54 orchard0 journal: ovirt-ha-agent
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.config.vm ERROR
> Failed scanning for OVF_STORE due to Command Volume.getInfo with args
> {'storagepoolID': '00000000-0000-0000-0000-000000000000',
> 'storagedomainID': 'd70b171e-7488-4d52-8cad-bbc581dbf16e', 'volumeID':
> u'2632f423-ed89-43d9-93a9-36738420b866', 'imageID':
> u'd909dc74-5bbd-4e39-b9b5-755c167a6ee8'} failed:#012(code=201,
> message=Volume does not exist: (u'2632f423-ed89-43d9-93a9-36738420b866',))
> Dec 13 00:57:54 orchard0 journal: ovirt-ha-agent
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.config.vm ERROR
> Unable to identify the OVF_STORE volume, falling back to initial vm.conf.
> Please ensure you already added your first data domain for regular VMs
>
>
> 2. Most of my gluster volumes still have un-healed entires which I can't
> seem to heal.  I'm not sure what the answer is here.
>
> On Fri, Dec 13, 2019 at 12:33 AM Jayme <[email protected]> wrote:
>
> I was able to get the hosted engine started manually via Virsh after
> re-creating a missing symlink in /var/run/vdsm/storage -- I later shut it
> down and am still having the same problem with ha broker starting.  It
> appears that the problem *might* be with a corrupt ha metadata file,
> although gluster is not stating there is split brain on the engine volume
>
> I'm seeing this:
>
> ls -al
> /rhev/data-center/mnt/glusterSD/orchard0\:_engine/d70b171e-7488-4d52-8cad-bbc581dbf16e/ha_agent/
> ls: cannot access
> /rhev/data-center/mnt/glusterSD/orchard0:_engine/d70b171e-7488-4d52-8cad-bbc581dbf16e/ha_agent/hosted-engine.metadata:
> Input/output error
> total 0
> drwxr-xr-x. 2 vdsm kvm  67 Dec 13 00:30 .
> drwxr-xr-x. 6 vdsm kvm  64 Aug  6  2018 ..
> lrwxrwxrwx. 1 vdsm kvm 132 Dec 13 00:30 hosted-engine.lockspace ->
> /var/run/vdsm/storage/d70b171e-7488-4d52-8cad-bbc581dbf16e/03a8ee8e-91f5-4e06-904b-9ed92a9706eb/db2699ce-6349-4020-b52d-8ab11d01e26d
> l?????????? ? ?    ?     ?            ? hosted-engine.metadata
>
> Clearly showing some sort of issue with hosted-engine.metadata on the
> client mount.
>
> on each node in /gluster_bricks I see this:
>
> # ls -al
> /gluster_bricks/engine/engine/d70b171e-7488-4d52-8cad-bbc581dbf16e/ha_agent/
> total 0
> drwxr-xr-x. 2 vdsm kvm  67 Dec 13 00:31 .
> drwxr-xr-x. 6 vdsm kvm  64 Aug  6  2018 ..
> lrwxrwxrwx. 2 vdsm kvm 132 Dec 13 00:31 hosted-engine.lockspace ->
> /var/run/vdsm/storage/d70b171e-7488-4d52-8cad-bbc581dbf16e/03a8ee8e-91f5-4e06-904b-9ed92a9706eb/db2699ce-6349-4020-b52d-8ab11d01e26d
> lrwxrwxrwx. 2 vdsm kvm 132 Dec 12 16:30 hosted-engine.metadata ->
> /var/run/vdsm/storage/d70b171e-7488-4d52-8cad-bbc581dbf16e/66bf05fa-bf50-45ec-98d8-d00002040317/a2250415-5ff0-42ab-8071-cd9d67c3048c
>
>  ls -al
> /var/run/vdsm/storage/d70b171e-7488-4d52-8cad-bbc581dbf16e/66bf05fa-bf50-45ec-98d8-d00002040317/a2250415-5ff0-42ab-8071-cd9d67c3048c
> -rw-rw----. 1 vdsm kvm 1073741824 Dec 12 16:48
> /var/run/vdsm/storage/d70b171e-7488-4d52-8cad-bbc581dbf16e/66bf05fa-bf50-45ec-98d8-d00002040317/a2250415-5ff0-42ab-8071-cd9d67c3048c
>
>
> I'm not sure how to proceed at this point.  Do I have data corruption, a
> gluster split-brain issue or something else?  Maybe I just need to
> re-generate metadata for the hosted engine?
>
> On Thu, Dec 12, 2019 at 6:36 PM Jayme <[email protected]> wrote:
>
> I'm running a three server HCI.  Up and running on 4.3.7 with no
> problems.  Today I updated to 4.3.8.  Engine upgraded fine, rebooted.
> First host updated fine, rebooted and let all gluster volumes heal.  Put
> second host in maintenance, upgraded successfully, rebooted.  Waited for
> gluster volumes to heal for over an hour but the heal process was not
> completing.  I tried restarting gluster services as well as the host with
> no success.
>
> I'm in a state right now where there are pending heals on almost all of my
> volumes.  Nothing is reporting split-brain, but the heals are not
> completing.
>
> All vms are still currently running except hosted engine.  Hosted engine
> was running but on the 2nd host I upgraded I was seeing errors such as:
>
> Dec 12 16:34:39 orchard2 journal: ovirt-ha-agent
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.config.vm ERROR
> Failed scanning for OVF_STORE due to Command Volume.getInfo with args
> {'storagepoolID': '00000000-0000-0000-0000-000000000000',
> 'storagedomainID': 'd70b171e-7488-4d52-8cad-bbc581dbf16e', 'volumeID':
> u'2632f423-ed89-43d9-93a9-36738420b866', 'imageID':
> u'd909dc74-5bbd-4e39-b9b5-755c167a6ee8'} failed:#012(code=201,
> message=Volume does not exist: (u'2632f423-ed89-43d9-93a9-36738420b866',))
>
> I shut down the engine VM and attempted a manual heal on the engine
> volume.  I cannot start the engine on any host now.  I get:
>
> The hosted engine configuration has not been retrieved from shared
> storage. Please ensure that ovirt-ha-agent is running and the storage
> server is reachable.
>
> I'm seeing ovirt-ha-agent crashing on all three nodes:
>
> Dec 12 18:30:48 orchard0 python: detected unhandled Python exception in
> '/usr/share/ovirt-hosted-engine-ha/ovirt-ha-broker'
> Dec 12 18:30:48 orchard0 abrt-server: Duplicate: core backtrace
> Dec 12 18:30:48 orchard0 abrt-server: DUP_OF_DIR:
> /var/tmp/abrt/Python-2019-03-14-21:02:52-44318
> Dec 12 18:30:48 orchard0 abrt-server: Deleting problem directory
> Python-2019-12-12-18:30:48-23193 (dup of Python-2019-03-14-21:02:52-44318)
> Dec 12 18:30:49 orchard0 vdsm[6087]: ERROR failed to retrieve Hosted
> Engine HA score '[Errno 2] No such file or directory'Is the Hosted Engine
> setup finished?
> Dec 12 18:30:49 orchard0 systemd: ovirt-ha-broker.service: main process
> exited, code=exited, status=1/FAILURE
> Dec 12 18:30:49 orchard0 systemd: Unit ovirt-ha-broker.service entered
> failed state.
> Dec 12 18:30:49 orchard0 systemd: ovirt-ha-broker.service failed.
> Dec 12 18:30:49 orchard0 systemd: ovirt-ha-broker.service holdoff time
> over, scheduling restart.
> Dec 12 18:30:49 orchard0 systemd: Cannot add dependency job for unit
> lvm2-lvmetad.socket, ignoring: Unit is masked.
> Dec 12 18:30:49 orchard0 systemd: Stopped oVirt Hosted Engine High
> Availability Communications Broker.
>
>
> Here is what gluster volume heal info on engine looks like, it's similar
> on other volumes as well (although more heals pending on some of those):
>
>  gluster volume heal engine info
> Brick gluster0:/gluster_bricks/engine/engine
>
> /d70b171e-7488-4d52-8cad-bbc581dbf16e/images/d909dc74-5bbd-4e39-b9b5-755c167a6ee8/2632f423-ed89-43d9-93a9-36738420b866.meta
>
> /d70b171e-7488-4d52-8cad-bbc581dbf16e/images/053171e4-f782-42d7-9115-c602beb3c826/627b8f93-5373-48bb-bd20-a308a455e082.meta
>
> /d70b171e-7488-4d52-8cad-bbc581dbf16e/master/tasks/a9b11e33-9b93-46a0-a36e-85063fd53ebe.backup
> /d70b171e-7488-4d52-8cad-bbc581dbf16e/dom_md/ids
> Status: Connected
> Number of entries: 4
>
> Brick gluster1:/gluster_bricks/engine/engine
>
> /d70b171e-7488-4d52-8cad-bbc581dbf16e/images/d909dc74-5bbd-4e39-b9b5-755c167a6ee8/2632f423-ed89-43d9-93a9-36738420b866.meta
>
> /d70b171e-7488-4d52-8cad-bbc581dbf16e/master/tasks/a9b11e33-9b93-46a0-a36e-85063fd53ebe.backup
>
> /d70b171e-7488-4d52-8cad-bbc581dbf16e/images/053171e4-f782-42d7-9115-c602beb3c826/627b8f93-5373-48bb-bd20-a308a455e082.meta
> /d70b171e-7488-4d52-8cad-bbc581dbf16e/dom_md/ids
> Status: Connected
> Number of entries: 4
>
> Brick gluster2:/gluster_bricks/engine/engine
> Status: Connected
> Number of entries: 0
>
> I don't see much in vdsm.log and gluster logs look fairly normal to me,
> I'm not seeing any obvious errors in the gluster logs.
>
> As far as I can tell the underlying storage is fine.  Why are my gluster
> volumes not healing and why is self-hosted engine failing to start?
>
> More agent and broker logs:
>
> ==> agent.log <==
> MainThread::ERROR::2019-12-12
> 18:36:09,056::hosted_engine::559::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_broker)
> Failed to start necessary monitors
> MainThread::ERROR::2019-12-12
> 18:36:09,058::agent::144::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
> Traceback (most recent call last):
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 131, in _run_agent
>     return action(he)
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 55, in action_proper
>     return he.start_monitoring()
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 432, in start_monitoring
>     self._initialize_broker()
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 556, in _initialize_broker
>     m.get('options', {}))
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
> line 89, in start_monitor
>     ).format(t=type, o=options, e=e)
> RequestError: brokerlink - failed to start monitor via ovirt-ha-broker:
> [Errno 2] No such file or directory, [monitor: 'network', options:
> {'tcp_t_address': None, 'network_test': None, 'tcp_t_port': None, 'addr':
> '10.11.0.254'}]
>
> MainThread::ERROR::2019-12-12
> 18:36:09,058::agent::145::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
> Trying to restart agent
> MainThread::ERROR::2019-12-12
> 18:36:19,619::hosted_engine::559::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_broker)
> Failed to start necessary monitors
> MainThread::ERROR::2019-12-12
> 18:36:19,619::agent::144::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
> Traceback (most recent call last):
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 131, in _run_agent
>     return action(he)
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 55, in action_proper
>     return he.start_monitoring()
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 432, in start_monitoring
>     self._initialize_broker()
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 556, in _initialize_broker
>     m.get('options', {}))
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
> line 89, in start_monitor
>     ).format(t=type, o=options, e=e)
> RequestError: brokerlink - failed to start monitor via ovirt-ha-broker:
> [Errno 2] No such file or directory, [monitor: 'network', options:
> {'tcp_t_address': None, 'network_test': None, 'tcp_t_port': None, 'addr':
> '10.11.0.254'}]
>
> MainThread::ERROR::2019-12-12
> 18:36:19,619::agent::145::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
> Trying to restart agent
> MainThread::ERROR::2019-12-12
> 18:36:30,568::hosted_engine::559::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_broker)
> Failed to start necessary monitors
> MainThread::ERROR::2019-12-12
> 18:36:30,570::agent::144::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
> Traceback (most recent call last):
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 131, in _run_agent
>     return action(he)
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 55, in action_proper
>     return he.start_monitoring()
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 432, in start_monitoring
>     self._initialize_broker()
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 556, in _initialize_broker
>     m.get('options', {}))
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
> line 89, in start_monitor
>     ).format(t=type, o=options, e=e)
> RequestError: brokerlink - failed to start monitor via ovirt-ha-broker:
> [Errno 2] No such file or directory, [monitor: 'network', options:
> {'tcp_t_address': None, 'network_test': None, 'tcp_t_port': None, 'addr':
> '10.11.0.254'}]
>
> MainThread::ERROR::2019-12-12
> 18:36:30,570::agent::145::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
> Trying to restart agent
> MainThread::ERROR::2019-12-12
> 18:36:41,581::hosted_engine::559::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_broker)
> Failed to start necessary monitors
> MainThread::ERROR::2019-12-12
> 18:36:41,583::agent::144::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
> Traceback (most recent call last):
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 131, in _run_agent
>     return action(he)
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 55, in action_proper
>     return he.start_monitoring()
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 432, in start_monitoring
>     self._initialize_broker()
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 556, in _initialize_broker
>     m.get('options', {}))
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
> line 89, in start_monitor
>     ).format(t=type, o=options, e=e)
> RequestError: brokerlink - failed to start monitor via ovirt-ha-broker:
> [Errno 2] No such file or directory, [monitor: 'network', options:
> {'tcp_t_address': None, 'network_test': None, 'tcp_t_port': None, 'addr':
> '10.11.0.254'}]
>
> MainThread::ERROR::2019-12-12
> 18:36:41,583::agent::145::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
> Trying to restart agent
>
>
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/U5YFDWCQJYNALSVNPZG4FLUO7KB2Z2XI/
>
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/A2WDQBXRNXL3UFH67WA7RQ7ODW6JTMOM/

Reply via email to