On Tue, Aug 3, 2021 at 10:27 PM Michal Skrivanek <
michal.skriva...@redhat.com> wrote:

>
>
> On 3. 8. 2021, at 11:43, Yedidyah Bar David <d...@redhat.com> wrote:
>
> On Tue, Aug 3, 2021 at 10:05 AM Yedidyah Bar David <d...@redhat.com>
> wrote:
>
>
> On Tue, Aug 3, 2021 at 7:50 AM <jenk...@jenkins.phx.ovirt.org> wrote:
>
>
> Project:
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/
> Build:
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2126/
> Build Number: 2126
> Build Status:  Failure
> Triggered By: Started by timer
>
> -------------------------------------
> Changes Since Last Success:
> -------------------------------------
> Changes for Build #2126
> [Michal Skrivanek] basic: skipping just the VNC console part of
> test_virtual_machines
>
>
>
>
> -----------------
> Failed Tests:
> -----------------
> 2 tests failed.
> FAILED:
>  
> he-basic-suite-master.test-scenarios.test_012_local_maintenance_sdk.test_local_maintenance
>
> Error Message:
> ovirtsdk4.Error: Failed to read response: [(<pycurl.Curl object at
> 0x5555faf11228>, 7, 'Failed to connect to 192.168.200.99 port 443:
> Connection refused')]
>
>
> This looks very similar to the issue we have with dns/dig failures
> that cause the engine VM to go down, and it's similar, but different.
>
> dig didn't fail (it now uses TCP), but something else caused the agent
> to stop the engine VM - a combination of high cpu load and low free
> memory, after restarting the engine VM as part of test_008.
>
>
> https://jenkins.ovirt.org/job/ovirt-system-tests_he-basic-suite-master/2126/artifact/exported-artifacts/test_logs/ost-he-basic-suite-master-host-0/var/log/ovirt-hosted-engine-ha/agent.log
> :
>
>
> =====================================================================================
> MainThread::INFO::2021-08-03
>
> 06:46:55,068::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state ReinitializeFSM (score: 0)
> MainThread::INFO::2021-08-03
>
> 06:47:04,089::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Global maintenance detected
> MainThread::INFO::2021-08-03
>
> 06:47:04,169::brokerlink::73::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
> Success, was notification of state_transition
> (ReinitializeFSM-GlobalMaintenance) sent? ignored
> MainThread::INFO::2021-08-03
>
> 06:47:05,249::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state GlobalMaintenance (score: 3400)
> MainThread::INFO::2021-08-03
>
> 06:47:14,439::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Global maintenance detected
> MainThread::INFO::2021-08-03
>
> 06:47:25,526::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
> Penalizing score by 814 due to cpu load
> MainThread::INFO::2021-08-03
>
> 06:47:25,527::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state GlobalMaintenance (score: 2586)
> MainThread::INFO::2021-08-03
>
> 06:47:25,537::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Global maintenance detected
> MainThread::INFO::2021-08-03
>
> 06:47:26,029::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state GlobalMaintenance (score: 2586)
> MainThread::INFO::2021-08-03
>
> 06:47:35,050::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Global maintenance detected
> MainThread::INFO::2021-08-03
>
> 06:47:35,576::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state GlobalMaintenance (score: 2586)
> MainThread::INFO::2021-08-03
>
> 06:47:45,597::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Global maintenance detected
> MainThread::INFO::2021-08-03
>
> 06:47:46,521::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state GlobalMaintenance (score: 2586)
> MainThread::INFO::2021-08-03
>
> 06:47:55,577::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Global maintenance detected
> MainThread::INFO::2021-08-03
>
> 06:47:56,559::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state GlobalMaintenance (score: 2586)
> MainThread::INFO::2021-08-03
>
> 06:47:56,559::hosted_engine::525::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Best remote host ost-he-basic-suite-master-host-1 (id: 2, score: 3400)
> MainThread::INFO::2021-08-03
>
> 06:48:05,633::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Global maintenance detected
> MainThread::INFO::2021-08-03
>
> 06:48:06,188::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
> Penalizing score by 820 due to cpu load
> MainThread::INFO::2021-08-03
>
> 06:48:06,188::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state GlobalMaintenance (score: 2580)
> MainThread::INFO::2021-08-03
>
> 06:48:16,256::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Global maintenance detected
> MainThread::INFO::2021-08-03
>
> 06:48:16,950::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
> Penalizing score by 831 due to cpu load
> MainThread::INFO::2021-08-03
>
> 06:48:16,951::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state GlobalMaintenance (score: 2569)
> MainThread::INFO::2021-08-03
>
> 06:48:26,053::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Global maintenance detected
> MainThread::INFO::2021-08-03
>
> 06:48:26,999::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
> Penalizing score by 839 due to cpu load
> MainThread::INFO::2021-08-03
>
> 06:48:26,999::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state GlobalMaintenance (score: 2561)
> MainThread::INFO::2021-08-03
>
> 06:48:36,026::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Global maintenance detected
> MainThread::INFO::2021-08-03
>
> 06:48:36,802::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
> Penalizing score by 844 due to cpu load
> MainThread::INFO::2021-08-03
>
> 06:48:36,802::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state GlobalMaintenance (score: 2556)
> MainThread::INFO::2021-08-03
>
> 06:48:45,827::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Global maintenance detected
> MainThread::INFO::2021-08-03
>
> 06:48:46,401::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
> Penalizing score by 849 due to cpu load
> MainThread::INFO::2021-08-03
>
> 06:48:46,401::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state GlobalMaintenance (score: 2551)
> MainThread::INFO::2021-08-03
>
> 06:48:56,588::brokerlink::73::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
> Success, was notification of state_transition
> (GlobalMaintenance-ReinitializeFSM) sent? ignored
> MainThread::INFO::2021-08-03
>
> 06:48:58,685::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state ReinitializeFSM (score: 0)
> MainThread::INFO::2021-08-03
>
> 06:49:05,729::brokerlink::73::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
> Success, was notification of state_transition
> (ReinitializeFSM-EngineStarting) sent? ignored
> MainThread::INFO::2021-08-03
>
> 06:49:06,146::states::176::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(score)
> Penalizing score by 830 due to cpu load
> MainThread::INFO::2021-08-03
>
> 06:49:06,146::states::72::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_penalize_memory)
> Penalizing score by 400 due to free memory 1782 being lower than
> required 3171
> MainThread::INFO::2021-08-03
>
> 06:49:06,146::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state EngineStarting (score: 2170)
> MainThread::INFO::2021-08-03
>
> 06:49:06,150::state_decorators::95::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Timeout cleared while transitioning <class
> 'ovirt_hosted_engine_ha.agent.states.EngineStarting'> -> <class
> 'ovirt_hosted_engine_ha.agent.states.EngineUp'>
> MainThread::INFO::2021-08-03
>
> 06:49:06,172::brokerlink::73::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
> Success, was notification of state_transition
> (EngineStarting-EngineUp) sent? ignored
> MainThread::INFO::2021-08-03
>
> 06:49:06,178::hosted_engine::517::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state EngineUp (score: 2570)
> MainThread::ERROR::2021-08-03
>
> 06:49:16,197::states::398::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(consume)
> Host ost-he-basic-suite-master-host-1 (id 2) score is significantly
> better than local score, shutting down VM on this host
>
> =====================================================================================
>
> I looked a bit at /var/log/messages of the host, and while there is
> quite some noise there, can't tell specifically what might have caused
> the high cpu load.
>
> Also looked at logs of previous runs, and there is indeed a similar
> pattern, where the cpuload on the host causes the agent to penalize
> the score, but according to the agent log the load goes down faster,
> up to the point of being quite low when the engine is up, and only
> then we exit global maintenance, allowing the agent to take actions.
>
> I now ran it again, but think that otherwise I'll ignore this for now,
> unless we see more similar failures. If we do, we might want to
> check/monitor/log the cpu load on the hosts, and/or change test_008 to
> wait, after the engine is up, until the cpu load on the host goes down
> a bit. Before the port to pytest we waited a hard-coded 5 minutes and
> I
> changed that then to only wait until the engine VM is not migrating,
> and this worked more-or-less ok. We might need to refine that,
> although I'd rather not introduce another arbitrary long delays but
> wait for some condition.
>
> I also noticed that this ran was started at "03:04 AM" (UTC), by
> timer, whereas all previous timer-based runs started at 01:55 AM,
> somewhat earlier - perhaps there are some other things that run at
> that time that cause load.
>
>
> Checked some more and found another difference: The total cpu use was
> not very different between the failed and the last successful runs,
> but load caused by the engine vm was different between them - in the
> failed run, engine kept being '0.0' for a longer time, compared with
> the successful run. Since the agent tries to ignore the load by the
> engine vm, the final result is significantly different. The load-by
> -engine-vm is calculated based on the result of a VDSM getStats call
> on the engine VM. Indeed, in vdsm.log, cpuUser and cpuSys are both
> starting with '0.00' for both runs, but the failed one is '0.00' for
> a longer time - 45 seconds between VM.create and first-non-zero
> cpuUser on the successful run, vs 93 seconds for the failed one.
> Does this make sense (from VDSM's POV)? If so, perhaps there is
> a way to get from VDSM the answer to this question: "Is this 0.00
> CPU usage a result of still-initial, non-accurate, non-complete
> monitoring? Or is it really not using the CPU at all?" and use it
> in the agent somehow.
>
>
> I think it’s futile to try to understand that on jenkins.ovirt.org runs.
> It’s too unpredictable since it runs in a shared environment, you can’t
> know what’s going on on the same physical host.
>

I understand that I can't get/expect stable performance/load there. That's
not my question. I asked about VDSM. I don't know the code there well. I
assume that when it starts a VM, it initializes something somewhere with
'cpuUser: 0.00', and once its gets _some_ data, whatever that might be, it
starts reporting (based on) this data. My question was if there is a
reliable way to know if '0.00' is just real data, or initial-not-yet-real
data.


> One of the reasons we try to move away from it to beaker-based runs that
> run dedicated.
> The problem is you have no metrics for the actual host, so the numbers OST
> or vdsm see can be very misleading
> We do have sar monitoring of OST VMs, we do not capture anything from the
> host (the host running OST code), but I would guess that even if we would,
> then for mock env it’s probably going to be misleading anyway.
>

Again: Perhaps the actual reported load would be misleading, because it's
on a VM, which does not own its CPUs. I'd still expect '0.00' to mean that,
or have a way to know that it actually means "I don't know yet".

Best regards,
-- 
Didi
_______________________________________________
Devel mailing list -- devel@ovirt.org
To unsubscribe send an email to devel-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/devel@ovirt.org/message/GFIQJ2LBV76W46E2Y4SNS2FPTPSACWIJ/

Reply via email to