[ovirt-users] Re: New oVirt setup with OVN : Hypervisor with LACP bond : queries

2022-01-23 Thread Gianluca Cecchi
On Sat, Jan 22, 2022 at 11:41 PM ravi k  wrote:

> Hello team,
>
Hi,

Thank you for all the wonderful work you've been doing. I'm starting out
> new with oVirt and OVN. So please excuse me if the questions are too naive.
> We intend to do a POC to check if we can migrate VMs off our current
> VMware to oVirt. The intention is to migrate the VMs with the same IP into
> oVirt. We've setup oVirt with three hypervisors. All of them have four
> ethernet adapters. We have SDN implemented in our network and LACP bonds
> are created at the switch level. So we've created two bonds, bond0 and
> bond1 in each hypervisor. bond0 has the logical networks with vlan tagging
> created like bond0.101, bond0.102 etc.
>

Can you give some more details about your current vSphere infrastructure?
What about the level of downtime you could give when migrating?
Have you already planned the strategy to transfer your VMs from vSphere to
oVirt?
Take care that probably on your VMware side your VMs have virtual hw for
nics defined as vmxnet, so when you migrate to oVirt, it will change and so
depending on your OS type (Windows based or Linux based) and in case of
Linux, depending on your distro and version, some manual operations could
be required to remap vnic assignments and definitions.

One possible first way to proceed could be to make a clone of one running
VM into one disconnected from the vSphere infra and then test on it the
steps to port to oVirt and so analyze times and impacts


> As a part of the POC we also want to explore OVN as well to check if we
> can implement a zero trust security policy. Here are the questions now :)
>
> 1. We would like to migrate VMs with the current IP into oVirt. Is it
> possible to achieve this? I've been reading notes and pages that mention
> about extending the physical network into OVN. But it's a bit confusing on
> how to implement it.
> How do we connect OVN to the physical network? Does the fact that we have
> a SDN make it easier to get this done?
>

The downstream (RHV) documentation to do it is here:
https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.4/html/administration_guide/sect-adding_external_providers#Connecting_an_OVN_Network_to_a_Physical_Network

the upstream one is here:
https://www.ovirt.org/documentation/administration_guide/#Adding_OVN_as_an_External_Network_Provider

Take care that in RHV this feature is still considered Technology Preview,
so not recommended for production. It could apply to oVirt even more, so...
BTW, what do you mean with "... the fact that we have a SDN..."? Do you
mean standard virtual networking in contrast with physical one or do you
have any kind of special networking in vSphere now (NSX or such...)?



>
>
> 2. We have the IP for the hypervisor assigned on a logical
> network(ovirtmgmt) in bond0. I read in
> https://lists.ovirt.org/archives/list/users@ovirt.org/thread/CIE6MZ47GRCEX4Z6GWRLFSERCEODADJY/
> that oVirt does not care about how the IP is configured when creating the
> tunnels.
>

That was a thread originated by me... ;-)
But please consider that it is 5 years old now! At that time we were at 4.1
stage, while now we are at very different 4.4, so refer in case to recent
threads and better recent upstream (oVirt) and downstream (RHV) official
documentation pointed above
Also, at that time ansible was not very much in place, while now in many
configuration tasks it is deeply involved.
The main concern in that thread was the impact of having OVN tunneling on
the ovirtmgmt management network, that is the default choice when you
configure OVN, in contrast with creating a dedicated network for it.


> 3. Once we have OVN setup, ovn logical networks created and VMs
> created/migrated, how do we establish the zero trust policy? From what I've
> read there are ACLs and security groups. Any pointers on where to explore
> more about implementing it.
>

The downstream documentation and notes for this is here:
https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.4/html/administration_guide/sect-external_provider_networks#Assigning_Security_Groups_to_Logical_Networks

and upstream here:
https://www.ovirt.org/documentation/administration_guide/#Assigning_Security_Groups_to_Logical_Networks

some manual undocumented steps through OpenStack Networking API or Ansible
could be required depending on your needs

BTW: both upstream and downstream docs refer here to 4.2.7 :
"
In oVirt 4.2.7, security groups are disabled by default.
"
and
"
In Red Hat Virtualization 4.2.7, security groups are disabled by default.
"

They should be changed with the corresponding version, or into something
like "in 4.2.7 and above..." if that applies and is intended



> If you've read till here, thank you for your patience.
>

no problem ;-)

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: 

[ovirt-users] CentOS 8.4 Linux hosts from 4.4.8 to Rocky Linux 4.4.10

2022-01-21 Thread Gianluca Cecchi
Hello,
after updating the external engine from CentOS 8.4 and 4.4.8 to Rocky Linux
8.5 and 4.4.9 as outlined here:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YUDJRC22SQPAPAIURQIVSEMGITDRQOOM/
I went further and updated also the hosts.
Environment is with an external engine and 3 CentOS Linux 8.4 hosts in
4.4.8 with iSCSI storage domain.

Preliminarily I upgraded the engine to 4.4.10 (not yet the just released
async) without problems.
Then, one host at a time:

. put host into maintenance from web admin UI
Management --> Maintenance

. In a terminal f host set proxy for my environment needs
export https_proxy=http://my_proxy:my_proxy_port
export http_proxy=http://my_proxy:my_proxy_port (not sure if this
necessary...)

. in the same terminal execute migration script
./migrate2rocky.sh -r

. executed Management --> SSH Management --> SSH Restart from web admin ui
the host comes on in maintenance mode

. selected Installation --> Check for Upgrade but the host is detected as
already update

. for further security and to be sure that all upgrade steps are applied I
executed
Installation --> Reinstall
I deselected
- activate host after install
- reboot host after install
It went ok so

. executed Management --> SSH Management --> SSH Restart from web admin ui
the host comes on in maintenance mode

. Management --> Activate

. Empty another host moving its VMs to the just updated host and continue
in the same way, also electing as new SPM the updated host

All went smoothly and without VMs disruption.
Let's see how it goes next days with the light workload I have on this
testing environment.
Currently the async 1 of 4.4.10 is not catched up by engine-upgrade-check
command.. I'm going to retry applying again during the next few days.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MNOD5FQUL2OBV7PANMWKUCPT2P7LXRGN/


[ovirt-users] Re: oVirt 4.4.10 is now generally available

2022-01-20 Thread Gianluca Cecchi
On Wed, Jan 19, 2022 at 8:28 AM Sandro Bonazzola 
wrote:

> oVirt 4.4.10 is now generally available
>

thanks, Sandro!


>-
>
>CentOS Stream 8 latest updates
>-
>
>Full list of changes:
>
> --- ovirt-node-ng-image-4.4.9.3.manifest-rpm 2021-12-15
> 15:40:13.501764699 +0100
>
> +++ ovirt-node-ng-image-4.4.10.manifest-rpm 2022-01-19 08:09:36.668868583
> +0100
>
> @@ -2,7 +2,6 @@
>
>
>
I don't see any qemu related change in ovirt-node-ng packages' list. Does
this mean that the expected version to be used is still 6.0, in particulare
what shipped in 4.4.9-2 such as qemu-kvm-6.0.0-33.el8s.x86_64?

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/R6UDRF3ZA7IDHYALQPN3YQZX4SO25KLS/


[ovirt-users] External engine CentOS 8.4 and 4.4.8 to Rocky Linux 8.5 and 4.4.9

2022-01-14 Thread Gianluca Cecchi
Just wanted to share what in subject and feedback/comments welcome.
Starting env is an external engine on CentOS 8.4, updated up to late
September, with 4.4.8 async 2 (ovirt-engine-4.4.8.6-1.el8.noarch.rpm) and
kernel 4.18.0-305.19.1.el8_4

[g.cecchi@ovmgr1 ~]$ cat /etc/centos-release
CentOS Linux release 8.4.2105

I migrated to Rocky Linux 8.5 and then run engine-setup to take it to
current 4.4.9 level.
The system is a VM on a vSphere environment.
Currently there are 3 hosts with CentOS 8.4, that I'm going then to migrate
to Rocky Linux too.

Downloaded the current migration script from

https://raw.githubusercontent.com/rocky-linux/rocky-tools/main/migrate2rocky/migrate2rocky.sh
chmod +x migrate2rocky.sh

My engine needs a proxy to exit on the internet and the script doesn't
check my /etc/yum.conf configuration

so I set
export https_proxy=http://my_proxy:my_proxy_port
export http_proxy=http://my_proxy:my_proxy_port (not sure if this
necessary...)

otherwise when you run the script you get
curl: (7) Failed to connect to dl.rockylinux.org port 443: Connection timed
out
Error downloading the Rocky Linux signing key.

then as root

./migrate2rocky.sh -r

notable things observed:

. some packages from ovirt repos are gathered too... I don't know if
correct in this stage...

Running dnf update before we attempt the migration.
Last metadata expiration check: 0:00:47 ago on Thu Jan 13 11:56:21 2022.
Dependencies resolved.

. . .
 otopi-common   noarch  1.9.6-1.el8
 ovirt-4.4   94 k
 ovirt-ansible-collection   noarch  1.6.5-1.el8
 ovirt-4.4  289 k
 ovirt-engine-dwh-grafana-integration-setup noarch  4.4.9.1-1.el8
 ovirt-4.4   88 k
 ovirt-engine-dwh-setup noarch  4.4.9.1-1.el8
 ovirt-4.4   95 k
 ovirt-engine-extension-aaa-ldapnoarch  1.4.5-1.el8
 ovirt-4.4  130 k
 ovirt-engine-metrics   noarch  1.4.4-1.el8
 ovirt-4.4   92 k
 ovirt-engine-setup noarch  4.4.9.5-1.el8
 ovirt-4.4   18 k
 ovirt-engine-setup-basenoarch  4.4.9.5-1.el8
 ovirt-4.4  116 k
 ovirt-engine-setup-plugin-cinderlibnoarch  4.4.9.5-1.el8
 ovirt-4.4   39 k
 ovirt-engine-setup-plugin-imageio  noarch  4.4.9.5-1.el8
 ovirt-4.4   27 k
 ovirt-engine-setup-plugin-ovirt-engine noarch  4.4.9.5-1.el8
 ovirt-4.4  203 k
 ovirt-engine-setup-plugin-ovirt-engine-common  noarch  4.4.9.5-1.el8
 ovirt-4.4  122 k
 ovirt-engine-setup-plugin-vmconsole-proxy-helper   noarch  4.4.9.5-1.el8
 ovirt-4.4   38 k
 ovirt-engine-setup-plugin-websocket-proxy  noarch  4.4.9.5-1.el8
 ovirt-4.4   39 k
 ovirt-engine-vmconsole-proxy-helpernoarch  4.4.9.5-1.el8
 ovirt-4.4   26 k
 ovirt-engine-websocket-proxy   noarch  4.4.9.5-1.el8
 ovirt-4.4   33 k
 ovirt-imageio-client   x86_64  2.3.0-1.el8
 ovirt-4.4   21 k
 ovirt-imageio-common   x86_64  2.3.0-1.el8
 ovirt-4.4  158 k
 ovirt-imageio-daemon   x86_64  2.3.0-1.el8
 ovirt-4.4   15 k
 ovirt-release44noarch  4.4.9.3-1.el8
 ovirt-4.4   20 k
 ovirt-web-ui   noarch  1.7.2-1.el8
 ovirt-4.4   11 M
. . .
 qemu-img   x86_64  15:6.0.0-33.el8
 ovirt-4.4-advanced-virtualization  1.9 M
. . .
Transaction Summary
=

Install5 Packages
Upgrade  393 Packages
Remove 3 Packages

Total download size: 689 M
. . .
Generating grub configuration file ...
File descriptor 3 (/dev/pts/1) leaked on vgs invocation. Parent PID
4056644: /usr/sbin/grub2-probe
File descriptor 4 (/dev/pts/1) leaked on vgs 

[ovirt-users] Re: possible actions on host remaining as nonresponsive

2021-12-23 Thread Gianluca Cecchi
On Thu, Dec 23, 2021 at 7:44 PM Darrell Budic 
wrote:

> Try restarting libvirtd. It will also restart vdsmd, sometimes that fixes
> things for me when there has been a storage hiccup.
>

Thanks for the suggestion, but with the "ssh host restart" action actually
the server has restarted completely, so I doubt it can be that


> If it’s a HA Engine host, I’ve also had to restart the ha-agent/ha-broker
> combo is some situations as well.
>
> No, it's an external engine

On the host:
[root@ov300 vdsm]# nodectl check
Status: OK
Bootloader ... OK
  Layer boot entries ... OK
  Valid boot entries ... OK
Mount points ... OK
  Separate /var ... OK
  Discard is used ... OK
Basic storage ... OK
  Initialized VG ... OK
  Initialized Thin Pool ... OK
  Initialized LVs ... OK
Thin storage ... OK
  Checking available space in thinpool ... OK
  Checking thinpool auto-extend ... OK
vdsmd ... OK
[root@rhvh300 vdsm]#

I already had the idea to try restarting the engine server, and then I
found this similar bugzilla (even if for an older release):
https://access.redhat.com/solutions/4222911
RHV host in "not responding" state until ovirt-engine service restarted

I rebooted the engine server (that is a VM inside a vSphere environment)
and all came back good with the host set as up, together with the other
ones.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/F4XG3ZJK6PXHOH5CVTWBWU3BYAMEYMOH/


[ovirt-users] possible actions on host remaining as nonresponsive

2021-12-23 Thread Gianluca Cecchi
Hello,
I have a 4.4.8 host that results as nonresponsive.
The DC is FC based
Tried to restart some daemons without effect (vdsmd, mom-vdsmd wdmd)
Then I executed a ssh host reboot but it seems it continues this way after
rebooting

>From storage and network point of view it seems all ok on the host.

In vdsm.log of the host I see every 5 seconds:

2021-12-23 18:54:53,053+0100 INFO  (vmrecovery) [vdsm.api] START
getConnectedStoragePoolsList() from=internal,
task_id=916bc455-ce37-4b50-9f38-b69e3b03807f (api:48)
2021-12-23 18:54:53,053+0100 INFO  (vmrecovery) [vdsm.api] FINISH
getConnectedStoragePoolsList return={'poollist': []} from=internal,
task_id=916bc455-ce37-4b50-9f38-b69e3b03807f (api:54)
2021-12-23 18:54:53,053+0100 INFO  (vmrecovery) [vds] recovery: waiting for
storage pool to go up (clientIF:735)
2021-12-23 18:54:53,444+0100 INFO  (periodic/0) [vdsm.api] START
repoStats(domains=()) from=internal,
task_id=eb5540e0-0f90-4996-bc9a-7c73949f390f (api:48)
2021-12-23 18:54:53,445+0100 INFO  (periodic/0) [vdsm.api] FINISH repoStats
return={} from=internal, task_id=eb5540e0-0f90-4996-bc9a-7c73949f390f
(api:54)

In engine.log

2021-12-23 18:54:38,745+01 INFO
 [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
(EE-ManagedScheduledExecutorService-engineThreadMonitoringThreadPool-Thread-1)
[] Thread pool 'hostUpdatesChecker' is using 0 threads out of 5, 5 threads
waiting for tasks.
2021-12-23 18:55:27,479+01 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-73) []
EVENT_ID: VDS_BROKER_COMMAND_FAILURE(10,802), VDSM ov300 command Get Host
Capabilities failed: Message timeout which can be caused by communication
issues
2021-12-23 18:55:27,479+01 ERROR
[org.ovirt.engine.core.vdsbroker.monitoring.HostMonitoring]
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-73) []
Unable to RefreshCapabilities: VDSNetworkException: VDSGenericException:
VDSNetworkException: Message timeout which can be caused by communication
issues

I would like to try to put into maintenance the host and then activate, or
reinstall, but there is a power action still in place since 1 hour ago
(when I executed ssh host reboot attempt that got host rebooted but not
connected apparently) that prevents it... what is its timeout?

WHat can I check to understand the source of these supposed communication
problems?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZUUP2VEHKSJB7XDAUZZ2UUGG3UMFU6AC/


[ovirt-users] Re: Help installing oVirt on single machine, without cockpit

2021-12-16 Thread Gianluca Cecchi
On Fri, Dec 17, 2021 at 5:08 AM  wrote:

>
>
> > One problem I see is that now the update of a host is driven by the
> engine through playbooks and is not manually doable from the host itself as
> some time ago. And to have a container going through making modifications
> to the host where it is running doesn't seem a good idea
>
> For an hosted-engine setup, does the engine VM modify the host it's
> running on then? And you need the VM abstraction to update the host safely?
> If so, I'm guessing running the engine directly on the node isn't safe
> either then.
> Also, if the host/node itself gets updates though `dnf upgrade`, would it
> be possible to let dnf handle upgrades for it? and then rebuild/pull the
> latest container for engine updates? Or could the updates get out of sync
> of one another?


> Thanks!
> Cameron
>

I think this thread (even if the title is not so clear about the
discussions born inside) could be a good read regarding single host
limitations in terms of updating the environment, after the initial
deployment:
https://lists.ovirt.org/archives/list/users@ovirt.org/thread/2E6IOFCH5MJSPZ6MNSB5GIW7BGUZNXDS/#PI2LS3NCULH3FXQKBSB4IGXLKUBXE6UL

We were at 4.4.2 time and I think the point now is the same. See in
particular the last 9-10 comments, starting from Sandro one at "Sunday, 4
October 11:30 p.m."
They were related to single host and update questions.
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/Y3NCKS4J57ZXUHSUZ4FUU66WBCMTJ266/


[ovirt-users] Re: Help installing oVirt on single machine, without cockpit

2021-12-16 Thread Gianluca Cecchi
On Thu, Dec 16, 2021 at 9:05 AM Yedidyah Bar David  wrote:

>
>
>
>>
>> # With the engine inside a container
>> Finally got engine-setup to run inside a container! I need to figure out
>> the networking next, along with how to best hook up a postgresql db to it.
>> This part is going smoothly so far, I just wanted to update you all.
>>
>
> Thanks for the update, sounds promising!
>
> Just in case you want to have a look at the very old project I mentioned,
> it's here:
>
> https://github.com/mgoldboi/oVirt-Dockerized
>
> Best regards,
> --
> Didi
>
>
One problem I see is that now the update of a host is driven by the engine
through playbooks and is not manually doable from the host itself as some
time ago.
And to have a container going through making modifications to the host
where it is running doesn't seem a good idea.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/TALKTR6AGMYDOT2AS4PB6DVFLE54UAP5/


[ovirt-users] Re: oVirt and log4j vulnerability

2021-12-13 Thread Gianluca Cecchi
On Mon, Dec 13, 2021 at 2:37 PM Derek Atkins  wrote:

>
> On Mon, December 13, 2021 8:04 am, Gianluca Cecchi wrote:
> >>
> > If I understood correctly reading here:
> >
> https://blog.qualys.com/vulnerabilities-threat-research/2021/12/10/apache-log4j2-zero-day-exploited-in-the-wild-log4shell
> >
> > you are protected by the RCE if java is 1.8 and greater than 1.8.121
> > (released on 2017)
>
> Do you mean 1.8.0.121?  For example, my system has:
>
> java-1.8.0-openjdk-headless-1.8.0.252.b09-2.el7_8.x86_64
>
> -derek
>
>
Yes, what the link refers to as 8u121:
https://www.oracle.com/java/technologies/javase/8u121-relnotes.html

Your version: 8u252 (or anyway based on it).
On my 4.4.8 engine I have
java-1.8.0-openjdk-headless-1.8.0.302.b08-0.el8_4.x86_64 but I have also
java-11-openjdk-headless-11.0.12.0.7-0.el8_4.x86_64 that is what
ovirt-engine uses, based on:

[root@ovmgr1 ovirt-engine]# ll /proc/$(pidof ovirt-engine)/fd | grep jvm
lr-x--. 1 ovirt ovirt 64 Sep 24 09:02 3 ->
/usr/lib/jvm/java-11-openjdk-11.0.12.0.7-0.el8_4.x86_64/lib/modules
[root@ovmgr1 ovirt-engine]#

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HMHHRWIIEPX2HPQKUBL6UO2YJPT4ANFE/


[ovirt-users] Re: oVirt and log4j vulnerability

2021-12-13 Thread Gianluca Cecchi
On Mon, Dec 13, 2021 at 1:38 PM Sandro Bonazzola 
wrote:

> So far we can't confirm whether oVirt engine systems are affected or not:
> the oVirt infra team is digging into this.
> I can confirm that ovirt-engine-wildfly is shipping a log4j version which
> is affected by the vulnerability and we are monitoring Wildfly project so
> we'll be able to ship an update as soon as a fix will be available (we are
> just repackaging the binary build they provide).
> But I got no report so far confirming if the way we run Wildfly exposes
> the vulnerable system to potential attackers yet.
>
>
>
If I understood correctly reading here:
https://blog.qualys.com/vulnerabilities-threat-research/2021/12/10/apache-log4j2-zero-day-exploited-in-the-wild-log4shell

you are protected by the RCE if java is 1.8 and greater than 1.8.121
(released on 2017)

"
If the server has Java runtimes later than 8u121, then it is protected
against remote code execution by defaulting
“com.sun.jndi.rmi.object.trustURLCodebase” and
“com.sun.jndi.cosnaming.object.trustURLCodebase” to “false”(see
https://www.oracle.com/java/technologies/javase/8u121-relnotes.html).
"

It is not clear to me if it means that Java 11 (and 17) also maintained
that setting.
In one of my oVirt with 4.4.8 it seems that engine is using
java-11-openjdk-headless-11.0.12.0.7-0.el8_4.x86_64 package

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WH3WZLRM6NYC7MJVWSTA4LY5YWDF57VW/


[ovirt-users] Re: Creating VMs from templates with their own disks

2021-11-28 Thread Gianluca Cecchi
On Sun, Nov 28, 2021 at 3:44 AM Sina Owolabi  wrote:

>
> - name: Print vm attributes
>   ovirt.ovirt.ovirt_vm_info:
> auth: "{{ ovirt_auth }}"
> pattern: name="{{ vm_fqdn }}"
>   register: vm_info
> - debug:
> msg: "{{ vm_info.ovirt_vms[0] }}"
>
>
here above you get an empty disk_attachments field.
You have to add fetch_nested: true to traverse and get the disk attachments
values

Eg:
- name: Print vm attributes
  ovirt.ovirt.ovirt_vm_info:
auth: "{{ ovirt_auth }}"
pattern: name="{{ vm_fqdn }}"
fetch_nested: true
  register: vm_info
- debug:
msg: "{{ vm_info.ovirt_vms[0] }}"


> - name: Rename disk
>   ovirt.ovirt.ovirt_disk:
> auth: "{{ ovirt_auth }}"
> id: "{{ vm_info.ovirt_vms[0].id }}"
> storage_domain: lrg0-ovirt-mydom-internal-Local
> name: "{{ vm_fqdn }}-osdisk0"
> vm_name: "{{ vm_fqdn }}"
>
>
Here, supposing from pattern filter above you get only 1 VM so you can use
index 0 and you get only 1 disk inside the VM and you can use index 0 also
for disk, you can use:
See also "ansible-doc ovirt_vm_info" help page, where you can filter the
nested_attributes...
Then the rename disk task would become

- name: Rename disk
  ovirt.ovirt.ovirt_disk:
auth: "{{ ovirt_auth }}"
id: "{{ vm_info.ovirt_vms[0].disk_attachments[0].id }}"
storage_domain: lrg0-ovirt-mydom-internal-Local
name: "{{ vm_fqdn }}-osdisk0"
vm_name: "{{ vm_fqdn }}"

This works for me in a 4.4.7 test environment

HIH,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KAS7RHBOSQ2YQ37DU3KHWNPH6XCSELML/


[ovirt-users] Suggested upgrading path from CentOS based 4.4.8 to 4.4.9

2021-11-08 Thread Gianluca Cecchi
I have a lab with an environment based on 4.4.8.6-1, with 3 CentOS Linux
8.4 hosts and a CentOS 8.4 external engine system (that is a VM on vSphere,
so that I can leverage a snapshot methodology for the process...).
I would like to pass to 4.4.9 and retain a full plain OS on hosts for the
moment, without going through oVirt nodes, but standing the repo problems
and CentOS 8.x going through EOL this is what I'm planning to do:

1. stop engine service on engine system

2. convert engine to CentOS Stream
This step needs some confirmation.
Could you provide an official link about the process?
I'm not able to find it again. Is it a problem of mine or all (CentOS
website, RHEL website) seem to point only to conversion from CentOS Linux
to RHEL??
Apart external websites provided workflows, I was only able to find a mid
January youtube video, when CentOS was based on 8.3, with these steps:
yum install centos-release-stream
yum swap centos-{linux,stream}-repos
yum repolist
yum distro-sync
reboot
The video link is here:
https://www.youtube.com/watch?v=Ba2ytp_8x7s

No mention at
https://www.redhat.com/en/blog/faq-centos-stream-updates

And on CentOS page I only found this:
https://centos.org/distro-faq/
with Q7 containing only the two instructions:
dnf swap centos-linux-repos centos-stream-repos
dnf distro-sync

What to use safely?
Is it possible to include some sort of documentation or links on oVirt
page, to migrate from CentOS Linux to CentOS Stream for oVirt upgrade
purposes?

3. After reboot implied, I think, in step 2., use the usual steps to update
engine to 4.4.9

4. update the first out of three hosts from CentOS Linux to CentOS Stream
and to 4.4.9.

4.a follow the same approach of engine (when defined) and pass it to Stream
retaining the 4.4.8.
4.b upgrade from the web admin gui to 4.4.9

5. Do the same for second host and third hosts

Any hints, comments, limitations in having mixed 4.4.8 and 4.4.9 hosts for
a while and such?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/L3CRTNR2IUGNRZNVQMROQABHHBKMEYPP/


[ovirt-users] Re: export to export domain concurrency

2021-11-04 Thread Gianluca Cecchi
On Thu, Nov 4, 2021 at 12:32 PM Benny Zlotnik  wrote:

>  yes, it should work, did you run into issues?
>
>
No, I was in doubt that it could be a serial task.

I can confirm the parallelism, as I tried with 2 VMS, both with 3 disks:
20Gb+20Gb+90Gb the first VM1 and 20Gb+20Gb+80Gb the second one VM2.
The source storage domain is based on iSCSI Equallogic with 10Gbit/s
connection.
The export domain is a vSphere VM (on 10Gb/s adapter) connected through the
ovirtmgmt channel (that is based on 10Gb/s adapters in LACP)

Going one by one I got
Starting export Vm VM1 to xfer 11/4/21 12:14:49 PM
Vm VM1 was exported successfully to xfer 11/4/21 12:21:54 PM
--> total 7' 5"

Starting export Vm VM2 to xfer 11/4/21 12:25:23 PM
Vm VM2 was exported successfully to xfer 11/4/21 12:32:19 PM
--> total 6' 56"

Going then in parallel (delay of 7 seconds, not using the method of
multi-select and export):
Starting export Vm VM1 to xfer 11/4/21 3:15:29 PM
Starting export Vm VM2 to xfer 11/4/21 3:15:36 PM
Vm VM2 was exported successfully to xfer 11/4/21 3:26:32 PM
Vm VM1 was exported successfully to xfer 11/4/21 3:26:46 PM

---> total 11' 17"

So I got about a 20% gain.
Next week I have to move some VMs with 900Gb of storage each, so I can get
optimizations going in parallel

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/OCCHHALG54GTSCT7H7DVIOO2HOER4LR7/


[ovirt-users] export to export domain concurrency

2021-11-04 Thread Gianluca Cecchi
Hello,
the export to export domain task is managed asynchronously.
Does it support concurrency, so that I can export more VMs at the same time
to the same export domain? If not by default, is there a parameter to
enable it?
The need would be on 4.4.7/4.4.8 based environments

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UWH4QB4DVUUKRFNDUCDZ6CA4AOW3OKZ5/


[ovirt-users] Re: upgrade dependency issues

2021-10-26 Thread Gianluca Cecchi
On Tue, Oct 26, 2021 at 12:12 PM Sandro Bonazzola 
wrote:

> Thanks for the report, my team is looking into the dedependency failures.
> oVirt 4.4.9 has been developed on CentOS Stream 8 and some dependencies
> are not yet available on RHEL 8.4 and derivatives.
>

Ok, fair enough you only test on CentOS Stream 8, but at least I think you
should change what you are going to write in the next release notes,
putting only what actually tested.

For 4.4.9 there was:

"

This release is available now on x86_64 architecture for:

   -

   Red Hat Enterprise Linux 8.4
   -

   CentOS Linux (or similar) 8.4
   -

   CentOS Stream 8


This release supports Hypervisor Hosts on x86_64 and ppc64le architectures
for:

   -

   Red Hat Enterprise Linux 8.4
   -

   CentOS Linux (or similar) 8.4
   -

   oVirt Node NG (based on CentOS Stream 8)
   -

   CentOS Stream 8


"
So one understands that at least installation/upgrade from 4.4.8 to 4.4.9
has been validated when the hosts are in CentOS 8.4 or in RH EL 8.4, that
currently are the latest 8.4 level released, while it seems both fails
right now, correct?

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MQ7SHH3BK6JLBL7OKYC6DAOUE2J4FET2/


[ovirt-users] Re: HA VM and vm leases usage with site failure

2021-10-17 Thread Gianluca Cecchi
On Tue, Aug 10, 2021 at 1:02 PM Klaas Demter  wrote:

> Okay, so the SPM is only blocking me if I have thin provisioned block
> storage that needs extending. This is luckily not the case because we
> primarily use NFS. It is something though I have never thought about, in
> my head the storage leases completely solved the "host crashes and power
> management does not answer" until I read your mail :)
>
>
> Thanks for the detailed explanation Nir!
>
>
> Greetings
>
> Klaas
>
>
In the meantime I opened a case (number 03004698 with also logs) regarding
expected behavior and the Disaster Recovery Guide not being so clear in
this scenario.
A documentation bugzilla has been opened so feel free to read and/ or
contribute to it.
https://bugzilla.redhat.com/show_bug.cgi?id=2014790
Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3AXUGHK22TYOWO22F7IINKZE3TSS43YF/


[ovirt-users] Re: How to add a note to a VM

2021-10-12 Thread Gianluca Cecchi
On Tue, Oct 12, 2021 at 3:29 PM Nir Soffer  wrote:

> On Tue, Oct 12, 2021 at 3:24 PM Gianluca Cecchi
>  wrote:
> >
> > Hello,
> > I know there are the "Comment" and "Description" columns available in
> many areas of the Webadmin Gui.
> > But there are some tasks, like "shutdown VM" or "Management ->
> Maintenance" for a host, where I can specify a "reason" for doing that task
> and then a note icon appears, aside the object, with the mouse over showing
> the note text, like in this image:
> >
> https://drive.google.com/file/d/1v3Yd2t7AtuRFMT6HPFYYZqYUmJLMHYMY/view?usp=sharing
> >
> > Is there a way to do it in general? So for example I have a VM and I
> want to put a note (for some colleague, or to remind me to do an action
> tomorrow, ecc...)
> >
> > And btw: how can I manually remove the note? Eg I shutdown a VM and fill
> in the "Reason" field and then in a second moment I want to remove it
>
> The "comment" field was designed exactly for this purpose.
>
> Maybe this is not documented?
>
> Nir
>
>
Probably yes, but it results less visible than having a tooltip with the
icon of a note...
And also, sometimes the "Comment" column is not one of the first, so you
have to arrange / order it so that it comes to the left...
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ANFKPM53FOCFPZR6AZXXD7VJRTFXNNMK/


[ovirt-users] How to add a note to a VM

2021-10-12 Thread Gianluca Cecchi
Hello,
I know there are the "Comment" and "Description" columns available in many
areas of the Webadmin Gui.
But there are some tasks, like "shutdown VM" or "Management -> Maintenance"
for a host, where I can specify a "reason" for doing that task and then a
note icon appears, aside the object, with the mouse over showing the note
text, like in this image:
https://drive.google.com/file/d/1v3Yd2t7AtuRFMT6HPFYYZqYUmJLMHYMY/view?usp=sharing

Is there a way to do it in general? So for example I have a VM and I want
to put a note (for some colleague, or to remind me to do an action
tomorrow, ecc...)

And btw: how can I manually remove the note? Eg I shutdown a VM and fill in
the "Reason" field and then in a second moment I want to remove it

Thanks in advance,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ABXZEZBLDF5PYBKTRTFTQQKJLB2RNX32/


[ovirt-users] Re: Upgrading from 4.3 to 4.4.8

2021-10-03 Thread Gianluca Cecchi
On Mon, Oct 4, 2021 at 12:32 AM Jason Harris  wrote:

> Engine is updated to the latest 4.3 release and rebooted.  Still getting
> this after re-enabling the 4.4 repo:
>
>
oVirt 4.4 requires engine and hosts based on RH EL / CentOS 8, while your
current env in 4.3 has them based on RH EL 7 / CentOS 7. You have to
reinstall the node that you isolated with the engine.
So after taing the engine at its latest 4.3.x level and rebooting the 4.3
engine OS, you have to follow the steps described in 6.5: stop the
ovirt-engine service, take a backup and copy the generated file outside of
the env, scratch the engine and related node and reinstall it with 4.4
ovirt node and then on the host, now in 4.4, run the command

hosted-engine --deploy --restore-from-file=backup.bck

and so on...

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/7QHHXUQPNXOZJ5QMLOM3PLFIPMJVVA3U/


[ovirt-users] Re: Upgrading from 4.3 to 4.4.8

2021-10-02 Thread Gianluca Cecchi
On Sat, Oct 2, 2021 at 4:45 PM  wrote:

> You guys scare me.  Seriously.
>
> I'm trying to update from oVirt 4.3 to 4.4 using a self-hosted engine.
> All VMs on the host are migrated off, other than the hosted engine, from
> the first host I intend to update to EL 8.4, per the documentation.
>
> I installed the RPM from here:
> https://resources.ovirt.org/pub/yum-repo/ovirt-release44.rpm
>
> I'm configuring the Engine following the upgrade document located here:
>
> https://www.ovirt.org/documentation/upgrade_guide/index.html#SHE_Upgrading_from_4-3
>
> Step 6.4, after installing the oVirt 4.4 RPM on the engine:
> engine-upgrade-check - I get this response:
>
>
Step 6.4 is where you update your self hosted engine to the latest 4.3.x
version available. Not 4.4 yet.
So if you are still in 4.3.7, with step 6.4 you take it at 4.3.10, so that
the backup you take in 4.3.10 (step 6.5) is tested and correctly restorable
in 4.4.8 (better 4.4.8 than 4.4.7 you cited..)
If you are already at 4.3.10 level you can skip step 6.4.

HIH,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/K6T5ZDTBAI5KHSOOZPNR5JE5P6IQWZAR/


[ovirt-users] Balancing actions not shown in events list

2021-10-01 Thread Gianluca Cecchi
Hello,
I have a cluster composed of 4 hosts, with 2 hosts in site A and 2 hosts in
site B.
Version of engine and hosts is latest 4.4.8-6.
Site A is the primary site and its hosts have SPM priority high, while site
B hosts have SPM priority low.
For critical VMs I create a cluster affinity group so that they preferably
run on hosts in site A.
If I migrate a VM from one host in site A to one host in site B, the
migration completes but suddenly, after a few seconds (ranging from 10 to
30) the VM comes back again (live migrates) to one host of the site A pool.

Two considerations:

. when the VM comes back to site A and I'm connected to the web admin gui I
see in bottom right the pop-up message regarding the balancing operation:
https://drive.google.com/file/d/1lfm0AVwYKyyRL1qHh94AySpr3XAtV7lO/view?usp=sharing
But then if I go in the VM, or cluster, or general events pane I don't see
any direct feedback regarding this balancing that took place.
I only see the VM migration events:
Oct 1, 2021, 2:47:01 PM Migration completed (VM: impoldsrvdbpbi, Source:
, Destination: , Duration: 15 seconds, Total: 27 seconds, Actual
downtime: 67ms)
Oct 1, 2021, 2:46:34 PM Migration initiated by system (VM: impoldsrvdbpbi,
Source: , Destination: , Reason: Affinity rules enforcement).
Oct 1, 2021, 2:45:45 PM Migration completed (VM: impoldsrvdbpbi, Source:
, Destination: , Duration: 2 seconds, Total: 14 seconds, Actual
downtime: (N/A))
Oct 1, 2021, 2:45:30 PM Migration started (VM: impoldsrvdbpbi, Source:
, Destination: , User: gian@internal).

That indeed contain some information (Reason: Affinity rules enforcement)
but only in the VM migration line.
Could it be useful to add an independent line regarding the balancing
trigger that implies then a migration?

. In this case could it be useful to give the user a warning that the VM
will be suddenly migrated back so that  he/she can think about it before
having at the end two migrations with a final stage that is the starting
point itself...?

If I leave only one host in site A and put it into maintenance, the VMs are
correctly migrated to hosts in site B and even when the host in site A
comes back available, the coming back operation is not triggered. Is this
something expected or should the live migrate to hosts in site A?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WFBGV4THNYH27CBWTHUBXTGUW6I6WUWZ/


[ovirt-users] Re: Hosted Engine cluster version compatib.

2021-09-24 Thread Gianluca Cecchi
On Fri, Sep 24, 2021 at 11:05 AM Michal Skrivanek <
michal.skriva...@redhat.com> wrote:

>
>
> On 24. 9. 2021, at 9:43, Gianluca Cecchi 
> wrote:
>
> BTW: there are two things that are not so clear to me:
> 1. Is this only impacting SHE environments or is it a general problem also
> with external standalone engine ones?
>
>
> it’s affecting all VMs, in standalone envs engine is not a VM
>

Of course... I meant if I have standalone external engine that manages VM1
and VM2 and I update from 4.4.7 to 4.4.8-5, will VM1 and VM2 impacted?
I presume yes from your answer


> 2. Is it correct to say that if I already upgraded in the past to 4.4.7
> and at that time I updated my cluster level from 4.5 to 4.6 (both in case
> of SHE and external engine), then I shouldn't have this kind of problems if
> then I updated to the impacted 4.4.8-5 version? And then I can go and
> continue updating to 4.4.8-6 without any risk/problems?
>
>
> the cluster level update problem is just a side effect of the time zone
> missing. You will see problems in other flows, or next time you upgrade to
> 4.7 if we ever have one.
>
>
So I don't understand if I need to rollback or if I can anyway update to
4.4.8-6 and eventually modify from database.
By the way in my env all my VMs in 4.4.7 had timezone set to the default
value "Etc/GMT". Does this imply no impact at all?

Eg on another environment I see that after updating from 4.4.7 to 4.4.8-5
and then 4.4.8-6, in Web Admin GUI if I select a VM, in general subtab I
have for 2 of them

Hardware Clock Time Offset: Etc/GMT

for all the other ones there isn't the "Hardware Clock Time Offset" item...
Fot these VMs under edit virtual machine -> system -> Hardware Clock Time
Offset I see they have the value
default:(GMTZ) Greenwich Standard Time

and from database point of view:

engine=# select time_zone,count(*) from vm_static group by time_zone;
 time_zone | count
---+---
   |18
 Etc/GMT   | 2
(2 rows)

engine=#

engine=# select vm_name from vm_static where time_zone='Etc/GMT';
  vm_name
---
 testcl1
 c76client
(2 rows)

engine=#

So I should use the db statement too, correct?
Any need to restart engine service after or before that?

Thanks
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/NWGZ4IGAPBOJU2GAJQZIJXCBJKNN4EXS/


[ovirt-users] Re: Managed Block Storage and Templates

2021-09-24 Thread Gianluca Cecchi
On Wed, Sep 22, 2021 at 2:30 PM Shantur Rathore 
wrote:

> Hi all,
>
> Anyone tried using Templates with Managed Block Storage?
> I created a VM on MBS and then took a snapshot.
> This worked but as soon as I created a Template from snapshot, the
> template got created but there is no disk attached to the template.
>
> Anyone seeing something similar?
>
> Thanks
>
>
Are you using an external ceph cluster? Or what other cinder volume driver
have you configured for the MBS storage domain?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XOOT3MEALKWAIZFBJGJMQEUBEM32M6FI/


[ovirt-users] Re: Hosted Engine cluster version compatib.

2021-09-24 Thread Gianluca Cecchi
BTW: there are two things that are not so clear to me:
1. Is this only impacting SHE environments or is it a general problem also
with external standalone engine ones?
2. Is it correct to say that if I already upgraded in the past to 4.4.7 and
at that time I updated my cluster level from 4.5 to 4.6 (both in case of
SHE and external engine), then I shouldn't have this kind of problems if
then I updated to the impacted 4.4.8-5 version? And then I can go and
continue updating to 4.4.8-6 without any risk/problems?

Thanks,
Gianluca

On Thu, Sep 23, 2021 at 9:45 PM Diggy Mc  wrote:

> The only VM that my cluster compatibility upgrade complains about is
> "HostedEngine".  I'm not about to test my SQL knowledge by writing my own
> SQL command and I see no reason to touch VMs that don't upset the cluster
> upgrade.  Can you please provide a SQL command that corrects ONLY the
> HostedEngine VM ???  Much appreciated.  NOTE: All my servers' OS (physical
> and VM) are set to "Etc/UCT" wherever possible.
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/TNM4W25ECQKXRXCDVMXEWAD3OA3B5IDE/
>
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/72BIXUGUE44Y2XGSNXD652VXYSQHTIEC/


[ovirt-users] Re: Hosted Engine cluster version compatib.

2021-09-23 Thread Gianluca Cecchi
On Thu, Sep 23, 2021 at 7:29 PM Diggy Mc  wrote:

> I just upgraded the HE to 4.4.8.6 and rebooted it.  I still cannot upgrade
> the cluster compatibility level.  Cannot edit properties of the HE either.
>
>
If I understood correctly, the fix is in the sense that if not already
updated to 4.4.8, the flow should be ok now.
But probably to solve the problems for the guys that already updated to the
"broken" 4.4.8, more work still has to be done by the developers...

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LBUHSVAZIFCAT3EOSCX5BH4ZNCDRIEYG/


[ovirt-users] Re: oVirt Node - master - CentOS Stream 9 based ISO available for testing

2021-09-16 Thread Gianluca Cecchi
On Thu, Sep 16, 2021 at 5:35 PM Sandro Bonazzola 
wrote:

> Hi,
> I'm still working on it but I have a first ISO ready for giving a first
> run at
>
> https://resources.ovirt.org/pub/ovirt-master-snapshot-static/iso/ovirt-node-ng-installer/4.5.0-2021091610/el9/ovirt-node-ng-installer-4.5.0-2021091610.el9.iso
>
> Known limitations:
> - No hosted engine setup available
>
>
Nice!
If SHE not available, what would be the procedure to install the standalone
engine before deploying the host?
Or could I try to deploy the node using a 4.4.8 standalone engine in its
own DC/Cluster?

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/PJY465PTKKOJMCPUQGJEKTUSUFZJGBJ4/


[ovirt-users] Re: problems testing 4.3.10 to 4.4.8 upgrade SHE

2021-09-12 Thread Gianluca Cecchi
On Sun, Sep 12, 2021 at 10:35 AM Yedidyah Bar David  wrote:

>
> >>
> >> It was the step I suspect there was a regression for in 4.4.8
> (comparing with 4.4.7) when updating the first hosted-engine host during
> the upgrade flow and retaining its hostname details.
>
> What's the regression?
>

I thought that in 4.4.7 there was not this problem if you use the same
hostname but with different (real or virtual) hw as the first host during
your SHE upgrade from 4.3.10 to 4.4.7.
But probably it was not so and I didn't remember correctly


> >> I'm going to test with latest async 2 4.4.8 and see if it solves the
> problem. Otherwise I'm going to open a bugzilla sending the logs.
>
> Can you clarify what the bug is?
>

The automatic mgmt of host adding during the "hosted-engine --deploy
--restore-from-file=backup.bck" step if you have different hw and you want
to recycle your previous hostname.
In the past it often happened to me to combine upgrades of systems with hw
refreshing (with standalone hosts, rhcs clusters, also ovirt/rhv from 4.2
to 4.3 if I remember correctly, ecc.) where you re-use an existing hostname
on new hardware
More than a bug it would be an RFE perhaps



> > As novirt2 and novirt1 (in 4.3) are VMS running on the same hypervisor I
> see that in their hw details I have the same serial number and the usual
> random uuid
>
> Same serial number? Doesn't sound right. Any idea why it's the same?
>

My env is nested oVirt and my hypervisors are Vms.
I notice that in oVirt if you clone a VM it changes the uuid in the clone
but it retains the serial number...

> Unfortunately I cannot try at the moment the scenario where I deploy the
> new novirt2 on the same virtual hw, because in the first 4.3 install I
> configured the OS disk as 50Gb and with this size 4.4.8 complains about
> insufficient space. And having the snapshot active in preview I cannot
> resize the disk
> > Eventually I can reinstall 4.3 on an 80Gb disk and try the same,
> maintaining the same hw ... but this would imply that in general I cannot
> upgrade using different hw and reusing the same hostnames correct?
>
> Yes. Either reuse a host and keep its name (what we recommend in the
> upgrade guide) or use a new host and a new name (backup/restore
> guide).
>
> The condition to remove the host prior to adding it is based on
> unique_id_out, which is set in (see also bz 1642440, 1654697):
>
>   - name: Get host unique id
> shell: |
>   if [ -e /etc/vdsm/vdsm.id ];
>   then cat /etc/vdsm/vdsm.id;
>   elif [ -e /proc/device-tree/system-id ];
>   then cat /proc/device-tree/system-id; #ppc64le
>   else dmidecode -s system-uuid;
>   fi;
> environment: "{{ he_cmd_lang }}"
> changed_when: true
> register: unique_id_out
>
> So if you want to "make this work", you can set the uuid (either in
> your (virtual) BIOS, to affect the /proc value, or in
> /etc/vdsm/vdsm.id) to match the one of the old host (the one you want
> to reuse its name). I didn't test this myself, though.
>
>
I confirm that I reverted the snapshots of the 2 VMs used as hypervisors
taking them again at initial 4.3 status and remade all the steps, but right
after the install of the OS of 4.4.8 oVirt node I created /etc/vdsm/vdsm.id
inside novirt2 with the old 4.3 value (the file was not there at that
moment) and then all the flow went as expected and I was then able to reach
the final 4.4.8 async 2 env with both hosts at 4.4.8, cluster and DC
updated to 4.6 compatibility level and no downtime for the VMs inside the
env, because I was able to execute live migration after upgrading the first
host


Perhaps, if you do want to open a bug, it should say something like:
> "HE deploy should remove the old host based on its name, and not its
> UUID". However, it's not completely clear to me that this won't
> introduce new regressions.
>
> I admit I didn't completely understand your flow, and especially your
> considerations there. If you think the current behavior prevents an
> important flow, please clarify.
>
> Best regards,
> --
> Didi
>
>
My considerations, as explained at the beginning, were to give the chance
to reuse the hostname (often the oVirt admin is not responsible for
hostname creation/mgmt) if you want to leverage new hw in combination with
the upgrade process.

Thanks for all the other considerations you put into your answer.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KXSZHEPQSHTFS3VSB25TUZ7DNFFVBHYB/


[ovirt-users] Re: how to remove a failed backup operation

2021-09-09 Thread Gianluca Cecchi
On Thu, Sep 9, 2021 at 2:21 PM Nir Soffer  wrote:

> On Thu, Sep 9, 2021 at 12:53 PM Nir Soffer  wrote:
> ...
> >> Any insight for finding the scratch disks ids in engine.log?
> >> See here my engine.log and timestamp of backup (as seen in database
> above) is 15:31 on 03 September:
> >>
> >>
> https://drive.google.com/file/d/1Ao1CIA2wlFCqMMKeXbxKXrWZXUrnJN2h/view?usp=sharing
> >
> >
> > To find the scratch disks the best way is to use the UI - open the
> storage > disks tab
> > and change the content type to "Backup scratch disks"
> > (see attached screenshot)
>

I confirm no scratch disks has been left in my case


> Regardless, it is useful to understand engine log, here are the
> relevant events in
> your log:
>
>
[snip]

11. Error in the backup command - not sure why...
>
> [snip]
>
> 12. Errors writing to database - no space left
>
>
>
[snip]


> This seems to be the root cause for the engine failure - engine cannot
> write to the
> database, so it cannot complete handling of the backup command.
>

[snip]


>
> So both scratch disks were removed as expected, and the only issue is the
> backup
> stuck in the finalizing state.
>
> Because the root cause is no space on the database disk, caused by user
> error
> (filling up engine disk by mistake), I don't think we can do much about
> this.
>
> Nir
>


Indeed I didn't recall my filesystem layout. The full was in my home dir
but as I have no dedicated /home filesystem, it generated a / filesystem
full and so impacting also with PostgreSQL database for the engine that
uses /var/lib/pgsql/data/base.
This goes and confirms your recommendation of not using the engine for
running the backup.

currently in fact the layout of filesystems of my external engine is :

[g.cecchi@ovmgr1 ~]$ df -h
Filesystem  Size  Used Avail Use% Mounted on
devtmpfs4.9G 0  4.9G   0% /dev
tmpfs   4.9G   24K  4.9G   1% /dev/shm
tmpfs   4.9G   25M  4.9G   1% /run
tmpfs   4.9G 0  4.9G   0% /sys/fs/cgroup
/dev/mapper/cl-root  43G  5.1G   36G  13% /
/dev/sda2   976M  199M  710M  22% /boot
/dev/sda1   599M  7.3M  592M   2% /boot/efi
tmpfs   998M 0  998M   0% /run/user/1000
[g.cecchi@ovmgr1 ~]$

Thanks very much for the detailed analysis.

Ok also for the closing of the bugzilla.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3F2MU7CM2LXSYREAGXRX6OMZZBF2B5M6/


[ovirt-users] Re: Time Drift Issues

2021-09-08 Thread Gianluca Cecchi
On Wed, Sep 8, 2021 at 1:19 PM Nur Imam Febrianto 
wrote:

> Hi Marcos,
>
>
>
> Want to clarify one thing. If I’m using and oVirt Node based host (not EL
> Linux Based), does it already configured with ntp client or can I just
> configure any ntp client on the host ? If I just configure ntp client on
> the host either chronyc or ntpd, will it persist after upgrading the node ?
>
>
>
> Thanks before.
>
>
>
> Regards,
>
> Nur Imam Febrianto
>
>
>
>
Just to clarify: when using oVirt node 4.4, during installation you have
the usual way to configure ntp inside anaconda. In 4.4, based on CentOS
(stream) 8, when done that way it will configure chronyd.
In the recent past there was a bug where sometimes installation crashed
when configuring ntp in certain conditions.
If you don't configure ntp at install time you can configure it later by
modifying /etc/chrony.conf and starting/enabling the service from the
cockpit web interface (or command line with usual commands). It will remain
across reboots and/or image updates.
HIH,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QNURIQHGR75WTB35UE7SZ7YSA6SFJNGI/


[ovirt-users] Re: how to remove a failed backup operation

2021-09-08 Thread Gianluca Cecchi
On Sun, Sep 5, 2021 at 6:00 PM Pavel Bar  wrote:

> Hi,
> Please try the instructions below and update whether it helped.
>
> Thank you!
>
> Pavel
>
>
Thanks for input.
If I understand it correctly I have to complete the steps described by Nir
and then work at db level.

Right now what I see in the table is:

engine=# \x
Expanded display is on.
engine=# select * from vm_backups;
-[ RECORD 1 ]--+-
backup_id  | 68f83141-9d03-4cb0-84d4-e71fdd8753bb
from_checkpoint_id |
to_checkpoint_id   | d31e35b6-bd16-46d2-a053-eabb26d283f5
vm_id  | dc386237-1e98-40c8-9d3d-45658163d1e2
phase  | Finalizing
_create_date   | 2021-09-03 15:31:11.447+02
host_id| cc241ec7-64fc-4c93-8cec-9e0e7005a49d

engine=#

see below my doubts...

On Sun, 5 Sept 2021 at 18:41, Nir Soffer  wrote:
>
>> On Sat, Sep 4, 2021 at 1:08 AM Gianluca Cecchi
>>  wrote:
>> ...
>> >>> ovirt_imageio._internal.nbd.ReplyError: Writing to file failed:
>> [Error 28] No space left on device
>> >> This error is expected if you don't have space to write the data.
>> > ok.
>>
>> I forgot to mention that running backup on engine host is not recommended.
>> It is better to run the backup on the hypervisor, speeding up the data
>> copy.
>>
>
OK, I will take care of it, thanks.

>>> How can I clean the situation?
>> >>
>> >> 1. Stop the current backup
>>
>>

> >> If stopping the backup failed, stopping the VM will stop the backup.
>>
>
OK, I will try to fix it with the VM running if possible, before going and
stopping it.


>> > But if I try the stop command I get the error
>> >
>> > [g.cecchi@ovmgr1 ~]$ python3
>> /usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py -c ovmgr1
>> stop dc386237-1e98-40c8-9d3d-45658163d1e2
>> 68f83141-9d03-4cb0-84d4-e71fdd8753bb
>> > [   0.0 ] Finalizing backup '68f83141-9d03-4cb0-84d4-e71fdd8753bb'
>> > Traceback (most recent call last):
>> ...
>> > ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is
>> "[Cannot stop VM backup. The VM backup is not in READY phase, backup phase
>> is FINALIZING. Please try again when the backup is in READY phase.]". HTTP
>> response code is 409.
>>
>> So your backup was already finalized, and it is stuck in "finalizing"
>> phase.
>>
>> Usually this means the backup on libvirt side was already stopped, but
>> engine
>> failed to detect this and failed to complete the finalize step
>> (ovirt-engine bug).
>>
>> You need to ensure if the backup was stopped on vdsm side.
>>
>> - If the vm was stopped, the bacukp is not running
>> - If the vm is running, we can make sure the backup is stopped using
>>
>> vdsm-client VM stop_backup
>> vmID=dc386237-1e98-40c8-9d3d-45658163d1e2
>> backup_id=68f83141-9d03-4cb0-84d4-e71fdd8753bb
>>
>
The VM is still running.
The host (I see it in its events with relation to backup errors) is ov200.
BTW: how can I see the mapping between host id and hostname (from the db
and/or api)?

[root@ov200 ~]# vdsm-client VM stop_backup
vmID=dc386237-1e98-40c8-9d3d-45658163d1e2
backup_id=68f83141-9d03-4cb0-84d4-e71fdd8753bb
{
"code": 0,
"message": "Done"
}
[root@ov200 ~]#


>> If this succeeds, the backup is not running on vdsm side.
>>
>
I preseum from the output above that the command succeeded, correct?

If this fails, you may need stop the VM to end the backup.
>>
>> If the backup was stopped, you may need to delete the scratch disks
>> used in this backup.
>> You can find the scratch disks ids in engine logs, and delete them
>> from engine UI.
>>
>
Any insight for finding the scratch disks ids in engine.log?
See here my engine.log and timestamp of backup (as seen in database above)
is 15:31 on 03 September:

https://drive.google.com/file/d/1Ao1CIA2wlFCqMMKeXbxKXrWZXUrnJN2h/view?usp=sharing


>> Finally, after you cleaned up vdsm side, you can delete the backup
>> from engine database,
>> and unlock the disks.
>>
>> Pavel, can you provide instructions on how to clean up engine db after
>> stuck backup?
>>
>
> Can you please try manually updating the 'phase" of the problematic
> backup entry in the "vm_backups" DB table to 1 of the final phases, which
> are either "Succeeded" or "Failed"?
> This should allow creating a new backup.
> [image: image.png]
>
>
>>
>> After vdsm and engine were cleaned, new backup should work normally.
>>
>
OK, so I wait f

[ovirt-users] Re: how to remove a failed backup operation

2021-09-03 Thread Gianluca Cecchi
On Fri, Sep 3, 2021 at 9:35 PM Nir Soffer  wrote:

> On Fri, Sep 3, 2021 at 4:45 PM Gianluca Cecchi 
> wrote:
>
>> Hello,
>> I was trying incremental backup with the provided
>> /usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py and began
>> using the "full" option.
>> But I specified an incorrect dir and during backup I got error due to
>> filesystem full
>>
>> [ 156.7 ] Creating image transfer for disk
>> '33b0f6fb-a855-465d-a628-5fce9b64496a'
>>
>> [snip]

> ovirt_imageio._internal.nbd.ReplyError: Writing to file failed: [Error 28]
>> No space left on device
>>
>
> This error is expected if you don't have space to write the data.
>

ok.


>
>>
>> Now if I try the same backup command (so with "full" option) and I get
>>
>> ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is
>> "[Cannot backup VM. The VM is during a backup operation.]". HTTP response
>> code is 409.
>>
>
> This looks like a bug in the backup script - the backup should be finalized
> even if the image transfer failed, but the error you get say the vm is
> still
> in backup mode.
>
>
>>
>> How can I clean the situation?
>>
>
> 1. Stop the current backup
>
> If you still have the output from the command, we log the backup UUID.
>
> If you lost the backup id, you can get it using the API - visit this
> address in your browser:
>
> https://myengine/ovirt-engine/api/vms/{vm-id}/backups/
>
> Then stop the current backup using:
>
> /usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py stop
> vm-id backup-id
>
> If stopping the backup failed, stopping the VM will stop the backup.
> I hope you are running recent enough version, since in early versions there
> was a bug when you cannot stop the vm during a backup.
>

It is the latest 4.4.7. I run the backup_vm.py script from the engine:

ovirt-engine-4.4.7.7-1.el8.noarch
ovirt-engine-setup-plugin-imageio-4.4.7.7-1.el8.noarch
ovirt-imageio-common-2.2.0-1.el8.x86_64
ovirt-imageio-client-2.2.0-1.el8.x86_64
ovirt-imageio-daemon-2.2.0-1.el8.x86_64
python3-ovirt-engine-sdk4-4.4.13-1.el8.x86_64

But if I try the stop command I get the error

[g.cecchi@ovmgr1 ~]$ python3
/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py -c ovmgr1
stop dc386237-1e98-40c8-9d3d-45658163d1e2
68f83141-9d03-4cb0-84d4-e71fdd8753bb
[   0.0 ] Finalizing backup '68f83141-9d03-4cb0-84d4-e71fdd8753bb'
Traceback (most recent call last):
  File "/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py",
line 493, in 
main()
  File "/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py",
line 176, in main
args.command(args)
  File "/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py",
line 262, in cmd_stop
stop_backup(connection, args.backup_uuid, args)
  File "/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py",
line 345, in stop_backup
backup_service.finalize()
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py", line
33869, in finalize
return self._internal_action(action, 'finalize', None, headers, query,
wait)
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line 299,
in _internal_action
return future.wait() if wait else future
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line 55,
in wait
return self._code(response)
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line 296,
in callback
self._check_fault(response)
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line 132,
in _check_fault
self._raise_error(response, body)
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line 118,
in _raise_error
raise error
ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is
"[Cannot stop VM backup. The VM backup is not in READY phase, backup phase
is FINALIZING. Please try again when the backup is in READY phase.]". HTTP
response code is 409.
[g.cecchi@ovmgr1 ~]$



>
> 2. File a bug about this
>


Filed this one, hope its is correct; I chose ovirt-imageio as the product
and Client as the component:

https://bugzilla.redhat.com/show_bug.cgi?id=2001136

I put information also about the error received with the stop command



>
>
>>
>> BTW: the parameter to put into ovirt.conf is backup-dir or backup_dir or
>> what?
>>
>
> ovirt.conf do not include the backup dir, only details about engine.
> Adding backup-dir
> to ovirt.conf or to backup specific configuration sounds like a good idea.
>
> Nir
>

I agree

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/OESNSO7MWVWZR2MS374ATPGYQRM2AXC3/


[ovirt-users] how to remove a failed backup operation

2021-09-03 Thread Gianluca Cecchi
Hello,
I was trying incremental backup with the provided
/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py and began
using the "full" option.
But I specified an incorrect dir and during backup I got error due to
filesystem full

[ 156.7 ] Creating image transfer for disk
'33b0f6fb-a855-465d-a628-5fce9b64496a'
[ 157.8 ] Image transfer 'ccc386d3-9f9d-4727-832a-56d355d60a95' is ready
--- Logging error ---, 105.02 seconds, 147.48 MiB/s

Traceback (most recent call last):
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/io.py",
line 242, in _run
handler.copy(req)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/io.py",
line 286, in copy
self._src.write_to(self._dst, req.length, self._buf)
  File
"/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/backends/http.py",
line 216, in write_to
writer.write(view[:n])
  File
"/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/backends/nbd.py",
line 118, in write
self._client.write(self._position, buf)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py",
line 445, in write
self._recv_reply(cmd)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py",
line 980, in _recv_reply
if self._recv_reply_chunk(cmd):
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py",
line 1031, in _recv_reply_chunk
self._handle_error_chunk(length, flags)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py",
line 1144, in _handle_error_chunk
raise ReplyError(code, message)
ovirt_imageio._internal.nbd.ReplyError: Writing to file failed: [Error 28]
No space left on device

Now if I try the same backup command (so with "full" option) and I get

ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is
"[Cannot backup VM. The VM is during a backup operation.]". HTTP response
code is 409.

How can I clean the situation?

BTW: the parameter to put into ovirt.conf is backup-dir or backup_dir or
what?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/6OZ7ZNH5GSNFCHDSDOPBNVXMN7WLWUXC/


[ovirt-users] Re: problems testing 4.3.10 to 4.4.8 upgrade SHE

2021-09-01 Thread Gianluca Cecchi
On Wed, Sep 1, 2021 at 4:00 PM Yedidyah Bar David  wrote:

>
> >
> > So I think there was something wrong with my system or probably a
> regression on this in 4.4.8.
> >
> > I see these lines in ansible steps of deploy of RHV 4.3 -> 4.4
> >
> > [ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Remove host used to
> redeploy]
> > [ INFO  ] changed: [localhost -> 192.168.222.170]
> >
> > possibly this step should remove the host that I'm reinstalling...?
>
> It should. From the DB, before adding it again. Matches on the uuid
> (search the code for unique_id_out if you want the details). Why?
>
> (I didn't follow all this thread, ignoring the rest for now...)
>
> Best regards,
>
>
>
It was the step I suspect there was a regression for in 4.4.8 (comparing
with 4.4.7) when updating the first hosted-engine host during the upgrade
flow and retaining its hostname details.
I'm going to test with latest async 2 4.4.8 and see if it solves the
problem. Otherwise I'm going to open a bugzilla sending the logs.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/P2CJRLA7INKW2RIH7HRRPUKIFSJ3NH7J/


[ovirt-users] Re: problems testing 4.3.10 to 4.4.8 upgrade SHE

2021-08-29 Thread Gianluca Cecchi
On Fri, Aug 27, 2021 at 7:57 PM Gianluca Cecchi 
wrote:

>
>
> Next step will be to try removing one of the two hosts while still all env
> in 4.3.10 and then take backup of engine and then install the second as
> 4.4.8 and see if it goes ok.
> I'm going to revert the 4.3.10 snapshot consistent env and try...
>
>
Actually this step is the same as what performed (new host, because I
pre-remove the existing one...).
And in fact I remember at the beginning of July I made a similar test on
the same test env with ovirt 4.4.7 async 2 node iso and didn't have this
kind of problem.

In the weekend I had to do similar steps with two different SHE
environments with RHV and was able to do as expected, without the
fingerprint error and using the same hostname for the first host I
redeployed in 4.4.
Passed from latest RHV 4.3 to latest 4.4, that currently is iso
4.4.7.4-0.20210804 and is based on 4.4.7 async 2 (Hypervisor Image for RHV
4.4.z batch#6 (oVirt-4.4.7-2) Async #1)

So I think there was something wrong with my system or probably a
regression on this in 4.4.8.

I see these lines in ansible steps of deploy of RHV 4.3 -> 4.4

[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Remove host used to
redeploy]
[ INFO  ] changed: [localhost -> 192.168.222.170]

possibly this step should remove the host that I'm reinstalling...?

I will redo the same again on ovirt and eventually open bugzilla with all
log files generated.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MAW5C3TS5QCBCJ4PA4BSQOFIZHBWXURA/


[ovirt-users] Re: problems testing 4.3.10 to 4.4.8 upgrade SHE

2021-08-27 Thread Gianluca Cecchi
On Fri, Aug 27, 2021 at 4:38 PM Gianluca Cecchi 
wrote:

> On Fri, Aug 27, 2021 at 4:10 PM Gianluca Cecchi 
> wrote:
>
>>
>>
>> no mention about ssh fingerprint reissue
>>
>> Any hint on how to do it?
>>
>> Gianluca
>>
>
> I found this link, related to 4.3 to 4.4 for RHV, that seems to somehow
> confirm the need of a "spare" host
> https://www.frangarcia.me/posts/notes-on-upgrading-rhv-43-to-rhv-44/
>
>
OK, so next step tried  has been (thanks Sandro for the input!):
. power down novirt2 where hosted-engine deploy was stuck (and related
still local vm), scratching it
. install the same host but with name novirt3.localdomain.local and
different ip
. run the
hosted-engine --deploy --restore-from-file=backup.bck
. now all goes ok and novirt3 has been added to engine and novirt1 results
as up, while novirt2 nonresponsive (it doesn't exist any more...)
all flow completes

[ INFO  ] changed: [localhost]
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Undefine local
storage-pool 84f8abd5-31ec-4c62-8130-521bb55c41e6]
[ INFO  ] changed: [localhost]
[ INFO  ] Generating answer file
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20210827182404.conf'
[ INFO  ] Generating answer file '/etc/ovirt-hosted-engine/answers.conf'
[ INFO  ] Stage: Pre-termination
[ INFO  ] Stage: Termination
[ INFO  ] Hosted Engine successfully deployed
[ INFO  ] Other hosted-engine hosts have to be reinstalled in order to
update their storage configuration. From the engine, host by host, please
set maintenance mode and then click on reinstall button ensuring you choose
DEPLOY in hosted engine tab.
[ INFO  ] Please note that the engine VM ssh keys have changed. Please
remove the engine VM entry in ssh known_hosts on your clients.
[root@novirt3 ~]#

Next step will be to try removing one of the two hosts while still all env
in 4.3.10 and then take backup of engine and then install the second as
4.4.8 and see if it goes ok.
I'm going to revert the 4.3.10 snapshot consistent env and try...

Problems so far before the next test:

. After the host deploy of the first 4.4 host (novirt3) I see that still
the current config results with novirt1 as the SPM and old hosted engine
storage as the master domain (I have iSCSI based SHE)
--> all as expected?

. the detach of the old engine storage gives at gui events
OVFs update was ignored - nothing to update for storage domain
'hosted_storage_old_20210827T173854'
Aug 27, 2021, 6:36:03 PM
Storage Domain hosted_storage_old_20210827T173854 (Data Center Default) was
deactivated and has moved to 'Preparing for maintenance' until it will no
longer be accessed by any Host of the Data Center.
8/27/216:36:11 PM
and the task seems completed, but continues to remain with the lock so I
cannot indeed deactivate (even after novirt1 put into maintenance to be
updated)

It didn't complain that the old hosted engine storage was master and
apparently switched another domain (DATA in my case) to master role I
don't know if related or not

. one not required network was not setup automatically on the new host (in
practice in this lab I only have ovirtmgmt and this network...), so when I
tried to live migrate VM of next host to be updated (novirt1) I got no
hosts available due to that
I went into setup host networks of novirt3 and added the network and then
all went fine: I was able to live migrate and install/add the next hosted
engine host (that is the last in my case, no ordinary hosts). Also on the
second host I had to go through the setup host networks button.

BTW: I was then able to update cluster and DC from 4.3 to 4.6 and
shutdown/boot the VM after that
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XKZ3MMFJYP5S2I3ACNKKB2OJKNNECW45/


[ovirt-users] Re: problems testing 4.3.10 to 4.4.8 upgrade SHE

2021-08-27 Thread Gianluca Cecchi
On Fri, Aug 27, 2021 at 4:10 PM Gianluca Cecchi 
wrote:

>
>
> no mention about ssh fingerprint reissue
>
> Any hint on how to do it?
>
> Gianluca
>

I found this link, related to 4.3 to 4.4 for RHV, that seems to somehow
confirm the need of a "spare" host
https://www.frangarcia.me/posts/notes-on-upgrading-rhv-43-to-rhv-44/

I don't know if the author reads the ML.
But this goes completely against what seems to be present inside the RHV
docs, where, as described in my previous post:
"
If you decide to use a new host, you must assign a unique name to the new
host and then add it to the existing cluster before you begin the upgrade
procedure.
"

And also the oVirt docs that are substantially based on RHV ones:
https://www.ovirt.org/documentation/upgrade_guide/index.html#SHE_Upgrading_from_4-3
"
When upgrading oVirt Engine, it is recommended that you use one of the
existing hosts. If you decide to use a new host, you must assign a unique
name to the new host and then add it to the existing cluster before you
begin the upgrade procedure.
. . .

It is recommended that you use one of the existing hosts. If you decide to
use a new host, you must assign a unique name to the new host and then add
it to the existing cluster before you begin the upgrade procedure.
"

So I would have in this case the same problem related to wrong fingerprint,
or I should be forced to copy fingerprint before scratching the host and
reuse it (if supported passing from node in version 7 to node in version 8
of the OS).
It seems strange this problem didn't arise before...

the phrase:
"
The upgraded host with the 4.4 self-hosted engine reports that HA mode is
active,...
"

lets think that the host name remains consistent with a pre-existing one
and as a reinstall is mandatory (7-->8) I don't see how it could work...
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XSZOJLM5BCDYPK57CEET5QA4TSYCLKTY/


[ovirt-users] Re: problems testing 4.3.10 to 4.4.8 upgrade SHE

2021-08-27 Thread Gianluca Cecchi
On Fri, Aug 27, 2021 at 3:25 PM Sandro Bonazzola 
wrote:

> Gianluca, after reinstalling the host with 4.4.8 ISO, did you update the
> ssh fingerprint of the fresh install within the ovirt engine? I'm assuming
> you didn't remove the host before reinstalling it and you didn't re-attach
> it to the engine after the upgrade.
>
>
No, I didn't do it... I would expect oVirt to manage it...
Can I do it now?
If I try to do I get:

Error while executing action: Cannot switch Host to Maintenance mode.
Host still has running VMs on it and is in Non Responsive state.

because the temporary engine is on it

It seems in 4.2 -> 4.3 update it was not necessary...

Also, for example on RHV guides I see:
https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.4/html/upgrade_guide/she_upgrading_from_4-3

"
When upgrading Red Hat Virtualization Manager, it is recommended that you
use one of the existing hosts. If you decide to use a new host, you must
assign a unique name to the new host and then add it to the existing
cluster before you begin the upgrade procedure.
"

and
https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.4/html/upgrade_guide/upgrading_the_manager_to_4-4_4-3_she
"
Install RHVH 4.4 or Red Hat Enterprise Linux 8.2 or later on the existing
node currently running the Manager virtual machine to use it as the
self-hosted engine deployment host. See Installing the Self-hosted Engine
Deployment Host for more information.
Note

It is recommended that you use one of the existing hosts. If you decide to
use a new host, you must assign a unique name to the new host and then add
it to the existing cluster before you begin the upgrade procedure.

. . .

-

The deployment script automatically disables global maintenance mode and
calls the HA agent to start the self-hosted engine virtual machine. The
upgraded host with the 4.4 self-hosted engine reports that HA mode is
active, but the other hosts report that global maintenance mode is still
enabled as they are still connected to the old self-hosted engine storage.
- Detach the storage domain that hosts the Manager 4.3 machine. For
details, see Detaching a Storage Domain from a Data Center

in the *Administration Guide*.
"

no mention about ssh fingerprint reissue

Any hint on how to do it?

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/N26KCVJGXW3QQXE7YYCGLXWFPVKRN5PG/


[ovirt-users] Re: problems testing 4.3.10 to 4.4.8 upgrade SHE

2021-08-27 Thread Gianluca Cecchi
On Wed, Aug 25, 2021 at 4:34 PM Gianluca Cecchi 
wrote:

> file /var/log/messages of novirt2
>
> https://drive.google.com/file/d/1hMcLeF3okJizLX4Gxj3jTG5bAPaAAFfK/view?usp=sharing
>
> Gianluca
>
>
Same problem with 4.4.8 async 1.

I'm deploying/restoring from novirt2 and the other host (still in 4.3.10)
is novirt1.

I arrive at

[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Check actual cluster
location]
[ INFO  ] skipping: [localhost]
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Enable GlusterFS at
cluster level]
[ INFO  ] skipping: [localhost]
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Set VLAN ID at datacenter
level]
[ INFO  ] skipping: [localhost]
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Get active list of active
firewalld zones]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Configure libvirt
firewalld zone]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Add host]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Include after_add_host
tasks files]
[ INFO  ] You can now connect to
https://novirt2.localdomain.local:6900/ovirt-engine/ and check the status
of this host and eventually remediate it, please continue only when the
host is listed as 'up'
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Create temporary lock
file]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Pause execution until
/tmp/ansible.5f702qq5_he_setup_lock is removed, delete it once ready to
proceed]

But then I'm able to connect to local engine web admin UI and novirt1
results up while novirt2 not responsive.

Every 3 seconds iinside engine.log I see these 3 lines

2021-08-27 11:05:54,065+02 INFO
 [org.ovirt.vdsm.jsonrpc.client.reactors.ReactorClient] (SSL Stomp Reactor)
[] Connecting to novirt2.localdomain.local/172.19.0.232
2021-08-27 11:05:54,067+02 ERROR
[org.ovirt.engine.core.vdsbroker.monitoring.HostMonitoring]
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-60) []
Unable to RefreshCapabilities: ConnectException: Connection refused
2021-08-27 11:05:54,068+02 ERROR
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetCapabilitiesAsyncVDSCommand]
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-60) []
Command 'GetCapabilitiesAsyncVDSCommand(HostName =
novirt2.localdomain.local,
VdsIdAndVdsVDSCommandParametersBase:{hostId='ca9ff6f7-5a7c-4168-9632-998c52f76cfa',
vds='Host[novirt2.localdomain.local,ca9ff6f7-5a7c-4168-9632-998c52f76cfa]'})'
execution failed: java.net.ConnectException: Connection refused


Can anyone telling what I can check?
Did you test SHE upgrade from 4.3.10 to 4.4.8 in your check flows?

BTW: I'm not using DNS but entries in /etc/hosts

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/TLPZINBZRS3TSVCFHH25VFUI3JJICKET/


[ovirt-users] Re: about the hugepage setting of the KVM server

2021-08-26 Thread Gianluca Cecchi
On Thu, Aug 26, 2021 at 9:50 AM Tommy Sway  wrote:

>
>
> Hi:
>
>
>
> I am running virtual machines on the KVM servers of oVirt env. The virtual
> machine is running  database, which needs to set the memory huge page on
> the OPERATING system of the virtual machine.
>
> If I need to set the huge page on the VIRTUAL machine, do I also need to
> set huge page on the KVM physical machine? Or just on the virtual machine?
>
>
>
> What best practices do you recommend?
>
>
>
> Thank you very much!
>

I think that this is still the reference for 4.4
https://mpolednik.github.io/2017/06/26/hugepages-and-ovirt/

The downside, if I'm correct, is that you will allocate as huge pages all
the memory of the virtual machine and not only what potentially needed.
Coming to your example of database VM: I want to allocate 32Gb as huge
pages in a VM because that would be the memory assigned to the DB instance.
But the VM for its work (user processes of the connections and other needs)
will be configured as a 64Gb ram VM.
In that case you "waste" 32Gb of huge pages allocation on the host, that
you would prefer to be instead as normal pages

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WSXXOE2NSFWWSDNQMRBHYBUZ6UJPBI5B/


[ovirt-users] Re: problems testing 4.3.10 to 4.4.8 upgrade SHE

2021-08-25 Thread Gianluca Cecchi
file /var/log/messages of novirt2
https://drive.google.com/file/d/1hMcLeF3okJizLX4Gxj3jTG5bAPaAAFfK/view?usp=sharing

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/266UGL6XDWJARJ6K6ZDAD4UM5G6VYPVM/


[ovirt-users] Re: problems testing 4.3.10 to 4.4.8 upgrade SHE

2021-08-25 Thread Gianluca Cecchi
On Wed, Aug 25, 2021 at 2:18 PM Gianluca Cecchi 
wrote:
[snip]

> I selected to pause an d I arrived here with local vm engine completing
>> its setup:
>>
>>  INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Add host]
>> [ INFO  ] changed: [localhost]
>> [ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Include after_add_host
>> tasks files]
>> [ INFO  ] You can now connect to
>> https://novirt2.localdomain.local:6900/ovirt-engine/ and check the
>> status of this host and eventually remediate it, please continue only when
>> the host is listed as 'up'
>> [ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks]
>> [ INFO  ] ok: [localhost]
>> [ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Create temporary lock
>> file]
>> [ INFO  ] changed: [localhost]
>> [ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Pause execution until
>> /tmp/ansible.4_o6a2wo_he_setup_lock is removed, delete it once ready to
>> proceed]
>>
>> But connecting t the provided
>> https://novirt2.localdomain.local:6900/ovirt-engine/ url
>> I see that only the still 4.3.10 host results up while novirt2 is not
>> resp[onsive
>>
>>
It is not clear the sense of the phrase above "check the status of this
host and eventually remediate it, please continue only when the host is
listed as 'up'"...
Does it refer to the novirt2 host (that is the first I'm installing while
nnovirt1 is still in 4.3.10 with a VM running), or novirt1?

Because if I go to the engine vm under /var/log/ovirt-engine I see:

 [root@novmgr ovirt-engine]# cd host-deploy/
[root@novmgr host-deploy]# ll
total 348
-rw-r--r--. 1 ovirt ovirt 354888 Aug 25 09:41
ovirt-host-mgmt-ansible-check-20210825094043-novirt1.localdomain.local.log
[root@novmgr host-deploy]#

So there is only the log file related to the deploy of novirt1 (that I see
as up) no log for novirt2.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3E2TSGSQUQQLEO66KJ7S6WS3ZFBP3V2A/


[ovirt-users] Re: problems testing 4.3.10 to 4.4.8 upgrade SHE

2021-08-25 Thread Gianluca Cecchi
On Wed, Aug 25, 2021 at 12:35 PM Gianluca Cecchi 
wrote:

> Hello,
> I'm testing what in object in a test env with novirt1 and novirt2 as hosts.
> First reinstalled host is novirt2
> For this I downloaded the 4.4.8 iso of the node:
>
> https://resources.ovirt.org/pub/ovirt-4.4/iso/ovirt-node-ng-installer/4.4.8-2021081816/el8/ovirt-node-ng-installer-4.4.8-2021081816.el8.iso
>
> before running the restore command for the first scratched node I
> pre-installed the appliance rpm on it and I got:
> ovirt-engine-appliance-4.4-20210818155544.1.el8.x86_64
>
> I selected to pause an d I arrived here with local vm engine completing
> its setup:
>
>  INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Add host]
> [ INFO  ] changed: [localhost]
> [ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Include after_add_host
> tasks files]
> [ INFO  ] You can now connect to
> https://novirt2.localdomain.local:6900/ovirt-engine/ and check the status
> of this host and eventually remediate it, please continue only when the
> host is listed as 'up'
> [ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks]
> [ INFO  ] ok: [localhost]
> [ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Create temporary lock
> file]
> [ INFO  ] changed: [localhost]
> [ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Pause execution until
> /tmp/ansible.4_o6a2wo_he_setup_lock is removed, delete it once ready to
> proceed]
>
> But connecting t the provided
> https://novirt2.localdomain.local:6900/ovirt-engine/ url
> I see that only the still 4.3.10 host results up while novirt2 is not
> resp[onsive
>
> vm situation:
>
> https://drive.google.com/file/d/1OwHHzK0owU2HWZqvHFaLLbHVvjnBhRRX/view?usp=sharing
>
> storage situation:
>
> https://drive.google.com/file/d/1D-rmlpGsKfRRmYx2avBk_EYCG7XWMXNq/view?usp=sharing
>
> hosts situation:
>
> https://drive.google.com/file/d/1yrmfYF6hJFzKaG54Xk0Rhe2kY-TIcUvA/view?usp=sharing
>
> In engine.log I see
>
> 2021-08-25 09:14:38,548+02 ERROR
> [org.ovirt.engine.core.vdsbroker.HostDevListByCapsVDSCommand]
> (EE-ManagedThreadFactory-engine-Thread-4) [5f4541ee] Command
> 'HostDevListByCapsVDSCommand(HostName = novirt2.localdomain.local,
> VdsIdAndVdsVDSCommandParametersBase:{hostId='ca9ff6f7-5a7c-4168-9632-998c52f76cfa',
> vds='Host[novirt2.localdomain.local,ca9ff6f7-5a7c-4168-9632-998c52f76cfa]'})'
> execution failed: java.net.ConnectException: Connection refused
>
> and continuouslly this message...
>
> I also tried to restart vdsmd on novit2 but nothing changed.
>
> Do I have to restart the HA daemons on novirt2?
>
> Any insight?
>
> Thanks
> Gianluca
>


it seems it has not been able to configure networks on
novirt2.localdomain.local, as I see no ovirtmgmt bridge...
During setup it asked for network card and I specified enp1s0 (default
proposed in square brackets was enp2s0)

172.19.0 is for mgmt network and ip of novirt2, 172.24.0 is for iscsi

[root@novirt2 ~]# ip a
1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group
default qlen 1000
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
   valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
   valid_lft forever preferred_lft forever
2: enp1s0:  mtu 1500 qdisc mq state UP
group default qlen 1000
link/ether 56:6f:bc:9a:00:5b brd ff:ff:ff:ff:ff:ff
inet 172.19.0.232/24 brd 172.19.0.255 scope global noprefixroute enp1s0
   valid_lft forever preferred_lft forever
inet6 fe80::546f:bcff:fe9a:5b/64 scope link noprefixroute
   valid_lft forever preferred_lft forever
3: enp2s0:  mtu 1500 qdisc mq state UP
group default qlen 1000
link/ether 56:6f:bc:9a:00:5c brd ff:ff:ff:ff:ff:ff
inet 172.24.0.232/24 brd 172.24.0.255 scope global noprefixroute enp2s0
   valid_lft forever preferred_lft forever
inet6 fe80::546f:bcff:fe9a:5c/64 scope link noprefixroute
   valid_lft forever preferred_lft forever
4: enp3s0:  mtu 1500 qdisc mq state UP
group default qlen 1000
link/ether 56:6f:bc:9a:00:5d brd ff:ff:ff:ff:ff:ff
6: virbr0:  mtu 1500 qdisc noqueue state
UP group default qlen 1000
link/ether 52:54:00:8b:b3:3a brd ff:ff:ff:ff:ff:ff
inet 192.168.222.1/24 brd 192.168.222.255 scope global virbr0
   valid_lft forever preferred_lft forever
7: vnet0:  mtu 1500 qdisc noqueue master
virbr0 state UNKNOWN group default qlen 1000
link/ether fe:16:3e:78:35:42 brd ff:ff:ff:ff:ff:ff
inet6 fe80::fc16:3eff:fe78:3542/64 scope link
   valid_lft forever preferred_lft forever
[root@novirt2 ~]#

[root@novirt2 network-scripts]# ll
total 12
-rw-r--r--. 1 root root 368 Aug 25 00:43 ifcfg-enp1s0
-rw-r--r--. 1 root root 277 Aug 25 00:51 ifcfg-enp2s0
-rw-r--r--. 1 root root 247 Aug 25 00:43 ifcfg-enp3s0
[root@novirt2 network-scripts]#

the strange thing is that if I go t

[ovirt-users] fencing and (virtual) power button pression reaction

2021-08-07 Thread Gianluca Cecchi
Hello,
in RHCS we have the fencing concept for targets very similar to the oVirt
ones: avoid data corruption and also fast react to problematic hosts
situations.
The implementation is quite similar to the oVirt one, with several fencing
agents sometimes common, like fence_ipmilan.
In RHCS documentation there is a chapter describing how to configure hosts
so that they don't react to power button pressure. This guarantees that
failover is as fast as possible, and also that the same host to be fenced,
could create more damages if it reacts and begins the shutdown procedure
instead of simply powering off.
See:
https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/8/html/configuring_and_managing_high_availability_clusters/assembly_configuring-fencing-configuring-and-managing-high-availability-clusters#proc_configuring-acpi-for-fence-devices-configuring-fencing

It seems to me that oVirt / RHV lack this feature.
Inside oVirt NGN and also RHVH-H, the /etc/systemd/logind.conf file is not
configured with an entry of type
HandlePowerKey=ignore

So in some cases of tests I'm doing, I see in its virtual console that the
to-be-fenced host begins its OS shutdown flow when it detects the power
button pressure. Typically after 2-3 seconds the system powers off and then
on again, but I also saw about 10 seconds delays in one case.

I have opened a case (number 03002278) for my RHV products, but I would
also like to get your comments here if I'm wrong with my considerations.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/TLOOLLF5GBMVXKKE25QCX4H57WJPNHON/


[ovirt-users] HA VM and vm leases usage with site failure

2021-08-05 Thread Gianluca Cecchi
Hello,
supposing latest 4.4.7 environment installed with an external engine and
two hosts, one in one site and one in another site.
For storage I have one FC storage domain.
I try to simulate a sort of "site failure scenario" to see what kind of HA
I should expect.

The 2 hosts have power mgmt configured through fence_ipmilan.

I have 2 VMs, one configured as HA with lease on storage (Resume Behavior:
kill) and one not marked as HA.

Initially host1 is SPM and it is the host that runs the two VMs.

Fencing of host1 from host2 initially works ok. I can test also from
command line:
# fence_ipmilan -a 10.10.193.152 -P -l my_fence_user -A password -L
operator -S /usr/local/bin/pwd.sh -o status
Status: ON

On host2 I then prevent reaching host1 iDRAC:
firewall-cmd --direct --add-rule ipv4 filter OUTPUT 0 -d 10.10.193.152 -p
udp --dport 623 -j DROP
firewall-cmd --direct --add-rule ipv4 filter OUTPUT 1 -j ACCEPT

so that:

# fence_ipmilan -a 10.10.193.152 -P -l my_fence_user -A password -L
operator -S /usr/local/bin/pwd.sh -o status
2021-08-05 15:06:07,254 ERROR: Failed: Unable to obtain correct plug status
or plug is not available

On host1 I generate panic:
# date ; echo 1 > /proc/sys/kernel/sysrq ; echo c > /proc/sysrq-trigger
Thu Aug  5 15:06:24 CEST 2021

host1 correctly completes its crash dump (kdump integration is enabled) and
reboots, but I stop it at grub prompt so that host1 is unreachable from
host2 point of view and also power fencing not determined

At this point I thought that VM lease functionality would have come in
place and host2 would be able to re-start the HA VM, as it is able to see
that the lease is not taken from the other host and so it can acquire the
lock itself
Instead it goes through the attempt to power fence loop
I wait about 25 minutes without any effect but continuous attempts.

After 2 minutes host2 correctly becomes SPM and VMs are marked as unknown

At a certain point after the failures in power fencing host1, I see the
event:

Failed to power fence host host1. Please check the host status and it's
power management settings, and then manually reboot it and click "Confirm
Host Has Been Rebooted"

If I select host and choose "Confirm Host Has Been Rebooted", then the two
VMs are marked as down and the HA one is correctly booted by host2.

But this requires my manual intervention.

Is the behavior above the expected one or the use of VM leases should have
allowed host2 to bypass fencing inability and start the HA VM with lease?
Otherwise I don't understand the reason to have the lease itself at all

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/FK254O4WOPWV56F753BVSK5GYQFZ4E5Q/


[ovirt-users] Re: Terrible Disk Performance on Windows 10 VM

2021-07-30 Thread Gianluca Cecchi
On Fri, Jul 30, 2021 at 4:17 PM  wrote:

>
> I think I'll deploy a new Windows guest and try the 'VirtIO-SCSI'
> interface and see if my performance is any better. It's just a default
> install of Windows at this point, so that'll be easy. :)
>
>
>
I think it should work also this way:
. hot add a disk to your Windows VM with size small as you want (eg 1Gb)
and configure it as virtio-scsi.
. in windows os when it asks for a driver, you provide what you need (the
virtio-win iso) and let windows configure
. eventually restart windows to verify all is ok
. shutdown the vm and set the boot disk as virtio-scsi
. power on the vm
. windows should auto configure it and perhaps reboot and you are then done
. remove the second disk

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/T4WMPVNIL7QQ54OITFZQ7KW6DOUBB5WR/


[ovirt-users] Change lease storage domain while VM running

2021-07-30 Thread Gianluca Cecchi
Hello,
on a 4.4.6 environment where I'm moving storage I want to change the lease
domain for an HA VM.
The VM is running and if I go and edit, change sd and then press ok button,
I get a window with:

"
Changes that can be applied immediately:

VM Lease
"

I don't select the "apply later" check box. But then instead when I
confirm, I get an error window with:

"
vm_name:

Cannot edit VM. Hot swapping of VM leases is not supported, please
remove the current lease and then set a new one.
"

Possibly it is possible to hot add a lease but not swap? So I think the
first window should be changed. Is this a known issue?
No problem if I change the lease storage domain while the VM is powered off.

I don't know if it is already fixed in 4.4.7... I'm going to check next
days..

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/2T75KL33773P3QVF7ZHITEPOZYEIKTUH/


[ovirt-users] Re: Requirements to put to cluster version 4.6

2021-07-27 Thread Gianluca Cecchi
On Tue, Jul 27, 2021 at 7:24 AM Yedidyah Bar David  wrote:

> On Mon, Jul 26, 2021 at 7:38 PM Gianluca Cecchi
>  wrote:
> >
> > On Wed, Jun 9, 2021 at 7:25 AM Yedidyah Bar David 
> wrote:
> >>
> >>
> >>
> >> To answer your direct question: RHV 4.4.6 is based on RHEL+AV 8.4, and
> >> the release dates of each part were coordinated internally for this.
> >>
> >> Best regards,
> >> --
> >> Didi
> >>
> >
> > Just to confirm that today I updated one of my DCs from oVirt 4.4.6 to
> the latest async of 4.4.7 and I was able to set Cluster and DC levels to
> 4.6.
>
> Thanks for the update!
>
> Best regards,
> --
> Didi
>
>
I forgot to better detail my configuration: external engine (installed on a
CentOS 8 system) and 3 hosts, all configured as plain CentOS 8.4 at the
moment

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/THKERCV5R6ZPPJE7KUFMQHN772E7CCN5/


[ovirt-users] Re: Requirements to put to cluster version 4.6

2021-07-26 Thread Gianluca Cecchi
On Wed, Jun 9, 2021 at 7:25 AM Yedidyah Bar David  wrote:

>
>
> To answer your direct question: RHV 4.4.6 is based on RHEL+AV 8.4, and
> the release dates of each part were coordinated internally for this.
>
> Best regards,
> --
> Didi
>
>
Just to confirm that today I updated one of my DCs from oVirt 4.4.6 to the
latest async of 4.4.7 and I was able to set Cluster and DC levels to 4.6.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KOJVIPRUUAFDY3YGF2G4CZC5UIAQCJ6G/


[ovirt-users] Re: Removing Direct Mapped LUNs

2021-07-16 Thread Gianluca Cecchi
On Fri, Jul 16, 2021 at 1:59 PM Vojtech Juranek  wrote:

> On Friday, 16 July 2021 12:31:34 CEST Gianluca Cecchi wrote:
> > On Fri, Jul 16, 2021 at 11:15 AM Vojtech Juranek 
> >
> > wrote:
> > > > What to do to crosscheck what is using the device and so preventing
> the
> > > > "-f" to complete?
> > >
> > > can you try
> > >
> > > dmsetup info  /dev/mapper/36090a0d800851c9d2195d5b837c9e328
> > >
> > > and check "Open count" filed to see if there is still anything open?
> > >
> > > Also, you can try
> > >
> > > fuser /dev/dm-2
> > >
> > > to see which process is using the device
> >
> > [root@ov301 ~]# dmsetup info
> /dev/mapper/36090a0d800851c9d2195d5b837c9e328
> > Name:  36090a0d800851c9d2195d5b837c9e328
> > State: ACTIVE
> > Read Ahead:256
> > Tables present:LIVE
> > Open count:1
>
> This means there's some open connection. As lsof or fuser doesn't show
> anything I wonder how this could happen.
>
> Theoretically (not tested as I actually don't know how to reproduce this)
> and
> on your own risk:-), you can try
>
> dmsetup suspend /dev/mapper/36090a0d800851c9d2195d5b837c9e328
> dmsetup clear /dev/mapper/36090a0d800851c9d2195d5b837c9e328
> dmsetup wipe_table /dev/mapper/36090a0d800851c9d2195d5b837c9e328
>
> which should remove any stale connection. After that dmsetup info should
> show
> Open count 0 and multipath -f 36090a0d800851c9d2195d5b837c9e328 should work
>
>
The host doesn't see the storage any more, and anyway it's a test system
where I try with oVirt, before going with oVirt itself or RHV in production.

[root@ov301 ~]# dmsetup suspend
/dev/mapper/36090a0d800851c9d2195d5b837c9e328
[root@ov301 ~]# dmsetup clear /dev/mapper/36090a0d800851c9d2195d5b837c9e328
[root@ov301 ~]#  dmsetup wipe_table
/dev/mapper/36090a0d800851c9d2195d5b837c9e328

But still
[root@ov301 ~]# dmsetup info  /dev/mapper/36090a0d800851c9d2195d5b837c9e328
Name:  36090a0d800851c9d2195d5b837c9e328
State: ACTIVE
Read Ahead:256
Tables present:LIVE
Open count:1
Event number:  0
Major, minor:  253, 2
Number of targets: 1
UUID: mpath-36090a0d800851c9d2195d5b837c9e328

Anyway the removal operation now goes ok:
[root@ov301 ~]# multipath -f 36090a0d800851c9d2195d5b837c9e328
[root@ov301 ~]# echo $?
0

and no multipath device in my output

[root@ov301 ~]# multipath -l
364817197c52f98316900666e8c2b0b2b dm-14 EQLOGIC,100E-00
size=2.0T features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
`-+- policy='round-robin 0' prio=0 status=active
  |- 16:0:0:0 sde 8:64 active undef running
  `- 17:0:0:0 sdf 8:80 active undef running
[root@ov301 ~]#

In /var/log/messages, during the sequence of commands above I see:

Jul 16 14:08:20 ov301 multipathd[1580]: 36090a0d800851c9d2195d5b837c9e328:
removing map by alias
Jul 16 14:08:20 ov301 multipath[2229532]: dm-2 is not a multipath map
Jul 16 14:09:03 ov301 multipathd[1580]: 36090a0d800851c9d2195d5b837c9e328:
remove map (operator)
Jul 16 14:09:03 ov301 multipathd[1580]: 36090a0d800851c9d2195d5b837c9e328:
devmap not registered, can't remove

Thanks for the moment...

I'm going to do similar storage moving and decommissioning of the old one
for 4 other storage domains (two of them iSCSI -> iSCSI, two of them iSCSI
-> FC) belonging to RHV environments (4.4.6 at the moment) in the next
weeks, so in case I'm going to open a case for them if I find the same
strange behavior.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/64ZAZZFKDYI37LXEGULJYOKP5RP6FUFZ/


[ovirt-users] Re: Removing Direct Mapped LUNs

2021-07-16 Thread Gianluca Cecchi
On Fri, Jul 16, 2021 at 11:15 AM Vojtech Juranek 
wrote:

>
> >
> > What to do to crosscheck what is using the device and so preventing the
> > "-f" to complete?
>
> can you try
>
> dmsetup info  /dev/mapper/36090a0d800851c9d2195d5b837c9e328
>
> and check "Open count" filed to see if there is still anything open?
>
> Also, you can try
>
> fuser /dev/dm-2
>
> to see which process is using the device
>
>
>
[root@ov301 ~]# dmsetup info  /dev/mapper/36090a0d800851c9d2195d5b837c9e328
Name:  36090a0d800851c9d2195d5b837c9e328
State: ACTIVE
Read Ahead:256
Tables present:LIVE
Open count:1
Event number:  0
Major, minor:  253, 2
Number of targets: 1
UUID: mpath-36090a0d800851c9d2195d5b837c9e328

[root@ov301 ~]# fuser /dev/dm-2
[root@ov301 ~]# echo $?
1
[root@ov301 ~]# ll /dev/dm-2
brw-rw. 1 root disk 253, 2 Jul 15 11:28 /dev/dm-2

I'm still unable to remove it:
[root@ov301 ~]# multipath -f 36090a0d800851c9d2195d5b837c9e328
Jul 16 12:25:11 | 36090a0d800851c9d2195d5b837c9e328: map in use
[root@ov301 ~]#
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SD67AU7Q3GX254V6JTOJBZNBZJTPY5O3/


[ovirt-users] Re: Removing Direct Mapped LUNs

2021-07-15 Thread Gianluca Cecchi
On Fri, Apr 23, 2021 at 7:15 PM Nir Soffer  wrote:

>
> >> > 1) Is this the expected behavior?
> >>
> >> yes, before removing multipath devices, you need to unzone LUN on
> storage
> >> server. As oVirt doesn't manage storage server in case of iSCSI, it has
> to be
> >> done by storage sever admin and therefore oVirt cannot manage whole
> flow.
> >>
> > Thank you for the information. Perhaps you can expand then on how the
> volumes are picked up once mapped from the Storage system?  Traditionally
> when mapping storage from an iSCSI or Fibre Channel storage we have to
> initiate a LIP or iSCSI login. How is it that oVirt doesn't need to do this?
> >
> >> > 2) Are we supposed to go to each KVM host and manually remove the
> >> > underlying multipath devices?
> >>
> >> oVirt provides ansible script for it:
> >>
> >> https://github.com/oVirt/ovirt-ansible-collection/blob/master/examples/
> >> remove_mpath_device.yml
> >>
> >> Usage is as follows:
> >>
> >> ansible-playbook --extra-vars "lun=" remove_mpath_device.yml
> >
>

I had to decommission one iSCSI based storage domain, after having added
one new iSCSI one (with another portal) and moved all the objects into the
new one (vm disks, template disks, iso disks, leases).
The Environment is based on 4.4.6, with 3 hosts, external engine.
So I tried the ansible playbook way to verify it.

Initial situation is this below; the storage domain to decommission is the
ovsd3750, based on the 5Tb LUN.

$ sudo multipath -l
364817197c52f98316900666e8c2b0b2b dm-13 EQLOGIC,100E-00
size=2.0T features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
`-+- policy='round-robin 0' prio=0 status=active
  |- 16:0:0:0 sde 8:64 active undef running
  `- 17:0:0:0 sdf 8:80 active undef running
36090a0d800851c9d2195d5b837c9e328 dm-2 EQLOGIC,100E-00
size=5.0T features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
`-+- policy='round-robin 0' prio=0 status=active
  |- 13:0:0:0 sdb 8:16 active undef running
  `- 14:0:0:0 sdc 8:32 active undef running

Connections are using iSCSI multipathing (iscsi1 and iscs2 in web admin
gui), so that I have two paths to each LUN:

$sudo  iscsiadm -m node
10.10.100.7:3260,1
iqn.2001-05.com.equallogic:0-8a0906-9d1c8500d-28e3c937b8d59521-ovsd3750
10.10.100.7:3260,1
iqn.2001-05.com.equallogic:0-8a0906-9d1c8500d-28e3c937b8d59521-ovsd3750
10.10.100.9:3260,1
iqn.2001-05.com.equallogic:4-771816-31982fc59-2b0b2b8c6e660069-ovsd3920
10.10.100.9:3260,1
iqn.2001-05.com.equallogic:4-771816-31982fc59-2b0b2b8c6e660069-ovsd3920

$ sudo iscsiadm -m session
tcp: [1] 10.10.100.7:3260,1
iqn.2001-05.com.equallogic:0-8a0906-9d1c8500d-28e3c937b8d59521-ovsd3750
(non-flash)
tcp: [2] 10.10.100.7:3260,1
iqn.2001-05.com.equallogic:0-8a0906-9d1c8500d-28e3c937b8d59521-ovsd3750
(non-flash)
tcp: [4] 10.10.100.9:3260,1
iqn.2001-05.com.equallogic:4-771816-31982fc59-2b0b2b8c6e660069-ovsd3920
(non-flash)
tcp: [5] 10.10.100.9:3260,1
iqn.2001-05.com.equallogic:4-771816-31982fc59-2b0b2b8c6e660069-ovsd3920
(non-flash)

One point not taken in consideration inside the previously opened bugs in
my opinion is the deletion of iSCSI connections and node at host side
(probably to be done by the os admin, but it could be taken in charge by
the ansible playbook...)
The bugs I'm referring are:
Bug 1310330 - [RFE] Provide a way to remove stale LUNs from hypervisors
Bug 1928041 - Stale DM links after block SD removal

Actions done:
put storage domain into maintenance
detach storage domain
remove storage domain
remove access from equallogic admin gui

I have a group named ovirt in ansible inventory composed by my 3 hosts:
ov200, ov300 and ov301
executed
$ ansible-playbook -b -l ovirt --extra-vars
"lun=36090a0d800851c9d2195d5b837c9e328" remove_mpath_device.yml

it went all ok with ov200 and ov300, but for ov301 I got

fatal: [ov301: FAILED! => {"changed": true, "cmd": "multipath -f
\"36090a0d800851c9d2195d5b837c9e328\"", "delta": "0:00:00.009003", "end":
"2021-07-15 11:17:37.340584", "msg": "non-zero return code", "rc": 1,
"start": "2021-07-15 11:17:37.331581", "stderr": "Jul 15 11:17:37 |
36090a0d800851c9d2195d5b837c9e328: map in use", "stderr_lines": ["Jul 15
11:17:37 | 36090a0d800851c9d2195d5b837c9e328: map in use"], "stdout": "",
"stdout_lines": []}

the complete output:

$ ansible-playbook -b -l ovirt --extra-vars
"lun=36090a0d800851c9d2195d5b837c9e328" remove_mpath_device.yml

PLAY [Cleanly remove unzoned storage devices (LUNs)]
*

TASK [Gathering Facts]
***
ok: [ov200]
ok: [ov300]
ok: [ov301]

TASK [Get underlying disks (paths) for a multipath device and turn them
into a list.] 
changed: [ov300]
changed: [ov200]
changed: [ov301]

TASK [Remove from multipath device.]
*
changed: [ov200]
changed: [ov300]
fatal: [ov301]: FAILED! => {"changed": true, "cmd": 

[ovirt-users] Re: Any way to terminate stuck export task

2021-07-06 Thread Gianluca Cecchi
On Tue, Jul 6, 2021 at 2:52 PM Nir Soffer  wrote:


>
> Too bad.
>
> You can evaluate how ovirt 4.4. will work with this appliance using
> this dd command:
>
> dd if=/dev/zero bs=8M count=38400 of=/path/to/new/disk
> oflag=direct conv=fsync
>
> We don't use dd for this, but the operation is the same on NFS < 4.2.
>
>
I confirm I'm able to saturate the 1Gb/s link. tried creating a 10Gb file
on the StoreOnce appliance
 # time dd if=/dev/zero bs=8M count=1280 of=/rhev/data-center/mnt/
172.16.1.137\:_nas_EXPORT-DOMAIN/ansible_ova/test.img oflag=direct
conv=fsync
1280+0 records in
1280+0 records out
10737418240 bytes (11 GB) copied, 98.0172 s, 110 MB/s

real 1m38.035s
user 0m0.003s
sys 0m2.366s

So are you saying that after upgrading to 4.4.6 (or just released 4.4.7) I
should be able to export with this speed? Or anyway I do need NFS v4.2?
BTW: is there any capping put in place by oVirt to the export phase (the
qemu-img command in practice)? Designed for example not to perturbate the
activity of hypervisor?Or do you think that if I have a 10Gb/s network
backend and powerful disks on oVirt and powerful NFS server processing
power  I should have much more speed?


> Based on the 50 MiB/s rate you reported earlier, I guess you have a
> 1Gbit network to
> this appliance, so zeroing can do up to 128 MiB/s, which will take
> about 40 minutes
> for 300G.
>
> Using NFS 4.2, fallocate will complete in less than a second.
>

I can sort of confirm this also for 4.3.10.
I have a test CentOS 7.4 VM configured as NFS server and, if I configure it
as an export domain using the default autonegotiate option, it is
(strangely enough) mounted as NFS v4.1 and the initial fallocate takes some
minutes (55Gb disk).
If I reconfigure it forcing NFS v4.2, it does it and the initial fallocate
is immediate, in the sense that "ls -l" on the export domain becomes quite
immediately the size of the virtual disk.

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YPUUKAKGVIL53JIW3EG5EOFXQJATJDUM/


[ovirt-users] Re: Any way to terminate stuck export task

2021-07-06 Thread Gianluca Cecchi
On Mon, Jul 5, 2021 at 5:06 PM Nir Soffer  wrote:


>
> qemu-img is busy in posix_fallocate(), wiring one byte to every 4k block.
>
> If you add -tt -T (as I suggested), we can see how much time each write
> takes,
> which may explain why this takes so much time.
>
> strace -f -p 14342 --tt -T
>
>
It seems I missed part of your suggestion... i didn't get the "-tt -T" (or
I didn't see it...)

With it I get this during the export (in networking of host console 4
mbit/s):

# strace -f -p 25243 -tt -T
strace: Process 25243 attached with 2 threads
[pid 25243] 09:17:32.503907 ppoll([{fd=9, events=POLLIN|POLLERR|POLLHUP}],
1, NULL, NULL, 8 
[pid 25244] 09:17:32.694207 pwrite64(12, "\0", 1, 3773509631) = 1 <0.59>
[pid 25244] 09:17:32.694412 pwrite64(12, "\0", 1, 3773513727) = 1 <0.78>
[pid 25244] 09:17:32.694608 pwrite64(12, "\0", 1, 3773517823) = 1 <0.56>
[pid 25244] 09:17:32.694729 pwrite64(12, "\0", 1, 3773521919) = 1 <0.24>
[pid 25244] 09:17:32.694796 pwrite64(12, "\0", 1, 3773526015) = 1 <0.20>
[pid 25244] 09:17:32.694855 pwrite64(12, "\0", 1, 3773530111) = 1 <0.15>
[pid 25244] 09:17:32.694908 pwrite64(12, "\0", 1, 3773534207) = 1 <0.14>
[pid 25244] 09:17:32.694950 pwrite64(12, "\0", 1, 3773538303) = 1 <0.16>
[pid 25244] 09:17:32.694993 pwrite64(12, "\0", 1, 3773542399) = 1 <0.200032>
[pid 25244] 09:17:32.895140 pwrite64(12, "\0", 1, 3773546495) = 1 <0.34>
[pid 25244] 09:17:32.895227 pwrite64(12, "\0", 1, 3773550591) = 1 <0.29>
[pid 25244] 09:17:32.895296 pwrite64(12, "\0", 1, 3773554687) = 1 <0.24>
[pid 25244] 09:17:32.895353 pwrite64(12, "\0", 1, 3773558783) = 1 <0.16>
[pid 25244] 09:17:32.895400 pwrite64(12, "\0", 1, 3773562879) = 1 <0.15>
[pid 25244] 09:17:32.895443 pwrite64(12, "\0", 1, 3773566975) = 1 <0.15>
[pid 25244] 09:17:32.895485 pwrite64(12, "\0", 1, 3773571071) = 1 <0.15>
[pid 25244] 09:17:32.895527 pwrite64(12, "\0", 1, 3773575167) = 1 <0.17>
[pid 25244] 09:17:32.895570 pwrite64(12, "\0", 1, 3773579263) = 1 <0.199493>
[pid 25244] 09:17:33.095147 pwrite64(12, "\0", 1, 3773583359) = 1 <0.31>
[pid 25244] 09:17:33.095262 pwrite64(12, "\0", 1, 3773587455) = 1 <0.61>
[pid 25244] 09:17:33.095378 pwrite64(12, "\0", 1, 3773591551) = 1 <0.27>
[pid 25244] 09:17:33.095445 pwrite64(12, "\0", 1, 3773595647) = 1 <0.21>
[pid 25244] 09:17:33.095498 pwrite64(12, "\0", 1, 3773599743) = 1 <0.16>
[pid 25244] 09:17:33.095542 pwrite64(12, "\0", 1, 3773603839) = 1 <0.14>
. . .

BTW: it seems my NAS appliance doesn't support 4.2 version of NFS, because
if I force it, I then get an error in mount and in engine.log this error
for both nodes as they try to mount:

2021-07-05 17:01:56,082+02 ERROR
[org.ovirt.engine.core.bll.storage.connection.FileStorageHelper]
(EE-ManagedThreadFactory-engine-Thread-2554190) [642eb6be] The connection
with details '172.16.1.137:/nas/EXPORT-DOMAIN' failed because of error code
'477' and error message is: problem while trying to mount target


and in vdsm.log:
MountError: (32, ';mount.nfs: Protocol not supported\n')

With NFSv3 I get apparently the same command:

vdsm 19702  3036  7 17:15 ?00:00:02 /usr/bin/qemu-img convert
-p -t none -T none -f raw
/rhev/data-center/mnt/blockSD/679c0725-75fb-4af7-bff1-7c447c5d789c/images/530b3e7f-4ce4-4051-9cac-1112f5f9e8b5/d2a89b5e-7d62-4695-96d8-b762ce52b379
-O raw -o preallocation=falloc /rhev/data-center/mnt/172.16.1.137:
_nas_EXPORT-DOMAIN/20433d5d-9d82-4079-9252-0e746ce54106/images/530b3e7f-4ce4-4051-9cac-1112f5f9e8b5/d2a89b5e-7d62-4695-96d8-b762ce52b379

The file size seems bigger but anyway very low throughput as with NFS v4.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QBPOJTMVBV6DXBYK4DDG3CX3SCJM54IZ/


[ovirt-users] Re: Any way to terminate stuck export task

2021-07-05 Thread Gianluca Cecchi
On Mon, Jul 5, 2021 at 2:13 PM Nir Soffer  wrote:

>
> >
> > vdsm 14342  3270  0 11:17 ?00:00:03 /usr/bin/qemu-img
> convert -p -t none -T none -f raw
> /rhev/data-center/mnt/blockSD/679c0725-75fb-4af7-bff1-7c447c5d789c/images/530b3e7f-4ce4-4051-9cac-1112f5f9e8b5/d2a89b5e-7d62-4695-96d8-b762ce52b379
> -O raw -o preallocation=falloc /rhev/data-center/mnt/172.16.1.137:
> _nas_EXPORT-DOMAIN/20433d5d-9d82-4079-9252-0e746ce54106/images/530b3e7f-4ce4-4051-9cac-1112f5f9e8b5/d2a89b5e-7d62-4695-96d8-b762ce52b379
>
> -o preallocation + NFS 4.0 + very slow NFS is your problem.
>
> qemu-img is using posix-fallocate() to preallocate the entire image at
> the start of the copy. With NFS 4.2
> this uses fallocate() linux specific syscall that allocates the space
> very efficiently in no time. With older
> NFS versions, this becomes a very slow loop, writing one byte for
> every 4k block.
>
> If you see -o preallocation, it means you are using an old vdsm
> version, we stopped using -o preallocation
> in 4.4.2, see https://bugzilla.redhat.com/1850267.
>

OK. As I said at the beginning the environment is latest 4.3
We are going to upgrade to 4.4 and we are making some complimentary
backups, for safeness.


> > On the hypervisor the ls commands quite hang, so from another hypervisor
> I see that the disk size seems to remain at 4Gb even if timestamp updates...
> >
> > # ll /rhev/data-center/mnt/172.16.1.137
> \:_nas_EXPORT-DOMAIN/20433d5d-9d82-4079-9252-0e746ce54106/images/530b3e7f-4ce4-4051-9cac-1112f5f9e8b5/
> > total 4260941
> > -rw-rw. 1 nobody nobody 4363202560 Jul  5 11:23
> d2a89b5e-7d62-4695-96d8-b762ce52b379
> > -rw-r--r--. 1 nobody nobody261 Jul  5 11:17
> d2a89b5e-7d62-4695-96d8-b762ce52b379.meta
> >
> > On host console I see a throughput of 4mbit/s...
> >
> > # strace -p 14342
>
> This shows only the main thread use -f use -f to show all threads.
>

 # strace -f -p 14342
strace: Process 14342 attached with 2 threads
[pid 14342] ppoll([{fd=9, events=POLLIN|POLLERR|POLLHUP}], 1, NULL, NULL, 8

[pid 14343] pwrite64(12, "\0", 1, 16474968063) = 1
[pid 14343] pwrite64(12, "\0", 1, 16474972159) = 1
[pid 14343] pwrite64(12, "\0", 1, 16474976255) = 1
[pid 14343] pwrite64(12, "\0", 1, 16474980351) = 1
[pid 14343] pwrite64(12, "\0", 1, 16474984447) = 1
[pid 14343] pwrite64(12, "\0", 1, 16474988543) = 1
[pid 14343] pwrite64(12, "\0", 1, 16474992639) = 1
[pid 14343] pwrite64(12, "\0", 1, 16474996735) = 1
[pid 14343] pwrite64(12, "\0", 1, 16475000831) = 1
[pid 14343] pwrite64(12, "\0", 1, 16475004927) = 1
. . . and so on . . .


>
> > This is a test oVirt env so I can wait and eventually test something...
> > Let me know your suggestions
>
> I would start by changing the NFS storage domain to version 4.2.
>

I'm going to try. RIght now I have set it to the default of
autonegotiated...


> 1. kill the hang qemu-img (it will probably cannot be killed, but worth
> trying)
> 2. deactivate the storage domain
> 3. fix the ownership on the storage domain (should be vdsm:kvm, not
> nobody:nobody)3.
>

Unfortunately it is an appliance. I have asked the guys that have it in
charge if we can set them.
Thanks for the other concepts explained.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XMHHYQO6IXDX37JOAMYEBTMFZESPFTN4/


[ovirt-users] Re: Any way to terminate stuck export task

2021-07-05 Thread Gianluca Cecchi
On Mon, Jul 5, 2021 at 11:56 AM Strahil Nikolov 
wrote:

> That NFS looks like it is not properly configured -> nobody:nobody is not
> suposed to be seen.
>
> Change the ownership from nfs side to 36:36. Also, you can define
> (all_squash,anonuid=36,anongid=36) as export options.
>
>
> Best Regards,
> Strahil Nikolov
>
>
I have those options in my test with a Linux box exporting via NFS. But
from the appliance point of view I have to check if it is possible... It is
not under my control and I don't know that appliance architecture. Anyone?

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LXHGYG535R3RENLMH7INRXFPBYSHHS7U/


[ovirt-users] Re: Any way to terminate stuck export task

2021-07-05 Thread Gianluca Cecchi
On Sun, Jul 4, 2021 at 1:01 PM Nir Soffer  wrote:

> On Sun, Jul 4, 2021 at 11:30 AM Strahil Nikolov 
> wrote:
> >
> > Isn't it better to strace it before killing qemu-img .
>
> It may be too late, but it may help to understand why this qemu-img
> run got stuck.
>
>
Hi, thanks for your answers and suggestions.
That env was a production one and so I was forced to power off the
hypervisor and power on it again (it was a maintenance window with all the
VMs powered down anyway). I was also unable to put the host into
maintenance because it replied that there were some tasks running, even
after the kill, because the 2 processes (the VM had 2 disks to export and
so two qemu-img processes) remained in defunct and after several minutes no
change in web admin feedback about the process

My first suspicion was something related to fw congestion because the
hypervisor network and the nas appliance were in different networks and I
wasn't sure if a fw was in place for it
But on a test oVirt environment with same oVirt version and with the same
network for hypervisors I was able to put a Linux server with the same
network as the nas and configure it as nfs server.
And the export went with a throughput of about 50MB/s, so no fw problem.
A VM with 55Gb disk exported in 19 minutes.

So I got the rights to mount the nas on the test env and mounted it as
export domain and now I have the same problems I can debug.
The same VM with only one disk (55Gb). The process:

vdsm 14342  3270  0 11:17 ?00:00:03 /usr/bin/qemu-img convert
-p -t none -T none -f raw
/rhev/data-center/mnt/blockSD/679c0725-75fb-4af7-bff1-7c447c5d789c/images/530b3e7f-4ce4-4051-9cac-1112f5f9e8b5/d2a89b5e-7d62-4695-96d8-b762ce52b379
-O raw -o preallocation=falloc /rhev/data-center/mnt/172.16.1.137:
_nas_EXPORT-DOMAIN/20433d5d-9d82-4079-9252-0e746ce54106/images/530b3e7f-4ce4-4051-9cac-1112f5f9e8b5/d2a89b5e-7d62-4695-96d8-b762ce52b379

On the hypervisor the ls commands quite hang, so from another hypervisor I
see that the disk size seems to remain at 4Gb even if timestamp updates...

# ll /rhev/data-center/mnt/172.16.1.137
\:_nas_EXPORT-DOMAIN/20433d5d-9d82-4079-9252-0e746ce54106/images/530b3e7f-4ce4-4051-9cac-1112f5f9e8b5/
total 4260941
-rw-rw. 1 nobody nobody 4363202560 Jul  5 11:23
d2a89b5e-7d62-4695-96d8-b762ce52b379
-rw-r--r--. 1 nobody nobody261 Jul  5 11:17
d2a89b5e-7d62-4695-96d8-b762ce52b379.meta

On host console I see a throughput of 4mbit/s...

# strace -p 14342
strace: Process 14342 attached
ppoll([{fd=9, events=POLLIN|POLLERR|POLLHUP}], 1, NULL, NULL, 8

# ll /proc/14342/fd
hangs...

# nfsstat -v
Client packet stats:
packetsudptcptcpconn
0  0  0  0

Client rpc stats:
calls  retransauthrefrsh
31171856   0  31186615

Client nfs v4:
null read writecommit   open open_conf

0 0% 2339179   7% 14872911 47% 7233  0% 74956 0% 2
0%
open_noatopen_dgrdclosesetattr  fsinfo   renew

2312347   7% 0 0% 2387293   7% 240% 230% 5
0%
setclntidconfirm  lock locktlockuaccess

3 0% 3 0% 8 0% 8 0% 5 0% 1342746
4%
getattr  lookup   lookup_root  remove   rename   link

3031001   9% 71551 0% 7 0% 74590 0% 6 0% 0
0%
symlink  create   pathconf statfs   readlink readdir

0 0% 9 0% 160% 4548231  14% 0 0% 98506
0%
server_caps  delegreturn  getacl   setacl   fs_locations
rel_lkowner
390% 140% 0 0% 0 0% 0 0% 0
0%
secinfo  exchange_id  create_ses   destroy_ses  sequence
get_lease_t
0 0% 0 0% 4 0% 2 0% 1 0% 0
0%
reclaim_comp layoutgetgetdevinfo   layoutcommit layoutreturn getdevlist

0 0% 2 0% 0 0% 0 0% 0 0% 0
0%
(null)
5 0%


# vmstat 3
procs ---memory-- ---swap-- -io -system--
--cpu-
 r  b   swpd   free   buff  cache   si   sobibo   in   cs us sy id
wa st
 3  1  0 82867112 437548 70665800054 100  0  0
100  0  0
 0  1  0 82867024 437548 706662000  1708 0 3720 8638  0  0
95  4  0
 4  1  0 82868728 437552 706661600   875 9 3004 8457  0  0
95  4  0
 0  1  0 82869600 437552 706663600  1785 6 2982 8359  0  0
95  4  0

I see the blocked process that is my qemu-img one...

In messages of hypervisor

Jul  5 11:33:06 node4 kernel: INFO: task qemu-img:14343 blocked for more
than 120 seconds.
Jul  5 11:33:06 node4 kernel: "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs" disables this message.
Jul  5 11:33:06 node4 kernel: qemu-imgD 9d960e7e1080 0
14343   3328 0x0080
Jul  5 11:33:06 node4 kernel: Call Trace:
Jul  5 11:33:06 node4 kernel: [] ?
sched_clock_cpu+0x85/0xc0

[ovirt-users] Any way to terminate stuck export task

2021-07-03 Thread Gianluca Cecchi
Hello,
in oVirt 4.3.10 an export job to export domain takes too long, probably due
to the NFS server slow.
How can I stop in a clean way the task?
I see the exported file remains always at 4,5Gb of size.
Command vmstat on host with qemu-img process gives no throughput but
blocked processes

procs ---memory-- ---swap-- -io -system--
--cpu-
 r  b   swpd   free   buff  cache   si   sobibo   in   cs us sy id
wa st
 1  2  0 170208752 474412 1698575200   71972 2948 5677  0
 0 96  4  0
 0  2  0 170207184 474412 1698578000  358099 5043 6790  0
 0 96  4  0
 0  2  0 170208800 474412 1698580400  137941 2332 5527  0
 0 96  4  0

and the generated file refreshes its timestamp but not the size

# ll -a  /rhev/data-center/mnt/172.16.1.137:
_nas_EXPORT-DOMAIN/20433d5d-9d82-4079-9252-0e746ce54106/images/125ad0f8-2672-468f-86a0-115a7be287f0/
total 4675651
drwxr-xr-x.  2 vdsm kvm   1024 Jul  3 14:10 .
drwxr-xr-x. 12 vdsm kvm   1024 Jul  3 14:10 ..
-rw-rw.  1 vdsm kvm 4787863552 Jul  3 14:33
bb94ae66-e574-432b-bf68-7497bb3ca9e6
-rw-r--r--.  1 vdsm kvm268 Jul  3 14:10
bb94ae66-e574-432b-bf68-7497bb3ca9e6.meta

# du -sh  /rhev/data-center/mnt/172.16.1.137:
_nas_EXPORT-DOMAIN/20433d5d-9d82-4079-9252-0e746ce54106/images/125ad0f8-2672-468f-86a0-115a7be287f0/
4.5G/rhev/data-center/mnt/172.16.1.137:
_nas_EXPORT-DOMAIN/20433d5d-9d82-4079-9252-0e746ce54106/images/125ad0f8-2672-468f-86a0-115a7be287f0/

The VM has two disks, 35Gb and 300GB, not full but quite occupied.

Can I simply kill the qemu-img processes on the chosen hypervisor (I
suppose the SPM one)?

Any way to track down why it is so slow?
Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LV6QCFVQO5F4WRMYFUBL7PNPIKG7Y53W/


[ovirt-users] Re: Strange Issue with imageio

2021-07-01 Thread Gianluca Cecchi
On Thu, May 27, 2021 at 7:43 AM Eyal Shenitzky  wrote:

> This bug is targeted to be fixed in 4.4.7 so 4.4.6 doesn't contain the fix.
>
>
But is there a workaround for this?
On a single host environment with external engine and local storage and
4.4.5 it seems that uploading an iso always gives OK without uploading
anything.
Both if selecting test connection or not...
Is it only related to the GUI or in generale even if I use the API?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/C5AD7OUCRSWG37QACRSY4434P5SRY5JM/


[ovirt-users] Expected behavior when link of required VM networks down

2021-06-24 Thread Gianluca Cecchi
Hello,
suppose I have a cluster with required VM logical network setup.
Suppose on hosts I have configured a bonding device with this logical
network mapped on it.
Suppose an active host for any reason loses the physical connectivity on
this bond (both links down or problems with new network config impacting
the bond, ecc)
So from the VMs point of view, those having a vnic on this logical network
will lose the connectivity, but I think they will not detect link down,
correct?
Is there any expected default action oVirt will take in respect of this
scenario or not?
Of course I expect to see events with messages about link down. But any
automatic action such as live migration of impacted VMs for example?
Any configurable one at oVirt level?
Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/X3OTQR32PTEWZKN5ZZ7AH5UPQKZ7NUQC/


[ovirt-users] Re: Update of plain CentOS hosts very slow

2021-06-09 Thread Gianluca Cecchi
On Wed, Jun 9, 2021 at 2:26 PM Dana Elfassy  wrote:

> You are correct. It was implemented this way taking in consideration the
> output when checking for updates on RHVH hosts.
> This implementation is indeed going to be changed, so that upgrade will
> become faster
>
> Dana
>
>
>
Ok, thanks for confirmation.
Indeed in case of ovirt-node-ng (or RHV-H for RHV) the update is only one
1Gb sized package containing the new image, but for plain CentOS Stream /
CentOS (or RH EL 8 hosts for RHV) it can contain a hundred of packages and
so a hundred of transactions vs one.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SWLTPRLG6ECVSV2GFJIZ7NOKR4GCO3WS/


[ovirt-users] Re: Update of plain CentOS hosts very slow

2021-06-09 Thread Gianluca Cecchi
On Wed, Jun 9, 2021 at 10:41 AM Dana Elfassy  wrote:

> Hi Gianluca,
> Upgrade host time depends on the upgrading that needs to be done
>

For sure. This is always true. In my case part of the update is updating
the host from 8.3 to 8.4, so many packages involved.
And my observation is that it seems as if the process updates all the
packages one by one as subsequent commands, with an unneeded long time in
my opinion.
Each time it seems it goes through transaction test and so on...
Very very suboptimal if it is true and confirmed.

>From an ansible point of view it seems that it happens something like this

- name: ensure a list of packages installed
  yum:
name: "{{ item }}"
   state: latest
  loop: "{{ packages }}"

instead of

- name: ensure a list of packages installed
  yum:
name: "{{ packages }}"
state: latest

Where {{ packages }} is a pre-computed (probably in a previous step) list
of packages to be updated.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JGBB6YZ7KYXU6DRREUZQX7RJTAD3V753/


[ovirt-users] Re: Requirements to put to cluster version 4.6

2021-06-08 Thread Gianluca Cecchi
On Tue, Jun 8, 2021 at 5:07 PM Martin Perina  wrote:

>
>
>> "
>> Have you the id of the bugs related to these new features provided in
>> 4.4.6?
>>
>
> Relevant bugs are linked in Blocks field in Bugzilla:
>
> *1669178*  - [RFE]
> Q35 SecureBoot - Add ability to preserve variable store certificates
> *1688177*  - [RFE]
> [SR-IOV] [Tech Preview RHV 4.4.6] Migration should not require downtime as
> of today
> *1821199*  - HP VM
> fails to migrate between identical hosts (the same cpu flags) not
> supporting TSC
> *1834250*  - CPU
> hotplug on UEFI VM causes VM reboot
> *1892800*  - NVDIMM:
> VM hangs in wait for launch for more than 7 minutes before VM is up
> *1906074*  - [RFE]
> Support disks copy between regular and managed block storage domains
> *1927718*  - [RFE]
> Provide Reset option for VMs
> *1936163*  - Enable
> bochs-display for UEFI guests by default
> *1936164*  - Enable
> KVM Software TPM by default
> *1950752*  - [RFE][CBT]
> redefine only the checkpoint that the backup is taken from and not the
> entire chain
>
>
>
OK


>> BTW [OT]: is RHV-H 4.4.6 host image ready for cluster level 4.6 or not
>> yet?
>>
>
> Yes, oVirt Node 4.4.6 is based CentOS Stream:
>
> *1907833*  - Rebase
> oVirt Node on CentOS Stream 8
>
>
I was referring to the downstream product with RHV-H 4.4.6... it is because
of this that I used the [OT]
I'm not sure if this 4.6 level is ready only upstream right now and needs
further improvements to go downstream to RHV, or if it is already there
production ready.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DULPJ3R4KCP7D62UQI5YXMOYAMEGDHNN/


[ovirt-users] Re: Requirements to put to cluster version 4.6

2021-06-08 Thread Gianluca Cecchi
On Tue, Jun 8, 2021 at 3:53 PM Martin Perina  wrote:

> Hi Gianluca,
>
> As mentioned in below RFE, EL 8.4 is not enough, you also need Advanced
> Virtualization 8.4 (libvirt >= 7.0.0) and AFAIK this has not been released
> upstream:
>
> https://bugzilla.redhat.com/show_bug.cgi?id=1933974
>
> So right now you can use 4.6 cluster level only on hosts running on CentOS
> Stream.
>
> Also the error message tells you that all your Up hosts need to provide EL
> + AV 8.4 features before you upgrade cluster level. But of course you can
> upgrade them one by one, it's not required to put all hosts into
> Maintenance before upgrade.
>
> Regards,
> Martin
>
>
OK, thanks for the note.
Inside Doc Text of the bugzilla above there is:
"
New features available in compatibility 4.6 are tracked as separate bugs
depending on this bug.
"
Have you the id of the bugs related to these new features provided in 4.4.6?

BTW [OT]: is RHV-H 4.4.6 host image ready for cluster level 4.6 or not yet?

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RDOATPUSJWAMGO6RB63S3DT37E4OZRLF/


[ovirt-users] Requirements to put to cluster version 4.6

2021-06-08 Thread Gianluca Cecchi
Hello,
I updated all my 3 CentOS 4.4.5 hosts to 4.4.6.
Or at least upgrading them from the GUI I see now they are on CentOS 8.4
and:
[root@ov200 ~]# rpm -qa| grep 4.4.6
ovirt-release44-4.4.6.3-1.el8.noarch
ovirt-host-4.4.6-1.el8.x86_64
ovirt-host-dependencies-4.4.6-1.el8.x86_64
[root@ov200 ~]#

How to see from the webadmin GUI or from the Host Console that they are
indeed 4.4.6?

If I try to set cluster compatibility version to 4.6 I get:

Error while executing action: Cannot change Cluster Compatibility Version
to higher version when there are active Hosts with lower version.
-Please move Host ov300, ov301, ov200 with lower version to maintenance
first.

I don't remember the need of giving downtime to get new cluster version...

Or what are further requirements (and benefits) to upgrade to 4.6?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MKF7J6M2DC2H5YBDIDH455WAYBRBPPV6/


[ovirt-users] Re: Update of plain CentOS hosts very slow

2021-06-08 Thread Gianluca Cecchi
On Tue, Jun 8, 2021 at 11:15 AM Gianluca Cecchi 
wrote:

> Hello,
> I have a 4.4.5 environment that I'm upgrading to 4.4.6.
>
> I'm upgrading plain CentOS hosts from the GUI.
> They are in 4.4.5, so in particular CentOS 8.3 and as part of the upgrade
> they have to be put to 8.4.
>
> In the past I used "yum update" on the host but now it seems it is not the
> correct way.
>
> But the ansible part related to package updates seems to be very slow.
> It gives the impression that it is doing it one by one and not as a whole
> when you run "yum update"
> Now it is about 30 minutes that the update is going on and my internet
> speed is for sure very high.
>
> In messages of host I see every single line suche this ones:
>
> Jun  8 11:09:30 ov300 python3[3031815]: ansible-dnf Invoked with
> name=['rsyslog-relp.x86_64'] state=latest lock_timeout=300
> conf_file=/tmp/yum.conf allow_downgrade=False autoremove=False bugfix=False
> disable_gpg_check=False disable_plugin=[] disablerepo=[]
> download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/
> install_repoquery=True install_weak_deps=True security=False
> skip_broken=False update_cache=False update_only=False validate_certs=True
> disable_excludes=None download_dir=None list=None releasever=None
> Jun  8 11:09:32 ov300 python3[3031828]: ansible-dnf Invoked with
> name=['runc.x86_64'] state=latest lock_timeout=300 conf_file=/tmp/yum.conf
> allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False
> disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[]
> enablerepo=[] exclude=[] installroot=/ install_repoquery=True
> install_weak_deps=True security=False skip_broken=False update_cache=False
> update_only=False validate_certs=True disable_excludes=None
> download_dir=None list=None releasever=None
>
> Any clarification?
>
> Thanks,
> Gianluca
>

BTW the update had a duration of 33 minutes (I selected to not reboot the
host):
Jun 8, 2021, 10:40:35 AM Host ov300 upgrade was started (User:
tekka@mydomain).
Jun 8, 2021, 11:13:13 AM Host ov300 upgrade was completed successfully.

At the end if I open a terminal I see:

[root@ov300 ~]# rpm -q ansible
ansible-2.9.16-2.el8.noarch
[root@ov300 ~]#

and

[root@ov300 ~]# yum update
Last metadata expiration check: 0:39:50 ago on Tue 08 Jun 2021 10:41:09 AM
CEST.
Dependencies resolved.
===
 Package Architecture   Version
RepositorySize
===
Upgrading:
 ansible noarch 2.9.21-2.el8
 ovirt-4.4-centos-ovirt44  17 M

Transaction Summary
===
Upgrade  1 Package

Total download size: 17 M
Is this ok [y/N]:
Operation aborted.
[root@ov300 ~]#

Is it expected?
Currently in my yum.conf I have:

[main]
gpgcheck=1
installonly_limit=3
clean_requirements_on_remove=True
best=True
skip_if_unavailable=False

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/4P6YORNJVZG67EQT5TTVGCKMGXL7VRY5/


[ovirt-users] Update of plain CentOS hosts very slow

2021-06-08 Thread Gianluca Cecchi
Hello,
I have a 4.4.5 environment that I'm upgrading to 4.4.6.

I'm upgrading plain CentOS hosts from the GUI.
They are in 4.4.5, so in particular CentOS 8.3 and as part of the upgrade
they have to be put to 8.4.

In the past I used "yum update" on the host but now it seems it is not the
correct way.

But the ansible part related to package updates seems to be very slow.
It gives the impression that it is doing it one by one and not as a whole
when you run "yum update"
Now it is about 30 minutes that the update is going on and my internet
speed is for sure very high.

In messages of host I see every single line suche this ones:

Jun  8 11:09:30 ov300 python3[3031815]: ansible-dnf Invoked with
name=['rsyslog-relp.x86_64'] state=latest lock_timeout=300
conf_file=/tmp/yum.conf allow_downgrade=False autoremove=False bugfix=False
disable_gpg_check=False disable_plugin=[] disablerepo=[]
download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/
install_repoquery=True install_weak_deps=True security=False
skip_broken=False update_cache=False update_only=False validate_certs=True
disable_excludes=None download_dir=None list=None releasever=None
Jun  8 11:09:32 ov300 python3[3031828]: ansible-dnf Invoked with
name=['runc.x86_64'] state=latest lock_timeout=300 conf_file=/tmp/yum.conf
allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False
disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[]
enablerepo=[] exclude=[] installroot=/ install_repoquery=True
install_weak_deps=True security=False skip_broken=False update_cache=False
update_only=False validate_certs=True disable_excludes=None
download_dir=None list=None releasever=None

Any clarification?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/CEDGFFPZU4KSZBUFIWIMPXOPO5E7HWHC/


[ovirt-users] Re: unable to login cockpit using root after upgrading to 4.4.6

2021-05-19 Thread Gianluca Cecchi
On Wed, May 19, 2021 at 9:05 AM Yaning Wang  wrote:

> add p...@redhat.com
>
> Sandro Bonazzola  于2021年5月19日周三 下午2:40写道:
>
>>
>>
>> Il giorno lun 17 mag 2021 alle ore 07:48 dhanaraj.ramesh--- via Users <
>> users@ovirt.org> ha scritto:
>>
>>> Hi Team,
>>>
>>> after upgrading ovirt node from 4.4.5 to 4.4.6, unale to login to the
>>> cockpit using root, but could able to login via ssh. please check whether
>>> this is considered as bug
>>>
>>
>> Can you please open a bugzilla ticket at
>> https://bugzilla.redhat.com/enter_bug.cgi?product=ovirt-node ?
>> Please attach a sos report to the bug, it will help understanding the
>> issue better.
>> +Chen Shao  , +Yaning Wang  , +Meital
>> Avital  never seen this happening on testing, can
>> you reproduce?
>>
>>
>>
I've just updated the engine of one of my environments from 4.4.5 to 4.4.6
and then one of the managed hosts that are plain CentOS 8 hosts.
And I can still login to the cockpit dashboard both as a normal user (id
1000 in this case) and as root
I don't have at the moment an environment with oVirt node or RHVH node to
test the same on them.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KBBQACP3R6NRGZ57ZVMANAAHSH2OAJHE/


[ovirt-users] Re: unable to login cockpit using root after upgrading to 4.4.6

2021-05-18 Thread Gianluca Cecchi
On Tue, May 18, 2021 at 4:50 PM Glenn Farmer 
wrote:

> The current thread is about 4.4.6 - nice that you can login to your 4.4.5.
>

The subject of the thread says it all... ;-)
My point was to ask if you see differences in /etc/pam.d/cockpit in your
4.4.6, in respect with the version I pasted for my 4.4.5 or if they are the
same.
I cannot compare as I have not yet 4.4.6 installed


> I changed the admin password on the engine - still cannot access the
> Cockpit GUI on any of my hosts.
>

The cockpit gui for the host is accessed through users defined on the
hosts, not on engine side. It is not related to the admin engine web admi
gui...
I think you can configure a normal user on your hypervisor host and see if
you can use it to connect to the cockpit gui or if you receive error.
Do you need any particular functionality to use the root user?

HIH,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VSM4BLBD36MFNXR5OXS4QWWHHGQXXZIP/


[ovirt-users] Re: unable to login cockpit using root after upgrading to 4.4.6

2021-05-18 Thread Gianluca Cecchi
On Tue, May 18, 2021 at 7:39 AM  wrote:

> Hello.
> I'm having the same issue with cockpit on the nodes. I'm unable to login
> as root or local user. I went from 4.4.5 to 4.4.6. It worked fine before
> the upgrade. I know the password is correct because I can log into the node
> via console and ssh. On one of the nodes I created a local account and have
> the same issue. The admin account works fine on the hosted engine VM.
>
>
I have not 4.4.6 yet, but could it be a change in /etc/pam.d/cockpit file?

On my 4.4.5 CentOS 8.3 based host, where I can connect as root in cockpit
host console, I currently have this:

#%PAM-1.0
# this MUST be first in the "auth" stack as it sets PAM_USER
# user_unknown is definitive, so die instead of ignore to avoid subsequent
modules mess up the error code
-auth  [success=done new_authtok_reqd=done user_unknown=die
default=ignore]   pam_cockpit_cert.so
auth   required pam_sepermit.so
auth   substack password-auth
auth   include  postlogin
auth   optional pam_ssh_add.so
accountrequired pam_nologin.so
accountinclude  password-auth
password   include  password-auth
# pam_selinux.so close should be the first session rule
sessionrequired pam_selinux.so close
sessionrequired pam_loginuid.so
# pam_selinux.so open should only be followed by sessions to be executed in
the user context
sessionrequired pam_selinux.so open env_params
sessionoptional pam_keyinit.so force revoke
sessionoptional pam_ssh_add.so
sessioninclude  password-auth
sessioninclude  postlogin

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/L3MWSV5IDQ4IZCJIDPZ2NWEVRENEAXMJ/


[ovirt-users] Re: Strange Issue with imageio

2021-05-17 Thread Gianluca Cecchi
On Sat, Apr 17, 2021 at 6:27 AM Nur Imam Febrianto 
wrote:

> Hi,
>
>
>
> Already submit *Bug 1950593*
>  for this issue.
>
> Thanks before.
>
>
>
> Regards,
>
> Nur Imam Febrianto
>
>
>
>
It seems I have the same problem with my 4.4.5.
Any info if it is fixed in the latest 4.4.6? It seems no update inside the
bug page..

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AOYURYKBUNWDOIABVNYMU6IVLAE4H3YD/


[ovirt-users] Re: Issue upgrading from 4.3 (Centos 7) to 4.4 (Centos 8)

2021-05-07 Thread Gianluca Cecchi
On Fri, May 7, 2021 at 2:04 PM Derek Atkins  wrote:

> HI,
>
> On Fri, May 7, 2021 2:55 am, Gianluca Cecchi wrote:
> > On Fri, May 7, 2021 at 2:39 AM Derek Atkins  wrote:
> > [snip]
> >
> >>
> >> *THIS* truly answers my underlying question(s).  I was concerned that
> >> the
> >> engine would be like the node, but this assures me it is not, it's just
> >> a
> >> "faster" way to build the engine VM, but it turns into (effectively) a
> >> standard CentOS+Engine VM post-install...  Which SHOULD be good enough
> >> for
> >> what I want to do!  (I have the engine running a few extra services, and
> >> wanted to make sure I could continue to do so).
> >>
> >>
> > OK.
> > But in the mid term the problem will impact you anyway.
> > What to do after 31/12/2021 with this system, as no new updates from
> > CentOS
> > repos?
>
> Umm, according to
>
> https://www.zdnet.com/article/red-hat-resets-centos-linux-and-users-are-angry/
> which links back to
> https://access.redhat.com/support/policy/updates/errata/#Life_Cycle_Dates
> which claims CentOS 7 will retain support through 2024-06-30.
>
> I think your 2021-12-31 date refers to EL8, not EL7.
>
>
yes, but as the thread was related to upgrade to 4.4 that needs version 8,
I thought you were speaking about it.

Even if version 7 of CentOS is supported through  2024, you are almost not
getting any oVirt based update since June last year when the latest version
of oVirt 4.3.x (4.3.10) has been released.
RH EL for its 7.x version updates doesn't necessarily track
incompatibilities with oVirt packages.
So in my opinion iit s also risky to remain with a frozen 4.3 oVirt
installation and to go with os overall updates for 3 years
I would identify the reasons not allowing to pass to 4.4 and solve them in
the meantime.

YMMV
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XNQ72M53EJVRR4X5GJW7VCJ5PRO3W255/


[ovirt-users] Re: Issue upgrading from 4.3 (Centos 7) to 4.4 (Centos 8)

2021-05-07 Thread Gianluca Cecchi
On Fri, May 7, 2021 at 2:39 AM Derek Atkins  wrote:
[snip]

>
> *THIS* truly answers my underlying question(s).  I was concerned that the
> engine would be like the node, but this assures me it is not, it's just a
> "faster" way to build the engine VM, but it turns into (effectively) a
> standard CentOS+Engine VM post-install...  Which SHOULD be good enough for
> what I want to do!  (I have the engine running a few extra services, and
> wanted to make sure I could continue to do so).
>
>
OK.
But in the mid term the problem will impact you anyway.
What to do after 31/12/2021 with this system, as no new updates from CentOS
repos?

I think in the next months it would be nice to have at least a CentOS -->
CentOS Stream migration path for the Hosted Engine VM or for an external
engine that are now based on CentOS
Or to confirm that the standard path described at
https://centos.org/centos-stream/ works also for CentOS systems where oVirt
repos are configured and active and engine packages installed.
Currently it says:
dnf swap centos-linux-repos centos-stream-repos
dnf distro-sync

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/IWTDAEYIKM277O2XJL5IA6CQGN5ZTFPK/


[ovirt-users] Re: oVirt 2021 Spring survey questions

2021-04-27 Thread Gianluca Cecchi
Desired storage options
Desired single server setup with included maintenance and updates made easy
Backup improvements desiderata

Thanks,
Gianluca

On Tue, Apr 27, 2021 at 10:16 AM Sandro Bonazzola 
wrote:

> Hi,
> it's about the usual time of the year when we ask the community to provide
> feedback with a survey.
> Any questions you'd like to be asked?
>
> --
>
> Sandro Bonazzola
>
> MANAGER, SOFTWARE ENGINEERING, EMEA R RHV
>
> Red Hat EMEA 
>
> sbona...@redhat.com
> 
>
> *Red Hat respects your work life balance. Therefore there is no need to
> answer this email out of your office hours.*
>
>
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/S25LWSV7WLARKMJOYVQVSRLXV7O4LVUF/
>
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/GAJZQRV65SGQ42T3ILQF2RWDI4ATMFCC/


[ovirt-users] Re: HCI - oVirt for CEPH

2021-04-26 Thread Gianluca Cecchi
On Mon, Apr 26, 2021 at 4:30 PM penguin pages 
wrote:

>
> It was on a support ticket / call I was having.  I googled around and the
> only article I found was the one about features being removed.. But not
> sure if this effects oVirt / HCI.
>
> My ticket was about trying to deploy OCP on a full SSD cluster of three
> nodes and disk performance over 10Gb will too slow and RH support was " We
> don't support use of gluster for OCP.. and need you to move off gluster for
> CEPH.
>
> So I opened another ticket about CEPH on HCI .. and was told "not
> supported.. CEPH nodes must be external"  So my three server small work
> office and demo stack, now is rethinking having to go to anther stack /
> vendor such as VMWare and vSAN, just because I can't get a stack that meets
> needs for small HCI stack with Linux.
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https:/ 
>

Remaining in enterprise and products/solutions supported by Red Hat, here
there are two different use cases for Red Hat Hyperconverged
infrastructure, see:
https://access.redhat.com/products/red-hat-hyperconverged-infrastructure

1) Red Hat Hyperconverged Infrastructure for Cloud
that is for Openstack and in that case Ceph (RHCS) is the only storage
solution supported.

2) Red Hat Hyperconverged Infrastructure for Virtualization
that is for RHV and in that case Gluster (RHGS) is the only storage
solution supported

Then there is the use case of OCP where you want to use persistent storage,
and again the only supported solution is Ceph (RHCS).
See
https://docs.openshift.com/container-platform/4.7/storage/persistent_storage/persistent-storage-ocs.html
https://access.redhat.com/articles/4731161

HIH clarifying,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UT7SDCSFF2TJ27CYL73MMYQN5TDZUHGP/


[ovirt-users] Re: HCI - oVirt for CEPH

2021-04-26 Thread Gianluca Cecchi
On Mon, Apr 26, 2021 at 2:34 PM penguin pages 
wrote:

>
> I have been building out HCI stack with KVM/RHEV + oVirt with the HCI
> deployment process.  This is very nice for small / remote site use cases,
> but with Gluster being anounced as EOL in 18 months, what is the
> replacement plan?
>

Are you referring to this:
https://access.redhat.com/support/policy/updates/rhhiv
?
If so, possibly in the meantime there will be a new release?
or what?

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LF64UVVZFDLXOTRJ5KS3O3DLQUXPFD5J/


[ovirt-users] Re: How do I share a disk across multiple VMs?

2021-04-24 Thread Gianluca Cecchi
On Sat, Apr 24, 2021 at 3:31 PM David White via Users 
wrote:

>
> Off topic, but something to address: We need a stable ovirt-guest-agent
> package. This doesn't seem to be working for me, although I'll take a look
> at it more closely again when I have some time:
> https://launchpad.net/ubuntu/focal/+source/ovirt-guest-agent
>
>
ovirt-guest-agent is deprecated in 4.4. See also here the downstream
documentation:
https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.4/html/release_notes/deprecated_features_rhv

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JWJJSLK24YY4XNJ7NIFKCAD7M7FKQO5H/


[ovirt-users] Re: How do I share a disk across multiple VMs?

2021-04-23 Thread Gianluca Cecchi
On Fri, Apr 23, 2021 at 9:24 AM Thomas Hoberg  wrote:

> Thank you Gianluca, for supporting my claim: it's patchwork and not "a
> solution designed for the entire enterprise".
>
> Instead it's more of "a set of assets where two major combinations from a
> myriad of potential permutations have received a bit of testing and might
> be useful somewhere in your enterprise".
>
> As such, I see very little future for oVirt as anything that doesn't
> achieve scale these days is doomed to die.
>

Actually the meaning of my sentence was opposite, in the sense that if you
consider it now in 4.4 a "patchwork", it was always so; you had been here
for many years and I think you should have already abandoned it in 4.3 (or
4.2) days, based on your considerations.
How it is composed didn't change so much with 4.4 release.
It's opensource and it's a project. Most of your claims could be done
against RHV product, not oVirt as a project.
And for sure many problems are there in Gluster implementations, but for
NFS, FC or iSCSI based the situation in my opinion is quite better.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/PLQKS7N4POODY7KKPQXSTGB65TDZAYCU/


[ovirt-users] Re: How to Upgrade Node with Local Storage ?

2021-04-22 Thread Gianluca Cecchi
On Thu, Apr 22, 2021 at 7:58 AM Vojtech Juranek  wrote:

> On Wednesday, 21 April 2021 16:40:29 CEST Nur Imam Febrianto wrote:
> > Set global maintenance and then turn off all vm, do yum update but it
> > completed with failed. Am I missing something ?
>
> can you share the details? What failed, wthat was the error?
>
>
>
I posted similar questions in the past about "yum update" not being
sufficient anymore to update a host, due to ansible managed tasks for this
now, directed form the engine itself.
So at the end in thi type of environments I decided to migrate to an
external engine that gives me less headaches to update.
I converted an HCI single host environment with gluster based storage
domains to another with external engine and the host, with "local on host"
ones, passing from xfs to ext4 and converting domains one by one, moving in
between the existing VMs.
I think there is no interest (because of not so much use case at the end
justifying the effort) in managing updates in SHE with only one host

It's a pity because at the beginning there was the all-in-one option, then
the SHE with HCI on one host and now nothing.
And if one wants to experiment / test / learn oVirt having only one server
it is not possible.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/Q37XUDIX2Z2KWMO47JCUEBWLJQOZBMUY/


[ovirt-users] Re: Moving templates and their thin based VMs

2021-04-21 Thread Gianluca Cecchi
On Wed, Apr 21, 2021 at 3:55 PM Chris Adams  wrote:

> Once upon a time, Gianluca Cecchi  said:
> > But some of these VMs had disk with storage allocation of type "Thin" on
> > their related template.
> > So I was both unable to move these templates and VMs disks...
>
> You can copy (rather than move) a template to the new storage domain,
> then any thin-provisioned disks that rely on the template can be moved.
>
> --
> Chris
>

Thank, Chris!

I didn't notice this.
So I can select in Storage --> DIsks the disk corresponding to a template,
then I see I can copy it to another storage domain.
And then the template continues to be one, but in Compute --> Templates
select template, DIsks I see now one disk with a "+" sign and expanding it
I see the two domains.
Great!

This means that if I do this I create a break in the dependency of the VM
from the particular storage domain, correct?
And what if later I go and edit the template, changing its properties? I
think they are only metadata that doesn't imply any modification of the
disk itself, that is now present in more than one domain, correct?

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/TC2SPXK3LJV6B5TFMXH57WUN4CPZGURS/


[ovirt-users] Moving templates and their thin based VMs

2021-04-21 Thread Gianluca Cecchi
Hello,
I had to free up and reformat a storage domain where I had some templates
and many VMs.
I was able to move to other existing storage domains disks of VMs created
with "Clone" Storage Allocation
But some of these VMs had disk with storage allocation of type "Thin" on
their related template.
So I was both unable to move these templates and VMs disks...

I was able to export as OVA the templates and then import again them (with
another name "_Copy") on to another storage domain then remove VMs and
templates and rename templates' names...

Are there any chances to better solve these kind of problems if one has to
decommission a storage domain? Only export of all VMs and related templates?

Is there a way to "consolidate" the disks of an existing "Thin" based
storage allocation to Clone?
Possibly I can clone the VM?

Thanks in advance,

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3P57FIIJ5255QG32DX56MHT66XY2XOIK/


[ovirt-users] Re: How do I share a disk across multiple VMs?

2021-04-21 Thread Gianluca Cecchi
On Wed, Apr 21, 2021 at 11:02 AM Thomas Hoberg  wrote:

> >
> > You're welcome to help with oVirt project design and discuss with the
> > community the parts that you think should benefit from a re-design.
>
> I consider these pesky little comments part of the discussion, even if I
> know they are not the best style.
>
> But how much is there to discuss, if Redhat has already decided to switch
> to a beta base (CentOS stream) underneath oVirt?
>
> Nobody wants bleeding edge on a hypervisor, except those who develop that
> hypervisor.
>
> oVirt is supposed to deliver a higher reliability than bare metal
> hardware, by providing a fault tolerant design and automatic fault recovery.
>
>
>
Only to point out that between core components of this type of hypervisor
are for sure libvirt and qemu-kvm and these two components were never the
ones provided OOTB by the downstream RHEL version.
Also vdsm for example, that is another core component, was never part of
the downstream OS.

In 4.2 deps.repo:
[ovirt-4.2-epel]
mirrorlist=
https://mirrors.fedoraproject.org/metalink?repo=epel-7=$basearch
[ovirt-4.2-centos-gluster312]
baseurl=http://mirror.centos.org/centos/7/storage/$basearch/gluster-3.12/
[ovirt-4.2-virtio-win-latest]
baseurl=http://fedorapeople.org/groups/virt/virtio-win/repo/latest
[ovirt-4.2-centos-qemu-ev]
baseurl=http://mirror.centos.org/centos/7/virt/$basearch/kvm-common/
[ovirt-4.2-centos-opstools]
baseurl=http://mirror.centos.org/centos/7/opstools/$basearch/
[centos-sclo-rh-release]
baseurl=http://mirror.centos.org/centos/7/sclo/$basearch/rh/
[ovirt-4.2-centos-ovirt42]
baseurl=http://mirror.centos.org/centos/7/virt/$basearch/ovirt-4.2/

In 4.3 deps.repo:
[ovirt-4.3-epel]
mirrorlist=
https://mirrors.fedoraproject.org/metalink?repo=epel-7=$basearch
[ovirt-4.3-centos-gluster6]
baseurl=http://mirror.centos.org/centos/7/storage/$basearch/gluster-6/
[ovirt-4.3-virtio-win-latest]
baseurl=http://fedorapeople.org/groups/virt/virtio-win/repo/latest
[ovirt-4.3-centos-qemu-ev]
baseurl=http://mirror.centos.org/centos/7/virt/$basearch/kvm-common/
[ovirt-4.3-centos-ovirt43]
baseurl=http://mirror.centos.org/centos/7/virt/$basearch/ovirt-4.3/
[ovirt-4.3-centos-ovirt-common]
baseurl=http://mirror.centos.org/centos/7/virt/$basearch/ovirt-common/
[ovirt-4.3-centos-opstools]
baseurl=http://mirror.centos.org/centos/7/opstools/$basearch/
[centos-sclo-rh-release]
baseurl=http://mirror.centos.org/centos/7/sclo/$basearch/rh/

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XFCSBFPCQS6XFLHLUF7SUYH5BI3NNLH5/


[ovirt-users] ipxe support with Q35 chipset and UEFI

2021-04-12 Thread Gianluca Cecchi
Hello,
I'm doing some tests with iPXE (downloaded latest version cloning the git
repo some days ago).
I'm using oVirt 4.4.5 with VMs configured with different chipset/firmware
type.

It seems that using dhcpd.conf directive of type

if exists user-class and option user-class = "iPXE" {
  filename "http://my_http_server/...;;
}
else {
...
}

the VM boot catches it when I use Q35 Chipset with BIOS, while it goes
inside the "else" section if using Q35 Chipset with UEFI (not the
SecureBoot one)
Does this mean that the Q35 UEFI doesn't support iPXE?

BTW: if anyone has suggestions about an utility that can let me boot via
network and give the user a general menu from which he/she can go and
choose standard pxe with bios or uefi boot, to be able to install both
Linux based systems and other ones, such as ESXi hosts, both with BIOS and
UEFI, it is welcome

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/EVGNR55AV6TILA4HB4DFHP53LKLIQXVK/


[ovirt-users] Re: ova import

2021-04-08 Thread Gianluca Cecchi
On Thu, Apr 8, 2021 at 10:49 AM Arik Hadas  wrote:

>
> On Thu, Apr 8, 2021 at 3:50 AM  wrote:
>
>> I am running Ovirt 4.3 and exported machines as ova which includes disks.
>> When I re import them. the disks are thin provisioned and the vm will not
>> boot.
>> Any idea how to get the import to  preallocate disks?  I have vms that
>> will not boot.
>>
> It states no bootable disk is available.
>> It seems the virt2vm isn't working correctly.
>> ANy help would be appreciated.
>>
>
> There's no option to change the format of the imported disks at the moment
> so they'd keep being thin-provisioned as they are within the OVA.
> I doubt that's the reason for the failure of those VMs to boot though -
> can you please share the engine.log that shows the configuration that a VM
> that didn't manage to boot started with?
>
>
>> Eric
>>
>
There were some threads in July 2020 about ova export zero size disks, and
so unusable import.
See these bugzillas for more information and also see suggestions about how
to put a single line fix to your 4.3 install as there was no backport but
solution only on 4.4
https://bugzilla.redhat.com/show_bug.cgi?id=1862115
https://bugzilla.redhat.com/show_bug.cgi?id=1813028

HIH,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AIJC5SA6XSW737KNSZVPTFFO7AHWFSHT/


[ovirt-users] Re: 4.4.5 released? Fails to upgrade

2021-03-18 Thread Gianluca Cecchi
On Thu, Mar 18, 2021 at 12:35 PM Rik Theys 
wrote:

> Hi,
>
> My systems pulled in 4.4.5 packages last night, so I assume oVirt 4.4.5
> was released? The release notes page does not list the release and I
> also did not see any announcement.
>
> The packages are 4.4.5.10-1.el8.
>
> I ran an engine-setup and upgraded to this release but the upgrade has
> failed due to a failure to update the database schema and it seems the
> rollback was not successful as my instance failed to start afterwards.
>
> I've downgraded all packages and ran engine-setup --offline, which seems
> to at least bring back my engine to a working state.
>
> I'm confused: has 4.4.5 been released or did I pull in some intermediate
> version with known issues?
>
> Regards,
>
> Rik
>

I see here the iso for the ng node in the standard location
https://resources.ovirt.org/pub/ovirt-4.4/iso/ovirt-node-ng-installer/4.4.5-2021031723/el8/
and also the engine appliance rpm here
https://resources.ovirt.org/pub/ovirt-4.4/rpm/el8/x86_64/
has name ovirt-engine-appliance-4.4-20210317223637.1.el8.x86_64.rpm
so it seems somehow released but no announce yet.
Possibly you got a not yet fully synced mirror and some problems related to
that?
You could retry when and if the announce has come and the download page at
the web site has been updated.
Or directly point to the resource.ovirt.org and not the mirrors

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SWJKTIK2TCXIIO6M6P5UV3CUG2GGM5TC/


[ovirt-users] doubt about CentOS 8.3 hypervisor and ansible version

2021-03-15 Thread Gianluca Cecchi
Hello,
I have a CentOS 8.3 + updates host in a 4.4.4 environment.
The web admin gui shows that it needs to upgrade to some packages.
The check event in web admin gui says:
Check for available updates on host ov300 was completed successfully with
message 'nss.x86_64, grub2-tools-efi.x86_64, bind-utils.x86_64,
dracut-network.x86_64, libvirt-daemon-driver-qemu.x86_64,
qemu-kvm-common.x86_64, libvirt-daemon-driver-storage-scsi.x86_64,
libvirt-daemon-driver-nodedev.x86_64, nss-sysinit.x86_64,
grub2-tools-minimal.x86_64 and 74 others. To see all packages check
engine.log.'.

If I go to engine in file ovirt-host-mgmt-ansible-check-.log I get:

  "yum_result" :
"bind-export-libs.x86_64\nbind-libs.x86_64\nbind-libs-lite.x86_64\nbind-lic
ense.noarch\nbind-utils.x86_64\nbpftool.x86_64\nbuildah.x86_64\ncockpit-podman.noarch\nconmon.x86_64
\ncontainer-selinux.noarch\ncontainernetworking-plugins.x86_64\ncontainers-common.x86_64\ncriu.x86_6
4\ndbxtool.x86_64\ndracut.x86_64\ndracut-config-rescue.x86_64\ndracut-network.x86_64\ndracut-squash.
x86_64\nfuse-overlayfs.x86_64\ngrub2-common.noarch\ngrub2-efi-x64.x86_64\ngrub2-tools.x86_64\ngrub2-
tools-extra.x86_64\ngrub2-tools-minimal.x86_64\nkernel.x86_64\nkernel-core.x86_64\nkernel-modules.x8
6_64\nkernel-tools.x86_64\nkernel-tools-libs.x86_64\nkmod-megaraid_sas.x86_64\nlibtpms.x86_64\nlibvi
rt-admin.x86_64\nlibvirt-bash-completion.x86_64\nlibvirt-client.x86_64\nlibvirt-daemon.x86_64\nlibvi
rt-daemon-config-network.x86_64\nlibvirt-daemon-config-nwfilter.x86_64\nlibvirt-daemon-driver-interf
ace.x86_64\nlibvirt-daemon-driver-network.x86_64\nlibvirt-daemon-driver-nodedev.x86_64\nlibvirt-daem
on-driver-nwfilter.x86_64\nlibvirt-daemon-driver-qemu.x86_64\nlibvirt-daemon-driver-secret.x86_64\nl
ibvirt-daemon-driver-storage.x86_64\nlibvirt-daemon-driver-storage-core.x86_64\nlibvirt-daemon-drive
r-storage-disk.x86_64\nlibvirt-daemon-driver-storage-gluster.x86_64\nlibvirt-daemon-driver-storage-i
scsi.x86_64\nlibvirt-daemon-driver-storage-iscsi-direct.x86_64\nlibvirt-daemon-driver-storage-logica
l.x86_64\nlibvirt-daemon-driver-storage-mpath.x86_64\nlibvirt-daemon-driver-storage-rbd.x86_64\nlibv
irt-daemon-driver-storage-scsi.x86_64\nlibvirt-daemon-kvm.x86_64\nlibvirt-libs.x86_64\nlibvirt-lock-
sanlock.x86_64\nmicrocode_ctl.x86_64\nnss.x86_64\nnss-softokn.x86_64\nnss-softokn-freebl.x86_64\nnss
-sysinit.x86_64\nnss-tools.x86_64\nnss-util.x86_64\npodman.x86_64\npodman-catatonit.x86_64\npython3-
bind.noarch\npython3-libvirt.x86_64\npython3-perf.x86_64\nqemu-img.x86_64\nqemu-kvm.x86_64\nqemu-kvm
-block-curl.x86_64\nqemu-kvm-block-gluster.x86_64\nqemu-kvm-block-iscsi.x86_64\nqemu-kvm-block-rbd.x
86_64\nqemu-kvm-block-ssh.x86_64\nqemu-kvm-common.x86_64\nqemu-kvm-core.x86_64\nrunc.x86_64\nslirp4n
etns.x86_64\nswtpm.x86_64\nswtpm-libs.x86_64\nswtpm-tools.x86_64\nvirt-v2v.x86_64\ngrub2-tools.x86_64\ngrub2-tools-efi.x86_64\ngrub2-tools-extra.x86_64\ngrub2-tools-minimal.x86_64\n"

I upgrade the host from the gui (without restart).
So far so good.
But now if I connect to the host I get:

[root@ov300 ~]# yum update
Last metadata expiration check: 0:05:42 ago on Mon 15 Mar 2021 09:36:31 AM
CET.
Dependencies resolved.

 Package   Architecture VersionRepository
   Size

Upgrading:
 ansible   noarch   2.9.18-2.el8
ovirt-4.4-centos-ovirt4417 M

Transaction Summary

Upgrade  1 Package

Total size: 17 M
Is this ok [y/N]:

Current version:

[root@ov300 ~]# rpm -q ansible
ansible-2.9.16-2.el8.noarch
[root@ov300 ~]#

Any suggestions on what is the right thing to do? Why the ansible package
was not updated and why it was not put into a sort of black list if it is
correct to do so?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MUCZYZ3GBDT4WN4YEJ3N42XI7WJYFVL7/


[ovirt-users] Re: login to self hosted endine during deployment

2021-03-14 Thread Gianluca Cecchi
Il Dom 14 Mar 2021, 08:41 Yedidyah Bar David  ha scritto:

> On Sun, Mar 14, 2021 at 8:01 AM  wrote:
> >
> > Is it possible to ssh or any other way connect to self hosted engine
> during deployment, since anssible is using ssh? I would like to check
> something?
> >
> > something like this:
> >
> > ssh -C -o ControlMaster=auto -o ControlPersist=60s -o User="root" -o
> ConnectTimeout=10 -o StrictHostKeyChecking=no -o
> UserKnownHostsFile=/dev/null -o ControlPath=/root/.ansible/cp/45a9ce675c
> -tt olvman.example.local /bin/sh -c
>
> An early phase of the deploy process creates a local VM, managed by
> libvirt directly (not vdsm).
>
> The IP address of this VM is written to the log.
>
> I use the following command to ssh to it from the host:
>

It seems to me that the temporary ip of the engine vm is also written
inside /etc/hosts of the hypervisor so I simply connected via ssh to this
ip without any particular option...

>
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HCWPX7XPFSEM5M7ZMBLR7KEKXXCRIUVM/


[ovirt-users] Re: Commvault

2021-03-10 Thread Gianluca Cecchi
On Wed, Mar 10, 2021 at 7:57 AM Tony Brian Albers  wrote:

> I agree with Dan, however EMC NetWorker can also backup RHEV.
>

Can you give a pointer about Networker capabilities?
I know Netbackup included RHV support since its 8.2 version.
Eg for 8.3 you have this "NetBackup™ Web UI RHV Administrator's Guide" and
other documents:
https://www.veritas.com/content/support/en_US/doc/138617403-138789763-0/v141695751-138789763

Do you have a similar pointer for Netbackup?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MSEA5XCFCZVTCLM3RSI2QT4A7JYI7IOY/


[ovirt-users] Re: 4.4.4 Image Copying / Template Create Fails - No Such File

2021-03-07 Thread Gianluca Cecchi
On Mon, Mar 8, 2021 at 8:15 AM Alex McWhirter  wrote:

> I apologize for the spam, this seems to be a long standing gluster issue.
> https://github.com/gluster/glusterfs/issues/597
>
> sharding does not support SEEK_DATA/SEEK_HOLE, with a preallocated image
> you likely never see this issue as there are no holes. However with a
> sparse image, that's very much not the case. I'm not sure when qemu-img
> changed to use these syscalls, as this is not something i experience on 4.3
> / CentOS 7.
>
> I'd be interested if anyone else can replicate this image copy behavior
> using raw sparse (thin provision) disks as the source on a gluster volume
> with sharding enabled, on oVirt 4.4.4+ (possibly earlier is also affected)
>
> If this is something current qemu-img cannot handle, i don't think
> supporting sparse disks on sharded gluster volumes is wise.
>
>
>
I only remember for sure that with 4.4.0 (and perhaps up to 4.4.2) with
single host HCI with Gluster I couldn't use sharding in a lab where I had
to create, destroy and recreate many times an Openshift environment
composed by many VMs and having thin disks was a must to improve times.
I sent several messages regarding this to the list but no active resolution
and was forced to disable sharding.

I also wrote to Gluster guys.

I can try to search and send the links.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JLRNJGTHQGLWIEI2VW7DADTLFTL73PGR/


[ovirt-users] Difference between edit cluster and upgrade?

2021-03-04 Thread Gianluca Cecchi
Hello,
supposing I upgrade my env from 4.3 to 4.4.
Then I can update my clusters' compatibility from 4.3 to 4.5.

I can do it in two ways:

1) select cluster line, "edit" button and change "Compatibility Version"
from 4.3 to 4.5
2) select cluster line, "upgrade" button

correct?

What is the difference?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/62RJFS2SSGQA2FHVN5OW6T6P7RJIHXM4/


[ovirt-users] SHE upgrade from 4.3 to 4.4 doubts

2021-03-03 Thread Gianluca Cecchi
Hello,
I would like to get some clarifications about the SHE environment upgrade
from 4.3.10 to 4.4.4.
Example scenario two hosts, both configured as HE hosts.
Interested both in iSCSI environment and in FC environment.
Going through here:
https://www.ovirt.org/documentation/upgrade_guide/#Upgrading_the_Manager_to_4-4_4-3_SHE
I have some doubts:

1) general one
between step 5 and 6 in section 4.4, actually the current 4.3 engine VM is
still running its OS (even if with ovirt-engine service stopped) and the VM
is running on the host you are going to scratch. So I think before
scratching the host and reinstalling it (step 6), it should be clarified
some step like
- shutdown engine VM
- shutdown host (is it ok in clean way or do I have to simply power off
it?)
- install from scratch ovirt-ng node 4.4.4 on host

2) general one
why in external engine upgrade guide it is not specified to stop
ovirt-engine service before executing the backup, while in SHE it is
specified?
See steps on section 2.3 here:
https://www.ovirt.org/documentation/upgrade_guide/#Upgrading_from_4-3
expected different behaviour? Or any change needed in the doc?

3) SHE when iSCSI and FC
It is clear that one new domain is to be provided, intended for the new
engine 4.4 VM
It is clear that during install of the new host in 4.4 it is safe to have
shared LUNs masked / not accessible so that you don't risk to corrupt
current 4.3 data

It is not clear in my opinion what to do before running the step 8 of 4.4
section:
hosted-engine --deploy --restore-from-file=backup.bck

a) iSCSI
in general the configured iSCSI sessions are activated only when the host
is being activated. So for example when you put host into maintenance, they
are logged off.
In this stage of upgrade what is expected I execute before running the
hosted-engine --deploy command?
Will the command log in into iSCSI sessions during itsmexecution (because
of the restore) or do I have to create / log in the connections before
running it?

b) FC
what should I have on host before running the command? Only the current
LUns (+ the new one intended for new hosted engine VM) visible, so that I
see all of them as output of "multipath -l" command and then the
hosted-engine --deploy command would mount the related filesystems and so
on..?

Possibly it would be nice to have an example output of the
hosted-engine --deploy --restore-from-file=backup.bck
command inside the doc page to see its expected workflow

Thanks for reading,

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZRKOY5BMLBSDERSB2MTDRU74CYKDD2TV/


[ovirt-users] Re: guest-agent on windows10 vm

2021-02-23 Thread Gianluca Cecchi
On Tue, Feb 23, 2021 at 9:40 PM Fedele Stabile 
wrote:

> Thank you for the answer Gianluca,
> the windows vm is new,
> other linux vm don't have warning messages on the guest-agent.
> So the problem about shared folder is related to a bug on the old release
> of oVirt, if I understand.
> Fedele
>
> 
>

Guest tools configuration has changed between versions..
Did you install ovirt-guest-tools on the VM from the iso available at your
engine vm package?
Did the related service start in WIndows 10 OS?

You didn't reply why you are forced to use so an "old" oVirt version: it is
not old in absolute, but oVirt is rapidly evolving and you should stay at
maximum behind one version.
So if the latest is 4.3, stay no behind 4.2. Now, with 4.4 out since many
months, stay no behind 4.3 at least...
This way you get more correct help because more guys are on latest
versions, that got many enhancements and many bug fixes (eg with snapshot
management, illegal disks state some time, and so on)

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MQZ2MNGDNJNEJPECD4OWLSF3S33AZFBP/


[ovirt-users] Re: guest-agent on windows10 vm

2021-02-23 Thread Gianluca Cecchi
On Tue, Feb 23, 2021 at 1:33 PM Fedele Stabile 
wrote:

> Hello all,
> I 'm using oVirt 4.2.1.7
> and on windows VM i see an orange exclamation point  that warns me to
> install the latest guest agent.
> On the VM (windows 10) i have installed:
> QEMU guest agent version 101.1.0
> SPICE Guest Tools 0.141
> Virtio-win-driver-installer version 0.1.185
> UsbDk Runtime Libraries version 1.0.22
> Spice webdavd 2.4 (64-bit)
>
> I can share the clipboard but
> can not mount a USB stick other than FAT-formatted  or use a shared folder
>
> Is there a solution?
>
>
oVirt as an upstream project cannot support every single release. Normally
only the current one is under best (effort) support
4.2.1.7 released about 3 years ago. Any reason not to update in the mean
time to a newer and better supported release?
4.3.0 has been released on February 2019 and is not supported any more,
even if newer than yours.
4.4.0 has been released on May 2020 and now arrived at 4.4.4

I think that orange exclamation is there for a quite a long time... or did
you update recently in any way from an older release and got it?

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BJLFU3RDXOHZRK2VZFCUO4HJGSNH3DIM/


[ovirt-users] Re: oVirt Engine increase RAM and CPU

2021-02-22 Thread Gianluca Cecchi
On Mon, Feb 22, 2021 at 3:22 PM Yedidyah Bar David  wrote:

> On Mon, Feb 22, 2021 at 4:10 PM Vrgotic, Marko
>
> >
> > @oVirt does anyone know how long it usually takes to have the change
> written to VM image?
>
> up to 60 minutes :-).
>
> I think you can change this with engine-config, item
> OvfUpdateIntervalInMinutes .
>
>
Isn't there the setting in web admin gui where you select a storage domain,
three dots at top right and then "Update OVFs"?
Shouldn't it force the update? I see it in 4.3.11, also for the hosted
engine storage domain

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/TC52QRHJWUAFR77EU2DYP5RQNDFP4VYG/


[ovirt-users] Q35: disk type view not consistent in edit vm and new disk pages

2021-02-19 Thread Gianluca Cecchi
Hello,
if I create a VM and select Q35 machine type, still the General --> create
image option in edit phase doesn't show sata as an option.
Even if I create the VM without disks and then go to edit General -->
create, I only see IDE, VirtIO-SCSI and VirtIO.
Instead if I click VM name, DIsks --> New, I can select VirtIO-SCSI, VirtIO
and SATA
After the first disk (SATA in my case) has been created, now if I go and
edit VM --> General --> Instance Images + --> Create, I see consistent
options.

I think GUI experience could be improved.
Also because I can reproduce this workflow converging to an error (even if
IDE disks are not so common nowadays...):
New VM
In  general I create an IDE disk
In System I select in Advanced parameters, Custom Chipset/Firmware Type Q35
(EFI or BIOS based)
go ahead

At the final OK I get the error:

Cannot add Virtual Disk: The disk interface is not supported by the VM OS:
Other OS.

The result is VM created but without the IDE disk as not supported in Q35
type.

Tested in 4.4.4
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WRRI7FD2RTSLBR5KHIX6JNVEPTCOFJED/


[ovirt-users] Testing PXE uefi boot in oVirt

2021-02-16 Thread Gianluca Cecchi
Hello,
I'm testing a configuration with a CentOS 8.3 VM as a PXE server for UEFI
boot.
Its chipset/fw type is "I440FX Chipset with BIOS", I don't think it is
important.

And I have another (tobe, configured in hw as rhel 8.0) CentOS 8 VM
configured now as "Q35 Chipset with UEFI" as a client.
I already removed filters from vnics of both

It seems that the client successfully downloads the BOOTX64.EFI but then
fails fetching the netboot image with the message:

Fetching Netboot Image
Unable to fetch TFTP image: TFTP Error

under tftp root dir I created grub.cfg this way

set timeout=60
menuentry 'CentOS 8.3' {
  linuxefi images/c83/vmlinuz ip=dhcp inst.repo=http://172.20.0.1/c83/
  initrdefi images/c83/initrd.img
}

and also a link to it with the mac address of the client in the form

grub.cfg-00:1a:4a:19:01:54 -> grub.cfg

documentation for uefi pxe server config is not quite clear on where to put
it and naming
Any good docs? Also the latest rh el 8.3 advanced installation guide is not
so clear...

See here for a screenshot:
https://drive.google.com/file/d/1YJwgmg-0vTcXYWHbgCf81m8SLsN7Rwjf/view?usp=sharing

For instance, is it expected that I can test this functionality at all
using oVirt VMs?
So the UEFI implementation for PXE boot?
At the moment I'm using virtio as the vnic type for both the server and the
client. Do I have to change them to something different (eg e1000)?

Thanks in advance for any input.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/C6MO7TCJOIBF7BXFXBUTASXETY4GAU5K/


[ovirt-users] Re: OVN network and external access

2021-02-16 Thread Gianluca Cecchi
On Tue, Feb 16, 2021 at 11:34 AM Ales Musil  wrote:

>
>
> On Tue, Feb 16, 2021 at 11:22 AM Gianluca Cecchi <
> gianluca.cec...@gmail.com> wrote:
>
>> Hello,
>> can I have external access from a VM with vnic configured on OVN network,
>> still maintaining switch type of cluster to Linux Bridge? Or am I forced to
>> use switch type OVS (that seems still in Tech Preview)?
>>
>
> Hi,
> there are some options you could do. One of them is using one VM as a
> router, being connected to linux bridge and OVN network
> at the same time.
>

Yes, this was one of the possible options, even if not the most desirable.
I would also like to disable dhcp on the OVN one. Is it possible now in
4.4.4 with a REST API call?


You could use OVN router, oVirt provider supports it to some extent, check
> [0].
>

This is not related to the consideration above where the routing is created
at OS level of the dedicated VM, right? The link you provided seems to
describe how to connect to routers (Software Defined ones, right?) but not
how to create them in OVN provided by oVirt


>
> And as for the OVS switch type, is there anything you are missing for the
> functionality?
>

Only that still in 4.4.4 is marked as tech preview if you try to create a
cluster based on it: "OVS (Technology Preview)"
BTW: if I have an already existing environment based on Linux Bridge  (with
external engine) is there a quick and not disruptive path to convert to
OVS?

The state of OVS in 4.4.4 is complete feature parity with Linux Bridge
> except STP (should not be a huge problem) and QoS.
> QoS on OVS should technically work but the only issue is that the engine
> will display it as out of sync because vdsm is not reporting back QoS on
> OVS.
>

So perhaps time to remove the "Technology Preview" phrasing?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/PE3K2PSLR5S2D7NHRF2TXZAYIE43NHVD/


[ovirt-users] OVN network and external access

2021-02-16 Thread Gianluca Cecchi
Hello,
can I have external access from a VM with vnic configured on OVN network,
still maintaining switch type of cluster to Linux Bridge? Or am I forced to
use switch type OVS (that seems still in Tech Preview)?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/2FJI2DZS5ECW3PMAEGAM6GP4HF5ETJ2O/


[ovirt-users] Re: export a VM to export domain in 4.4.4

2021-02-12 Thread Gianluca Cecchi
On Fri, Feb 12, 2021 at 5:24 PM Gianluca Cecchi 
wrote:

> Hello,
> in 4.4.4 I see that I can import from export domain, but if I select
> "export VM" I can only export to a storage domain is this by design?
> What if I want to export from 4.4.4 and import into an older oVirt version?
>
> Thanks,
> Gianluca
>

OK, solved.
Now I have an "Export" button in the tab pane at top that can only "export"
to a data domain (through cloning) and also three dots at top right where I
can then select "Export to Export Domain" and this is the old style way.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JDQ6L4XDQHTW4F5USEY4ACF4XBSFJRPQ/


[ovirt-users] export a VM to export domain in 4.4.4

2021-02-12 Thread Gianluca Cecchi
Hello,
in 4.4.4 I see that I can import from export domain, but if I select
"export VM" I can only export to a storage domain is this by design?
What if I want to export from 4.4.4 and import into an older oVirt version?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3PFLVMVISELEOOSFJDT2FSSONMT33Y4K/


[ovirt-users] Re: uploading ang .OVA image

2021-02-04 Thread Gianluca Cecchi
On Thu, Feb 4, 2021 at 2:12 PM Ariez Ahito  wrote:

> hi guys, i just want to ask if you can just upload an .OVA file to ovirt
> just like uploading ISO?
> without setting up virt-v2v and export domain?
>
> thanks
>
>
Put the OVA file on one of your hypervisors in a certain path
Go to Web Admin Gui --> Compute --> Virtual Machines
Select the three vertical dots at top right (near the "Migrate" button) and
"Import"
Select "Virtual Appliance (OVA)" as source and choose the host and File
Path where you previously put the OVA
select the VMs to import using the horizontal arrows

In theory it should be supported oVirt as internal format and only vSphere
as external format of OVA.
But as far as your OVA complies with what oVirt expects, it could work too.

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/IUA6QJGTZYHVAHIDUBEEGCTHCIR3CI2R/


  1   2   3   4   5   6   7   8   9   10   >