[ovirt-users] Re: how to remove a failed backup operation

2021-09-03 Thread Gianluca Cecchi
On Fri, Sep 3, 2021 at 9:35 PM Nir Soffer  wrote:

> On Fri, Sep 3, 2021 at 4:45 PM Gianluca Cecchi 
> wrote:
>
>> Hello,
>> I was trying incremental backup with the provided
>> /usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py and began
>> using the "full" option.
>> But I specified an incorrect dir and during backup I got error due to
>> filesystem full
>>
>> [ 156.7 ] Creating image transfer for disk
>> '33b0f6fb-a855-465d-a628-5fce9b64496a'
>>
>> [snip]

> ovirt_imageio._internal.nbd.ReplyError: Writing to file failed: [Error 28]
>> No space left on device
>>
>
> This error is expected if you don't have space to write the data.
>

ok.


>
>>
>> Now if I try the same backup command (so with "full" option) and I get
>>
>> ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is
>> "[Cannot backup VM. The VM is during a backup operation.]". HTTP response
>> code is 409.
>>
>
> This looks like a bug in the backup script - the backup should be finalized
> even if the image transfer failed, but the error you get say the vm is
> still
> in backup mode.
>
>
>>
>> How can I clean the situation?
>>
>
> 1. Stop the current backup
>
> If you still have the output from the command, we log the backup UUID.
>
> If you lost the backup id, you can get it using the API - visit this
> address in your browser:
>
> https://myengine/ovirt-engine/api/vms/{vm-id}/backups/
>
> Then stop the current backup using:
>
> /usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py stop
> vm-id backup-id
>
> If stopping the backup failed, stopping the VM will stop the backup.
> I hope you are running recent enough version, since in early versions there
> was a bug when you cannot stop the vm during a backup.
>

It is the latest 4.4.7. I run the backup_vm.py script from the engine:

ovirt-engine-4.4.7.7-1.el8.noarch
ovirt-engine-setup-plugin-imageio-4.4.7.7-1.el8.noarch
ovirt-imageio-common-2.2.0-1.el8.x86_64
ovirt-imageio-client-2.2.0-1.el8.x86_64
ovirt-imageio-daemon-2.2.0-1.el8.x86_64
python3-ovirt-engine-sdk4-4.4.13-1.el8.x86_64

But if I try the stop command I get the error

[g.cecchi@ovmgr1 ~]$ python3
/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py -c ovmgr1
stop dc386237-1e98-40c8-9d3d-45658163d1e2
68f83141-9d03-4cb0-84d4-e71fdd8753bb
[   0.0 ] Finalizing backup '68f83141-9d03-4cb0-84d4-e71fdd8753bb'
Traceback (most recent call last):
  File "/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py",
line 493, in 
main()
  File "/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py",
line 176, in main
args.command(args)
  File "/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py",
line 262, in cmd_stop
stop_backup(connection, args.backup_uuid, args)
  File "/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py",
line 345, in stop_backup
backup_service.finalize()
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py", line
33869, in finalize
return self._internal_action(action, 'finalize', None, headers, query,
wait)
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line 299,
in _internal_action
return future.wait() if wait else future
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line 55,
in wait
return self._code(response)
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line 296,
in callback
self._check_fault(response)
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line 132,
in _check_fault
self._raise_error(response, body)
  File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line 118,
in _raise_error
raise error
ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is
"[Cannot stop VM backup. The VM backup is not in READY phase, backup phase
is FINALIZING. Please try again when the backup is in READY phase.]". HTTP
response code is 409.
[g.cecchi@ovmgr1 ~]$



>
> 2. File a bug about this
>


Filed this one, hope its is correct; I chose ovirt-imageio as the product
and Client as the component:

https://bugzilla.redhat.com/show_bug.cgi?id=2001136

I put information also about the error received with the stop command



>
>
>>
>> BTW: the parameter to put into ovirt.conf is backup-dir or backup_dir or
>> what?
>>
>
> ovirt.conf do not include the backup dir, only details about engine.
> Adding backup-dir
> to ovirt.conf or to backup specific configuration sounds like a good idea.
>
> Nir
>

I agree

Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/OESNSO7MWVWZR2MS374ATPGYQRM2AXC3/


[ovirt-users] Re: how to remove a failed backup operation

2021-09-03 Thread Nir Soffer
On Fri, Sep 3, 2021 at 4:45 PM Gianluca Cecchi 
wrote:

> Hello,
> I was trying incremental backup with the provided
> /usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py and began
> using the "full" option.
> But I specified an incorrect dir and during backup I got error due to
> filesystem full
>
> [ 156.7 ] Creating image transfer for disk
> '33b0f6fb-a855-465d-a628-5fce9b64496a'
> [ 157.8 ] Image transfer 'ccc386d3-9f9d-4727-832a-56d355d60a95' is ready
> --- Logging error ---, 105.02 seconds, 147.48 MiB/s
>
> Traceback (most recent call last):
>   File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/io.py",
> line 242, in _run
> handler.copy(req)
>   File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/io.py",
> line 286, in copy
> self._src.write_to(self._dst, req.length, self._buf)
>   File
> "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/backends/http.py",
> line 216, in write_to
> writer.write(view[:n])
>   File
> "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/backends/nbd.py",
> line 118, in write
> self._client.write(self._position, buf)
>   File
> "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py", line
> 445, in write
> self._recv_reply(cmd)
>   File
> "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py", line
> 980, in _recv_reply
> if self._recv_reply_chunk(cmd):
>   File
> "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py", line
> 1031, in _recv_reply_chunk
> self._handle_error_chunk(length, flags)
>   File
> "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py", line
> 1144, in _handle_error_chunk
> raise ReplyError(code, message)
> ovirt_imageio._internal.nbd.ReplyError: Writing to file failed: [Error 28]
> No space left on device
>

This error is expected if you don't have space to write the data.


>
> Now if I try the same backup command (so with "full" option) and I get
>
> ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is
> "[Cannot backup VM. The VM is during a backup operation.]". HTTP response
> code is 409.
>

This looks like a bug in the backup script - the backup should be finalized
even if the image transfer failed, but the error you get say the vm is still
in backup mode.


>
> How can I clean the situation?
>

1. Stop the current backup

If you still have the output from the command, we log the backup UUID.

If you lost the backup id, you can get it using the API - visit this
address in your browser:

https://myengine/ovirt-engine/api/vms/{vm-id}/backups/

Then stop the current backup using:

/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py stop
vm-id backup-id

If stopping the backup failed, stopping the VM will stop the backup.
I hope you are running recent enough version, since in early versions there
was a bug when you cannot stop the vm during a backup.

2. File a bug about this


>
> BTW: the parameter to put into ovirt.conf is backup-dir or backup_dir or
> what?
>

ovirt.conf do not include the backup dir, only details about engine. Adding
backup-dir
to ovirt.conf or to backup specific configuration sounds like a good idea.

Nir
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HSUZOKSHCEWQAHUOAQ6EVUHWACTGANH7/


[ovirt-users] Re: Update to 4.4.8 leaves cluster in a circular error-state

2021-09-03 Thread Paul-Erik Törrönen

On 9/2/21 9:16 AM, Lucia Jelinkova wrote:
Could you please share more details about the CPU problem you're facing? 
There shouldn't be any breaking change in that CPU definition in 4.4+ 
compatibility version.


Unfortunately not, I've already made irreversible changes to the cluster 
so that I can no longer reproduce the error which I got when I tried to 
activate one of the Dell hosts and which resulted in an error about the 
CPU family.


After having wiped out most of the configurations I still do get a 
related error:


"The host CPU does not match the Cluster CPU Type and is running in a 
degraded mode. It is missing the following CPU flags: vmx, nx, 
model_Westmere, aes. Please update the host CPU microcode or change the 
Cluster CPU Type."


This error is not quite accurate since lscpu does list all of the flags 
mentioned above, except for the model_Westmere.


IIRC there were some comments in the mailing list earlier this year WRT 
this flag-mismatch and being related to incompatible linux-firmware 
package. Currently the host that generates this error has this package 
installed:


Name: linux-firmware
Version : 20210702
Release : 103.gitd79c2677.el8

Will need to dig through the mailing list archives.

Poltsi
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SUGLF3ON72DBZCE3Z4BEJDSBYPHFX6GU/


[ovirt-users] Re: how to remove a failed backup operation

2021-09-03 Thread Strahil Nikolov via Users
This looks like a bug. It should have 'recovered' from the failure.
I'm not sure which logs would help identify the root cause.

Best Regards,Strahil Nikolov
 
 
  On Fri, Sep 3, 2021 at 16:45, Gianluca Cecchi 
wrote:   Hello,I was trying incremental backup with the provided 
/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py and began using 
the "full" option.But I specified an incorrect dir and during backup I got 
error due to filesystem full
[ 156.7 ] Creating image transfer for disk 
'33b0f6fb-a855-465d-a628-5fce9b64496a'
[ 157.8 ] Image transfer 'ccc386d3-9f9d-4727-832a-56d355d60a95' is ready
--- Logging error ---, 105.02 seconds, 147.48 MiB/s                            
Traceback (most recent call last):
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/io.py", line 
242, in _run
    handler.copy(req)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/io.py", line 
286, in copy
    self._src.write_to(self._dst, req.length, self._buf)
  File 
"/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/backends/http.py", 
line 216, in write_to
    writer.write(view[:n])
  File 
"/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/backends/nbd.py", 
line 118, in write
    self._client.write(self._position, buf)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py", 
line 445, in write
    self._recv_reply(cmd)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py", 
line 980, in _recv_reply
    if self._recv_reply_chunk(cmd):
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py", 
line 1031, in _recv_reply_chunk
    self._handle_error_chunk(length, flags)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py", 
line 1144, in _handle_error_chunk
    raise ReplyError(code, message)
ovirt_imageio._internal.nbd.ReplyError: Writing to file failed: [Error 28] No 
space left on device
Now if I try the same backup command (so with "full" option) and I get 

ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[Cannot 
backup VM. The VM is during a backup operation.]". HTTP response code is 409.
How can I clean the situation?
BTW: the parameter to put into ovirt.conf is backup-dir or backup_dir or what?
Thanks,Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/6OZ7ZNH5GSNFCHDSDOPBNVXMN7WLWUXC/
  
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VIVHL2UBPXM43SU7CQCFFV2O2IO73UTQ/


[ovirt-users] Re: NFS storage was locked for 45 minutes after I attempted a clone operation

2021-09-03 Thread Strahil Nikolov via Users
That's really odd. Maybe you can try to clone it and then experiment on the 
clone itself. Once the reason is found out, you can try with the original.
My first look is to check all logs on the engine and the SPM for clues.
Best Regards,Strahil Nikolov 
 
  On Fri, Sep 3, 2021 at 11:42, David White via Users wrote:   
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/6XHROFQDWZY4Y6Z5LWWORTEJKCDBYIPT/
  
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AHSTIYM54FND422YEJ6GRXPJ4RTGGNAF/


[ovirt-users] how to remove a failed backup operation

2021-09-03 Thread Gianluca Cecchi
Hello,
I was trying incremental backup with the provided
/usr/share/doc/python3-ovirt-engine-sdk4/examples/backup_vm.py and began
using the "full" option.
But I specified an incorrect dir and during backup I got error due to
filesystem full

[ 156.7 ] Creating image transfer for disk
'33b0f6fb-a855-465d-a628-5fce9b64496a'
[ 157.8 ] Image transfer 'ccc386d3-9f9d-4727-832a-56d355d60a95' is ready
--- Logging error ---, 105.02 seconds, 147.48 MiB/s

Traceback (most recent call last):
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/io.py",
line 242, in _run
handler.copy(req)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/io.py",
line 286, in copy
self._src.write_to(self._dst, req.length, self._buf)
  File
"/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/backends/http.py",
line 216, in write_to
writer.write(view[:n])
  File
"/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/backends/nbd.py",
line 118, in write
self._client.write(self._position, buf)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py",
line 445, in write
self._recv_reply(cmd)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py",
line 980, in _recv_reply
if self._recv_reply_chunk(cmd):
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py",
line 1031, in _recv_reply_chunk
self._handle_error_chunk(length, flags)
  File "/usr/lib64/python3.6/site-packages/ovirt_imageio/_internal/nbd.py",
line 1144, in _handle_error_chunk
raise ReplyError(code, message)
ovirt_imageio._internal.nbd.ReplyError: Writing to file failed: [Error 28]
No space left on device

Now if I try the same backup command (so with "full" option) and I get

ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is
"[Cannot backup VM. The VM is during a backup operation.]". HTTP response
code is 409.

How can I clean the situation?

BTW: the parameter to put into ovirt.conf is backup-dir or backup_dir or
what?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/6OZ7ZNH5GSNFCHDSDOPBNVXMN7WLWUXC/


[ovirt-users] oVirt Node 4.4.8.3 Async update

2021-09-03 Thread Sandro Bonazzola
oVirt Node 4.4.8.3 Async update

On September 3rd 2021 the oVirt project released an async update for oVirt
Node consuming the following packages:

   -

   ovirt-release44 4.4.8.3
   -

   ovirt-node-ng-image-update 4.4.8.3

oVirt Node respin also consumed most recent CentOS Stream and Advanced
Virtualization content.

It also includes fixes for:

   -

   Bug 1996602  - VM
   remains in paused state when trying to write on a resized disk resides on
   iscsi
   -

  Via qemu-kvm-6.0.0-29.el8s which shipped live.



Here’s the full list of changes:

--- ovirt-node-ng-image-4.4.8.2.manifest-rpm 2021-09-01 12:59:05.195037688
+0200

+++ ovirt-node-ng-image-4.4.8.3.manifest-rpm 2021-09-03 13:27:52.006368887
+0200

@@ -47 +46,0 @@

-boost-iostreams-1.66.0-10.el8.x86_64

@@ -112 +111 @@

-device-mapper-persistent-data-0.9.0-1.el8.x86_64

+device-mapper-persistent-data-0.9.0-4.el8.x86_64

@@ -355 +354 @@

-libcap-2.26-4.el8.x86_64

+libcap-2.26-5.el8.x86_64

@@ -637 +636 @@

-osinfo-db-20210215-1.el8.noarch

+osinfo-db-20210809-1.el8.noarch

@@ -648 +647 @@

-ovirt-node-ng-image-update-placeholder-4.4.8.2-1.el8.noarch

+ovirt-node-ng-image-update-placeholder-4.4.8.3-1.el8.noarch

@@ -656,2 +655,2 @@

-ovirt-release-host-node-4.4.8.2-1.el8.noarch

-ovirt-release44-4.4.8.2-1.el8.noarch

+ovirt-release-host-node-4.4.8.3-1.el8.noarch

+ovirt-release44-4.4.8.3-1.el8.noarch

@@ -768 +767 @@

-python3-eventlet-0.25.2-3.el8.noarch

+python3-eventlet-0.25.2-3.1.el8.noarch

@@ -814 +813 @@

-python3-os-brick-4.0.3-1.el8.noarch

+python3-os-brick-4.0.3-2.el8.noarch

@@ -890,14 +889,14 @@

-qemu-guest-agent-6.0.0-27.el8s.x86_64

-qemu-img-6.0.0-27.el8s.x86_64

-qemu-kvm-6.0.0-27.el8s.x86_64

-qemu-kvm-block-curl-6.0.0-27.el8s.x86_64

-qemu-kvm-block-gluster-6.0.0-27.el8s.x86_64

-qemu-kvm-block-iscsi-6.0.0-27.el8s.x86_64

-qemu-kvm-block-rbd-6.0.0-27.el8s.x86_64

-qemu-kvm-block-ssh-6.0.0-27.el8s.x86_64

-qemu-kvm-common-6.0.0-27.el8s.x86_64

-qemu-kvm-core-6.0.0-27.el8s.x86_64

-qemu-kvm-docs-6.0.0-27.el8s.x86_64

-qemu-kvm-hw-usbredir-6.0.0-27.el8s.x86_64

-qemu-kvm-ui-opengl-6.0.0-27.el8s.x86_64

-qemu-kvm-ui-spice-6.0.0-27.el8s.x86_64

+qemu-guest-agent-6.0.0-29.el8s.x86_64

+qemu-img-6.0.0-29.el8s.x86_64

+qemu-kvm-6.0.0-29.el8s.x86_64

+qemu-kvm-block-curl-6.0.0-29.el8s.x86_64

+qemu-kvm-block-gluster-6.0.0-29.el8s.x86_64

+qemu-kvm-block-iscsi-6.0.0-29.el8s.x86_64

+qemu-kvm-block-rbd-6.0.0-29.el8s.x86_64

+qemu-kvm-block-ssh-6.0.0-29.el8s.x86_64

+qemu-kvm-common-6.0.0-29.el8s.x86_64

+qemu-kvm-core-6.0.0-29.el8s.x86_64

+qemu-kvm-docs-6.0.0-29.el8s.x86_64

+qemu-kvm-hw-usbredir-6.0.0-29.el8s.x86_64

+qemu-kvm-ui-opengl-6.0.0-29.el8s.x86_64

+qemu-kvm-ui-spice-6.0.0-29.el8s.x86_64

@@ -934,2 +933,2 @@

-selinux-policy-3.14.3-78.el8.noarch

-selinux-policy-targeted-3.14.3-78.el8.noarch

+selinux-policy-3.14.3-79.el8.noarch

+selinux-policy-targeted-3.14.3-79.el8.noarch

@@ -1007 +1006 @@

-virt-v2v-1.42.0-14.el8s.x86_64

+virt-v2v-1.42.0-15.el8s.x86_64



-- 

Sandro Bonazzola

MANAGER, SOFTWARE ENGINEERING, EMEA R RHV

Red Hat EMEA 

sbona...@redhat.com


*Red Hat respects your work life balance. Therefore there is no need to
answer this email out of your office hours.*
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/N4PJ5WXMNJQU5YQY776WVE6QBDS7D7K3/


[ovirt-users] Re: posix storage mount path error when creating volumes

2021-09-03 Thread Konstantin Shalygin
We are use CephFS domains without issues from 4.3, currently on 4.4

[root@ovirt-host1 /]# ls -la /rhev/data-center/mnt/*ovirt*
'/rhev/data-center/mnt/172.16.16.2:3300,172.16.16.3:3300,172.16.16.4:3300:_ovirt__iso':
total 0
drwxrwxrwx. 3 root root   1 Sep  3 15:43 .
drwxr-xr-x. 5 vdsm kvm  182 Aug 19 14:35 ..
drwxr-xr-x. 4 vdsm kvm2 Aug 19 14:35 def09ff0-e986-44c4-ac1c-470668ec2822


[root@ovirt-host1 /]# mount | grep iso
172.16.16.2:3300,172.16.16.3:3300,172.16.16.4:3300:/ovirt_iso on 
/rhev/data-center/mnt/172.16.16.2:3300,172.16.16.3:3300,172.16.16.4:3300:_ovirt__iso
 type ceph 
(rw,relatime,seclabel,name=mds_ovirt_iso,secret=,ms_mode=prefer-crc,dirstat,acl)


k


> On 1 Sep 2021, at 23:52, Nir Soffer  wrote:
> 
> This was never supported.
> 
> We had this old fix that was rejected:
> https://gerrit.ovirt.org/c/vdsm/+/94027 
> 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/5WCJTIBY5GHNAJLCETP3RJOX2ITFOHD4/


[ovirt-users] Re: NFS storage was locked for 45 minutes after I attempted a clone operation

2021-09-03 Thread David White via Users
The save operation is going at a snail's pace, though.

Using "watch du -skh", I counted about 5-7 seconds per .1 GB (1/10 of 1GB).
It's a virtual disk, but I'm using over 200GB... so at this rate, it'll take a 
very long time.

I wonder if Pascal is on to something, and the export is happening over the 
frontend 1GB network?

I'm going to cancel this operation, as the VM has now been down for close to an 
hour.

Sent with ProtonMail Secure Email.

‐‐‐ Original Message ‐‐‐
On Friday, September 3rd, 2021 at 4:33 AM, David White 
 wrote:

> Update perhaps I have discovered a bug somewhere?
> 

> I started another export after hours (it's very early morning hours right 
> now, and I can tolerate a little downtime on this VM). I had the same 
> symptoms, but this time, I just left it alone. I waited about 45 minutes with 
> no progress.
> 

> I then ssh'd to the NFS destination (also on the 10Gbps storage network), and 
> running tcpdump, I didn't see any traffic coming across the wire.
> 

> So I then powered off my VM, and I immediately began to see a new backup 
> image appear in my NFS export. 
> 

> I wonder if the VM was trying to snapshot the memory and there wasn't enough 
> on the host or something? The VM has 16GB of RAM, and there are multiple VMs 
> on that host (although the host itself has 64GB of physical RAM, so should 
> have been plenty).
> 

> Sent with ProtonMail Secure Email.
> 

> ‐‐‐ Original Message ‐‐‐
> On Friday, September 3rd, 2021 at 4:10 AM, David White via Users 
>  wrote:
> 

> > In this particular case, I have 1 (one) 250GB virtual disk..
> > 

> > Sent with ProtonMail Secure Email.
> > 

> > ‐‐‐ Original Message ‐‐‐
> > On Tuesday, August 31st, 2021 at 11:21 PM, Strahil Nikolov 
> >  wrote:
> > 

> > > Hi David,
> > > 

> > > how big are your VM disks ?
> > > 

> > > I suppose you have several very large ones.
> > > 

> > > Best Regards,
> > > Strahil Nikolov
> > > 

> > > Sent from Yahoo Mail on Android
> > > 

> > > > On Thu, Aug 26, 2021 at 3:27, David White via Users
> > > >  wrote:
> > > > I have an HCI cluster running on Gluster storage. I exposed an NFS 
> > > > share into oVirt as a storage domain so that I could clone all of my 
> > > > VMs (I'm preparing to move physically to a new datacenter). I got 3-4 
> > > > VMs cloned perfectly fine yesterday. But then this evening, I tried to 
> > > > clone a big VM, and it caused the disk to lock up. The VM went totally 
> > > > unresponsive, and I didn't see a way to cancel the clone. Nagios NRPE 
> > > > (on the client VM) was reporting server load over 65+, but I was never 
> > > > able to establish an SSH connection. 
> > > > 

> > > > Eventually, I tried restarting the ovirt-engine, per 
> > > > https://access.redhat.com/solutions/396753. When that didn't work, I 
> > > > powered down the VM completely. But the disks were still locked. So I 
> > > > then tried to put the storage domain into maintenance mode, but that 
> > > > wound up putting the entire domain into a "locked" state. Finally, 
> > > > eventually, the disks unlocked, and I was able to power the VM back 
> > > > online.
> > > > 

> > > > From start to finish, my VM was down for about 45 minutes, including 
> > > > the time when NRPE was still sending data to Nagios.
> > > > 

> > > > What logs should I look at, and how can I troubleshoot what went wrong 
> > > > here, and hopefully avoid this from happening again?
> > > > 

> > > > Sent with ProtonMail Secure Email.
> > > > 

> > > > ___
> > > > Users mailing list -- users@ovirt.org
> > > > To unsubscribe send an email to users-le...@ovirt.org
> > > > Privacy Statement: https://www.ovirt.org/privacy-policy.html
> > > > oVirt Code of Conduct: 
> > > > https://www.ovirt.org/community/about/community-guidelines/
> > > > List Archives: 
> > > > https://lists.ovirt.org/archives/list/users@ovirt.org/message/ASEENELT4TRTXQ7MF4FKB6L75D3H75AN/

publickey - dmwhite823@protonmail.com - 0x320CD582.asc
Description: application/pgp-keys


signature.asc
Description: OpenPGP digital signature
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/6XHROFQDWZY4Y6Z5LWWORTEJKCDBYIPT/


[ovirt-users] Re: NFS storage was locked for 45 minutes after I attempted a clone operation

2021-09-03 Thread David White via Users
Update perhaps I have discovered a bug somewhere?

I started another export after hours (it's very early morning hours right now, 
and I can tolerate a little downtime on this VM). I had the same symptoms, but 
this time, I just left it alone. I waited about 45 minutes with no progress.

I then ssh'd to the NFS destination (also on the 10Gbps storage network), and 
running tcpdump, I didn't see any traffic coming across the wire.

So I then powered off my VM, and I immediately began to see a new backup image 
appear in my NFS export. 

I wonder if the VM was trying to snapshot the memory and there wasn't enough on 
the host or something? The VM has 16GB of RAM, and there are multiple VMs on 
that host (although the host itself has 64GB of physical RAM, so should have 
been plenty).

Sent with ProtonMail Secure Email.

‐‐‐ Original Message ‐‐‐

On Friday, September 3rd, 2021 at 4:10 AM, David White via Users 
 wrote:

> In this particular case, I have 1 (one) 250GB virtual disk..
> 

> Sent with ProtonMail Secure Email.
> 

> ‐‐‐ Original Message ‐‐‐
> 

> On Tuesday, August 31st, 2021 at 11:21 PM, Strahil Nikolov 
>  wrote:
> 

> > Hi David,
> > 

> > how big are your VM disks ?
> > 

> > I suppose you have several very large ones.
> > 

> > Best Regards,Strahil Nikolov
> > 

> > Sent from Yahoo Mail on Android
> > 

> > > On Thu, Aug 26, 2021 at 3:27, David White via Users 
> > > wrote:I have an HCI cluster running on Gluster storage. I exposed an NFS 
> > > share into oVirt as a storage domain so that I could clone all of my VMs 
> > > (I'm preparing to move physically to a new datacenter). I got 3-4 VMs 
> > > cloned perfectly fine yesterday. But then this evening, I tried to clone 
> > > a big VM, and it caused the disk to lock up. The VM went totally 
> > > unresponsive, and I didn't see a way to cancel the clone. Nagios NRPE (on 
> > > the client VM) was reporting server load over 65+, but I was never able 
> > > to establish an SSH connection. 
> > > 

> > > Eventually, I tried restarting the ovirt-engine, per 
> > > https://access.redhat.com/solutions/396753. When that didn't work, I 
> > > powered down the VM completely. But the disks were still locked. So I 
> > > then tried to put the storage domain into maintenance mode, but that 
> > > wound up putting the entire domain into a "locked" state. Finally, 
> > > eventually, the disks unlocked, and I was able to power the VM back 
> > > online.
> > > 

> > > From start to finish, my VM was down for about 45 minutes, including the 
> > > time when NRPE was still sending data to Nagios.
> > > 

> > > What logs should I look at, and how can I troubleshoot what went wrong 
> > > here, and hopefully avoid this from happening again?
> > > 

> > > Sent with ProtonMail Secure Email.
> > > 

> > > ___
> > > 

> > > Users mailing list -- users@ovirt.org
> > > 

> > > To unsubscribe send an email to users-le...@ovirt.org
> > > 

> > > Privacy Statement: https://www.ovirt.org/privacy-policy.html
> > > 

> > > oVirt Code of Conduct: 
> > > https://www.ovirt.org/community/about/community-guidelines/
> > > 

> > > List Archives: 
> > > https://lists.ovirt.org/archives/list/users@ovirt.org/message/ASEENELT4TRTXQ7MF4FKB6L75D3H75AN/

publickey - dmwhite823@protonmail.com - 0x320CD582.asc
Description: application/pgp-keys


signature.asc
Description: OpenPGP digital signature
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/PMKJX2YEAFHG574N375H3ASU3N3VR3UW/


[ovirt-users] Re: NFS storage was locked for 45 minutes after I attempted a clone operation

2021-09-03 Thread David White via Users
In this particular case, I have 1 (one) 250GB virtual disk..

Sent with ProtonMail Secure Email.

‐‐‐ Original Message ‐‐‐

On Tuesday, August 31st, 2021 at 11:21 PM, Strahil Nikolov 
 wrote:

> Hi David,
> 

> how big are your VM disks ?
> 

> I suppose you have several very large ones.
> 

> Best Regards,Strahil Nikolov
> 

> Sent from Yahoo Mail on Android
> 

> > On Thu, Aug 26, 2021 at 3:27, David White via Users 
> > wrote:I have an HCI cluster running on Gluster storage. I exposed an NFS 
> > share into oVirt as a storage domain so that I could clone all of my VMs 
> > (I'm preparing to move physically to a new datacenter). I got 3-4 VMs 
> > cloned perfectly fine yesterday. But then this evening, I tried to clone a 
> > big VM, and it caused the disk to lock up. The VM went totally 
> > unresponsive, and I didn't see a way to cancel the clone. Nagios NRPE (on 
> > the client VM) was reporting server load over 65+, but I was never able to 
> > establish an SSH connection. 
> > 

> > Eventually, I tried restarting the ovirt-engine, per 
> > https://access.redhat.com/solutions/396753. When that didn't work, I 
> > powered down the VM completely. But the disks were still locked. So I then 
> > tried to put the storage domain into maintenance mode, but that wound up 
> > putting the entire domain into a "locked" state. Finally, eventually, the 
> > disks unlocked, and I was able to power the VM back online.
> > 

> > From start to finish, my VM was down for about 45 minutes, including the 
> > time when NRPE was still sending data to Nagios.
> > 

> > What logs should I look at, and how can I troubleshoot what went wrong 
> > here, and hopefully avoid this from happening again?
> > 

> > Sent with ProtonMail Secure Email.
> > 

> > ___
> > 

> > Users mailing list -- users@ovirt.org
> > 

> > To unsubscribe send an email to users-le...@ovirt.org
> > 

> > Privacy Statement: https://www.ovirt.org/privacy-policy.html
> > 

> > oVirt Code of Conduct: 
> > https://www.ovirt.org/community/about/community-guidelines/
> > 

> > List Archives: 
> > https://lists.ovirt.org/archives/list/users@ovirt.org/message/ASEENELT4TRTXQ7MF4FKB6L75D3H75AN/

publickey - dmwhite823@protonmail.com - 0x320CD582.asc
Description: application/pgp-keys


signature.asc
Description: OpenPGP digital signature
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3CRWWNSNJTSATXRDIG7BHZDOQ3VCKQMT/