Hey there,

The CSI driver in this repository is built with OpenShift in mind, and does
not have an upstream that is intended to work with vanilla Kubernetes. Even
if you get it to work now, it may break in the future.

We have had some discussions around reviving the upstream for the CSI
driver, which is located here: https://github.com/oVirt/csi-driver

We are also doing a fair bit of work which will make the process easier
here: https://github.com/oVirt/go-ovirt-client

However, at this time we haven't made any progress that would be useful to
you. With that in mind, I would recommend reporting possible bugs either on
GitHub or on Bugzilla, that way they will reach us quicker, even if we
officially don't support vanilla Kubernetes.

Cheers,

Janos

On Mon, Aug 23, 2021 at 12:57 AM <ssarang...@gmail.com> wrote:

> Hi all,
>
> I deployed ovirt-csi in the k8s by applying yaml manually. I used the
> latest version of the container image.
> (https://github.com/openshift/ovirt-csi-driver-operator/tree/master/assets
> )
>
> After successfully creating pvc and pod, I tried to delete it.
> And the pod is deleted, but the pvc is not deleted. This is because
> deleting a pod does not unmap /dev/rbd0 attached to the ovirt vm.
>
> How can I delete the pvc successfully?
>
> oVirt engine version is 4.4.7.6-1.el8.
> Here is the engine log when deleting the pod:
>
> 2021-08-20 17:40:35,385+09 INFO
> [org.ovirt.engine.core.sso.service.AuthenticationService] (default
> task-149) [] User admin@internal-authz with profile [internal]
> successfully logged in with scopes: ovirt-app-api
> ovirt-ext=token-info:authz-search ovirt-ext=token-info:public-authz-search
> ovirt-ext=token-info:validate ovirt-ext=token:password-access
> 2021-08-20 17:40:35,403+09 INFO
> [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-149)
> [68ee3182] Running command: CreateUserSessionCommand internal: false.
> 2021-08-20 17:40:35,517+09 INFO
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> (default task-149) [68ee3182] EVENT_ID: USER_VDC_LOGIN(30), User
> admin@internal-authz connecting from '192.168.7.169' using session
> 'XfDgNkmAGnPiZahK5itLhHQTCNHZ3JwXMMzOiZrYL3C32+1TTys3xcjrAmCIKPu02hgN1sdVpfZXWd0FznaPCQ=='
> logged in.
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource",
> will return null
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource",
> will return null
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource",
> will return null
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource",
> will return null
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource",
> will return null
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource",
> will return null
> 2021-08-20 17:40:35,663+09 INFO
> [org.ovirt.engine.core.bll.storage.disk.DetachDiskFromVmCommand] (default
> task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Running command:
> DetachDiskFromVmCommand internal: false. Entities affected :  ID:
> 59a7461c-72fe-4e01-86a7-c70243f31596 Type: VMAction group
> CONFIGURE_VM_STORAGE with role type USER
> 2021-08-20 17:40:35,664+09 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand]
> (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] START,
> HotUnPlugDiskVDSCommand(HostName = host,
> HotPlugDiskVDSParameters:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094',
> vmId='59a7461c-72fe-4e01-86a7-c70243f31596',
> diskId='63a64445-1659-4d5f-8847-e7266e64b09e'}), log id: 506ff4a4
> 2021-08-20 17:40:35,678+09 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand]
> (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Disk hot-unplug:
> <?xml version="1.0" encoding="UTF-8"?><hotunplug>
>   <devices>
>     <disk>
>       <alias name="ua-63a64445-1659-4d5f-8847-e7266e64b09e"/>
>     </disk>
>   </devices>
> </hotunplug>
>
> 2021-08-20 17:40:35,749+09 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand]
> (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] FINISH,
> HotUnPlugDiskVDSCommand, return: , log id: 506ff4a4
> 2021-08-20 17:40:35,842+09 INFO
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] EVENT_ID:
> USER_DETACH_DISK_FROM_VM(2,018), Disk
> pvc-9845a0ff-e94c-497c-8c65-fc6a1e26db20 was successfully detached from VM
> centos by admin@internal-authz.
> 2021-08-20 17:40:35,916+09 ERROR
> [org.ovirt.engine.core.sso.service.SsoService] (default task-150) []
> OAuthException invalid_grant: The provided authorization grant for the auth
> code has expired.
> 2021-08-20 17:40:35,917+09 ERROR
> [org.ovirt.engine.core.aaa.filters.SsoRestApiAuthFilter] (default task-149)
> [] Cannot authenticate using authentication Headers: invalid_grant: The
> provided authorization grant for the auth code has expired.
> 2021-08-20 17:40:36,029+09 INFO
> [org.ovirt.engine.core.sso.service.AuthenticationService] (default
> task-149) [] User admin@internal-authz with profile [internal]
> successfully logged in with scopes: ovirt-app-api
> ovirt-ext=token-info:authz-search ovirt-ext=token-info:public-authz-search
> ovirt-ext=token-info:validate ovirt-ext=token:password-access
> 2021-08-20 17:40:36,046+09 INFO
> [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-149)
> [4c4bf441] Running command: CreateUserSessionCommand internal: false.
> 2021-08-20 17:40:36,114+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource",
> will return null
> 2021-08-20 17:40:36,114+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource",
> will return null
> 2021-08-20 17:40:36,114+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource",
> will return null
> 2021-08-20 17:40:36,114+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource",
> will return null
> 2021-08-20 17:40:49,241+09 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand]
> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) []
> START, DumpXmlsVDSCommand(HostName = host,
> Params:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094',
> vmIds='[59a7461c-72fe-4e01-86a7-c70243f31596]'}), log id: 7eb54202
> 2021-08-20 17:40:49,244+09 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand]
> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) []
> FINISH, DumpXmlsVDSCommand, return:
> {59a7461c-72fe-4e01-86a7-c70243f31596=<domain type='kvm' id='1' xmlns:qemu='
> http://libvirt.org/schemas/domain/qemu/1.0'>
>   <name>centos</name>
>   <uuid>59a7461c-72fe-4e01-86a7-c70243f31596</uuid>
>   <metadata xmlns:ns1="http://ovirt.org/vm/tune/1.0"; xmlns:ovirt-vm="
> http://ovirt.org/vm/1.0";>
>     <ns1:qos/>
>     <ovirt-vm:vm xmlns:ovirt-vm="http://ovirt.org/vm/1.0";>
>     <ovirt-vm:balloonTarget type="int">4194304</ovirt-vm:balloonTarget>
>     <ovirt-vm:ballooningEnabled>true</ovirt-vm:ballooningEnabled>
>     <ovirt-vm:clusterVersion>4.6</ovirt-vm:clusterVersion>
>     <ovirt-vm:destroy_on_reboot
> type="bool">False</ovirt-vm:destroy_on_reboot>
>     <ovirt-vm:launchPaused>false</ovirt-vm:launchPaused>
>     <ovirt-vm:memGuaranteedSize
> type="int">4096</ovirt-vm:memGuaranteedSize>
>     <ovirt-vm:minGuaranteedMemoryMb
> type="int">4096</ovirt-vm:minGuaranteedMemoryMb>
>     <ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior>
>     <ovirt-vm:startTime
> type="float">1628558564.8754532</ovirt-vm:startTime>
>     <ovirt-vm:device alias="ua-7c9f38e9-8889-46c8-83bb-92efb9272de9"
> mac_address="56:6f:16:a8:00:07">
>         <ovirt-vm:network>ovirtmgmt</ovirt-vm:network>
>         <ovirt-vm:custom>
>             <ovirt-vm:queues>2</ovirt-vm:queues>
>         </ovirt-vm:custom>
>     </ovirt-vm:device>
>     <ovirt-vm:device devtype="disk" name="sda">
>
> <ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
>         <ovirt-vm:guestName>/dev/sda</ovirt-vm:guestName>
>
> <ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
>         <ovirt-vm:managed type="bool">False</ovirt-vm:managed>
>
> <ovirt-vm:poolID>4ca6e0e8-e3a4-11eb-8830-480fcf63834f</ovirt-vm:poolID>
>
> <ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID>
>         <ovirt-vm:volumeChain>
>             <ovirt-vm:volumeChainNode>
>
> <ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
>
> <ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
>                 <ovirt-vm:leaseOffset type="int">0</ovirt-vm:leaseOffset>
>                 <ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:
> _home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003.lease</ovirt-vm:leasePath>
>                 <ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:
> _home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:path>
>
> <ovirt-vm:volumeID>4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:volumeID>
>             </ovirt-vm:volumeChainNode>
>             <ovirt-vm:volumeChainNode>
>
> <ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
>
> <ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
>                 <ovirt-vm:leaseOffset type="int">0</ovirt-vm:leaseOffset>
>                 <ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:
> _home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4.lease</ovirt-vm:leasePath>
>                 <ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:
> _home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:path>
>
> <ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID>
>             </ovirt-vm:volumeChainNode>
>         </ovirt-vm:volumeChain>
>     </ovirt-vm:device>
>     <ovirt-vm:device devtype="disk" name="sdc">
>         <ovirt-vm:managed type="bool">False</ovirt-vm:managed>
>     </ovirt-vm:device>
> </ovirt-vm:vm>
>   </metadata>
>   <maxMemory slots='16' unit='KiB'>16777216</maxMemory>
>   <memory unit='KiB'>4194304</memory>
>   <currentMemory unit='KiB'>4194304</currentMemory>
>   <vcpu placement='static' current='2'>16</vcpu>
>   <resource>
>     <partition>/machine</partition>
>   </resource>
>   <sysinfo type='smbios'>
>     <system>
>       <entry name='manufacturer'>oVirt</entry>
>       <entry name='product'>RHEL</entry>
>       <entry name='version'>8.4-1.2105.el8</entry>
>       <entry name='serial'>83e66af8-0500-11e6-9c43-bc00007c0000</entry>
>       <entry name='uuid'>59a7461c-72fe-4e01-86a7-c70243f31596</entry>
>       <entry name='family'>oVirt</entry>
>     </system>
>   </sysinfo>
>   <os>
>     <type arch='x86_64' machine='pc-q35-rhel8.4.0'>hvm</type>
>     <boot dev='hd'/>
>     <smbios mode='sysinfo'/>
>   </os>
>   <features>
>     <acpi/>
>   </features>
>   <cpu mode='custom' match='exact' check='full'>
>     <model fallback='forbid'>Skylake-Client-noTSX-IBRS</model>
>     <topology sockets='16' dies='1' cores='1' threads='1'/>
>     <feature policy='require' name='ssbd'/>
>     <feature policy='require' name='md-clear'/>
>     <feature policy='disable' name='mpx'/>
>     <feature policy='require' name='hypervisor'/>
>     <numa>
>       <cell id='0' cpus='0-15' memory='4194304' unit='KiB'/>
>     </numa>
>   </cpu>
>   <clock offset='variable' adjustment='0' basis='utc'>
>     <timer name='rtc' tickpolicy='catchup'/>
>     <timer name='pit' tickpolicy='delay'/>
>     <timer name='hpet' present='no'/>
>   </clock>
>   <on_poweroff>destroy</on_poweroff>
>   <on_reboot>restart</on_reboot>
>   <on_crash>destroy</on_crash>
>   <pm>
>     <suspend-to-mem enabled='no'/>
>     <suspend-to-disk enabled='no'/>
>   </pm>
>   <devices>
>     <emulator>/usr/libexec/qemu-kvm</emulator>
>     <disk type='file' device='cdrom'>
>       <driver name='qemu' error_policy='report'/>
>       <source startupPolicy='optional'/>
>       <target dev='sdc' bus='sata'/>
>       <readonly/>
>       <alias name='ua-df0ac774-3623-4868-8bd3-45c8f2aa3dc4'/>
>       <address type='drive' controller='0' bus='0' target='0' unit='2'/>
>     </disk>
>     <disk type='file' device='disk' snapshot='no'>
>       <driver name='qemu' type='qcow2' cache='none' error_policy='stop'
> io='threads'/>
>       <source 
> file='/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4'
> index='8'>
>         <seclabel model='dac' relabel='no'/>
>       </source>
>       <backingStore type='file' index='1'>
>         <format type='qcow2'/>
>         <source file='/rhev/data-center/mnt/192.168.7.18:
> _home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003'>
>           <seclabel model='dac' relabel='no'/>
>         </source>
>         <backingStore/>
>       </backingStore>
>       <target dev='sda' bus='scsi'/>
>       <serial>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</serial>
>       <alias name='ua-bee44276-234f-4ed7-8a8a-d90a5e3cb5b3'/>
>       <address type='drive' controller='0' bus='0' target='0' unit='0'/>
>     </disk>
>     <controller type='pci' index='1' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='1' port='0x10'/>
>       <alias name='pci.1'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x0' multifunction='on'/>
>     </controller>
>     <controller type='pci' index='2' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='2' port='0x11'/>
>       <alias name='pci.2'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x1'/>
>     </controller>
>     <controller type='pci' index='3' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='3' port='0x12'/>
>       <alias name='pci.3'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x2'/>
>     </controller>
>     <controller type='pci' index='4' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='4' port='0x13'/>
>       <alias name='pci.4'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x3'/>
>     </controller>
>     <controller type='pci' index='5' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='5' port='0x14'/>
>       <alias name='pci.5'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x4'/>
>     </controller>
>     <controller type='pci' index='6' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='6' port='0x15'/>
>       <alias name='pci.6'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x5'/>
>     </controller>
>     <controller type='pci' index='7' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='7' port='0x16'/>
>       <alias name='pci.7'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x6'/>
>     </controller>
>     <controller type='pci' index='8' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='8' port='0x17'/>
>       <alias name='pci.8'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x7'/>
>     </controller>
>     <controller type='pci' index='9' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='9' port='0x18'/>
>       <alias name='pci.9'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x0' multifunction='on'/>
>     </controller>
>     <controller type='pci' index='10' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='10' port='0x19'/>
>       <alias name='pci.10'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x1'/>
>     </controller>
>     <controller type='pci' index='11' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='11' port='0x1a'/>
>       <alias name='pci.11'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x2'/>
>     </controller>
>     <controller type='pci' index='12' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='12' port='0x1b'/>
>       <alias name='pci.12'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x3'/>
>     </controller>
>     <controller type='pci' index='13' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='13' port='0x1c'/>
>       <alias name='pci.13'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x4'/>
>     </controller>
>     <controller type='pci' index='14' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='14' port='0x1d'/>
>       <alias name='pci.14'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x5'/>
>     </controller>
>     <controller type='pci' index='15' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='15' port='0x1e'/>
>       <alias name='pci.15'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x6'/>
>     </controller>
>     <controller type='pci' index='16' model='pcie-root-port'>
>       <model name='pcie-root-port'/>
>       <target chassis='16' port='0x1f'/>
>       <alias name='pci.16'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x7'/>
>     </controller>
>     <controller type='pci' index='0' model='pcie-root'>
>       <alias name='pcie.0'/>
>     </controller>
>     <controller type='sata' index='0'>
>       <alias name='ide'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x1f'
> function='0x2'/>
>     </controller>
>     <controller type='virtio-serial' index='0' ports='16'>
>       <alias name='virtio-serial0'/>
>       <address type='pci' domain='0x0000' bus='0x04' slot='0x00'
> function='0x0'/>
>     </controller>
>     <controller type='scsi' index='0' model='virtio-scsi'>
>       <alias name='ua-82c49f93-c4e8-460b-bb7d-95db0e9d87a0'/>
>       <address type='pci' domain='0x0000' bus='0x02' slot='0x00'
> function='0x0'/>
>     </controller>
>     <controller type='usb' index='0' model='qemu-xhci' ports='8'>
>       <alias name='ua-ad56daea-edb1-45c7-a1ab-2a7db3aaeee2'/>
>       <address type='pci' domain='0x0000' bus='0x03' slot='0x00'
> function='0x0'/>
>     </controller>
>     <interface type='bridge'>
>       <mac address='56:6f:16:a8:00:07'/>
>       <source bridge='ovirtmgmt'/>
>       <target dev='vnet0'/>
>       <model type='virtio'/>
>       <driver name='vhost' queues='2'/>
>       <filterref filter='vdsm-no-mac-spoofing'/>
>       <link state='up'/>
>       <mtu size='1500'/>
>       <alias name='ua-7c9f38e9-8889-46c8-83bb-92efb9272de9'/>
>       <address type='pci' domain='0x0000' bus='0x01' slot='0x00'
> function='0x0'/>
>     </interface>
>     <channel type='unix'>
>       <source mode='bind'
> path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.ovirt-guest-agent.0'/>
>       <target type='virtio' name='ovirt-guest-agent.0'
> state='disconnected'/>
>       <alias name='channel0'/>
>       <address type='virtio-serial' controller='0' bus='0' port='1'/>
>     </channel>
>     <channel type='unix'>
>       <source mode='bind'
> path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.org.qemu.guest_agent.0'/>
>       <target type='virtio' name='org.qemu.guest_agent.0'
> state='connected'/>
>       <alias name='channel1'/>
>       <address type='virtio-serial' controller='0' bus='0' port='2'/>
>     </channel>
>     <channel type='spicevmc'>
>       <target type='virtio' name='com.redhat.spice.0'
> state='disconnected'/>
>       <alias name='channel2'/>
>       <address type='virtio-serial' controller='0' bus='0' port='3'/>
>     </channel>
>     <input type='tablet' bus='usb'>
>       <alias name='input0'/>
>       <address type='usb' bus='0' port='1'/>
>     </input>
>     <input type='mouse' bus='ps2'>
>       <alias name='input1'/>
>     </input>
>     <input type='keyboard' bus='ps2'>
>       <alias name='input2'/>
>     </input>
>     <graphics type='vnc' port='5900' autoport='yes' listen='192.168.7.18'
> keymap='en-us' passwdValidTo='1970-01-01T00:00:01'>
>       <listen type='network' address='192.168.7.18'
> network='vdsm-ovirtmgmt'/>
>     </graphics>
>     <graphics type='spice' port='5901' tlsPort='5902' autoport='yes'
> listen='192.168.7.18' passwdValidTo='1970-01-01T00:00:01'>
>       <listen type='network' address='192.168.7.18'
> network='vdsm-ovirtmgmt'/>
>       <channel name='main' mode='secure'/>
>       <channel name='display' mode='secure'/>
>       <channel name='inputs' mode='secure'/>
>       <channel name='cursor' mode='secure'/>
>       <channel name='playback' mode='secure'/>
>       <channel name='record' mode='secure'/>
>       <channel name='smartcard' mode='secure'/>
>       <channel name='usbredir' mode='secure'/>
>     </graphics>
>     <video>
>       <model type='qxl' ram='65536' vram='8192' vgamem='16384' heads='1'
> primary='yes'/>
>       <alias name='ua-799f065a-b2b9-4e37-a502-f86c7cc8dc51'/>
>       <address type='pci' domain='0x0000' bus='0x00' slot='0x01'
> function='0x0'/>
>     </video>
>     <memballoon model='virtio'>
>       <stats period='5'/>
>       <alias name='ua-c2bfe0b9-065a-46b7-9b0b-ef7e0f699611'/>
>       <address type='pci' domain='0x0000' bus='0x05' slot='0x00'
> function='0x0'/>
>     </memballoon>
>     <rng model='virtio'>
>       <backend model='random'>/dev/urandom</backend>
>       <alias name='ua-fa110e6b-5eed-4b4b-93d8-0ac5de08aa2e'/>
>       <address type='pci' domain='0x0000' bus='0x06' slot='0x00'
> function='0x0'/>
>     </rng>
>   </devices>
>   <seclabel type='dynamic' model='selinux' relabel='yes'>
>     <label>system_u:system_r:svirt_t:s0:c437,c650</label>
>     <imagelabel>system_u:object_r:svirt_image_t:s0:c437,c650</imagelabel>
>   </seclabel>
>   <seclabel type='dynamic' model='dac' relabel='yes'>
>     <label>+107:+107</label>
>     <imagelabel>+107:+107</imagelabel>
>   </seclabel>
>   <qemu:capabilities>
>     <qemu:add capability='blockdev'/>
>     <qemu:add capability='incremental-backup'/>
>   </qemu:capabilities>
> </domain>
> }, log id: 7eb54202
>
>
>
> Here is the engine log when deleting the pvc:
>
> 2021-08-20 17:43:12,964 - cinder.volume.driver - INFO - Driver hasn't
> implemented _init_vendor_properties()
> 2021-08-20 17:43:12,990 - cinderlib-client - INFO - Deleting volume
> '63a64445-1659-4d5f-8847-e7266e64b09e'
> [feefc62f-e7cb-435d-ae21-4b52b53fbdfa]
> 2021-08-20 17:43:28,856 - cinder.volume.drivers.rbd - WARNING - ImageBusy
> error raised while deleting rbd volume. This may have been caused by a
> connection from a client that has crashed and, if so, may be resolved by
> retrying the delete after 30 seconds has elapsed.
> 2021-08-20 17:43:28,900 - cinderlib-client - ERROR - Failure occurred when
> trying to run command 'delete_volume': ImageBusy error raised while
> deleting rbd volume. This may have been caused by a connection from a
> client that has crashed and, if so, may be resolved by retrying the delete
> after 30 seconds has elapsed. [feefc62f-e7cb-435d-ae21-4b52b53fbdfa]
> 2021-08-20 17:43:28,901 - cinder - CRITICAL - Unhandled error
> Traceback (most recent call last):
>   File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py",
> line 1195, in delete_volume
>     _try_remove_volume(client, volume_name)
>   File "/usr/lib/python3.6/site-packages/cinder/utils.py", line 683, in
> _wrapper
>     return r.call(f, *args, **kwargs)
>   File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 409,
> in call
>     do = self.iter(retry_state=retry_state)
>   File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 368,
> in iter
>     raise retry_exc.reraise()
>   File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 186,
> in reraise
>     raise self.last_attempt.result()
>   File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 425, in
> result
>     return self.__get_result()
>   File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 384, in
> __get_result
>     raise self._exception
>   File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 412,
> in call
>     result = fn(*args, **kwargs)
>   File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py",
> line 1190, in _try_remove_volume
>     self.RBDProxy().remove(client.ioctx, volume_name)
>   File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 190, in
> doit
>     result = proxy_call(self._autowrap, f, *args, **kwargs)
>   File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 148, in
> proxy_call
>     rv = execute(f, *args, **kwargs)
>   File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 129, in
> execute
>     six.reraise(c, e, tb)
>   File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise
>     raise value
>   File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 83, in
> tworker
>     rv = meth(*args, **kwargs)
>   File "rbd.pyx", line 767, in rbd.RBD.remove
> rbd.ImageBusy: [errno 16] RBD image is busy (error removing image)
>
> During handling of the above exception, another exception occurred:
>
> Traceback (most recent call last):
>   File "./cinderlib-client.py", line 170, in main
>     args.command(args)
>   File "./cinderlib-client.py", line 218, in delete_volume
>     vol.delete()
>   File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line
> 494, in delete
>     self._raise_with_resource()
>   File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line
> 222, in _raise_with_resource
>     six.reraise(*exc_info)
>   File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise
>     raise value
>   File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line
> 487, in delete
>     self.backend.driver.delete_volume(self._ovo)
>   File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py",
> line 1205, in delete_volume
>     raise exception.VolumeIsBusy(msg, volume_name=volume_name)
> cinder.exception.VolumeIsBusy: ImageBusy error raised while deleting rbd
> volume. This may have been caused by a connection from a client that has
> crashed and, if so, may be resolved by retrying the delete after 30 seconds
> has elapsed.
>
> During handling of the above exception, another exception occurred:
>
> Traceback (most recent call last):
>   File "./cinderlib-client.py", line 390, in <module>
>     sys.exit(main(sys.argv[1:]))
>   File "./cinderlib-client.py", line 176, in main
>     sys.stderr.write(traceback.format_exc(e))
>   File "/usr/lib64/python3.6/traceback.py", line 167, in format_exc
>     return "".join(format_exception(*sys.exc_info(), limit=limit,
> chain=chain))
>   File "/usr/lib64/python3.6/traceback.py", line 121, in format_exception
>     type(value), value, tb, limit=limit).format(chain=chain))
>   File "/usr/lib64/python3.6/traceback.py", line 498, in __init__
>     _seen=_seen)
>   File "/usr/lib64/python3.6/traceback.py", line 509, in __init__
>     capture_locals=capture_locals)
>   File "/usr/lib64/python3.6/traceback.py", line 338, in extract
>     if limit >= 0:
> TypeError: '>=' not supported between instances of 'VolumeIsBusy' and 'int'
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/AFYZX2HJZ3RBJS5P7ENIQ4UIAJGP5MGG/
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DOWSPABBAKHOZNRP4O5RZQDHJUS4CGKM/

Reply via email to