I can confirm that this worked. I had to shut down every single VM then change ownership to vdsm:kvm of the image file then start VM back up.
On Wed, Feb 13, 2019 at 3:08 PM Simone Tiraboschi <[email protected]> wrote: > > > On Wed, Feb 13, 2019 at 8:06 PM Jayme <[email protected]> wrote: > >> >> I might be hitting this bug: >> https://bugzilla.redhat.com/show_bug.cgi?id=1666795 >> > > Yes, you definitively are. > Fixing files ownership on file system side is a valid workaround. > > >> >> On Wed, Feb 13, 2019 at 1:35 PM Jayme <[email protected]> wrote: >> >>> This may be happening because I changed cluster compatibility to 4.3 >>> then immediately after changed data center compatibility to 4.3 (before >>> restarting VMs after cluster compatibility change). If this is the case I >>> can't fix by downgrading the data center compatibility to 4.2 as it won't >>> allow me to do so. What can I do to fix this, any VM I restart will break >>> (I am leaving the others running for now, but there are some down that I >>> can't start). >>> >>> Full error from VDSM: >>> >>> 2019-02-13 13:30:55,465-0400 ERROR (vm/d070ce80) >>> [storage.TaskManager.Task] (Task='d5c8e50a-0a6f-4fe7-be79-fd322b273a1e') >>> Unexpected error (task:875) >>> Traceback (most recent call last): >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line >>> 882, in _run >>> return fn(*args, **kargs) >>> File "<string>", line 2, in prepareImage >>> File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 50, >>> in method >>> ret = func(*args, **kwargs) >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line >>> 3198, in prepareImage >>> legality = dom.produceVolume(imgUUID, volUUID).getLegality() >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/sd.py", line 818, >>> in produceVolume >>> volUUID) >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/glusterVolume.py", >>> line 45, in __init__ >>> volUUID) >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line >>> 800, in __init__ >>> self._manifest = self.manifestClass(repoPath, sdUUID, imgUUID, >>> volUUID) >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/fileVolume.py", >>> line 71, in __init__ >>> volUUID) >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line >>> 86, in __init__ >>> self.validate() >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line >>> 112, in validate >>> self.validateVolumePath() >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/fileVolume.py", >>> line 131, in validateVolumePath >>> raise se.VolumeDoesNotExist(self.volUUID) >>> VolumeDoesNotExist: Volume does not exist: >>> (u'2d6d5f87-ccb0-48ce-b3ac-84495bd12d32',) >>> 2019-02-13 13:30:55,468-0400 ERROR (vm/d070ce80) [storage.Dispatcher] >>> FINISH prepareImage error=Volume does not exist: >>> (u'2d6d5f87-ccb0-48ce-b3ac-84495bd12d32',) (dispatcher:81) >>> 2019-02-13 13:30:55,469-0400 ERROR (vm/d070ce80) [virt.vm] >>> (vmId='d070ce80-e0bc-489d-8ee0-47d5926d5ae2') The vm start process failed >>> (vm:937) >>> Traceback (most recent call last): >>> File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 866, in >>> _startUnderlyingVm >>> self._run() >>> File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 2749, in >>> _run >>> self._devices = self._make_devices() >>> File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 2589, in >>> _make_devices >>> disk_objs = self._perform_host_local_adjustment() >>> File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 2662, in >>> _perform_host_local_adjustment >>> self._preparePathsForDrives(disk_params) >>> File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 1011, in >>> _preparePathsForDrives >>> drive['path'] = self.cif.prepareVolumePath(drive, self.id) >>> File "/usr/lib/python2.7/site-packages/vdsm/clientIF.py", line 415, in >>> prepareVolumePath >>> raise vm.VolumeError(drive) >>> VolumeError: Bad volume specification {'address': {'function': '0x0', >>> 'bus': '0x00', 'domain': '0x0000', 'type': 'pci', 'slot': '0x06'}, >>> 'serial': 'd81a6826-dc46-44db-8de7-405d30e44d57', 'index': 0, 'iface': >>> 'virtio', 'apparentsize': '64293699584', 'specParams': {}, 'cache': 'none', >>> 'imageID': 'd81a6826-dc46-44db-8de7-405d30e44d57', 'truesize': >>> '64293814272', 'type': 'disk', 'domainID': >>> '1f2e9989-9ab3-43d5-971d-568b8feca918', 'reqsize': '0', 'format': 'cow', >>> 'poolID': 'a45e442e-9989-11e8-b0e4-00163e4bf18a', 'device': 'disk', 'path': >>> '/rhev/data-center/a45e442e-9989-11e8-b0e4-00163e4bf18a/1f2e9989-9ab3-43d5-971d-568b8feca918/images/d81a6826-dc46-44db-8de7-405d30e44d57/2d6d5f87-ccb0-48ce-b3ac-84495bd12d32', >>> 'propagateErrors': 'off', 'name': 'vda', 'bootOrder': '1', 'volumeID': >>> '2d6d5f87-ccb0-48ce-b3ac-84495bd12d32', 'diskType': 'file', 'alias': >>> 'ua-d81a6826-dc46-44db-8de7-405d30e44d57', 'discard': False} >>> >>> On Wed, Feb 13, 2019 at 1:19 PM Jayme <[email protected]> wrote: >>> >>>> I may have made matters worse. So I changed to 4.3 compatible cluster >>>> then 4.3 compatible data center. All VMs were marked as requiring a >>>> reboot. I restarted a couple of them and none of them will start up, they >>>> are saying "bad volume specification". The ones running that I did not yet >>>> restart are still running ok. I need to figure out why the VMs aren't >>>> restarting. >>>> >>>> Here is an example from vdsm.log >>>> >>>> olumeError: Bad volume specification {'address': {'function': '0x0', >>>> 'bus': '0x00', 'domain': '0x0000', 'type': 'pci', 'slot': '0x06'}, >>>> 'serial': 'd81a6826-dc46-44db-8de7-405d30e44d57', 'index': 0, 'iface': >>>> 'virtio', 'apparentsize': '64293699584', 'specParams': {}, 'cache': 'none', >>>> 'imageID': 'd81a6826-dc46-44db-8de7-405d30e44d57', 'truesize': >>>> '64293814272', 'type': 'disk', 'domainID': >>>> '1f2e9989-9ab3-43d5-971d-568b8feca918', 'reqsize': '0', 'format': 'cow', >>>> 'poolID': 'a45e442e-9989-11e8-b0e4-00163e4bf18a', 'device': 'disk', 'path': >>>> '/rhev/data-center/a45e442e-9989-11e8-b0e4-00163e4bf18a/1f2e9989-9ab3-43d5-971d-568b8feca918/images/d81a6826-dc46-44db-8de7-405d30e44d57/2d6d5f87-ccb0-48ce-b3ac-84495bd12d32', >>>> 'propagateErrors': 'off', 'name': 'vda', 'bootOrder': '1', 'volumeID': >>>> '2d6d5f87-ccb0-48ce-b3ac-84495bd12d32', 'diskType': 'file', 'alias': >>>> 'ua-d81a6826-dc46-44db-8de7-405d30e44d57', 'discard': False} >>>> >>>> On Wed, Feb 13, 2019 at 1:01 PM Jayme <[email protected]> wrote: >>>> >>>>> I think I just figured out what I was doing wrong. On edit cluster >>>>> screen I was changing both the CPU type and cluster level 4.3. I tried it >>>>> again by switching to the new CPU type first (leaving cluster on 4.2) then >>>>> saving, then going back in and switching compat level to 4.3. It appears >>>>> that you need to do this in two steps for it to work. >>>>> >>>>> >>>>> >>>>> On Wed, Feb 13, 2019 at 12:57 PM Jayme <[email protected]> wrote: >>>>> >>>>>> Hmm interesting, I wonder how you were able to switch from >>>>>> SandyBridge IBRS to SandyBridge IBRS SSBD. I just attempted the same in >>>>>> both regular mode and in global maintenance mode and it won't allow me >>>>>> to, >>>>>> it says that all hosts have to be in maintenance mode (screenshots >>>>>> attached). Are you also running HCI/Gluster setup? >>>>>> >>>>>> >>>>>> >>>>>> On Wed, Feb 13, 2019 at 12:44 PM Ron Jerome <[email protected]> >>>>>> wrote: >>>>>> >>>>>>> > Environment setup: >>>>>>> > >>>>>>> > 3 Host HCI GlusterFS setup. Identical hosts, Dell R720s w/ Intel >>>>>>> E5-2690 >>>>>>> > CPUs >>>>>>> > >>>>>>> > 1 default data center (4.2 compat) >>>>>>> > 1 default cluster (4.2 compat) >>>>>>> > >>>>>>> > Situation: I recently upgraded my three node HCI cluster from >>>>>>> Ovirt 4.2 to >>>>>>> > 4.3. I did so by first updating the engine to 4.3 then upgrading >>>>>>> each >>>>>>> > ovirt-node host to 4.3 and rebooting. >>>>>>> > >>>>>>> > Currently engine and all hosts are running 4.3 and all is working >>>>>>> fine. >>>>>>> > >>>>>>> > To complete the upgrade I need to update cluster compatibility to >>>>>>> 4.3 and >>>>>>> > then data centre to 4.3. This is where I am stuck. >>>>>>> > >>>>>>> > The CPU type on cluster is "Intel SandyBridge IBRS Family". This >>>>>>> option is >>>>>>> > no longer available if I select 4.3 compatibility. Any other >>>>>>> option chosen >>>>>>> > such as SandyBridge IBRS SSBD will not allow me to switch to 4.3 >>>>>>> as all >>>>>>> > hosts must be in maintenance mode (which is not possible w/ self >>>>>>> hosted >>>>>>> > engine). >>>>>>> > >>>>>>> > I saw another post about this where someone else followed steps to >>>>>>> create a >>>>>>> > second cluster on 4.3 with new CPU type then move one host to it, >>>>>>> start >>>>>>> > engine on it then perform other steps to eventually get to 4.3 >>>>>>> > compatibility. >>>>>>> > >>>>>>> >>>>>>> I have the exact same hardware configuration and was able to change >>>>>>> to "SandyBridge IBRS SSBD" without creating a new cluster. How I made >>>>>>> that >>>>>>> happen, I'm not so sure, but the cluster may have been in "Global >>>>>>> Maintenance" mode when I changed it. >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Users mailing list -- [email protected] >>>>>>> To unsubscribe send an email to [email protected] >>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>> oVirt Code of Conduct: >>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>> List Archives: >>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/5B3TAXKO7IBTWRVNF2K4II472TDISO6P/ >>>>>>> >>>>>> _______________________________________________ >> Users mailing list -- [email protected] >> To unsubscribe send an email to [email protected] >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> List Archives: >> https://lists.ovirt.org/archives/list/[email protected]/message/PK7IR27DGLZRZSXVZEN66FL4O377GOHT/ >> >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/BPWTQIUEFH2IU3R6NQOGJDEWHVJWT3BP/

