Hi, all.

I currently operates several VMs on cockpit-machines, and I want to move those 
VMs to oVirt.

When I tried oVirt 4.5 last month, installation failed and I tried oVirt 
4.3.10, it works. I moved several qcow2 disks on hyper-converged gluster and I 
attached them on VMs and some disks boot and operate well, but some cannot boot 
due to XFS issues - https://access.redhat.com/solutions/4582401.

So, I installed newest ovirt-engine on CentOS 8 Stream. I installed ovirt-node 
via ovirt-node-ng-installer-4.5.0-2022051313.el8.iso. I applied 
vdsm-gluster-cli patch manually on ovirt nodes, that does not applied on 
ovirt-ng iso. 
https://github.com/oVirt/vdsm/commit/2da10debcea0b9f2e235dc18c2d567e6aa4254c0.

Then, I encountered the following boot error when I launched several VMs that 
worked on cockpit-machines and oVirt 4.3.10.

BdsDxe: failed to load Boot0001 "UEFI Misc Device" from PciRoot 
(0x0)/Pci(0x2,0x3)/Pci(0x0,0x0): Not Found
BdsDxe: No bootable option or device was found
BdsDxe: Press any key to enter the Boot Manager Menu.

Is there any patch that I should apply?

Best Regards
Minseok Jang.

----
PS.
FIY, I leave my current status.

I currently running 4 machines
engine.ovirt1.int - VM
n1~n3.ovirt1.int - HP DL380 Gen9, hyper-converged gluster applied.

engine.ovirt1.int # dnf info ovirt-engine.noarch
Installed Packages
Name         : ovirt-engine
Version      : 4.5.0.8

n1.ovirt1.int # dnf info ovirt-release-host-node
Installed Packages
Name         : ovirt-release-host-node
Version      : 4.5.0.2

n1.ovirt1.int # lscpu
Architecture:        x86_64
CPU(s):              24
On-line CPU(s) list: 0-23
Model name:          Intel(R) Xeon(R) CPU E5-2643 v4 @ 3.40GHz

n1.ovirt1.int # lsblk
NAME                                                     MAJ:MIN RM   SIZE RO 
TYPE MOUNTPOINT
sda                                                        8:0    0 223.6G  0 
disk
...
└─sda3                                                     8:3    0   222G  0 
part
  ├─onn-pool00_tmeta                                     253:0    0     1G  0 
lvm
  │ └─onn-pool00-tpool                                   253:2    0   173G  0 
lvm
...
  ├─onn-pool00_tdata                                     253:1    0   173G  0 
lvm
  │ └─onn-pool00-tpool                                   253:2    0   173G  0 
lvm
  ...
  └─onn-swap                                             253:4    0     4G  0 
lvm  [SWAP]
sdb                                                        8:16   0   3.5T  0 
disk
├─gluster_vg_sdb-gluster_thinpool_gluster_vg_sdb_tmeta   253:5    0  15.9G  0 
lvm
│ └─gluster_vg_sdb-gluster_thinpool_gluster_vg_sdb-tpool 253:7    0   3.5T  0 
lvm
│   ├─gluster_vg_sdb-gluster_thinpool_gluster_vg_sdb     253:8    0   3.5T  1 
lvm
│   └─gluster_vg_sdb-gluster_lv_vmstore1                 253:10   0   3.2T  0 
lvm  /gluster_bricks/vmstore1
└─gluster_vg_sdb-gluster_thinpool_gluster_vg_sdb_tdata   253:6    0   3.5T  0 
lvm
  └─gluster_vg_sdb-gluster_thinpool_gluster_vg_sdb-tpool 253:7    0   3.5T  0 
lvm
    ├─gluster_vg_sdb-gluster_thinpool_gluster_vg_sdb     253:8    0   3.5T  1 
lvm
    └─gluster_vg_sdb-gluster_lv_vmstore1                 253:10   0   3.2T  0 
lvm  /gluster_bricks/vmstore1

# gluster volume info vmstore1
Volume Name: vmstore1
Type: Replicate
Status: Started
Number of Bricks: 1 x 3 = 3
Bricks:
Brick1: n1.ovirt1.int:/gluster_bricks/vmstore1/vmstore1
Brick2: n2.ovirt1.int:/gluster_bricks/vmstore1/vmstore1
Brick3: n3.ovirt1.int:/gluster_bricks/vmstore1/vmstore1
...
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/IC5FIL6CRJ62J2DOWCHWITW2OFHA7SWG/

Reply via email to