On Tue, Jan 2, 2018 at 1:18 AM, Gianluca Cecchi <gianluca.cec...@gmail.com> wrote:
> Hello, > a system upgraded from 4.1.7 (with libgfapi not enabled) to 4.2. > 3 hosts in a HC configuration > > Now I try to enable libgfapi: > > Before a CentOS 6 VM booted with a qemu-kvm line of type: > > -drive file=/rhev/data-center/mnt/glusterSD/ovirt01.localdomain. > local:data/190f4096-003e-4908-825a-6c231e60276d/images/ > 02731d5e-c222-4697-8f1f-d26a6a23ec79/1836df76-835b- > 4625-9ce8-0856176dc30c,format=raw,if=none,id=drive-virtio- > disk0,serial=02731d5e-c222-4697-8f1f-d26a6a23ec79,cache= > none,werror=stop,rerror=stop,aio=thread > > Shutdown VM named centos6 > > Setup engine > root@ovengine log]# engine-config -s LibgfApiSupported=true > Please select a version: > 1. 3.6 > 2. 4.0 > 3. 4.1 > 4. 4.2 > 4 > [root@ovengine log]# engine-config -g LibgfApiSupported > LibgfApiSupported: false version: 3.6 > LibgfApiSupported: false version: 4.0 > LibgfApiSupported: false version: 4.1 > LibgfApiSupported: true version: 4.2 > > Restart engine > > [root@ovengine log]# systemctl restart ovirt-engine > [root@ovengine log]# > > reconnect to web admin portal > > Power on the centos6 VM > I get "Failed to run VM" on all the 3 configured hosts > > Jan 1, 2018, 11:53:35 PM Failed to run VM centos6 (User: > admin@internal-authz). > Jan 1, 2018, 11:53:35 PM Failed to run VM centos6 on Host > ovirt02.localdomain.local. > Jan 1, 2018, 11:53:35 PM Failed to run VM centos6 on Host > ovirt03.localdomain.local. > Jan 1, 2018, 11:53:35 PM Failed to run VM centos6 on Host > ovirt01.localdomain.local. > > In engine.log > 2018-01-01 23:53:34,996+01 ERROR [org.ovirt.engine.core. > vdsbroker.vdsbroker.CreateBrokerVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-2885) > [8d7c68c6-b236-4e76-b7b2-f000e2b07425] Failed in 'CreateBrokerVDS' > method, for vds: 'ovirt01.localdomain.local'; host: > 'ovirt01.localdomain.local': 1 > 2018-01-01 23:53:34,996+01 ERROR [org.ovirt.engine.core. > vdsbroker.vdsbroker.CreateBrokerVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-2885) > [8d7c68c6-b236-4e76-b7b2-f000e2b07425] Command > 'CreateBrokerVDSCommand(HostName > = ovirt01.localdomain.local, CreateVDSCommandParameters:{ > hostId='e5079118-1147-469e-876f-e20013276ece', > vmId='64da5593-1022-4f66-ae3f-b273deda4c22', vm='VM [centos6]'})' > execution failed: 1 > 2018-01-01 23:53:34,996+01 INFO [org.ovirt.engine.core. > vdsbroker.vdsbroker.CreateBrokerVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-2885) > [8d7c68c6-b236-4e76-b7b2-f000e2b07425] FINISH, CreateBrokerVDSCommand, > log id: e3bbe56 > 2018-01-01 23:53:34,996+01 ERROR > [org.ovirt.engine.core.vdsbroker.CreateVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-2885) > [8d7c68c6-b236-4e76-b7b2-f000e2b07425] > Failed to create VM: 1 > 2018-01-01 23:53:34,997+01 ERROR > [org.ovirt.engine.core.vdsbroker.CreateVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-2885) > [8d7c68c6-b236-4e76-b7b2-f000e2b07425] > Command 'CreateVDSCommand( CreateVDSCommandParameters:{ > hostId='e5079118-1147-469e-876f-e20013276ece', > vmId='64da5593-1022-4f66-ae3f-b273deda4c22', vm='VM [centos6]'})' > execution failed: java.lang.ArrayIndexOutOfBoundsException: 1 > 2018-01-01 23:53:34,997+01 INFO > [org.ovirt.engine.core.vdsbroker.CreateVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-2885) > [8d7c68c6-b236-4e76-b7b2-f000e2b07425] > FINISH, CreateVDSCommand, return: Down, log id: ab299ce > 2018-01-01 23:53:34,997+01 WARN [org.ovirt.engine.core.bll.RunVmCommand] > (EE-ManagedThreadFactory-engine-Thread-2885) > [8d7c68c6-b236-4e76-b7b2-f000e2b07425] > Failed to run VM 'centos6': EngineException: java.lang.RuntimeException: > java.lang.ArrayIndexOutOfBoundsException: 1 (Failed with error ENGINE and > code 5001) > > All engine.log file here: > https://drive.google.com/file/d/1UZ9dWnGrBaFVnfx1E_ > Ch52CtYDDtzT3p/view?usp=sharing > > The VM fails to start on all 3 hosts, but I don't see particular error on > them; eg on ovirt01 vdsm.log.1.xz here: > https://drive.google.com/file/d/1yIlKtRtvftJVzWNlzV3WhJ3DaP4ks > Qvw/view?usp=sharing > I'm indeed missing even the create command on the VDSM... strange. Can you file a bug about this? Y. > > The domain where the VM disk is on seems ok; > [root@ovirt01 vdsm]# gluster volume info data > > Volume Name: data > Type: Replicate > Volume ID: 2238c6db-48c5-4071-8929-879cedcf39bf > Status: Started > Snapshot Count: 0 > Number of Bricks: 1 x (2 + 1) = 3 > Transport-type: tcp > Bricks: > Brick1: ovirt01.localdomain.local:/gluster/brick2/data > Brick2: ovirt02.localdomain.local:/gluster/brick2/data > Brick3: ovirt03.localdomain.local:/gluster/brick2/data (arbiter) > Options Reconfigured: > transport.address-family: inet > performance.readdir-ahead: on > performance.quick-read: off > performance.read-ahead: off > performance.io-cache: off > performance.stat-prefetch: off > cluster.eager-lock: enable > network.remote-dio: off > cluster.quorum-type: auto > cluster.server-quorum-type: server > storage.owner-uid: 36 > storage.owner-gid: 36 > features.shard: on > features.shard-block-size: 512MB > performance.low-prio-threads: 32 > cluster.data-self-heal-algorithm: full > cluster.locking-scheme: granular > cluster.shd-wait-qlength: 10000 > cluster.shd-max-threads: 6 > network.ping-timeout: 30 > user.cifs: off > nfs.disable: on > performance.strict-o-direct: on > [root@ovirt01 vdsm]# > > [root@ovirt01 vdsm]# gluster volume heal data info > Brick ovirt01.localdomain.local:/gluster/brick2/data > Status: Connected > Number of entries: 0 > > Brick ovirt02.localdomain.local:/gluster/brick2/data > Status: Connected > Number of entries: 0 > > Brick ovirt03.localdomain.local:/gluster/brick2/data > Status: Connected > Number of entries: 0 > > [root@ovirt01 vdsm]# > > I followed here: > https://www.ovirt.org/develop/release-management/features/ > storage/glusterfs-storage-domain/ > > Any other action to do at host side? > The cluster and DC are already at 4.2 compatibility version. > > > Thanks > Gianluca > > > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users