Redirecting to /bin/systemctl reconfigure vdsmd.service Unknown operation 'reconfigure'.
... seems to me, I should get rid of the ovirt-node iso installation and move to a rpm based install? Thanks, Oliver -----Ursprüngliche Nachricht----- Von: Dafna Ron [mailto:[email protected]] Gesendet: Freitag, 03. Jänner 2014 15:51 An: Albl, Oliver Cc: [email protected] Betreff: Re: AW: AW: AW: AW: AW: [Users] Host cannot access storage domains can you run: service vdsmd reconfigure on the second host? On 01/03/2014 02:43 PM, Albl, Oliver wrote: > Dafna, > > yes, the VM starts on the first node, the issues are on the second node > only. > > /etc/libvirt/qemu-sanlock.conf is identical on on both nodes: > > auto_disk_leases=0 > require_lease_for_disks=0 > > yum updates reports "Using yum is not supported"... > > Thanks, > Oliver > > -----Ursprüngliche Nachricht----- > Von: Dafna Ron [mailto:[email protected]] > Gesendet: Freitag, 03. Jänner 2014 15:39 > An: Albl, Oliver > Cc: [email protected] > Betreff: Re: AW: AW: AW: AW: [Users] Host cannot access storage > domains > > ok, let's try to zoom in on the issue... > can you run vm's on the first host or do you have issues only on the second > host you added? > can you run on both hosts? > # egrep -v ^# /etc/libvirt/qemu-sanlock.conf > > can you run yum update on one of the hosts and see if there are newer > packages? > > Thanks, > > Dafna > > On 01/03/2014 02:30 PM, Albl, Oliver wrote: >> I installed both hosts using the oVirt Node ISO image: >> >> OS Version: oVirt Node - 3.0.3 - 1.1.fc19 Kernel Version: 3.11.9 - >> 200.fc19.x86_64 KVM Version: 1.6.1 - 2.fc19 LIBVIRT Version: >> libvirt-1.1.3.1-2.fc19 VDSM Version: vdsm-4.13.0-11.fc19 >> >> Thanks, >> Oliver >> -----Ursprüngliche Nachricht----- >> Von: Dafna Ron [mailto:[email protected]] >> Gesendet: Freitag, 03. Jänner 2014 15:24 >> An: Albl, Oliver >> Cc: [email protected] >> Betreff: Re: AW: AW: AW: [Users] Host cannot access storage domains >> >> ignore the link :) >> >> so searching for this error I hit an old bug and it seemed to be an issue >> between libvirt/sanlock. >> >> https://bugzilla.redhat.com/show_bug.cgi?id=828633 >> >> are you using latest packages? >> >> >> >> >> On 01/03/2014 02:15 PM, Albl, Oliver wrote: >>> Dafna, >>> >>> Libvirtd.log shows no errors, but VM log shows the following: >>> >>> 2014-01-03 13:52:11.296+0000: starting up LC_ALL=C >>> PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin >>> QEMU_AUDIO_DRV=spice /usr/bin/qemu-kvm -name OATEST2 -S -machine >>> pc-1.0,accel=kvm,usb=off -cpu SandyBridge -m 1024 -realtime >>> mlock=off -smp 1,sockets=1,cores=1,threads=1 -uuid >>> d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 -smbios >>> type=1,manufacturer=oVirt,product=oVirt >>> Node,version=3.0.3-1.1.fc19,serial=30313436-3631-5A43-4A33-3332304C3 >>> 8 >>> 4 >>> C,uuid=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 -no-user-config >>> -nodefaults -chardev >>> socket,id=charmonitor,path=/var/lib/libvirt/qemu/OATEST2.monitor,ser >>> v e r,nowait -mon chardev=charmonitor,id=monitor,mode=control -rtc >>> base=2014-01-03T13:52:11,driftfix=slew -no-shutdown -device >>> piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device >>> virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device >>> virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x3 -drive >>> if=none,id=drive-ide0-1-0,readonly=on,format=raw,serial= -device >>> ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -drive >>> file=/rhev/data-center/mnt/blockSD/7841a1c0-181a-4d43-9a25-b707accb5 >>> c >>> 4 >>> b/images/de7ca992-b1c1-4cb8-9470-2494304c9b69/cbf1f376-23e8-40f3-838 >>> 7 >>> - >>> ed299ee62607,if=none,id=drive-virtio-disk0,format=raw,serial=de7ca99 >>> 2 >>> - >>> b1c1-4cb8-9470-2494304c9b69,cache=none,werror=stop,rerror=stop,aio=n >>> a >>> t >>> ive -device >>> virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0, >>> i >>> d >>> =virtio-disk0,bootindex=1 -chardev >>> socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/d2bddcdb- >>> a >>> 2 c8-4c77-b0cf-b83fa3c2a0b6.com.redhat.rhevm.vdsm,server,nowait >>> -device >>> virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=cha >>> n n el0,name=com.redhat.rhevm.vdsm -chardev >>> socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/d2bddcdb- >>> a >>> 2 c8-4c77-b0cf-b83fa3c2a0b6.org.qemu.guest_agent.0,server,nowait >>> -device >>> virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=cha >>> n >>> n >>> el1,name=org.qemu.guest_agent.0 -chardev >>> spicevmc,id=charchannel2,name=vdagent -device >>> virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=cha >>> n >>> n >>> el2,name=com.redhat.spice.0 -spice >>> tls-port=5900,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,tls-channe >>> l >>> = >>> main,tls-channel=display,tls-channel=inputs,tls-channel=cursor,tls-c >>> h >>> a >>> nnel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=u >>> s b redir,seamless-migration=on -k en-us -device >>> qxl-vga,id=video0,ram_size=67108864,vram_size=33554432,bus=pci.0,add >>> r >>> = >>> 0x2 -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x5 >>> libvirt: Lock Driver error : unsupported configuration: Read/write, >>> exclusive access, disks were present, but no leases specified >>> 2014-01-03 13:52:11.306+0000: shutting down >>> >>> Not sure what you mean with this >>> http://forums.opensuse.org/english/get-technical-help-here/virtualization/492483-cannot-start-libvert-kvm-guests-after-update-tumbleweed.html. >>> Do you want me to update libvirt with these repos on the oVirt-Node based >>> installation? >>> >>> Thanks, >>> Oliver >>> -----Ursprüngliche Nachricht----- >>> Von: Dafna Ron [mailto:[email protected]] >>> Gesendet: Freitag, 03. Jänner 2014 15:10 >>> An: Albl, Oliver >>> Cc: [email protected] >>> Betreff: Re: AW: AW: [Users] Host cannot access storage domains >>> >>> actually, looking at this again, it's a libvirt error and it can be related >>> to selinux or sasl. >>> can you also, look at libvirt log and the vm log under /var/log/libvirt? >>> >>> On 01/03/2014 02:00 PM, Albl, Oliver wrote: >>>> Dafna, >>>> >>>> please find the logs below: >>>> >>>> ERRORs in vdsm.log on host02: >>>> >>>> Thread-61::ERROR::2014-01-03 >>>> 13:51:48,956::sdc::137::Storage.StorageDomainCache::(_findDomain) >>>> looking for unfetched domain f404398a-97f9-474c-af2c-e8887f53f688 >>>> Thread-61::ERROR::2014-01-03 >>>> 13:51:48,959::sdc::154::Storage.StorageDomainCache::(_findUnfetched >>>> D >>>> o >>>> m >>>> ain) looking for domain f404398a-97f9-474c-af2c-e8887f53f688 >>>> Thread-323::ERROR::2014-01-03 >>>> 13:52:11,527::vm::2132::vm.Vm::(_startUnderlyingVm) >>>> vmId=`d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6`::The vm start process failed >>>> Traceback (most recent call last): >>>> File "/usr/share/vdsm/vm.py", line 2092, in _startUnderlyingVm >>>> self._run() >>>> File "/usr/share/vdsm/vm.py", line 2959, in _run >>>> self._connection.createXML(domxml, flags), >>>> File "/usr/lib64/python2.7/site-packages/vdsm/libvirtconnection.py", >>>> line 76, in wrapper >>>> ret = f(*args, **kwargs) >>>> File "/usr/lib64/python2.7/site-packages/libvirt.py", line >>>> 2920, in createXML >>>> libvirtError: Child quit during startup handshake: Input/output >>>> error >>>> Thread-60::ERROR::2014-01-03 >>>> 13:52:23,111::sdc::137::Storage.StorageDomainCache::(_findDomain) >>>> looking for unfetched domain 52cf84ce-6eda-4337-8c94-491d94f5a18d >>>> Thread-60::ERROR::2014-01-03 >>>> 13:52:23,111::sdc::154::Storage.StorageDomainCache::(_findUnfetched >>>> D >>>> o >>>> m >>>> ain) looking for domain 52cf84ce-6eda-4337-8c94-491d94f5a18d >>>> Thread-62::ERROR::2014-01-03 >>>> 13:52:26,353::sdc::137::Storage.StorageDomainCache::(_findDomain) >>>> looking for unfetched domain 7841a1c0-181a-4d43-9a25-b707accb5c4b >>>> Thread-62::ERROR::2014-01-03 >>>> 13:52:26,355::sdc::154::Storage.StorageDomainCache::(_findUnfetched >>>> D >>>> o >>>> m >>>> ain) looking for domain 7841a1c0-181a-4d43-9a25-b707accb5c4b >>>> >>>> engine.log: >>>> >>>> 2014-01-03 14:52:06,976 INFO >>>> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] START, >>>> IsVmDuringInitiatingVDSCommand( vmId = >>>> d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6), log id: 5940cf72 >>>> 2014-01-03 14:52:06,976 INFO >>>> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH, >>>> IsVmDuringInitiatingVDSCommand, return: false, log id: 5940cf72 >>>> 2014-01-03 14:52:07,057 INFO >>>> [org.ovirt.engine.core.bll.RunVmOnceCommand] >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] Running command: RunVmOnceCommand >>>> internal: false. >>>> Entities affected : ID: d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 Type: VM, >>>> ID: >>>> d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 Type: VM >>>> 2014-01-03 14:52:07,151 INFO >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.IsoPrefixVDSCommand] >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] START, >>>> IsoPrefixVDSCommand(HostName = host02, HostId = >>>> 6dc7fac6-149e-4445-ace1-3c334a24d52a, >>>> storagePoolId=b33d1793-252b-44ac-9685-3fe56b83c4c9), log id: >>>> 1705b611 >>>> 2014-01-03 14:52:07,152 INFO >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.IsoPrefixVDSCommand] >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH, IsoPrefixVDSCommand, return: >>>> /rhev/data-center/mnt/vmmgmt:_var_lib_exports_iso/f74f052e-0dc6-456 >>>> d >>>> - a f95-248c2227c2e5/images/11111111-1111-1111-1111-111111111111, >>>> log >>>> id: >>>> 1705b611 >>>> 2014-01-03 14:52:07,170 INFO >>>> [org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand] >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] START, >>>> CreateVmVDSCommand(HostName = host02, HostId = >>>> 6dc7fac6-149e-4445-ace1-3c334a24d52a, >>>> vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6, vm=VM [TEST2]), log id: >>>> 27b504de >>>> 2014-01-03 14:52:07,190 INFO >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand] >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] START, CreateVDSCommand(HostName >>>> = host02, HostId = 6dc7fac6-149e-4445-ace1-3c334a24d52a, >>>> vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6, vm=VM [TEST2]), log id: >>>> 6ad0220 >>>> 2014-01-03 14:52:08,472 INFO >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand] >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] >>>> org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand >>>> spiceSslCipherSuite=DEFAULT,memSize=1024,kvmEnable=true,smp=1,vmTyp >>>> e >>>> = >>>> k >>>> vm,emulatedMachine=pc-1.0,keyboardLayout=en-us,memGuaranteedSize=10 >>>> 2 >>>> 4 >>>> , >>>> pitReinjection=false,nice=0,display=qxl,smartcardEnable=false,smpCo >>>> r >>>> e >>>> s >>>> PerSocket=1,spiceSecureChannels=smain,sinputs,scursor,splayback,sre >>>> c >>>> o >>>> r >>>> d,sdisplay,susbredir,ssmartcard,timeOffset=0,transparentHugePages=t >>>> r >>>> u >>>> e >>>> ,vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6,devices=[Ljava.util.Hash >>>> M >>>> a >>>> p >>>> ;@3692311a,acpiEnable=true,vmName=TEST2,cpuType=SandyBridge,custom= >>>> { >>>> } >>>> 2014-01-03 14:52:08,476 INFO >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand] >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH, CreateVDSCommand, log id: >>>> 6ad0220 >>>> 2014-01-03 14:52:08,484 INFO >>>> [org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand] >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH, CreateVmVDSCommand, return: >>>> WaitForLaunch, log id: 27b504de >>>> 2014-01-03 14:52:08,497 INFO >>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] Correlation ID: 2ab5cd2, Job ID: >>>> 2913133b-1301-484e-9887-b110841c8078, Call Stack: null, Custom Event ID: >>>> -1, Message: VM TEST2 was started by oliver.albl (Host: host02). >>>> 2014-01-03 14:52:14,728 INFO >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] START, >>>> DestroyVDSCommand(HostName = host02, HostId = >>>> 6dc7fac6-149e-4445-ace1-3c334a24d52a, >>>> vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6, force=false, >>>> secondsToWait=0, gracefully=false), log id: 6a95ffd5 >>>> 2014-01-03 14:52:15,783 INFO >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] FINISH, >>>> DestroyVDSCommand, log id: 6a95ffd5 >>>> 2014-01-03 14:52:15,804 INFO >>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] Correlation ID: null, Call >>>> Stack: null, Custom Event ID: -1, Message: VM TEST2 is down. Exit message: >>>> Child quit during startup handshake: Input/output error. >>>> 2014-01-03 14:52:15,805 INFO >>>> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] Running on vds during >>>> rerun failed vm: null >>>> 2014-01-03 14:52:15,805 INFO >>>> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] vm TEST2 running in db >>>> and not running in vds - add to rerun treatment. vds host02 >>>> 2014-01-03 14:52:15,808 ERROR >>>> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] Rerun vm >>>> d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6. Called from vds host02 >>>> 2014-01-03 14:52:15,810 INFO >>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>> (pool-6-thread-40) [24696b3e] Correlation ID: 2ab5cd2, Job ID: >>>> 2913133b-1301-484e-9887-b110841c8078, Call Stack: null, Custom Event ID: >>>> -1, Message: Failed to run VM TEST2 on Host host02. >>>> 2014-01-03 14:52:15,823 INFO >>>> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] >>>> (pool-6-thread-40) [24696b3e] START, >>>> IsVmDuringInitiatingVDSCommand( vmId = >>>> d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6), log id: 35e1eec >>>> 2014-01-03 14:52:15,824 INFO >>>> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] >>>> (pool-6-thread-40) [24696b3e] FINISH, >>>> IsVmDuringInitiatingVDSCommand, >>>> return: false, log id: 35e1eec >>>> 2014-01-03 14:52:15,858 WARN >>>> [org.ovirt.engine.core.bll.RunVmOnceCommand] (pool-6-thread-40) >>>> [24696b3e] CanDoAction of action RunVmOnce failed. >>>> Reasons:VAR__ACTION__RUN,VAR__TYPE__VM,VAR__ACTION__RUN,VAR__TYPE__ >>>> V >>>> M >>>> , >>>> SCHEDULING_ALL_HOSTS_FILTERED_OUT >>>> 2014-01-03 14:52:15,862 INFO >>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>> (pool-6-thread-40) [24696b3e] Correlation ID: 2ab5cd2, Job ID: >>>> 2913133b-1301-484e-9887-b110841c8078, Call Stack: null, Custom Event ID: >>>> -1, Message: Failed to run VM TEST2 (User: oliver.albl). >>>> >>>> Thanks, >>>> Oliver >>>> -----Ursprüngliche Nachricht----- >>>> Von: Dafna Ron [mailto:[email protected]] >>>> Gesendet: Freitag, 03. Jänner 2014 14:51 >>>> An: Albl, Oliver >>>> Cc: [email protected] >>>> Betreff: Re: AW: [Users] Host cannot access storage domains >>>> >>>> Thanks for reporting the issue :) >>>> >>>> As for the vm, can you please find the error in vdsm.log and in engine and >>>> paste it? >>>> >>>> Thanks, >>>> >>>> Dafna >>>> >>>> >>>> On 01/03/2014 01:49 PM, Albl, Oliver wrote: >>>>> Dafna, >>>>> >>>>> you were right, it seems to be a caching issue. Rebooting the host >>>>> did the job: >>>>> >>>>> Before Reboot: >>>>> >>>>> [root@host01 log]# vdsClient -s 0 getStorageDomainsList >>>>> 52cf84ce-6eda-4337-8c94-491d94f5a18d >>>>> f404398a-97f9-474c-af2c-e8887f53f688 >>>>> 7841a1c0-181a-4d43-9a25-b707accb5c4b >>>>> >>>>> [root@host02 log]# vdsClient -s 0 getStorageDomainsList >>>>> 52cf84ce-6eda-4337-8c94-491d94f5a18d >>>>> f404398a-97f9-474c-af2c-e8887f53f688 >>>>> 7841a1c0-181a-4d43-9a25-b707accb5c4b >>>>> 925ee53a-69b5-440f-b145-138ada5b452e >>>>> >>>>> After Reboot: >>>>> >>>>> [root@host02 admin]# vdsClient -s 0 getStorageDomainsList >>>>> 52cf84ce-6eda-4337-8c94-491d94f5a18d >>>>> f404398a-97f9-474c-af2c-e8887f53f688 >>>>> 7841a1c0-181a-4d43-9a25-b707accb5c4b >>>>> >>>>> So now I have both hosts up and running but when I try to start a VM on >>>>> the second host, I receive the following messages in the events pane: >>>>> >>>>> VM TEST2 was started by oliver.albl (Host: host02) VM TEST2 is down. >>>>> Exit message: Child quit during startup handshake: Input/output error. >>>>> >>>>> Thanks again for your help! >>>>> Oliver >>>>> >>>>> -----Ursprüngliche Nachricht----- >>>>> Von: Dafna Ron [mailto:[email protected]] >>>>> Gesendet: Freitag, 03. Jänner 2014 14:22 >>>>> An: Albl, Oliver >>>>> Cc: [email protected] >>>>> Betreff: Re: [Users] Host cannot access storage domains >>>>> >>>>> yes, please attach the vdsm log >>>>> also, can you run vdsClient 0 getStorageDomainsList and vdsClient 0 >>>>> getDeviceList on both hosts? >>>>> >>>>> It might be a cache issue, so can you please restart the host and if it >>>>> helps attach output before and after the reboot? >>>>> >>>>> Thanks, >>>>> >>>>> Dafna >>>>> >>>>> >>>>> On 01/03/2014 01:12 PM, Albl, Oliver wrote: >>>>>> Hi, >>>>>> >>>>>> I am starting with oVirt 3.3.2 and I have an issue adding a host >>>>>> to a cluster. >>>>>> >>>>>> I am using oVirt Engine Version 3.3.2-1.el6 >>>>>> >>>>>> There is a cluster with one host (installed with oVirt Node - >>>>>> 3.0.3 >>>>>> - >>>>>> 1.1.fc19 ISO image) up and running. >>>>>> >>>>>> I installed a second host using the same ISO image. >>>>>> >>>>>> I approved the host in the cluster. >>>>>> >>>>>> When I try to activate the second host, I receive the following >>>>>> messages in the events pane: >>>>>> >>>>>> State was set to Up for host host02. >>>>>> >>>>>> Host host02 reports about one of the Active Storage Domains as >>>>>> Problematic. >>>>>> >>>>>> Host host02 cannot access one of the Storage Domains attached to >>>>>> the Data Center Test303. Stetting Host state to Non-Operational. >>>>>> >>>>>> Failed to connect Host host02 to Storage Pool Test303 >>>>>> >>>>>> There are 3 FC Storage Domains configured and visible to both hosts. >>>>>> >>>>>> multipath -ll shows all LUNs on both hosts. >>>>>> >>>>>> The engine.log reports the following about every five minutes: >>>>>> >>>>>> 2014-01-03 13:50:15,408 ERROR >>>>>> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand] >>>>>> (pool-6-thread-44) Domain 7841a1c0-181a-4d43-9a25-b707accb5c4b: >>>>>> LUN_105 check timeot 69.7 is too big >>>>>> >>>>>> 2014-01-03 13:50:15,409 ERROR >>>>>> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand] >>>>>> (pool-6-thread-44) Domain 52cf84ce-6eda-4337-8c94-491d94f5a18d: >>>>>> LUN_103 check timeot 59.6 is too big >>>>>> >>>>>> 2014-01-03 13:50:15,410 ERROR >>>>>> [org.ovirt.engine.core.bll.InitVdsOnUpCommand] (pool-6-thread-44) >>>>>> Storage Domain LUN_105 of pool Test303 is in problem in host >>>>>> host02 >>>>>> >>>>>> 2014-01-03 13:50:15,411 ERROR >>>>>> [org.ovirt.engine.core.bll.InitVdsOnUpCommand] (pool-6-thread-44) >>>>>> Storage Domain LUN_103 of pool Test030 is in problem in host >>>>>> host02 >>>>>> >>>>>> Please let me know if there are any log files I should attach. >>>>>> >>>>>> Thank you for your help! >>>>>> >>>>>> All the best, >>>>>> >>>>>> Oliver Albl >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Users mailing list >>>>>> [email protected] >>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>> -- >>>>> Dafna Ron >>>> -- >>>> Dafna Ron >>> -- >>> Dafna Ron >> -- >> Dafna Ron >> > > -- > Dafna Ron -- Dafna Ron _______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

