On Fri, Apr 29, 2016 at 4:44 AM, Luiz Claudio Prazeres Goncalves <luiz...@gmail.com> wrote: > Hi Simone, I was reviewing the changelog of 3.6.6, on the link below, but i > was not able to find the bug (https://bugzilla.redhat.com/1327516) as fixed > on the list. According to Bugzilla the target is really 3.6.6, so what's > wrong? > > > http://www.ovirt.org/release/3.6.6/
' oVirt 3.6.6 first release candidate' so it's still not the GA. > Thanks > Luiz > > Em qui, 28 de abr de 2016 11:33, Luiz Claudio Prazeres Goncalves > <luiz...@gmail.com> escreveu: >> >> Nice!... so, I'll survive a bit more with these issues until the version >> 3.6.6 gets released... >> >> >> Thanks >> -Luiz >> >> 2016-04-28 4:50 GMT-03:00 Simone Tiraboschi <stira...@redhat.com>: >>> >>> On Thu, Apr 28, 2016 at 8:32 AM, Sahina Bose <sab...@redhat.com> wrote: >>> > This seems like issue reported in >>> > https://bugzilla.redhat.com/show_bug.cgi?id=1327121 >>> > >>> > Nir, Simone? >>> >>> The issue is here: >>> MainThread::INFO::2016-04-27 >>> >>> 03:26:27,185::storage_server::229::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(disconnect_storage_server) >>> Disconnecting storage server >>> MainThread::INFO::2016-04-27 >>> >>> 03:26:27,816::upgrade::983::ovirt_hosted_engine_ha.lib.upgrade.StorageServer::(fix_storage_path) >>> Fixing storage path in conf file >>> >>> And it's tracked here: https://bugzilla.redhat.com/1327516 >>> >>> We already have a patch, it will be fixed with 3.6.6 >>> >>> As far as I saw this issue will only cause a lot of mess in the logs >>> and some false alert but it's basically harmless >>> >>> > On 04/28/2016 05:35 AM, Luiz Claudio Prazeres Goncalves wrote: >>> > >>> > >>> > Hi everyone, >>> > >>> > Until today my environment was fully updated (3.6.5+centos7.2) with 3 >>> > nodes >>> > (kvm1,kvm2 and kvm3 hosts) . I also have 3 external gluster nodes >>> > (gluster-root1,gluster1 and gluster2 hosts ) , replica 3, which the >>> > engine >>> > storage domain is sitting on top (3.7.11 fully updated+centos7.2) >>> > >>> > For some weird reason i've been receiving emails from oVirt with >>> > EngineUnexpectedDown (attached picture) on a daily basis more or less, >>> > but >>> > the engine seems to be working fine and my vm's are up and running >>> > normally. >>> > I've never had any issue to access the User Interface to manage the >>> > vm's >>> > >>> > Today I run "yum update" on the nodes and realised that vdsm was >>> > outdated, >>> > so I updated the kvm hosts and they are now , again, fully updated. >>> > >>> > >>> > Reviewing the logs It seems to be an intermittent connectivity issue >>> > when >>> > trying to access the gluster engine storage domain as you can see >>> > below. I >>> > don't have any network issue in place and I'm 100% sure about it. I >>> > have >>> > another oVirt Cluster using the same network and using a engine storage >>> > domain on top of an iSCSI Storage Array with no issues. >>> > >>> > Here seems to be the issue: >>> > >>> > Thread-1111::INFO::2016-04-27 >>> > 23:01:27,864::fileSD::357::Storage.StorageDomain::(validate) >>> > sdUUID=03926733-1872-4f85-bb21-18dc320560db >>> > >>> > Thread-1111::DEBUG::2016-04-27 >>> > 23:01:27,865::persistentDict::234::Storage.PersistentDict::(refresh) >>> > read >>> > lines (FileMetadataRW)=[] >>> > >>> > Thread-1111::DEBUG::2016-04-27 >>> > 23:01:27,865::persistentDict::252::Storage.PersistentDict::(refresh) >>> > Empty >>> > metadata >>> > >>> > Thread-1111::ERROR::2016-04-27 >>> > 23:01:27,865::task::866::Storage.TaskManager.Task::(_setError) >>> > Task=`d2acf575-1a60-4fa0-a5bb-cd4363636b94`::Unexpected error >>> > >>> > Traceback (most recent call last): >>> > >>> > File "/usr/share/vdsm/storage/task.py", line 873, in _run >>> > >>> > return fn(*args, **kargs) >>> > >>> > File "/usr/share/vdsm/logUtils.py", line 49, in wrapper >>> > >>> > res = f(*args, **kwargs) >>> > >>> > File "/usr/share/vdsm/storage/hsm.py", line 2835, in >>> > getStorageDomainInfo >>> > >>> > dom = self.validateSdUUID(sdUUID) >>> > >>> > File "/usr/share/vdsm/storage/hsm.py", line 278, in validateSdUUID >>> > >>> > sdDom.validate() >>> > >>> > File "/usr/share/vdsm/storage/fileSD.py", line 360, in validate >>> > >>> > raise se.StorageDomainAccessError(self.sdUUID) >>> > >>> > StorageDomainAccessError: Domain is either partially accessible or >>> > entirely >>> > inaccessible: (u'03926733-1872-4f85-bb21-18dc320560db',) >>> > >>> > Thread-1111::DEBUG::2016-04-27 >>> > 23:01:27,865::task::885::Storage.TaskManager.Task::(_run) >>> > Task=`d2acf575-1a60-4fa0-a5bb-cd4363636b94`::Task._run: >>> > d2acf575-1a60-4fa0-a5bb-cd4363636b94 >>> > ('03926733-1872-4f85-bb21-18dc320560db',) {} failed - stopping task >>> > >>> > Thread-1111::DEBUG::2016-04-27 >>> > 23:01:27,865::task::1246::Storage.TaskManager.Task::(stop) >>> > Task=`d2acf575-1a60-4fa0-a5bb-cd4363636b94`::stopping in state >>> > preparing >>> > (force False) >>> > >>> > Thread-1111::DEBUG::2016-04-27 >>> > 23:01:27,865::task::993::Storage.TaskManager.Task::(_decref) >>> > Task=`d2acf575-1a60-4fa0-a5bb-cd4363636b94`::ref 1 aborting True >>> > >>> > Thread-1111::INFO::2016-04-27 >>> > 23:01:27,865::task::1171::Storage.TaskManager.Task::(prepare) >>> > Task=`d2acf575-1a60-4fa0-a5bb-cd4363636b94`::aborting: Task is aborted: >>> > 'Domain is either partially accessible or entirely inaccessible' - code >>> > 379 >>> > >>> > Thread-1111::DEBUG::2016-04-27 >>> > 23:01:27,866::task::1176::Storage.TaskManager.Task::(prepare) >>> > Task=`d2acf575-1a60-4fa0-a5bb-cd4363636b94`::Prepare: aborted: Domain >>> > is >>> > either partially accessible or entirely inaccessible >>> > >>> > >>> > Question: Anyone know what might be happening? I have several gluster >>> > config's, as you can see below. All the storage domain are using the >>> > same >>> > config's >>> > >>> > >>> > More information: >>> > >>> > I have the "engine" storage domain, "vmos1" storage domain and "master" >>> > storage domain, so everything looks good. >>> > >>> > [root@kvm1 vdsm]# vdsClient -s 0 getStorageDomainsList >>> > >>> > 03926733-1872-4f85-bb21-18dc320560db >>> > >>> > 35021ff4-fb95-43d7-92a3-f538273a3c2e >>> > >>> > e306e54e-ca98-468d-bb04-3e8900f8840c >>> > >>> > >>> > Gluster config: >>> > >>> > [root@gluster-root1 ~]# gluster volume info >>> > >>> > >>> > >>> > Volume Name: engine >>> > >>> > Type: Replicate >>> > >>> > Volume ID: 64b413d2-c42e-40fd-b356-3e6975e941b0 >>> > >>> > Status: Started >>> > >>> > Number of Bricks: 1 x 3 = 3 >>> > >>> > Transport-type: tcp >>> > >>> > Bricks: >>> > >>> > Brick1: gluster1.xyz.com:/gluster/engine/brick1 >>> > >>> > Brick2: gluster2.xyz.com:/gluster/engine/brick1 >>> > >>> > Brick3: gluster-root1.xyz.com:/gluster/engine/brick1 >>> > >>> > Options Reconfigured: >>> > >>> > performance.cache-size: 1GB >>> > >>> > performance.write-behind-window-size: 4MB >>> > >>> > performance.write-behind: off >>> > >>> > performance.quick-read: off >>> > >>> > performance.read-ahead: off >>> > >>> > performance.io-cache: off >>> > >>> > performance.stat-prefetch: off >>> > >>> > cluster.eager-lock: enable >>> > >>> > cluster.quorum-type: auto >>> > >>> > network.remote-dio: enable >>> > >>> > cluster.server-quorum-type: server >>> > >>> > cluster.data-self-heal-algorithm: full >>> > >>> > performance.low-prio-threads: 32 >>> > >>> > features.shard-block-size: 512MB >>> > >>> > features.shard: on >>> > >>> > storage.owner-gid: 36 >>> > >>> > storage.owner-uid: 36 >>> > >>> > performance.readdir-ahead: on >>> > >>> > >>> > Volume Name: master >>> > >>> > Type: Replicate >>> > >>> > Volume ID: 20164808-7bbe-4eeb-8770-d222c0e0b830 >>> > >>> > Status: Started >>> > >>> > Number of Bricks: 1 x 3 = 3 >>> > >>> > Transport-type: tcp >>> > >>> > Bricks: >>> > >>> > Brick1: gluster1.xyz.com:/home/storage/master/brick1 >>> > >>> > Brick2: gluster2.xyz.com:/home/storage/master/brick1 >>> > >>> > Brick3: gluster-root1.xyz.com:/home/storage/master/brick1 >>> > >>> > Options Reconfigured: >>> > >>> > performance.readdir-ahead: on >>> > >>> > performance.quick-read: off >>> > >>> > performance.read-ahead: off >>> > >>> > performance.io-cache: off >>> > >>> > performance.stat-prefetch: off >>> > >>> > cluster.eager-lock: enable >>> > >>> > network.remote-dio: enable >>> > >>> > cluster.quorum-type: auto >>> > >>> > cluster.server-quorum-type: server >>> > >>> > storage.owner-uid: 36 >>> > >>> > storage.owner-gid: 36 >>> > >>> > features.shard: on >>> > >>> > features.shard-block-size: 512MB >>> > >>> > performance.low-prio-threads: 32 >>> > >>> > cluster.data-self-heal-algorithm: full >>> > >>> > performance.write-behind: off >>> > >>> > performance.write-behind-window-size: 4MB >>> > >>> > performance.cache-size: 1GB >>> > >>> > >>> > Volume Name: vmos1 >>> > >>> > Type: Replicate >>> > >>> > Volume ID: ea8fb50e-7bc8-4de3-b775-f3976b6b4f13 >>> > >>> > Status: Started >>> > >>> > Number of Bricks: 1 x 3 = 3 >>> > >>> > Transport-type: tcp >>> > >>> > Bricks: >>> > >>> > Brick1: gluster1.xyz.com:/gluster/vmos1/brick1 >>> > >>> > Brick2: gluster2.xyz.com:/gluster/vmos1/brick1 >>> > >>> > Brick3: gluster-root1.xyz.com:/gluster/vmos1/brick1 >>> > >>> > Options Reconfigured: >>> > >>> > network.ping-timeout: 60 >>> > >>> > performance.readdir-ahead: on >>> > >>> > performance.quick-read: off >>> > >>> > performance.read-ahead: off >>> > >>> > performance.io-cache: off >>> > >>> > performance.stat-prefetch: off >>> > >>> > cluster.eager-lock: enable >>> > >>> > network.remote-dio: enable >>> > >>> > cluster.quorum-type: auto >>> > >>> > cluster.server-quorum-type: server >>> > >>> > storage.owner-uid: 36 >>> > >>> > storage.owner-gid: 36 >>> > >>> > features.shard: on >>> > >>> > features.shard-block-size: 512MB >>> > >>> > performance.low-prio-threads: 32 >>> > >>> > cluster.data-self-heal-algorithm: full >>> > >>> > performance.write-behind: off >>> > >>> > performance.write-behind-window-size: 4MB >>> > >>> > performance.cache-size: 1GB >>> > >>> > >>> > >>> > Attached goes all the logs... >>> > >>> > >>> > >>> > Thanks >>> > >>> > -Luiz >>> > >>> > >>> > >>> > >>> > _______________________________________________ >>> > Users mailing list >>> > Users@ovirt.org >>> > http://lists.ovirt.org/mailman/listinfo/users >>> > >>> > >> >> > _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users