Hi,

On a new host, I am running into exactly the same scenario.

I have a host with an oVirt-managed GlusterFS volume (single brick on
local disk in distribute mode) on an XFS file system.

I think I have found the root cause, but I doubt I can fix it.

Around the time of the VMs going paused, there seemed to be a glusterfsd
restart:

> [2014-12-18 01:43:27.272235] W [glusterfsd.c:1194:cleanup_and_exit] (--> 0-: 
> received signum (15), shutting down
> [2014-12-18 01:43:27.272279] I [fuse-bridge.c:5599:fini] 0-fuse: Unmounting 
> '/rhev/data-center/mnt/glusterSD/onode3.isaac.local:data02'.
> [2014-12-18 01:49:36.854339] I [MSGID: 100030] [glusterfsd.c:2018:main] 
> 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.6.1 
> (args: /usr/sbin/glusterfs -
> -volfile-server=onode3.isaac.local --volfile-id=data02 
> /rhev/data-center/mnt/glusterSD/onode3.isaac.local:data02)
> [2014-12-18 01:49:36.862887] I [dht-shared.c:337:dht_init_regex] 
> 0-data02-dht: using regex rsync-hash-regex = ^\.(.+)\.[^.]+$
> [2014-12-18 01:49:36.863749] I [client.c:2280:notify] 0-data02-client-0: 
> parent translators are ready, attempting connect on transport

So I thought I'd check /var/log/messages for potential sources of the
SIGTERM, and I found this:

> Dec 18 02:43:26 onode3 kernel: supervdsmServer[1960]: segfault at 18 ip 
> 00007faa89951bca sp 00007fa355b80f40 error 4 in 
> libgfapi.so.0.0.0[7faa8994c000+18000]
> Dec 18 02:43:27 onode3 systemd: supervdsmd.service: main process exited, 
> code=killed, status=11/SEGV
> Dec 18 02:43:27 onode3 systemd: Unit supervdsmd.service entered failed state.
> Dec 18 02:43:27 onode3 journal: vdsm jsonrpc.JsonRpcServer ERROR Internal 
> server error
> Traceback (most recent call last):
>   File "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 486, in 
> _serveRequest
>     res = method(**params)
>   File "/usr/share/vdsm/rpc/Bridge.py", line 266, in _dynamicMethod
>     result = fn(*methodArgs)
>   File "/usr/share/vdsm/gluster/apiwrapper.py", line 106, in status
>     return self._gluster.volumeStatus(volumeName, brick, statusOption)
>   File "/usr/share/vdsm/gluster/api.py", line 54, in wrapper
>     rv = func(*args, **kwargs)
>   File "/usr/share/vdsm/gluster/api.py", line 221, in volumeStatus
>     data = self.svdsmProxy.glusterVolumeStatvfs(volumeName)
>   File "/usr/share/vdsm/supervdsm.py", line 50, in __call__
>     return callMethod()
>   File "/usr/share/vdsm/supervdsm.py", line 48, in <lambda>
>     **kwargs)
>   File "<string>", line 2, in glusterVolumeStatvfs
>   File "/usr/lib64/python2.7/multiprocessing/managers.py", line 759, in 
> _callmethod
>     kind, result = conn.recv()
> EOFError
> Dec 18 02:43:27 onode3 systemd: supervdsmd.service holdoff time over, 
> scheduling restart.
> Dec 18 02:43:27 onode3 systemd: Stopping Virtual Desktop Server Manager...
> Dec 18 02:43:27 onode3 systemd: Stopping "Auxiliary vdsm service for running 
> helper functions as root"...
> Dec 18 02:43:27 onode3 systemd: Starting "Auxiliary vdsm service for running 
> helper functions as root"...
> Dec 18 02:43:27 onode3 systemd: Started "Auxiliary vdsm service for running 
> helper functions as root".
> Dec 18 02:43:27 onode3 journal: vdsm IOProcessClient ERROR IOProcess failure
> Traceback (most recent call last):
>   File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 107, in 
> _communicate
>     raise Exception("FD closed")
> Exception: FD closed


I guess I'll file a bug report.

Best regards,
Martijn Grendelman






Punit Dambiwal schreef op 12-12-2014 om 3:44:
> Hi Dan,
> 
> Yes..it's glusterfs....
> 
> glusterfs logs :- http://ur1.ca/j3b5f
> 
> OS Version: RHEL - 7 - 0.1406.el7.centos.2.3
> Kernel Version: 3.10.0 - 123.el7.x86_64
> KVM Version: 1.5.3 - 60.el7_0.2
> LIBVIRT Version: libvirt-1.1.1-29.el7_0.3
> VDSM Version: vdsm-4.16.7-1.gitdb83943.el7
> GlusterFS Version: glusterfs-3.6.1-1.el7
> Qemu Version : QEMU emulator version 1.5.3 (qemu-kvm-1.5.3-60.el7_0.2)
> 
> Thanks,
> punit
> 
> 
> 
> 
> On Thu, Dec 11, 2014 at 5:47 PM, Dan Kenigsberg <dan...@redhat.com
> <mailto:dan...@redhat.com>> wrote:
> 
>     On Thu, Dec 11, 2014 at 03:41:01PM +0800, Punit Dambiwal wrote:
>     > Hi,
>     >
>     > Suddenly all of my VM on one host paused with the following error :-
>     >
>     > vm has paused due to unknown storage error
>     >
>     > I am using glusterfs storage with distributed replicate
>     replica=2....my
>     > storage and compute both running on the same node...
>     >
>     > engine logs :- http://ur1.ca/j31iu
>     > Host logs :- http://ur1.ca/j31kk    (I grep it for one Failed VM)
> 
>     libvirtEventLoop::INFO::2014-12-11
>     15:00:48,627::vm::4780::vm.Vm::(_onIOError)
>     vmId=`e84bb987-a817-436a-9417-8eab9148e57e`::abnormal vm stop device
>     virtio-disk0 error eother
> 
>     Which type of storage is it? gluster? Do you have anything in particular
>     on glusterfs logs?
> 
>     Which glusterfs/qemu/libvirt/vdsm versions do you have installed?
> 
> 
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to