Hi,

gdb doesn't accept the coredump file ...

root@nfs05 /var/crash # gdb /usr/sbin/glusterfsd /var/crash/_usr_sbin_glusterfsd.0.crash
GNU gdb (Ubuntu 7.7.1-0ubuntu5~14.04.2) 7.7.1
Copyright (C) 2014 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law.  Type "show copying"
and "show warranty" for details.
This GDB was configured as "x86_64-linux-gnu".
Type "show configuration" for configuration details.
For bug reporting instructions, please see:
<http://www.gnu.org/software/gdb/bugs/>.
Find the GDB manual and other documentation resources online at:
<http://www.gnu.org/software/gdb/documentation/>.
For help, type "help".
Type "apropos word" to search for commands related to "word"...
Reading symbols from /usr/sbin/glusterfsd...(no debugging symbols found)...done. "/var/crash/_usr_sbin_glusterfsd.0.crash" is not a core dump: File format not recognized
(gdb) bt
No stack.

Yann


Le 29/06/2016 08:39, Anoop C S a écrit :
On Tue, 2016-06-28 at 10:49 +0200, Yann LEMARIE wrote:
Hi,

I found the coredump file, but it's a 15Mo file (zipped), I can't
post it on this mailling list.

Great. In order to exactly pin point the crash location, can you please
attach gdb to extracted coredump file and share us the complete back
trace by executing `bt` command in gdb shell? Apart from gdb you may be
instructed to install some debug-info packages for extracting a useful
back trace while attaching gdb as follows:

# gdb /usr/sbin/glusterfsd <path-to-coredump-file>

If prompted install required packages and reattach the coredump file.
When you are inside (gdb) prompt type 'bt' and paste the back trace.

Here is some parts of the repport :

ProblemType: Crash
Architecture: amd64
Date: Sun Jun 26 11:27:44 2016
DistroRelease: Ubuntu 14.04
ExecutablePath: /usr/sbin/glusterfsd
ExecutableTimestamp: 1460982898
ProcCmdline: /usr/sbin/glusterfsd -s nfs05 --volfile-id
cdn.nfs05.srv-cdn -p /var/lib/glusterd/vols/cdn/run/nfs05-srv-
cdn.pid -S /var/run/gluster/d52ac3e6c0a3fa316a9e8360976f3af5.socket
--brick-name /srv/cdn -l /var/log/glusterfs/bricks/srv-cdn.log --
xlator-option *-posix.glusterd-uuid=6af63b78-a3da-459d-a909-
c010e6c9072c --brick-port 49155 --xlator-option cdn-server.listen-
port=49155
ProcCwd: /
ProcEnviron:
  PATH=(custom, no user)
  TERM=linux
ProcMaps:
  7f25f18d9000-7f25f18da000 ---p 00000000 00:00 0
  7f25f18da000-7f25f19da000 rw-p 00000000 00:00
0                          [stack:849]
  7f25f19da000-7f25f19db000 ---p 00000000 00:00 0
  ...
ProcStatus:
  Name:  glusterfsd
  State: D (disk sleep)
  Tgid:  7879
  Ngid:  0
  Pid:   7879
  PPid:  1
  TracerPid:     0
  Uid:   0       0       0       0
  Gid:   0       0       0       0
  FDSize:        64
  Groups:        0
  VmPeak:          878404 kB
  VmSize:          878404 kB
  VmLck:        0 kB
  VmPin:        0 kB
  VmHWM:    96104 kB
  VmRSS:    90652 kB
  VmData:          792012 kB
  VmStk:      276 kB
  VmExe:       84 kB
  VmLib:     7716 kB
  VmPTE:      700 kB
  VmSwap:           20688 kB
  Threads:       22
  SigQ:  0/30034
  SigPnd:        0000000000000000
  ShdPnd:        0000000000000000
  SigBlk:        0000000000004a01
  SigIgn:        0000000000001000
  SigCgt:        00000001800000fa
  CapInh:        0000000000000000
  CapPrm:        0000001fffffffff
  CapEff:        0000001fffffffff
  CapBnd:        0000001fffffffff
  Seccomp:       0
  Cpus_allowed:  7fff
  Cpus_allowed_list:     0-14
  Mems_allowed:  00000000,00000001
  Mems_allowed_list:     0
  voluntary_ctxt_switches:       3
  nonvoluntary_ctxt_switches:    1
Signal: 11
Uname: Linux 3.13.0-44-generic x86_64
UserGroups:
CoreDump: base64
  ...

Yann

Le 28/06/2016 09:31, Anoop C S a écrit :
On Mon, 2016-06-27 at 15:05 +0200, Yann LEMARIE wrote:
  @Anoop,

Where can I find the coredump file ?

You will get hints about the crash from entries inside
/var/log/messages(for example pid of the process, location of
coredump
etc).

The crash occurs 2 times last 7 days, each time a sunday morning
with
no reason, no increase of traffic or something like this, the
volume
was mounted since 15 days.

The bricks are used as a CDN like, distributting small images and
css
files with a nginx https service (with a load balancer and 2
EC2), on
a sunday morning there is not a lot of activity ...

 From the very minimal back trace that we have from brick logs I
would
assume that a truncate operation was being handled by trash
translator
and it crashed.

Volume infos:
root@nfs05 /var/log/glusterfs # gluster volume info cdn
Volume Name: cdn
Type: Replicate
Volume ID: c53b9bae-5e12-4f13-8217-53d8c96c302c
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: nfs05:/srv/cdn
Brick2: nfs06:/srv/cdn
Options Reconfigured:
performance.readdir-ahead: on
features.trash: on
features.trash-max-filesize: 20MB
I don't know if there is a link with this crash problem, but I
have
another problem with my 2 servers that make GluserFS's clients
disconnected (from another volume) :
Jun 24 02:28:04 nfs05 kernel: [2039468.818617] xen_netfront:
xennet: skb rides the rocket: 19 slots
Jun 24 02:28:11 nfs05 kernel: [2039475.744086] net_ratelimit:
66
callbacks suppressed
  It seem to be a network interface problem :
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1317811

Yann

Le 27/06/2016 12:59, Anoop C S a écrit :
On Mon, 2016-06-27 at 09:47 +0200, Yann LEMARIE wrote:
Hi,

I'm using GlusterFS since many years and never see this
problem,
but
this is the second time in one week ...

I have 3 volumes with 2 bricks and 1 volume crash with no
reason,
Did you observe the crash while mounting the volume? Or can you
be
more
specific on what were you doing just before you saw the crash?
Can
you
please share the output of `gluster volume info <VOLNAME>`?

  I just have to stop/start the volume to make it up again.
The only logs I can find are in syslog :

Jun 26 11:27:44 nfs05 srv-cdn[7879]: pending frames:
Jun 26 11:27:44 nfs05 srv-cdn[7879]: frame : type(0) op(10)
Jun 26 11:27:44 nfs05 srv-cdn[7879]: patchset:
git://git.gluster.com/glusterfs.git
Jun 26 11:27:44 nfs05 srv-cdn[7879]: signal received: 11
Jun 26 11:27:44 nfs05 srv-cdn[7879]: time of crash:
Jun 26 11:27:44 nfs05 srv-cdn[7879]: 2016-06-26 09:27:44
Jun 26 11:27:44 nfs05 srv-cdn[7879]: configuration details:
Jun 26 11:27:44 nfs05 srv-cdn[7879]: argp 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: backtrace 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: dlfcn 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: libpthread 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: llistxattr 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: setfsid 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: spinlock 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: epoll.h 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: xattr.h 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: st_atim.tv_nsec 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: package-string:
glusterfs
3.7.11
Jun 26 11:27:44 nfs05 srv-cdn[7879]: ---------

Jun 26 11:27:44 nfs06 srv-cdn[1787]: pending frames:
Jun 26 11:27:44 nfs06 srv-cdn[1787]: frame : type(0) op(10)
Jun 26 11:27:44 nfs06 srv-cdn[1787]: patchset:
git://git.gluster.com/glusterfs.git
Jun 26 11:27:44 nfs06 srv-cdn[1787]: signal received: 11
Jun 26 11:27:44 nfs06 srv-cdn[1787]: time of crash:
Jun 26 11:27:44 nfs06 srv-cdn[1787]: 2016-06-26 09:27:44
Jun 26 11:27:44 nfs06 srv-cdn[1787]: configuration details:
Jun 26 11:27:44 nfs06 srv-cdn[1787]: argp 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: backtrace 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: dlfcn 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: libpthread 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: llistxattr 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: setfsid 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: spinlock 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: epoll.h 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: xattr.h 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: st_atim.tv_nsec 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: package-string:
glusterfs
3.7.11
Jun 26 11:27:44 nfs06 srv-cdn[1787]: ---------

Thanks for your help


Regards
--
Yann Lemarié
iRaiser - Support Technique
[email protected]
_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users
--
Yann Lemarié
iRaiser - Support Technique
[email protected]


_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users
--
Yann Lemarié
iRaiser - Support Technique
[email protected]
_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users

--
Yann Lemarié
iRaiser - Support Technique
iRaiser Logotype <http://www.iraiser.eu>
[email protected] <mailto:[email protected]>
_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to