Re: [Gluster-devel] Core generated by trash.t
+Krutika - Original Message - > From: "Anoop C S" <anoo...@redhat.com> > To: "Atin Mukherjee" <amukh...@redhat.com> > Cc: "Pranith Kumar Karampuri" <pkara...@redhat.com>, "Ravishankar N" > <ravishan...@redhat.com>, "Anuradha Talur" > <ata...@redhat.com>, gluster-devel@gluster.org > Sent: Friday, April 22, 2016 2:14:28 PM > Subject: Re: [Gluster-devel] Core generated by trash.t > > On Wed, 2016-04-20 at 16:24 +0530, Atin Mukherjee wrote: > > I should have said the regression link is irrelevant here. Try > > running > > this test on your local setup multiple times on mainline. I do > > believe > > you should see the crash. > > > > I could see coredump on running trash.t multiple times in a while loop. > Info from coredump: > > Core was generated by `/usr/local/sbin/glusterfs -s localhost -- > volfile-id gluster/glustershd -p /var/'. > Program terminated with signal SIGSEGV, Segmentation fault. > #0 0x0040bd31 in glusterfs_handle_translator_op > (req=0x7feab8001dec) at glusterfsd-mgmt.c:590 > 590 any = active->first; > [Current thread is 1 (Thread 0x7feac1657700 (LWP 12050))] > (gdb) l > 585 goto out; > 586 } > 587 > 588 ctx = glusterfsd_ctx; > 589 active = ctx->active; > 590 any = active->first; > 591 input = dict_new (); > 592 ret = dict_unserialize (xlator_req.input.input_val, > 593 xlator_req.input.input_len, > 594 ); > (gdb) p ctx > $1 = (glusterfs_ctx_t *) 0x7fa010 > (gdb) p ctx->active > $2 = (glusterfs_graph_t *) 0x0 I think this is because the request came to shd even before the graph is intialized? Thanks for the test case. I will take a look at this. Pranith > (gdb) p *req > $1 = {trans = 0x7feab8000e20, svc = 0x83ca50, prog = 0x874810, xid = 1, > prognum = 4867634, progver = 2, procnum = 3, type = 0, uid = 0, gid = > 0, pid = 0, lk_owner = {len = 4, > data = '\000' }, gfs_id = 0, auxgids = > 0x7feab800223c, auxgidsmall = {0 }, auxgidlarge = > 0x0, auxgidcount = 0, msg = {{iov_base = 0x7feacc253840, > iov_len = 488}, {iov_base = 0x0, iov_len = 0} times>}, count = 1, iobref = 0x7feab8000c40, rpc_status = 0, rpc_err = > 0, auth_err = 0, txlist = {next = 0x7feab800256c, > prev = 0x7feab800256c}, payloadsize = 0, cred = {flavour = 390039, > datalen = 24, authdata = '\000' , "\004", '\000' > }, verf = {flavour = 0, > datalen = 0, authdata = '\000' }, synctask = > _gf_true, private = 0x0, trans_private = 0x0, hdr_iobuf = 0x82b038, > reply = 0x0} > (gdb) p req->procnum > $3 = 3 <== GLUSTERD_BRICK_XLATOR_OP > (gdb) t a a bt > > Thread 6 (Thread 0x7feabf178700 (LWP 12055)): > #0 0x7feaca522043 in epoll_wait () at ../sysdeps/unix/syscall- > template.S:84 > #1 0x7feacbe5076f in event_dispatch_epoll_worker (data=0x878130) > at event-epoll.c:664 > #2 0x7feacac4560a in start_thread (arg=0x7feabf178700) at > pthread_create.c:334 > #3 0x7feaca521a4d in clone () at > ../sysdeps/unix/sysv/linux/x86_64/clone.S:109 > > Thread 5 (Thread 0x7feac2659700 (LWP 12048)): > #0 do_sigwait (sig=0x7feac2658e3c, set=) at > ../sysdeps/unix/sysv/linux/sigwait.c:64 > #1 __sigwait (set=, sig=0x7feac2658e3c) at > ../sysdeps/unix/sysv/linux/sigwait.c:96 > #2 0x00409895 in glusterfs_sigwaiter (arg=0x7ffe3debbf00) at > glusterfsd.c:2032 > #3 0x7feacac4560a in start_thread (arg=0x7feac2659700) at > pthread_create.c:334 > #4 0x7feaca521a4d in clone () at > ../sysdeps/unix/sysv/linux/x86_64/clone.S:109 > > Thread 4 (Thread 0x7feacc2b4780 (LWP 12046)): > #0 0x7feacac466ad in pthread_join (threadid=140646205064960, > thread_return=0x0) at pthread_join.c:90 > #1 0x7feacbe509bb in event_dispatch_epoll (event_pool=0x830b80) at > event-epoll.c:758 > #2 0x7feacbe17a91 in event_dispatch (event_pool=0x830b80) at > event.c:124 > #3 0x0040a3c8 in main (argc=13, argv=0x7ffe3debd0f8) at > glusterfsd.c:2376 > > Thread 3 (Thread 0x7feac2e5a700 (LWP 12047)): > #0 0x7feacac4e27d in nanosleep () at ../sysdeps/unix/syscall- > template.S:84 > #1 0x7feacbdfc152 in gf_timer_proc (ctx=0x7fa010) at timer.c:188 > #2 0x7feacac4560a in start_thread (arg=0x7feac2e5a700) at > pthread_create.c:334 > #3 0x7feaca521a4d in clone () at > ../sysdeps/unix/sysv/linux/x86_64/clone.S:109 > > Thread 2 (Thread 0x7feac1e58700 (LWP 12049)): > #0 pthread_cond_timedwait@@GLIBC_2.3.2 () at > ../sysdeps/unix/sysv/linux/x86_64/pthread_cond_timedwait.S:225 > #
Re: [Gluster-devel] Core generated by trash.t
On Wed, 2016-04-20 at 16:24 +0530, Atin Mukherjee wrote: > I should have said the regression link is irrelevant here. Try > running > this test on your local setup multiple times on mainline. I do > believe > you should see the crash. > I could see coredump on running trash.t multiple times in a while loop. Info from coredump: Core was generated by `/usr/local/sbin/glusterfs -s localhost -- volfile-id gluster/glustershd -p /var/'. Program terminated with signal SIGSEGV, Segmentation fault. #0 0x0040bd31 in glusterfs_handle_translator_op (req=0x7feab8001dec) at glusterfsd-mgmt.c:590 590 any = active->first; [Current thread is 1 (Thread 0x7feac1657700 (LWP 12050))] (gdb) l 585 goto out; 586 } 587 588 ctx = glusterfsd_ctx; 589 active = ctx->active; 590 any = active->first; 591 input = dict_new (); 592 ret = dict_unserialize (xlator_req.input.input_val, 593 xlator_req.input.input_len, 594 ); (gdb) p ctx $1 = (glusterfs_ctx_t *) 0x7fa010 (gdb) p ctx->active $2 = (glusterfs_graph_t *) 0x0 (gdb) p *req $1 = {trans = 0x7feab8000e20, svc = 0x83ca50, prog = 0x874810, xid = 1, prognum = 4867634, progver = 2, procnum = 3, type = 0, uid = 0, gid = 0, pid = 0, lk_owner = {len = 4, data = '\000' }, gfs_id = 0, auxgids = 0x7feab800223c, auxgidsmall = {0 }, auxgidlarge = 0x0, auxgidcount = 0, msg = {{iov_base = 0x7feacc253840, iov_len = 488}, {iov_base = 0x0, iov_len = 0} }, count = 1, iobref = 0x7feab8000c40, rpc_status = 0, rpc_err = 0, auth_err = 0, txlist = {next = 0x7feab800256c, prev = 0x7feab800256c}, payloadsize = 0, cred = {flavour = 390039, datalen = 24, authdata = '\000' , "\004", '\000' }, verf = {flavour = 0, datalen = 0, authdata = '\000' }, synctask = _gf_true, private = 0x0, trans_private = 0x0, hdr_iobuf = 0x82b038, reply = 0x0} (gdb) p req->procnum $3 = 3 <== GLUSTERD_BRICK_XLATOR_OP (gdb) t a a bt Thread 6 (Thread 0x7feabf178700 (LWP 12055)): #0 0x7feaca522043 in epoll_wait () at ../sysdeps/unix/syscall- template.S:84 #1 0x7feacbe5076f in event_dispatch_epoll_worker (data=0x878130) at event-epoll.c:664 #2 0x7feacac4560a in start_thread (arg=0x7feabf178700) at pthread_create.c:334 #3 0x7feaca521a4d in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:109 Thread 5 (Thread 0x7feac2659700 (LWP 12048)): #0 do_sigwait (sig=0x7feac2658e3c, set=) at ../sysdeps/unix/sysv/linux/sigwait.c:64 #1 __sigwait (set=, sig=0x7feac2658e3c) at ../sysdeps/unix/sysv/linux/sigwait.c:96 #2 0x00409895 in glusterfs_sigwaiter (arg=0x7ffe3debbf00) at glusterfsd.c:2032 #3 0x7feacac4560a in start_thread (arg=0x7feac2659700) at pthread_create.c:334 #4 0x7feaca521a4d in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:109 Thread 4 (Thread 0x7feacc2b4780 (LWP 12046)): #0 0x7feacac466ad in pthread_join (threadid=140646205064960, thread_return=0x0) at pthread_join.c:90 #1 0x7feacbe509bb in event_dispatch_epoll (event_pool=0x830b80) at event-epoll.c:758 #2 0x7feacbe17a91 in event_dispatch (event_pool=0x830b80) at event.c:124 #3 0x0040a3c8 in main (argc=13, argv=0x7ffe3debd0f8) at glusterfsd.c:2376 Thread 3 (Thread 0x7feac2e5a700 (LWP 12047)): #0 0x7feacac4e27d in nanosleep () at ../sysdeps/unix/syscall- template.S:84 #1 0x7feacbdfc152 in gf_timer_proc (ctx=0x7fa010) at timer.c:188 #2 0x7feacac4560a in start_thread (arg=0x7feac2e5a700) at pthread_create.c:334 #3 0x7feaca521a4d in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:109 Thread 2 (Thread 0x7feac1e58700 (LWP 12049)): #0 pthread_cond_timedwait@@GLIBC_2.3.2 () at ../sysdeps/unix/sysv/linux/x86_64/pthread_cond_timedwait.S:225 #1 0x7feacbe2d73d in syncenv_task (proc=0x838310) at syncop.c:603 #2 0x7feacbe2d9dd in syncenv_processor (thdata=0x838310) at syncop.c:695 #3 0x7feacac4560a in start_thread (arg=0x7feac1e58700) at pthread_create.c:334 #4 0x7feaca521a4d in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:109 Thread 1 (Thread 0x7feac1657700 (LWP 12050)): #0 0x0040bd31 in glusterfs_handle_translator_op (req=0x7feab8001dec) at glusterfsd-mgmt.c:590 #1 0x7feacbe2cf04 in synctask_wrap (old_task=0x7feab80031c0) at syncop.c:375 #2 0x7feaca467f30 in ?? () from /lib64/libc.so.6 #3 0x in ?? () Looking at the core, crash was seen from glusterfs_handle_translator_op() routine while doing a 'volume heal' command. I could then easily create a small test case to re-produce the issue. Please find the attachment for the same. --Anoop C S. core-reprod.t Description: Perl program ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] Core generated by trash.t
I should have said the regression link is irrelevant here. Try running this test on your local setup multiple times on mainline. I do believe you should see the crash. ~Atin On 04/20/2016 03:36 PM, Anoop C S wrote: > On Wed, 2016-04-20 at 13:25 +0530, Anoop C S wrote: >> On Wed, 2016-04-20 at 01:21 +0530, Atin Mukherjee wrote: >>> >>> Regression run [1] failed from trash.t, however the same doesn't >>> talk >>> about any core file, but when I run it with and with out my changes >>> the same generates a core. >> on which platform you ran trash.t, NetBSD or Linux? >> >>> >>> >>> [1] >>> https://build.gluster.org/job/rackspace-netbsd7-regression-triggere >>> d/ >>> 15971/consoleFull >>> >> not ok 62 , LINENUM:245 >> FAILED COMMAND: start_vol patchy /mnt/glusterfs/0 >> /mnt/glusterfs/0/abc >> >> start_vol function basically performs volume start command and >> repeatedly checks for the presence of directory named 'abc' under >> root >> of the volume. I see the following glusterd errors from archived >> logs(build-install-etc-glusterfs-glusterd.vol.log): >> >>> >>> [2016-04-19 15:56:48.722283] W [common- >>> utils.c:1805:gf_string2boolean] (-->0xb9c31f47 >>>at >>> /build/install/lib/glusterfs/3.8dev/xlator/mgmt/glusterd.so -- 0xbb733b56 at >>> /build/install/lib/libglusterfs.so.0 ) 0-management: argument >>> invalid >>> [Invalid argument] >>> [2016-04-19 15:56:48.766453] I [MSGID: 106144] [glusterd- >>> pmap.c:270:pmap_registry_remove] 0-pmap: removing brick >>> /d/backends/patchy11 on port 49153 >>> [2016-04-19 15:56:48.771041] E [MSGID: 106005] [glusterd- >>> utils.c:4689:glusterd_brick_start] 0-management: Unable to start >>> brick nbslave75.cloud.gluster.org:/d/backends/patchy1 >>> [2016-04-19 15:56:48.771132] E [MSGID: 106123] [glusterd- >>> mgmt.c:306:gd_mgmt_v3_commit_fn] 0-management: Volume start commit >>> failed. >>> [2016-04-19 15:56:48.771161] E [MSGID: 106123] [glusterd- >>> mgmt.c:1423:glusterd_mgmt_v3_commit] 0-management: Commit failed >>> for >>> operation Start on local node >>> [2016-04-19 15:56:48.771188] E [MSGID: 106123] [glusterd- >>> mgmt.c:2014:glusterd_mgmt_v3_initiate_all_phases] 0-management: >>> Commit Op Failed >> Brick errors from bricks/d-backends-patchy1.log: >> >>> >>> [2016-04-19 15:56:48.763066] I >>> [rpcsvc.c:2218:rpcsvc_set_outstanding_rpc_limit] 0-rpc-service: >>> Configured rpc.outstanding-rpc-limit with value 64 >>> [2016-04-19 15:56:48.763127] W [MSGID: 101002] >>> [options.c:954:xl_opt_validate] 0-patchy-server: option 'listen- >>> port' >>> is deprecated, preferred is 'transport.socket.listen-port', >>> continuing with correction >>> [2016-04-19 15:56:48.763273] E [socket.c:765:__socket_server_bind] >>> 0- >>> tcp.patchy-server: binding to failed: Address already in use >>> [2016-04-19 15:56:48.763293] E [socket.c:768:__socket_server_bind] >>> 0- >>> tcp.patchy-server: Port is already in use >>> [2016-04-19 15:56:48.763314] W >>> [rpcsvc.c:1600:rpcsvc_transport_create] 0-rpc-service: listening on >>> transport failed >>> [2016-04-19 15:56:48.763332] W [MSGID: 115045] [server.c:1061:init] >>> 0-patchy-server: creation of listener failed >>> [2016-04-19 15:56:48.763351] E [MSGID: 101019] >>> [xlator.c:430:xlator_init] 0-patchy-server: Initialization of >>> volume >>> 'patchy-server' failed, review your volfile again >>> [2016-04-19 15:56:48.763368] E [MSGID: 101066] >>> [graph.c:324:glusterfs_graph_init] 0-patchy-server: initializing >>> translator failed >>> [2016-04-19 15:56:48.763383] E [MSGID: 101176] >>> [graph.c:670:glusterfs_graph_activate] 0-graph: init failed >>> [2016-04-19 15:56:48.766235] W [glusterfsd.c:1265:cleanup_and_exit] >>> (-->0x8050b83 at >>> /build/install/sbin/glusterfsd -->0x804e8e7 >>> at /build/install/sbin/glusterfsd ) 0-: received signum (0), >>> shutting >>> down > > I found the following BZ with exactly similar brick error messages. Can > you please confirm whether they are related or not? > > https://bugzilla.redhat.com/show_bug.cgi?id=1322805 > >> Is this something related to the >> patch(http://review.gluster.org/#/c/10 >> 785/) for which the this regression was run against? Because I don't >> expect volume start to fail. >> >> If you have the coredump, can you please share the back trace for >> further analysis? >> >> Thanks, >> --Anoop C S. >> >>> >>> ~Atin >>> ___ >>> Gluster-devel mailing list >>> Gluster-devel@gluster.org >>> http://www.gluster.org/mailman/listinfo/gluster-devel >> ___ >> Gluster-devel mailing list >> Gluster-devel@gluster.org >> http://www.gluster.org/mailman/listinfo/gluster-devel ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] Core generated by trash.t
On Wed, 2016-04-20 at 13:25 +0530, Anoop C S wrote: > On Wed, 2016-04-20 at 01:21 +0530, Atin Mukherjee wrote: > > > > Regression run [1] failed from trash.t, however the same doesn't > > talk > > about any core file, but when I run it with and with out my changes > > the same generates a core. > on which platform you ran trash.t, NetBSD or Linux? > > > > > > > [1] > > https://build.gluster.org/job/rackspace-netbsd7-regression-triggere > > d/ > > 15971/consoleFull > > > not ok 62 , LINENUM:245 > FAILED COMMAND: start_vol patchy /mnt/glusterfs/0 > /mnt/glusterfs/0/abc > > start_vol function basically performs volume start command and > repeatedly checks for the presence of directory named 'abc' under > root > of the volume. I see the following glusterd errors from archived > logs(build-install-etc-glusterfs-glusterd.vol.log): > > > > > [2016-04-19 15:56:48.722283] W [common- > > utils.c:1805:gf_string2boolean] (-->0xb9c31f47 > >at > > /build/install/lib/glusterfs/3.8dev/xlator/mgmt/glusterd.so -- > > > > > > 0xbb733b56 at > > /build/install/lib/libglusterfs.so.0 ) 0-management: argument > > invalid > > [Invalid argument] > > [2016-04-19 15:56:48.766453] I [MSGID: 106144] [glusterd- > > pmap.c:270:pmap_registry_remove] 0-pmap: removing brick > > /d/backends/patchy11 on port 49153 > > [2016-04-19 15:56:48.771041] E [MSGID: 106005] [glusterd- > > utils.c:4689:glusterd_brick_start] 0-management: Unable to start > > brick nbslave75.cloud.gluster.org:/d/backends/patchy1 > > [2016-04-19 15:56:48.771132] E [MSGID: 106123] [glusterd- > > mgmt.c:306:gd_mgmt_v3_commit_fn] 0-management: Volume start commit > > failed. > > [2016-04-19 15:56:48.771161] E [MSGID: 106123] [glusterd- > > mgmt.c:1423:glusterd_mgmt_v3_commit] 0-management: Commit failed > > for > > operation Start on local node > > [2016-04-19 15:56:48.771188] E [MSGID: 106123] [glusterd- > > mgmt.c:2014:glusterd_mgmt_v3_initiate_all_phases] 0-management: > > Commit Op Failed > Brick errors from bricks/d-backends-patchy1.log: > > > > > [2016-04-19 15:56:48.763066] I > > [rpcsvc.c:2218:rpcsvc_set_outstanding_rpc_limit] 0-rpc-service: > > Configured rpc.outstanding-rpc-limit with value 64 > > [2016-04-19 15:56:48.763127] W [MSGID: 101002] > > [options.c:954:xl_opt_validate] 0-patchy-server: option 'listen- > > port' > > is deprecated, preferred is 'transport.socket.listen-port', > > continuing with correction > > [2016-04-19 15:56:48.763273] E [socket.c:765:__socket_server_bind] > > 0- > > tcp.patchy-server: binding to failed: Address already in use > > [2016-04-19 15:56:48.763293] E [socket.c:768:__socket_server_bind] > > 0- > > tcp.patchy-server: Port is already in use > > [2016-04-19 15:56:48.763314] W > > [rpcsvc.c:1600:rpcsvc_transport_create] 0-rpc-service: listening on > > transport failed > > [2016-04-19 15:56:48.763332] W [MSGID: 115045] [server.c:1061:init] > > 0-patchy-server: creation of listener failed > > [2016-04-19 15:56:48.763351] E [MSGID: 101019] > > [xlator.c:430:xlator_init] 0-patchy-server: Initialization of > > volume > > 'patchy-server' failed, review your volfile again > > [2016-04-19 15:56:48.763368] E [MSGID: 101066] > > [graph.c:324:glusterfs_graph_init] 0-patchy-server: initializing > > translator failed > > [2016-04-19 15:56:48.763383] E [MSGID: 101176] > > [graph.c:670:glusterfs_graph_activate] 0-graph: init failed > > [2016-04-19 15:56:48.766235] W [glusterfsd.c:1265:cleanup_and_exit] > > (-->0x8050b83 at > > /build/install/sbin/glusterfsd -->0x804e8e7 > > at /build/install/sbin/glusterfsd ) 0-: received signum (0), > > shutting > > down I found the following BZ with exactly similar brick error messages. Can you please confirm whether they are related or not? https://bugzilla.redhat.com/show_bug.cgi?id=1322805 > Is this something related to the > patch(http://review.gluster.org/#/c/10 > 785/) for which the this regression was run against? Because I don't > expect volume start to fail. > > If you have the coredump, can you please share the back trace for > further analysis? > > Thanks, > --Anoop C S. > > > > > ~Atin > > ___ > > Gluster-devel mailing list > > Gluster-devel@gluster.org > > http://www.gluster.org/mailman/listinfo/gluster-devel > ___ > Gluster-devel mailing list > Gluster-devel@gluster.org > http://www.gluster.org/mailman/listinfo/gluster-devel ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] Core generated by trash.t
On Wed, 2016-04-20 at 01:21 +0530, Atin Mukherjee wrote: > Regression run [1] failed from trash.t, however the same doesn't talk > about any core file, but when I run it with and with out my changes > the same generates a core. on which platform you ran trash.t, NetBSD or Linux? > > [1] > https://build.gluster.org/job/rackspace-netbsd7-regression-triggered/ > 15971/consoleFull > not ok 62 , LINENUM:245 FAILED COMMAND: start_vol patchy /mnt/glusterfs/0 /mnt/glusterfs/0/abc start_vol function basically performs volume start command and repeatedly checks for the presence of directory named 'abc' under root of the volume. I see the following glusterd errors from archived logs(build-install-etc-glusterfs-glusterd.vol.log): > [2016-04-19 15:56:48.722283] W [common- > utils.c:1805:gf_string2boolean] (-->0xb9c31f47 >at > /build/install/lib/glusterfs/3.8dev/xlator/mgmt/glusterd.so -- > >0xbb733b56 at > /build/install/lib/libglusterfs.so.0 ) 0-management: argument invalid > [Invalid argument] > [2016-04-19 15:56:48.766453] I [MSGID: 106144] [glusterd- > pmap.c:270:pmap_registry_remove] 0-pmap: removing brick > /d/backends/patchy11 on port 49153 > [2016-04-19 15:56:48.771041] E [MSGID: 106005] [glusterd- > utils.c:4689:glusterd_brick_start] 0-management: Unable to start > brick nbslave75.cloud.gluster.org:/d/backends/patchy1 > [2016-04-19 15:56:48.771132] E [MSGID: 106123] [glusterd- > mgmt.c:306:gd_mgmt_v3_commit_fn] 0-management: Volume start commit > failed. > [2016-04-19 15:56:48.771161] E [MSGID: 106123] [glusterd- > mgmt.c:1423:glusterd_mgmt_v3_commit] 0-management: Commit failed for > operation Start on local node > [2016-04-19 15:56:48.771188] E [MSGID: 106123] [glusterd- > mgmt.c:2014:glusterd_mgmt_v3_initiate_all_phases] 0-management: > Commit Op Failed Brick errors from bricks/d-backends-patchy1.log: > [2016-04-19 15:56:48.763066] I > [rpcsvc.c:2218:rpcsvc_set_outstanding_rpc_limit] 0-rpc-service: > Configured rpc.outstanding-rpc-limit with value 64 > [2016-04-19 15:56:48.763127] W [MSGID: 101002] > [options.c:954:xl_opt_validate] 0-patchy-server: option 'listen-port' > is deprecated, preferred is 'transport.socket.listen-port', > continuing with correction > [2016-04-19 15:56:48.763273] E [socket.c:765:__socket_server_bind] 0- > tcp.patchy-server: binding to failed: Address already in use > [2016-04-19 15:56:48.763293] E [socket.c:768:__socket_server_bind] 0- > tcp.patchy-server: Port is already in use > [2016-04-19 15:56:48.763314] W > [rpcsvc.c:1600:rpcsvc_transport_create] 0-rpc-service: listening on > transport failed > [2016-04-19 15:56:48.763332] W [MSGID: 115045] [server.c:1061:init] > 0-patchy-server: creation of listener failed > [2016-04-19 15:56:48.763351] E [MSGID: 101019] > [xlator.c:430:xlator_init] 0-patchy-server: Initialization of volume > 'patchy-server' failed, review your volfile again > [2016-04-19 15:56:48.763368] E [MSGID: 101066] > [graph.c:324:glusterfs_graph_init] 0-patchy-server: initializing > translator failed > [2016-04-19 15:56:48.763383] E [MSGID: 101176] > [graph.c:670:glusterfs_graph_activate] 0-graph: init failed > [2016-04-19 15:56:48.766235] W [glusterfsd.c:1265:cleanup_and_exit] > (-->0x8050b83 at > /build/install/sbin/glusterfsd -->0x804e8e7 > at /build/install/sbin/glusterfsd ) 0-: received signum (0), shutting > down Is this something related to the patch(http://review.gluster.org/#/c/10 785/) for which the this regression was run against? Because I don't expect volume start to fail. If you have the coredump, can you please share the back trace for further analysis? Thanks, --Anoop C S. > ~Atin > ___ > Gluster-devel mailing list > Gluster-devel@gluster.org > http://www.gluster.org/mailman/listinfo/gluster-devel ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel