I could see that glusterd process is up from the log file. Aren't you able to run any gluster commands on the same node?
~Atin On Tue, Jul 5, 2016 at 4:05 PM, Atul Yadav <[email protected]> wrote: > Hi Team, > > Please go through the attachment > > Thank you > Atul Yadav > > On Tue, Jul 5, 2016 at 3:55 PM, Samikshan Bairagya <[email protected]> > wrote: > >> >> >> On 07/05/2016 03:15 PM, Atul Yadav wrote: >> >>> Hi Atin, >>> >>> please go through the attachment for log file. >>> >>> Waiting for your input.... >>> >>> >> Hi Atul, >> >> The glusterd log file can be found here: >> /var/log/glusterfs/usr-local-etc-glusterfs-glusterd.vol.log >> >> Sharing that instead of a RAR archive would make things easier for us. >> >> >> >> thank you >>> Atul Yadav >>> >>> On Tue, Jul 5, 2016 at 2:43 PM, Atin Mukherjee <[email protected]> >>> wrote: >>> >>> Why don't you share the glusterd log file? >>>> >>>> On Tue, Jul 5, 2016 at 12:53 PM, Atul Yadav <[email protected]> >>>> wrote: >>>> >>>> Hi , >>>>> >>>>> After restarting the service. service entered in to the fail state. >>>>> [root@master1 ~]# /etc/init.d/glusterd restart >>>>> Stopping glusterd: [FAILED] >>>>> Starting glusterd: [FAILED] >>>>> >>>>> Note: This behavior only happening over rdma network. But with ethernet >>>>> there is no issue. >>>>> >>>>> Thank you >>>>> Atul Yadav >>>>> >>>>> >>>>> >>>>> On Tue, Jul 5, 2016 at 11:28 AM, Atin Mukherjee <[email protected]> >>>>> wrote: >>>>> >>>>> >>>>>> >>>>>> On Tue, Jul 5, 2016 at 11:01 AM, Atul Yadav <[email protected]> >>>>>> wrote: >>>>>> >>>>>> Hi All, >>>>>>> >>>>>>> The glusterfs environment details are given below:- >>>>>>> >>>>>>> [root@master1 ~]# cat /etc/redhat-release >>>>>>> CentOS release 6.7 (Final) >>>>>>> [root@master1 ~]# uname -r >>>>>>> 2.6.32-642.1.1.el6.x86_64 >>>>>>> [root@master1 ~]# rpm -qa | grep -i gluster >>>>>>> glusterfs-rdma-3.8rc2-1.el6.x86_64 >>>>>>> glusterfs-api-3.8rc2-1.el6.x86_64 >>>>>>> glusterfs-3.8rc2-1.el6.x86_64 >>>>>>> glusterfs-cli-3.8rc2-1.el6.x86_64 >>>>>>> glusterfs-client-xlators-3.8rc2-1.el6.x86_64 >>>>>>> glusterfs-server-3.8rc2-1.el6.x86_64 >>>>>>> glusterfs-fuse-3.8rc2-1.el6.x86_64 >>>>>>> glusterfs-libs-3.8rc2-1.el6.x86_64 >>>>>>> [root@master1 ~]# >>>>>>> >>>>>>> Volume Name: home >>>>>>> Type: Replicate >>>>>>> Volume ID: 2403ddf9-c2e0-4930-bc94-734772ef099f >>>>>>> Status: Stopped >>>>>>> Number of Bricks: 1 x 2 = 2 >>>>>>> Transport-type: rdma >>>>>>> Bricks: >>>>>>> Brick1: master1-ib.dbt.au:/glusterfs/home/brick1 >>>>>>> Brick2: master2-ib.dbt.au:/glusterfs/home/brick2 >>>>>>> Options Reconfigured: >>>>>>> network.ping-timeout: 20 >>>>>>> nfs.disable: on >>>>>>> performance.readdir-ahead: on >>>>>>> transport.address-family: inet >>>>>>> config.transport: rdma >>>>>>> cluster.server-quorum-type: server >>>>>>> cluster.quorum-type: fixed >>>>>>> cluster.quorum-count: 1 >>>>>>> locks.mandatory-locking: off >>>>>>> cluster.enable-shared-storage: disable >>>>>>> cluster.server-quorum-ratio: 51% >>>>>>> >>>>>>> When my single master node is up only, but other nodes are still >>>>>>> showing connected mode .... >>>>>>> gluster pool list >>>>>>> UUID Hostname State >>>>>>> 89ccd72e-cb99-4b52-a2c0-388c99e5c7b3 master2-ib.dbt.au >>>>>>> Connected >>>>>>> d2c47fc2-f673-4790-b368-d214a58c59f4 compute01-ib.dbt.au >>>>>>> Connected >>>>>>> a5608d66-a3c6-450e-a239-108668083ff2 localhost >>>>>>> Connected >>>>>>> [root@master1 ~]# >>>>>>> >>>>>>> >>>>>>> Please advise us >>>>>>> Is this normal behavior Or This is issue. >>>>>>> >>>>>>> >>>>>> First of, we don't have any master slave configuration mode for >>>>>> gluster >>>>>> trusted storage pool i.e. peer list. Secondly, if master2 and >>>>>> compute01 are >>>>>> still reflecting as 'connected' even though they are down it means >>>>>> that >>>>>> localhost here didn't receive disconnect events for some reason. >>>>>> Could you >>>>>> restart glusterd service on this node and check the output of gluster >>>>>> pool >>>>>> list again? >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>> Thank You >>>>>>> Atul Yadav >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Gluster-users mailing list >>>>>>> [email protected] >>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>> >>>> >>> >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> [email protected] >>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >>> >
_______________________________________________ Gluster-users mailing list [email protected] http://www.gluster.org/mailman/listinfo/gluster-users
