Re: [DRBD-user] Testing DRBD9*.rpms: Not able to join the second node
Roland, thanks for the reply see my reply below On Wed, Dec 28, 2016 at 3:38 AM, Roland Kammererwrote: > On Tue, Dec 27, 2016 at 10:17:32AM -0600, T.J. Yang wrote: > > Hi > > > > I am testing out latest DRBD9 rpms, hoping to create bravo,alpha and > > charlie 3 nodes cluster. Need pointer on where I did wrong. > > > > 2. Trying add 2nd node alpha from bravo node. > > > > drbdmanage add-node alpha 192.168.174.136 > > > > The command failed after waiting on server (which server ?) > > Its a client/server application, the server is drbdmanaged. > > > I login into alpha to run the command manually. > > > > [root@alpha drbd9-rpms]# /bin/python /usr/bin/drbdmanage join -p 6999 > > 192.168.174.136 1 bravo 192.168.174.141 0 AMcP2OIcqCwO+iTer5Bx > > That was spit out by the failed autojoin or by "howto-join", or manually > crafted? > I copy it out of htop output on alpha node, this command was issued to alpha from bravo by drbdmanage init command. > > > Waiting for server: ... > > Error: Server currently not ready, please retry later > > drbdmanaged failed to start up, because it failed on DRBD level. > > > [ 9031.805838] drbd .drbdctrl alpha: Starting receiver thread (from > > drbd_w_.drbdctr [42225]) > > [ 9031.806840] drbd .drbdctrl alpha: conn( Unconnected -> Connecting ) > > [ 9031.807425] drbd .drbdctrl tcp:alpha: bind before listen failed, err = > > -99 > > [ 9031.808002] drbd .drbdctrl alpha: Failed to initiate connection, > err=-99 > > [ 9031.808551] drbd .drbdctrl alpha: conn( Connecting -> Disconnecting ) > > [ 9031.811144] drbd .drbdctrl alpha: Connection closed > > [ 9031.813713] drbd .drbdctrl alpha: conn( Disconnecting -> StandAlone ) > > [ 9031.814300] drbd .drbdctrl alpha: Terminating receiver thread > > [ 9166.872047] drbd .drbdctrl alpha: Terminating sender thread > > That is the real issue. DRBD creates a kernel socket and calls > sock->ops->bind(). This returns -99, which is -EADDRNOTAVAIL. So the > question is why wouldn't it be available? Because I had firealld.service enabled, my bad. > Everything right with your IP > addresses? If you have multiple you used the one you want to use for > cluster communication on "init"? The nodes can reach each other and > everything is nice and cozy from a network point of view? > I missed checking firewalld step. once I fix that I am all good to test out drbd9 ;) > > > [root@bravo log]# drbdmanage n > > +--- > --+ > > | Name | Pool Size | Pool Free | | > > State | > > |--- > --| > > | alpha | unknown | unknown | | offline/quorum vote ignored, > > pending actions: adjust connections | > > | bravo | 4088 | 4080 | | > > ok | > > +--- > --+ > > "normal", as the control volumes did not connect. > > Regards, rck > ___ > drbd-user mailing list > drbd-user@lists.linbit.com > http://lists.linbit.com/mailman/listinfo/drbd-user > -- T.J. Yang ___ drbd-user mailing list drbd-user@lists.linbit.com http://lists.linbit.com/mailman/listinfo/drbd-user
Re: [DRBD-user] Testing new DRBD9 dedicated repo for PVE
Il 27/12/2016 10:19, Roberto Resoli ha scritto: > All is up and running nicely, in any case. I said that too early: I tried to restore a kvm virtual machine and it fails, systematically. I attached here the command with relative output, and the fragment of syslog related to restore. It seems that resources (the vm had two disks) are being created correctly, but access to them is failing. I have tried with another vm dump (one disk only); same problem. Bye, rob qmrestore-error.tgz Description: application/compressed-tar ___ drbd-user mailing list drbd-user@lists.linbit.com http://lists.linbit.com/mailman/listinfo/drbd-user
Re: [DRBD-user] Testing DRBD9*.rpms: Not able to join the second node
On Tue, Dec 27, 2016 at 10:17:32AM -0600, T.J. Yang wrote: > Hi > > I am testing out latest DRBD9 rpms, hoping to create bravo,alpha and > charlie 3 nodes cluster. Need pointer on where I did wrong. > > 2. Trying add 2nd node alpha from bravo node. > > drbdmanage add-node alpha 192.168.174.136 > > The command failed after waiting on server (which server ?) Its a client/server application, the server is drbdmanaged. > I login into alpha to run the command manually. > > [root@alpha drbd9-rpms]# /bin/python /usr/bin/drbdmanage join -p 6999 > 192.168.174.136 1 bravo 192.168.174.141 0 AMcP2OIcqCwO+iTer5Bx That was spit out by the failed autojoin or by "howto-join", or manually crafted? > Waiting for server: ... > Error: Server currently not ready, please retry later drbdmanaged failed to start up, because it failed on DRBD level. > [ 9031.805838] drbd .drbdctrl alpha: Starting receiver thread (from > drbd_w_.drbdctr [42225]) > [ 9031.806840] drbd .drbdctrl alpha: conn( Unconnected -> Connecting ) > [ 9031.807425] drbd .drbdctrl tcp:alpha: bind before listen failed, err = > -99 > [ 9031.808002] drbd .drbdctrl alpha: Failed to initiate connection, err=-99 > [ 9031.808551] drbd .drbdctrl alpha: conn( Connecting -> Disconnecting ) > [ 9031.811144] drbd .drbdctrl alpha: Connection closed > [ 9031.813713] drbd .drbdctrl alpha: conn( Disconnecting -> StandAlone ) > [ 9031.814300] drbd .drbdctrl alpha: Terminating receiver thread > [ 9166.872047] drbd .drbdctrl alpha: Terminating sender thread That is the real issue. DRBD creates a kernel socket and calls sock->ops->bind(). This returns -99, which is -EADDRNOTAVAIL. So the question is why wouldn't it be available? Everything right with your IP addresses? If you have multiple you used the one you want to use for cluster communication on "init"? The nodes can reach each other and everything is nice and cozy from a network point of view? > [root@bravo log]# drbdmanage n > +-+ > | Name | Pool Size | Pool Free | | > State | > |-| > | alpha | unknown | unknown | | offline/quorum vote ignored, > pending actions: adjust connections | > | bravo | 4088 | 4080 | | > ok | > +-+ "normal", as the control volumes did not connect. Regards, rck ___ drbd-user mailing list drbd-user@lists.linbit.com http://lists.linbit.com/mailman/listinfo/drbd-user
Re: [DRBD-user] drbdmanage error message: "IndexError: list index out of range"
On Tue, Dec 27, 2016 at 01:04:13PM -0600, T.J. Yang wrote: > Following error message was from deleting my last node(alpha) in a 3 node > cluster. > > [root@alpha drbd.d]# drbdmanage nT > raceback (most recent call last): I guess the 'T' from the drbdmanage line belongs to the second line and is "Traceback", but what exactly was on the first line? Regards, rck ___ drbd-user mailing list drbd-user@lists.linbit.com http://lists.linbit.com/mailman/listinfo/drbd-user