[ 
https://issues.apache.org/jira/browse/CASSANDRA-10052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14725534#comment-14725534
 ] 

Tyler Hobbs edited comment on CASSANDRA-10052 at 9/1/15 3:20 PM:
-----------------------------------------------------------------

bq. I'm not sure why we should skip notifications altogether rather than still 
sending them using the Gossip endpoint address, in a similar way to what 
CASSANDRA-5899 does when rcp_address is set to 0.0.0.0? Tyler Hobbs any 
thoughts?

The python driver uniquely identifies nodes by their {{rpc_address}} (or 
{{broadcast_rpc_address}}).  I'm looking into what the other drivers do, but 
I'm guessing it's the same (EDIT: they are the same).  If we send a 
notification with the {{listen}}/{{broadcast_address}}, the drivers are likely 
to ignore it because the address won't be recognized.

Ultimately, I think the drivers may need to move to uniquely representing hosts 
by their {{broadcast_address}} to handle this kind of setup better.  That's a 
little tricky, because the initial contact points and load balancing policies 
all currently work on rpc addresses.  To help support this on the C* side, we 
should consider sending both the {{rpc_address}} and {{broadcast_address}} in 
push notifications.

bq. I see. Sounds like we should just special-case it and not send anything 
from onDown if a peer listening on localhost goes down.

I think this would work okay if we change the condition a bit.  Instead, don't 
send anything from {{onDown}} if a peer with the same 
({{broadcast_}}){{rpc_address}} that we have goes down.  That _should_ only 
happen when the cluster is set up like this, but will still allow setups like 
ccm's to work normally.


was (Author: thobbs):
bq. I'm not sure why we should skip notifications altogether rather than still 
sending them using the Gossip endpoint address, in a similar way to what 
CASSANDRA-5899 does when rcp_address is set to 0.0.0.0? Tyler Hobbs any 
thoughts?

The python driver uniquely identifies nodes by their {{rpc_address}} (or 
{{broadcast_rpc_address}}).  I'm looking into what the other drivers do, but 
I'm guessing it's the same.  If we send a notification with the 
{{listen}}/{{broadcast_address}}, the drivers are likely to ignore it because 
the address won't be recognized.

Ultimately, I think the drivers may need to move to uniquely representing hosts 
by their {{broadcast_address}} to handle this kind of setup better.  That's a 
little tricky, because the initial contact points and load balancing policies 
all currently work on rpc addresses.  To help support this on the C* side, we 
should consider sending both the {{rpc_address}} and {{broadcast_address}} in 
push notifications.

bq. I see. Sounds like we should just special-case it and not send anything 
from onDown if a peer listening on localhost goes down.

I think this would work okay if we change the condition a bit.  Instead, don't 
send anything from {{onDown}} if a peer with the same 
({{broadcast_}}){{rpc_address}} that we have goes down.  That _should_ only 
happen when the cluster is set up like this, but will still allow setups like 
ccm's to work normally.

> Bringing one node down, makes the whole cluster go down for a second
> --------------------------------------------------------------------
>
>                 Key: CASSANDRA-10052
>                 URL: https://issues.apache.org/jira/browse/CASSANDRA-10052
>             Project: Cassandra
>          Issue Type: Bug
>            Reporter: Sharvanath Pathak
>            Assignee: Stefania
>              Labels: client-impacting
>             Fix For: 2.1.x, 2.2.x
>
>
> When a node goes down, the other nodes learn that through the gossip.
> And I do see the log from (Gossiper.java):
> {code}
> private void markDead(InetAddress addr, EndpointState localState)
>    {
>        if (logger.isTraceEnabled())
>            logger.trace("marking as down {}", addr);
>        localState.markDead();
>        liveEndpoints.remove(addr);
>        unreachableEndpoints.put(addr, System.nanoTime());
>        logger.info("InetAddress {} is now DOWN", addr);
>        for (IEndpointStateChangeSubscriber subscriber : subscribers)
>            subscriber.onDead(addr, localState);
>        if (logger.isTraceEnabled())
>            logger.trace("Notified " + subscribers);
>    }
> {code}
> Saying: "InetAddress 192.168.101.1 is now Down", in the Cassandra's system 
> log.
> Now on all the other nodes the client side (java driver) says, " Cannot 
> connect to any host, scheduling retry in 1000 milliseconds". They eventually 
> do reconnect but some queries fail during this intermediate period.
> To me it seems like when the server pushes the nodeDown event, it call the 
> getRpcAddress(endpoint), and thus sends localhost as the argument in the 
> nodeDown event.  
> As in org.apache.cassandra.transport.Server.java
> {code}
>   public void onDown(InetAddress endpoint)
>        {      
>            
> server.connectionTracker.send(Event.StatusChange.nodeDown(getRpcAddress(endpoint),
>  server.socket.getPort()));
>        }
> {code}
> the getRpcAddress returns localhost for any endpoint if the cassandra.yaml is 
> using localhost as the configuration for rpc_address (which by the way is the 
> default).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to