arjunashok commented on code in PR #58:
URL: https://github.com/apache/cassandra-sidecar/pull/58#discussion_r1307686750


##########
adapters/base/src/main/java/org/apache/cassandra/sidecar/adapters/base/TokenRangeReplicaProvider.java:
##########
@@ -0,0 +1,249 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.cassandra.sidecar.adapters.base;
+
+import java.math.BigInteger;
+import java.net.UnknownHostException;
+import java.util.Collection;
+import java.util.HashSet;
+import java.util.List;
+import java.util.Map;
+import java.util.Objects;
+import java.util.Set;
+import java.util.function.Function;
+import java.util.stream.Collectors;
+import java.util.stream.Stream;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.cassandra.sidecar.common.JmxClient;
+import org.apache.cassandra.sidecar.common.data.GossipInfoResponse;
+import org.apache.cassandra.sidecar.common.data.TokenRangeReplicasResponse;
+import org.apache.cassandra.sidecar.common.utils.GossipInfoParser;
+import org.jetbrains.annotations.NotNull;
+
+import static 
org.apache.cassandra.sidecar.adapters.base.ClusterMembershipJmxOperations.FAILURE_DETECTOR_OBJ_NAME;
+import static 
org.apache.cassandra.sidecar.adapters.base.EndpointSnitchJmxOperations.ENDPOINT_SNITCH_INFO_OBJ_NAME;
+import static 
org.apache.cassandra.sidecar.adapters.base.StorageJmxOperations.STORAGE_SERVICE_OBJ_NAME;
+
+/**
+ * Aggregates the replica-set by token range
+ */
+public class TokenRangeReplicaProvider
+{
+    private final JmxClient jmxClient;
+
+    private static final Logger LOGGER = 
LoggerFactory.getLogger(TokenRangeReplicaProvider.class);
+
+    public TokenRangeReplicaProvider(JmxClient jmxClient)
+    {
+        this.jmxClient = jmxClient;
+    }
+
+    public TokenRangeReplicasResponse tokenRangeReplicas(String keyspace, 
Partitioner partitioner)
+    {
+        Objects.requireNonNull(keyspace, "keyspace must be non-null");
+
+        StorageJmxOperations storage = 
jmxClient.proxy(StorageJmxOperations.class, STORAGE_SERVICE_OBJ_NAME);
+
+        // Retrieve map of primary token ranges to endpoints that describe the 
ring topology
+        Map<List<String>, List<String>> rangeToEndpointMappings = 
storage.getRangeToEndpointWithPortMap(keyspace);
+        // Pending ranges include bootstrap tokens and leaving endpoints as 
represented in the Cassandra TokenMetadata
+        Map<List<String>, List<String>> pendingRangeMappings = 
storage.getPendingRangeToEndpointWithPortMap(keyspace);
+
+        Set<String> replicaSet = 
Stream.concat(rangeToEndpointMappings.values().stream().flatMap(List::stream),
+                                               
pendingRangeMappings.values().stream().flatMap(List::stream))
+                                       .collect(Collectors.toSet());
+
+        Map<String, String> hostToDatacenter = 
groupHostsByDatacenter(replicaSet);
+
+        // Retrieve map of all token ranges (pending & primary) to endpoints
+        List<TokenRangeReplicasResponse.ReplicaInfo> writeReplicas =
+        writeReplicasFromPendingRanges(rangeToEndpointMappings,
+                                       pendingRangeMappings,
+                                       hostToDatacenter,
+                                       partitioner,
+                                       keyspace);
+
+        Map<String, String> replicaToStateMap = replicaToStateMap(replicaSet, 
storage);
+
+        return new TokenRangeReplicasResponse(
+        replicaToStateMap,
+        writeReplicas,
+        mappingsToUnwrappedReplicaSet(rangeToEndpointMappings, 
hostToDatacenter, partitioner));
+    }
+
+    private Map<String, String> replicaToStateMap(Set<String> replicaSet, 
StorageJmxOperations storage)
+    {
+        List<String> joiningNodes = storage.getJoiningNodesWithPort();
+        List<String> leavingNodes = storage.getLeavingNodesWithPort();
+        List<String> movingNodes = storage.getMovingNodesWithPort();
+
+        String rawGossipInfo = getRawGossipInfo();
+        GossipInfoResponse gossipInfo = GossipInfoParser.parse(rawGossipInfo);
+
+        StateWithReplacement state = new StateWithReplacement(joiningNodes, 
leavingNodes, movingNodes, gossipInfo);
+
+        return replicaSet.stream()
+                         .collect(Collectors.toMap(Function.identity(), 
state::of));
+    }
+
+    private String getRawGossipInfo()
+    {
+        return jmxClient.proxy(ClusterMembershipJmxOperations.class, 
FAILURE_DETECTOR_OBJ_NAME)
+                        .getAllEndpointStatesWithPort();
+    }
+
+    private List<TokenRangeReplicasResponse.ReplicaInfo>
+    writeReplicasFromPendingRanges(Map<List<String>, List<String>> 
naturalReplicaMappings,
+                                   Map<List<String>, List<String>> 
pendingRangeMappings,
+                                   Map<String, String> hostToDatacenter,
+                                   Partitioner partitioner,
+                                   String keyspace)
+    {
+        LOGGER.debug("Pending token ranges for keyspace={}, 
pendingRangeMappings={}", keyspace, pendingRangeMappings);
+        // Merge natural and pending range replicas to generate candidates for 
write-replicas
+        List<TokenRangeReplicas> replicas = Stream.concat(
+                                                  
naturalReplicaMappings.entrySet().stream(),
+                                                  
pendingRangeMappings.entrySet().stream())
+                                                  .map(entry -> 
TokenRangeReplicas.generateTokenRangeReplicas(
+                                                  new 
BigInteger(entry.getKey().get(0)),
+                                                  new 
BigInteger(entry.getKey().get(1)),
+                                                  partitioner,
+                                                  new 
HashSet<>(entry.getValue())))
+                                                  .flatMap(Collection::stream)
+                                                  
.collect(Collectors.toList());
+
+        // Candidate write-replica mappings (merged from natural and pending 
ranges) are normalized
+        // by consolidating overlapping ranges
+        return TokenRangeReplicas.normalize(replicas).stream()
+                                 .map(range -> {
+                                     Map<String, List<String>> replicasByDc =
+                                     replicasByDataCenter(hostToDatacenter, 
range.replicaSet());
+                                     return new 
TokenRangeReplicasResponse.ReplicaInfo(range.start().toString(),
+                                                                               
        range.end().toString(),
+                                                                               
        replicasByDc);
+                                 })
+                                 .collect(Collectors.toList());
+    }
+
+    private List<TokenRangeReplicasResponse.ReplicaInfo>
+    mappingsToUnwrappedReplicaSet(Map<List<String>, List<String>> 
replicasByTokenRange,
+                                  Map<String, String> hostToDatacenter,
+                                  Partitioner partitioner)
+    {
+        return replicasByTokenRange.entrySet().stream()
+                                   .map(entry -> 
TokenRangeReplicas.generateTokenRangeReplicas(
+                                   new BigInteger(entry.getKey().get(0)),
+                                   new BigInteger(entry.getKey().get(1)),
+                                   partitioner,
+                                   new HashSet<>(entry.getValue())))
+                                   .flatMap(Collection::stream)
+                                   .sorted()
+                                   .map(rep -> {
+                                       Map<String, List<String>> replicasByDc =
+                                       replicasByDataCenter(hostToDatacenter, 
rep.replicaSet());
+                                       return new 
TokenRangeReplicasResponse.ReplicaInfo(rep.start().toString(),
+                                                                               
          rep.end().toString(),
+                                                                               
          replicasByDc);
+                                   })
+                                   .collect(Collectors.toList());
+    }
+
+    private Map<String, String> groupHostsByDatacenter(Set<String> replicaSet)

Review Comment:
   Will update. This method was grouping by DC is a previous iteration which I 
did not rename when modified. 



##########
adapters/base/src/main/java/org/apache/cassandra/sidecar/adapters/base/TokenRangeReplicaProvider.java:
##########
@@ -0,0 +1,249 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.cassandra.sidecar.adapters.base;
+
+import java.math.BigInteger;
+import java.net.UnknownHostException;
+import java.util.Collection;
+import java.util.HashSet;
+import java.util.List;
+import java.util.Map;
+import java.util.Objects;
+import java.util.Set;
+import java.util.function.Function;
+import java.util.stream.Collectors;
+import java.util.stream.Stream;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.cassandra.sidecar.common.JmxClient;
+import org.apache.cassandra.sidecar.common.data.GossipInfoResponse;
+import org.apache.cassandra.sidecar.common.data.TokenRangeReplicasResponse;
+import org.apache.cassandra.sidecar.common.utils.GossipInfoParser;
+import org.jetbrains.annotations.NotNull;
+
+import static 
org.apache.cassandra.sidecar.adapters.base.ClusterMembershipJmxOperations.FAILURE_DETECTOR_OBJ_NAME;
+import static 
org.apache.cassandra.sidecar.adapters.base.EndpointSnitchJmxOperations.ENDPOINT_SNITCH_INFO_OBJ_NAME;
+import static 
org.apache.cassandra.sidecar.adapters.base.StorageJmxOperations.STORAGE_SERVICE_OBJ_NAME;
+
+/**
+ * Aggregates the replica-set by token range
+ */
+public class TokenRangeReplicaProvider
+{
+    private final JmxClient jmxClient;
+
+    private static final Logger LOGGER = 
LoggerFactory.getLogger(TokenRangeReplicaProvider.class);
+
+    public TokenRangeReplicaProvider(JmxClient jmxClient)
+    {
+        this.jmxClient = jmxClient;
+    }
+
+    public TokenRangeReplicasResponse tokenRangeReplicas(String keyspace, 
Partitioner partitioner)
+    {
+        Objects.requireNonNull(keyspace, "keyspace must be non-null");
+
+        StorageJmxOperations storage = 
jmxClient.proxy(StorageJmxOperations.class, STORAGE_SERVICE_OBJ_NAME);
+
+        // Retrieve map of primary token ranges to endpoints that describe the 
ring topology
+        Map<List<String>, List<String>> rangeToEndpointMappings = 
storage.getRangeToEndpointWithPortMap(keyspace);
+        // Pending ranges include bootstrap tokens and leaving endpoints as 
represented in the Cassandra TokenMetadata
+        Map<List<String>, List<String>> pendingRangeMappings = 
storage.getPendingRangeToEndpointWithPortMap(keyspace);
+
+        Set<String> replicaSet = 
Stream.concat(rangeToEndpointMappings.values().stream().flatMap(List::stream),
+                                               
pendingRangeMappings.values().stream().flatMap(List::stream))
+                                       .collect(Collectors.toSet());
+
+        Map<String, String> hostToDatacenter = 
groupHostsByDatacenter(replicaSet);
+
+        // Retrieve map of all token ranges (pending & primary) to endpoints
+        List<TokenRangeReplicasResponse.ReplicaInfo> writeReplicas =
+        writeReplicasFromPendingRanges(rangeToEndpointMappings,
+                                       pendingRangeMappings,
+                                       hostToDatacenter,
+                                       partitioner,
+                                       keyspace);
+
+        Map<String, String> replicaToStateMap = replicaToStateMap(replicaSet, 
storage);
+
+        return new TokenRangeReplicasResponse(
+        replicaToStateMap,
+        writeReplicas,
+        mappingsToUnwrappedReplicaSet(rangeToEndpointMappings, 
hostToDatacenter, partitioner));
+    }
+
+    private Map<String, String> replicaToStateMap(Set<String> replicaSet, 
StorageJmxOperations storage)
+    {
+        List<String> joiningNodes = storage.getJoiningNodesWithPort();
+        List<String> leavingNodes = storage.getLeavingNodesWithPort();
+        List<String> movingNodes = storage.getMovingNodesWithPort();
+
+        String rawGossipInfo = getRawGossipInfo();
+        GossipInfoResponse gossipInfo = GossipInfoParser.parse(rawGossipInfo);
+
+        StateWithReplacement state = new StateWithReplacement(joiningNodes, 
leavingNodes, movingNodes, gossipInfo);
+
+        return replicaSet.stream()
+                         .collect(Collectors.toMap(Function.identity(), 
state::of));
+    }
+
+    private String getRawGossipInfo()
+    {
+        return jmxClient.proxy(ClusterMembershipJmxOperations.class, 
FAILURE_DETECTOR_OBJ_NAME)
+                        .getAllEndpointStatesWithPort();
+    }
+
+    private List<TokenRangeReplicasResponse.ReplicaInfo>
+    writeReplicasFromPendingRanges(Map<List<String>, List<String>> 
naturalReplicaMappings,
+                                   Map<List<String>, List<String>> 
pendingRangeMappings,
+                                   Map<String, String> hostToDatacenter,
+                                   Partitioner partitioner,
+                                   String keyspace)
+    {
+        LOGGER.debug("Pending token ranges for keyspace={}, 
pendingRangeMappings={}", keyspace, pendingRangeMappings);
+        // Merge natural and pending range replicas to generate candidates for 
write-replicas
+        List<TokenRangeReplicas> replicas = Stream.concat(
+                                                  
naturalReplicaMappings.entrySet().stream(),
+                                                  
pendingRangeMappings.entrySet().stream())
+                                                  .map(entry -> 
TokenRangeReplicas.generateTokenRangeReplicas(
+                                                  new 
BigInteger(entry.getKey().get(0)),
+                                                  new 
BigInteger(entry.getKey().get(1)),
+                                                  partitioner,
+                                                  new 
HashSet<>(entry.getValue())))
+                                                  .flatMap(Collection::stream)
+                                                  
.collect(Collectors.toList());
+
+        // Candidate write-replica mappings (merged from natural and pending 
ranges) are normalized
+        // by consolidating overlapping ranges
+        return TokenRangeReplicas.normalize(replicas).stream()
+                                 .map(range -> {
+                                     Map<String, List<String>> replicasByDc =
+                                     replicasByDataCenter(hostToDatacenter, 
range.replicaSet());
+                                     return new 
TokenRangeReplicasResponse.ReplicaInfo(range.start().toString(),
+                                                                               
        range.end().toString(),
+                                                                               
        replicasByDc);
+                                 })
+                                 .collect(Collectors.toList());
+    }
+
+    private List<TokenRangeReplicasResponse.ReplicaInfo>
+    mappingsToUnwrappedReplicaSet(Map<List<String>, List<String>> 
replicasByTokenRange,
+                                  Map<String, String> hostToDatacenter,
+                                  Partitioner partitioner)
+    {
+        return replicasByTokenRange.entrySet().stream()
+                                   .map(entry -> 
TokenRangeReplicas.generateTokenRangeReplicas(
+                                   new BigInteger(entry.getKey().get(0)),
+                                   new BigInteger(entry.getKey().get(1)),
+                                   partitioner,
+                                   new HashSet<>(entry.getValue())))
+                                   .flatMap(Collection::stream)
+                                   .sorted()
+                                   .map(rep -> {
+                                       Map<String, List<String>> replicasByDc =
+                                       replicasByDataCenter(hostToDatacenter, 
rep.replicaSet());
+                                       return new 
TokenRangeReplicasResponse.ReplicaInfo(rep.start().toString(),
+                                                                               
          rep.end().toString(),
+                                                                               
          replicasByDc);
+                                   })
+                                   .collect(Collectors.toList());
+    }
+
+    private Map<String, String> groupHostsByDatacenter(Set<String> replicaSet)
+    {
+        EndpointSnitchJmxOperations endpointSnitchInfo = 
jmxClient.proxy(EndpointSnitchJmxOperations.class,
+                                                                         
ENDPOINT_SNITCH_INFO_OBJ_NAME);
+
+        return replicaSet.stream()
+                         .collect(Collectors.toMap(Function.identity(),
+                                                   (String host) -> 
getDatacenter(endpointSnitchInfo, host)));
+    }
+
+    private String getDatacenter(EndpointSnitchJmxOperations 
endpointSnitchInfo, String host)
+    {
+        try
+        {
+            return endpointSnitchInfo.getDatacenter(host);
+        }
+        catch (UnknownHostException e)
+        {
+            throw new RuntimeException(e);
+        }
+    }
+
+    @NotNull
+    private static Map<String, List<String>> replicasByDataCenter(Map<String, 
String> hostToDatacenter,
+                                                                  
Collection<String> replicas)
+    {
+        return 
replicas.stream().collect(Collectors.groupingBy(hostToDatacenter::get));
+    }
+
+    /**
+     * We want to identity a joining node, to replace a dead node, differently 
from a newly joining node. To
+     * do this we analyze gossip info and set 'Replacing' state for node 
replacing a dead node.
+     * {@link StateWithReplacement} is used to set replacing state for a node.
+     *
+     * <p>We are adding this state for token range replica provider endpoint. 
To send out replicas for a
+     * range along with state of replicas including replacing state.
+     */
+    static class StateWithReplacement extends RingProvider.State
+    {
+        private static final String STATE_REPLACING = "Replacing";

Review Comment:
   Always in favor of using enums for grouped constants such as this case. Will 
update.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to