http://git-wip-us.apache.org/repos/asf/cassandra/blob/5fb4e58b/src/java/org/apache/cassandra/tools/NodeTool.java ---------------------------------------------------------------------- diff --cc src/java/org/apache/cassandra/tools/NodeTool.java index 10c0915,8e88f45..3cb51e2 --- a/src/java/org/apache/cassandra/tools/NodeTool.java +++ b/src/java/org/apache/cassandra/tools/NodeTool.java @@@ -336,26 -381,2319 +336,26 @@@ public class NodeToo } } - public static Map<String, SetHostStat> getOwnershipByDc(NodeProbe probe, boolean resolveIp, - Map<String, String> tokenToEndpoint, - Map<InetAddress, Float> ownerships) - @Command(name = "info", description = "Print node information (uptime, load, ...)") - public static class Info extends NodeToolCmd - { - @Option(name = {"-T", "--tokens"}, description = "Display all tokens") - private boolean tokens = false; - - @Override - public void execute(NodeProbe probe) - { - boolean gossipInitialized = probe.isInitialized(); - - System.out.printf("%-23s: %s%n", "ID", probe.getLocalHostId()); - System.out.printf("%-23s: %s%n", "Gossip active", gossipInitialized); - System.out.printf("%-23s: %s%n", "Thrift active", probe.isThriftServerRunning()); - System.out.printf("%-23s: %s%n", "Native Transport active", probe.isNativeTransportRunning()); - System.out.printf("%-23s: %s%n", "Load", probe.getLoadString()); - if (gossipInitialized) - System.out.printf("%-23s: %s%n", "Generation No", probe.getCurrentGenerationNumber()); - else - System.out.printf("%-23s: %s%n", "Generation No", 0); - - // Uptime - long secondsUp = probe.getUptime() / 1000; - System.out.printf("%-23s: %d%n", "Uptime (seconds)", secondsUp); - - // Memory usage - MemoryUsage heapUsage = probe.getHeapMemoryUsage(); - double memUsed = (double) heapUsage.getUsed() / (1024 * 1024); - double memMax = (double) heapUsage.getMax() / (1024 * 1024); - System.out.printf("%-23s: %.2f / %.2f%n", "Heap Memory (MB)", memUsed, memMax); - try - { - System.out.printf("%-23s: %.2f%n", "Off Heap Memory (MB)", getOffHeapMemoryUsed(probe)); - } - catch (RuntimeException e) - { - // offheap-metrics introduced in 2.1.3 - older versions do not have the appropriate mbeans - if (!(e.getCause() instanceof InstanceNotFoundException)) - throw e; - } - - // Data Center/Rack - System.out.printf("%-23s: %s%n", "Data Center", probe.getDataCenter()); - System.out.printf("%-23s: %s%n", "Rack", probe.getRack()); - - // Exceptions - System.out.printf("%-23s: %s%n", "Exceptions", probe.getStorageMetric("Exceptions")); - - CacheServiceMBean cacheService = probe.getCacheServiceMBean(); - - // Key Cache: Hits, Requests, RecentHitRate, SavePeriodInSeconds - System.out.printf("%-23s: entries %d, size %s, capacity %s, %d hits, %d requests, %.3f recent hit rate, %d save period in seconds%n", - "Key Cache", - probe.getCacheMetric("KeyCache", "Entries"), - FileUtils.stringifyFileSize((long) probe.getCacheMetric("KeyCache", "Size")), - FileUtils.stringifyFileSize((long) probe.getCacheMetric("KeyCache", "Capacity")), - probe.getCacheMetric("KeyCache", "Hits"), - probe.getCacheMetric("KeyCache", "Requests"), - probe.getCacheMetric("KeyCache", "HitRate"), - cacheService.getKeyCacheSavePeriodInSeconds()); - - // Row Cache: Hits, Requests, RecentHitRate, SavePeriodInSeconds - System.out.printf("%-23s: entries %d, size %s, capacity %s, %d hits, %d requests, %.3f recent hit rate, %d save period in seconds%n", - "Row Cache", - probe.getCacheMetric("RowCache", "Entries"), - FileUtils.stringifyFileSize((long) probe.getCacheMetric("RowCache", "Size")), - FileUtils.stringifyFileSize((long) probe.getCacheMetric("RowCache", "Capacity")), - probe.getCacheMetric("RowCache", "Hits"), - probe.getCacheMetric("RowCache", "Requests"), - probe.getCacheMetric("RowCache", "HitRate"), - cacheService.getRowCacheSavePeriodInSeconds()); - - // Counter Cache: Hits, Requests, RecentHitRate, SavePeriodInSeconds - System.out.printf("%-23s: entries %d, size %s, capacity %s, %d hits, %d requests, %.3f recent hit rate, %d save period in seconds%n", - "Counter Cache", - probe.getCacheMetric("CounterCache", "Entries"), - FileUtils.stringifyFileSize((long) probe.getCacheMetric("CounterCache", "Size")), - FileUtils.stringifyFileSize((long) probe.getCacheMetric("CounterCache", "Capacity")), - probe.getCacheMetric("CounterCache", "Hits"), - probe.getCacheMetric("CounterCache", "Requests"), - probe.getCacheMetric("CounterCache", "HitRate"), - cacheService.getCounterCacheSavePeriodInSeconds()); - - // check if node is already joined, before getting tokens, since it throws exception if not. - if (probe.isJoined()) - { - // Tokens - List<String> tokens = probe.getTokens(); - if (tokens.size() == 1 || this.tokens) - for (String token : tokens) - System.out.printf("%-23s: %s%n", "Token", token); - else - System.out.printf("%-23s: (invoke with -T/--tokens to see all %d tokens)%n", "Token", - tokens.size()); - } - else - { - System.out.printf("%-23s: (node is not joined to the cluster)%n", "Token"); - } - } - - /** - * Returns the total off heap memory used in MB. - * @return the total off heap memory used in MB. - */ - private static double getOffHeapMemoryUsed(NodeProbe probe) - { - long offHeapMemUsedInBytes = 0; - // get a list of column family stores - Iterator<Map.Entry<String, ColumnFamilyStoreMBean>> cfamilies = probe.getColumnFamilyStoreMBeanProxies(); - - while (cfamilies.hasNext()) - { - Entry<String, ColumnFamilyStoreMBean> entry = cfamilies.next(); - String keyspaceName = entry.getKey(); - String cfName = entry.getValue().getColumnFamilyName(); - - offHeapMemUsedInBytes += (Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "MemtableOffHeapSize"); - offHeapMemUsedInBytes += (Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "BloomFilterOffHeapMemoryUsed"); - offHeapMemUsedInBytes += (Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "IndexSummaryOffHeapMemoryUsed"); - offHeapMemUsedInBytes += (Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "CompressionMetadataOffHeapMemoryUsed"); - } - - return offHeapMemUsedInBytes / (1024d * 1024); - } - } - - @Command(name = "ring", description = "Print information about the token ring") - public static class Ring extends NodeToolCmd - { - @Arguments(description = "Specify a keyspace for accurate ownership information (topology awareness)") - private String keyspace = null; - - @Option(title = "resolve_ip", name = {"-r", "--resolve-ip"}, description = "Show node domain names instead of IPs") - private boolean resolveIp = false; - - @Override - public void execute(NodeProbe probe) - { - Map<String, String> tokensToEndpoints = probe.getTokenToEndpointMap(); - LinkedHashMultimap<String, String> endpointsToTokens = LinkedHashMultimap.create(); - boolean haveVnodes = false; - for (Map.Entry<String, String> entry : tokensToEndpoints.entrySet()) - { - haveVnodes |= endpointsToTokens.containsKey(entry.getValue()); - endpointsToTokens.put(entry.getValue(), entry.getKey()); - } - - int maxAddressLength = Collections.max(endpointsToTokens.keys(), new Comparator<String>() - { - @Override - public int compare(String first, String second) - { - return ((Integer) first.length()).compareTo(second.length()); - } - }).length(); - - String formatPlaceholder = "%%-%ds %%-12s%%-7s%%-8s%%-16s%%-20s%%-44s%%n"; - String format = format(formatPlaceholder, maxAddressLength); - - StringBuffer errors = new StringBuffer(); - boolean showEffectiveOwnership = true; - // Calculate per-token ownership of the ring - Map<InetAddress, Float> ownerships; - try - { - ownerships = probe.effectiveOwnership(keyspace); - } - catch (IllegalStateException ex) - { - ownerships = probe.getOwnership(); - errors.append("Note: " + ex.getMessage() + "%n"); - showEffectiveOwnership = false; - } - catch (IllegalArgumentException ex) - { - System.out.printf("%nError: " + ex.getMessage() + "%n"); - return; - } - - - System.out.println(); - for (Entry<String, SetHostStat> entry : getOwnershipByDc(probe, resolveIp, tokensToEndpoints, ownerships).entrySet()) - printDc(probe, format, entry.getKey(), endpointsToTokens, entry.getValue(),showEffectiveOwnership); - - if (haveVnodes) - { - System.out.println(" Warning: \"nodetool ring\" is used to output all the tokens of a node."); - System.out.println(" To view status related info of a node use \"nodetool status\" instead.\n"); - } - - System.out.printf("%n " + errors.toString()); - } - - private void printDc(NodeProbe probe, String format, - String dc, - LinkedHashMultimap<String, String> endpointsToTokens, - SetHostStat hoststats,boolean showEffectiveOwnership) - { - Collection<String> liveNodes = probe.getLiveNodes(); - Collection<String> deadNodes = probe.getUnreachableNodes(); - Collection<String> joiningNodes = probe.getJoiningNodes(); - Collection<String> leavingNodes = probe.getLeavingNodes(); - Collection<String> movingNodes = probe.getMovingNodes(); - Map<String, String> loadMap = probe.getLoadMap(); - - System.out.println("Datacenter: " + dc); - System.out.println("=========="); - - // get the total amount of replicas for this dc and the last token in this dc's ring - List<String> tokens = new ArrayList<>(); - String lastToken = ""; - - for (HostStat stat : hoststats) - { - tokens.addAll(endpointsToTokens.get(stat.endpoint.getHostAddress())); - lastToken = tokens.get(tokens.size() - 1); - } - - System.out.printf(format, "Address", "Rack", "Status", "State", "Load", "Owns", "Token"); - - if (hoststats.size() > 1) - System.out.printf(format, "", "", "", "", "", "", lastToken); - else - System.out.println(); - - for (HostStat stat : hoststats) - { - String endpoint = stat.endpoint.getHostAddress(); - String rack; - try - { - rack = probe.getEndpointSnitchInfoProxy().getRack(endpoint); - } - catch (UnknownHostException e) - { - rack = "Unknown"; - } - - String status = liveNodes.contains(endpoint) - ? "Up" - : deadNodes.contains(endpoint) - ? "Down" - : "?"; - - String state = "Normal"; - - if (joiningNodes.contains(endpoint)) - state = "Joining"; - else if (leavingNodes.contains(endpoint)) - state = "Leaving"; - else if (movingNodes.contains(endpoint)) - state = "Moving"; - - String load = loadMap.containsKey(endpoint) - ? loadMap.get(endpoint) - : "?"; - String owns = stat.owns != null && showEffectiveOwnership? new DecimalFormat("##0.00%").format(stat.owns) : "?"; - System.out.printf(format, stat.ipOrDns(), rack, status, state, load, owns, stat.token); - } - System.out.println(); - } - } - - @Command(name = "netstats", description = "Print network information on provided host (connecting node by default)") - public static class NetStats extends NodeToolCmd - { - @Option(title = "human_readable", - name = {"-H", "--human-readable"}, - description = "Display bytes in human readable form, i.e. KB, MB, GB, TB") - private boolean humanReadable = false; - - @Override - public void execute(NodeProbe probe) - { - System.out.printf("Mode: %s%n", probe.getOperationMode()); - Set<StreamState> statuses = probe.getStreamStatus(); - if (statuses.isEmpty()) - System.out.println("Not sending any streams."); - for (StreamState status : statuses) - { - System.out.printf("%s %s%n", status.description, status.planId.toString()); - for (SessionInfo info : status.sessions) - { - System.out.printf(" %s", info.peer.toString()); - // print private IP when it is used - if (!info.peer.equals(info.connecting)) - { - System.out.printf(" (using %s)", info.connecting.toString()); - } - System.out.printf("%n"); - if (!info.receivingSummaries.isEmpty()) - { - if (humanReadable) - System.out.printf(" Receiving %d files, %s total. Already received %d files, %s total%n", info.getTotalFilesToReceive(), FileUtils.stringifyFileSize(info.getTotalSizeToReceive()), info.getTotalFilesReceived(), FileUtils.stringifyFileSize(info.getTotalSizeReceived())); - else - System.out.printf(" Receiving %d files, %d bytes total. Already received %d files, %d bytes total%n", info.getTotalFilesToReceive(), info.getTotalSizeToReceive(), info.getTotalFilesReceived(), info.getTotalSizeReceived()); - for (ProgressInfo progress : info.getReceivingFiles()) - { - System.out.printf(" %s%n", progress.toString()); - } - } - if (!info.sendingSummaries.isEmpty()) - { - if (humanReadable) - System.out.printf(" Sending %d files, %s total. Already sent %d files, %s total%n", info.getTotalFilesToSend(), FileUtils.stringifyFileSize(info.getTotalSizeToSend()), info.getTotalFilesSent(), FileUtils.stringifyFileSize(info.getTotalSizeSent())); - else - System.out.printf(" Sending %d files, %d bytes total. Already sent %d files, %d bytes total%n", info.getTotalFilesToSend(), info.getTotalSizeToSend(), info.getTotalFilesSent(), info.getTotalSizeSent()); - for (ProgressInfo progress : info.getSendingFiles()) - { - System.out.printf(" %s%n", progress.toString()); - } - } - } - } - - if (!probe.isStarting()) - { - System.out.printf("Read Repair Statistics:%nAttempted: %d%nMismatch (Blocking): %d%nMismatch (Background): %d%n", probe.getReadRepairAttempted(), probe.getReadRepairRepairedBlocking(), probe.getReadRepairRepairedBackground()); - - MessagingServiceMBean ms = probe.msProxy; - System.out.printf("%-25s", "Pool Name"); - System.out.printf("%10s", "Active"); - System.out.printf("%10s", "Pending"); - System.out.printf("%15s%n", "Completed"); - - int pending; - long completed; - - pending = 0; - for (int n : ms.getCommandPendingTasks().values()) - pending += n; - completed = 0; - for (long n : ms.getCommandCompletedTasks().values()) - completed += n; - System.out.printf("%-25s%10s%10s%15s%n", "Commands", "n/a", pending, completed); - - pending = 0; - for (int n : ms.getResponsePendingTasks().values()) - pending += n; - completed = 0; - for (long n : ms.getResponseCompletedTasks().values()) - completed += n; - System.out.printf("%-25s%10s%10s%15s%n", "Responses", "n/a", pending, completed); - } - } - } - - @Command(name = "cfstats", description = "Print statistics on column families") - public static class CfStats extends NodeToolCmd ++ public static SortedMap<String, SetHostStat> getOwnershipByDc(NodeProbe probe, boolean resolveIp, ++ Map<String, String> tokenToEndpoint, ++ Map<InetAddress, Float> ownerships) { - Map<String, SetHostStat> ownershipByDc = Maps.newLinkedHashMap(); - @Arguments(usage = "[<keyspace.cfname>...]", description = "List of column families (or keyspace) names") - private List<String> cfnames = new ArrayList<>(); - - @Option(name = "-i", description = "Ignore the list of column families and display the remaining cfs") - private boolean ignore = false; - - @Option(title = "human_readable", - name = {"-H", "--human-readable"}, - description = "Display bytes in human readable form, i.e. KB, MB, GB, TB") - private boolean humanReadable = false; - - @Override - public void execute(NodeProbe probe) - { - OptionFilter filter = new OptionFilter(ignore, cfnames); - Map<String, List<ColumnFamilyStoreMBean>> cfstoreMap = new HashMap<>(); - - // get a list of column family stores - Iterator<Map.Entry<String, ColumnFamilyStoreMBean>> cfamilies = probe.getColumnFamilyStoreMBeanProxies(); - - while (cfamilies.hasNext()) - { - Map.Entry<String, ColumnFamilyStoreMBean> entry = cfamilies.next(); - String keyspaceName = entry.getKey(); - ColumnFamilyStoreMBean cfsProxy = entry.getValue(); - - if (!cfstoreMap.containsKey(keyspaceName) && filter.isColumnFamilyIncluded(entry.getKey(), cfsProxy.getColumnFamilyName())) - { - List<ColumnFamilyStoreMBean> columnFamilies = new ArrayList<>(); - columnFamilies.add(cfsProxy); - cfstoreMap.put(keyspaceName, columnFamilies); - } else if (filter.isColumnFamilyIncluded(entry.getKey(), cfsProxy.getColumnFamilyName())) - { - cfstoreMap.get(keyspaceName).add(cfsProxy); - } - } - - // make sure all specified kss and cfs exist - filter.verifyKeyspaces(probe.getKeyspaces()); - filter.verifyColumnFamilies(); - - // print out the table statistics - for (Map.Entry<String, List<ColumnFamilyStoreMBean>> entry : cfstoreMap.entrySet()) - { - String keyspaceName = entry.getKey(); - List<ColumnFamilyStoreMBean> columnFamilies = entry.getValue(); - long keyspaceReadCount = 0; - long keyspaceWriteCount = 0; - int keyspacePendingFlushes = 0; - double keyspaceTotalReadTime = 0.0f; - double keyspaceTotalWriteTime = 0.0f; - - System.out.println("Keyspace: " + keyspaceName); - for (ColumnFamilyStoreMBean cfstore : columnFamilies) - { - String cfName = cfstore.getColumnFamilyName(); - long writeCount = ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "WriteLatency")).getCount(); - long readCount = ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "ReadLatency")).getCount(); - - if (readCount > 0) - { - keyspaceReadCount += readCount; - keyspaceTotalReadTime += (long) probe.getColumnFamilyMetric(keyspaceName, cfName, "ReadTotalLatency"); - } - if (writeCount > 0) - { - keyspaceWriteCount += writeCount; - keyspaceTotalWriteTime += (long) probe.getColumnFamilyMetric(keyspaceName, cfName, "WriteTotalLatency"); - } - keyspacePendingFlushes += (long) probe.getColumnFamilyMetric(keyspaceName, cfName, "PendingFlushes"); - } - - double keyspaceReadLatency = keyspaceReadCount > 0 - ? keyspaceTotalReadTime / keyspaceReadCount / 1000 - : Double.NaN; - double keyspaceWriteLatency = keyspaceWriteCount > 0 - ? keyspaceTotalWriteTime / keyspaceWriteCount / 1000 - : Double.NaN; - - System.out.println("\tRead Count: " + keyspaceReadCount); - System.out.println("\tRead Latency: " + String.format("%s", keyspaceReadLatency) + " ms."); - System.out.println("\tWrite Count: " + keyspaceWriteCount); - System.out.println("\tWrite Latency: " + String.format("%s", keyspaceWriteLatency) + " ms."); - System.out.println("\tPending Flushes: " + keyspacePendingFlushes); - - // print out column family statistics for this keyspace - for (ColumnFamilyStoreMBean cfstore : columnFamilies) - { - String cfName = cfstore.getColumnFamilyName(); - if (cfName.contains(".")) - System.out.println("\t\tTable (index): " + cfName); - else - System.out.println("\t\tTable: " + cfName); - - System.out.println("\t\tSSTable count: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "LiveSSTableCount")); - - int[] leveledSStables = cfstore.getSSTableCountPerLevel(); - if (leveledSStables != null) - { - System.out.print("\t\tSSTables in each level: ["); - for (int level = 0; level < leveledSStables.length; level++) - { - int count = leveledSStables[level]; - System.out.print(count); - long maxCount = 4L; // for L0 - if (level > 0) - maxCount = (long) Math.pow(10, level); - // show max threshold for level when exceeded - if (count > maxCount) - System.out.print("/" + maxCount); - - if (level < leveledSStables.length - 1) - System.out.print(", "); - else - System.out.println("]"); - } - } - - Long memtableOffHeapSize = null; - Long bloomFilterOffHeapSize = null; - Long indexSummaryOffHeapSize = null; - Long compressionMetadataOffHeapSize = null; - - Long offHeapSize = null; - - try - { - memtableOffHeapSize = (Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "MemtableOffHeapSize"); - bloomFilterOffHeapSize = (Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "BloomFilterOffHeapMemoryUsed"); - indexSummaryOffHeapSize = (Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "IndexSummaryOffHeapMemoryUsed"); - compressionMetadataOffHeapSize = (Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "CompressionMetadataOffHeapMemoryUsed"); - - offHeapSize = memtableOffHeapSize + bloomFilterOffHeapSize + indexSummaryOffHeapSize + compressionMetadataOffHeapSize; - } - catch (RuntimeException e) - { - // offheap-metrics introduced in 2.1.3 - older versions do not have the appropriate mbeans - if (!(e.getCause() instanceof InstanceNotFoundException)) - throw e; - } - - System.out.println("\t\tSpace used (live): " + format((Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "LiveDiskSpaceUsed"), humanReadable)); - System.out.println("\t\tSpace used (total): " + format((Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "TotalDiskSpaceUsed"), humanReadable)); - System.out.println("\t\tSpace used by snapshots (total): " + format((Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "SnapshotsSize"), humanReadable)); - if (offHeapSize != null) - System.out.println("\t\tOff heap memory used (total): " + format(offHeapSize, humanReadable)); - System.out.println("\t\tSSTable Compression Ratio: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "CompressionRatio")); - System.out.println("\t\tNumber of keys (estimate): " + probe.getColumnFamilyMetric(keyspaceName, cfName, "EstimatedRowCount")); - System.out.println("\t\tMemtable cell count: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "MemtableColumnsCount")); - System.out.println("\t\tMemtable data size: " + format((Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "MemtableLiveDataSize"), humanReadable)); - if (memtableOffHeapSize != null) - System.out.println("\t\tMemtable off heap memory used: " + format(memtableOffHeapSize, humanReadable)); - System.out.println("\t\tMemtable switch count: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "MemtableSwitchCount")); - System.out.println("\t\tLocal read count: " + ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "ReadLatency")).getCount()); - double localReadLatency = ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "ReadLatency")).getMean() / 1000; - double localRLatency = localReadLatency > 0 ? localReadLatency : Double.NaN; - System.out.printf("\t\tLocal read latency: %01.3f ms%n", localRLatency); - System.out.println("\t\tLocal write count: " + ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "WriteLatency")).getCount()); - double localWriteLatency = ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "WriteLatency")).getMean() / 1000; - double localWLatency = localWriteLatency > 0 ? localWriteLatency : Double.NaN; - System.out.printf("\t\tLocal write latency: %01.3f ms%n", localWLatency); - System.out.println("\t\tPending flushes: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "PendingFlushes")); - System.out.println("\t\tBloom filter false positives: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "BloomFilterFalsePositives")); - System.out.printf("\t\tBloom filter false ratio: %s%n", String.format("%01.5f", probe.getColumnFamilyMetric(keyspaceName, cfName, "RecentBloomFilterFalseRatio"))); - System.out.println("\t\tBloom filter space used: " + format((Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "BloomFilterDiskSpaceUsed"), humanReadable)); - if (bloomFilterOffHeapSize != null) - System.out.println("\t\tBloom filter off heap memory used: " + format(bloomFilterOffHeapSize, humanReadable)); - if (indexSummaryOffHeapSize != null) - System.out.println("\t\tIndex summary off heap memory used: " + format(indexSummaryOffHeapSize, humanReadable)); - if (compressionMetadataOffHeapSize != null) - System.out.println("\t\tCompression metadata off heap memory used: " + format(compressionMetadataOffHeapSize, humanReadable)); - - System.out.println("\t\tCompacted partition minimum bytes: " + format((Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "MinRowSize"), humanReadable)); - System.out.println("\t\tCompacted partition maximum bytes: " + format((Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "MaxRowSize"), humanReadable)); - System.out.println("\t\tCompacted partition mean bytes: " + format((Long) probe.getColumnFamilyMetric(keyspaceName, cfName, "MeanRowSize"), humanReadable)); - JmxReporter.HistogramMBean histogram = (JmxReporter.HistogramMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "LiveScannedHistogram"); - System.out.println("\t\tAverage live cells per slice (last five minutes): " + histogram.getMean()); - System.out.println("\t\tMaximum live cells per slice (last five minutes): " + histogram.getMax()); - histogram = (JmxReporter.HistogramMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "TombstoneScannedHistogram"); - System.out.println("\t\tAverage tombstones per slice (last five minutes): " + histogram.getMean()); - System.out.println("\t\tMaximum tombstones per slice (last five minutes): " + histogram.getMax()); - - System.out.println(""); - } - System.out.println("----------------"); - } - } - - private String format(long bytes, boolean humanReadable) { - return humanReadable ? FileUtils.stringifyFileSize(bytes) : Long.toString(bytes); - } - - /** - * Used for filtering keyspaces and columnfamilies to be displayed using the cfstats command. - */ - private static class OptionFilter ++ SortedMap<String, SetHostStat> ownershipByDc = Maps.newTreeMap(); + EndpointSnitchInfoMBean epSnitchInfo = probe.getEndpointSnitchInfoProxy(); + try { - private Map<String, List<String>> filter = new HashMap<>(); - private Map<String, List<String>> verifier = new HashMap<>(); - private List<String> filterList = new ArrayList<>(); - private boolean ignoreMode; - - public OptionFilter(boolean ignoreMode, List<String> filterList) - { - this.filterList.addAll(filterList); - this.ignoreMode = ignoreMode; - - for (String s : filterList) - { - String[] keyValues = s.split("\\.", 2); - - // build the map that stores the ks' and cfs to use - if (!filter.containsKey(keyValues[0])) - { - filter.put(keyValues[0], new ArrayList<String>()); - verifier.put(keyValues[0], new ArrayList<String>()); - - if (keyValues.length == 2) - { - filter.get(keyValues[0]).add(keyValues[1]); - verifier.get(keyValues[0]).add(keyValues[1]); - } - } else - { - if (keyValues.length == 2) - { - filter.get(keyValues[0]).add(keyValues[1]); - verifier.get(keyValues[0]).add(keyValues[1]); - } - } - } - } - - public boolean isColumnFamilyIncluded(String keyspace, String columnFamily) - { - // supplying empty params list is treated as wanting to display all kss & cfs - if (filterList.isEmpty()) - return !ignoreMode; - - List<String> cfs = filter.get(keyspace); - - // no such keyspace is in the map - if (cfs == null) - return ignoreMode; - // only a keyspace with no cfs was supplied - // so ignore or include (based on the flag) every column family in specified keyspace - else if (cfs.size() == 0) - return !ignoreMode; - - // keyspace exists, and it contains specific cfs - verifier.get(keyspace).remove(columnFamily); - return ignoreMode ^ cfs.contains(columnFamily); - } - - public void verifyKeyspaces(List<String> keyspaces) - { - for (String ks : verifier.keySet()) - if (!keyspaces.contains(ks)) - throw new IllegalArgumentException("Unknown keyspace: " + ks); - } - - public void verifyColumnFamilies() + for (Entry<String, String> tokenAndEndPoint : tokenToEndpoint.entrySet()) { - for (String ks : filter.keySet()) - if (verifier.get(ks).size() > 0) - throw new IllegalArgumentException("Unknown column families: " + verifier.get(ks).toString() + " in keyspace: " + ks); + String dc = epSnitchInfo.getDatacenter(tokenAndEndPoint.getValue()); + if (!ownershipByDc.containsKey(dc)) + ownershipByDc.put(dc, new SetHostStat(resolveIp)); + ownershipByDc.get(dc).add(tokenAndEndPoint.getKey(), tokenAndEndPoint.getValue(), ownerships); } } - } - - @Command(name = "toppartitions", description = "Sample and print the most active partitions for a given column family") - public static class TopPartitions extends NodeToolCmd - { - @Arguments(usage = "<keyspace> <cfname> <duration>", description = "The keyspace, column family name, and duration in milliseconds") - private List<String> args = new ArrayList<>(); - @Option(name = "-s", description = "Capacity of stream summary, closer to the actual cardinality of partitions will yield more accurate results (Default: 256)") - private int size = 256; - @Option(name = "-k", description = "Number of the top partitions to list (Default: 10)") - private int topCount = 10; - @Option(name = "-a", description = "Comma separated list of samplers to use (Default: all)") - private String samplers = join(Sampler.values(), ','); - @Override - public void execute(NodeProbe probe) + catch (UnknownHostException e) { - checkArgument(args.size() == 3, "toppartitions requires keyspace, column family name, and duration"); - checkArgument(topCount < size, "TopK count (-k) option must be smaller then the summary capacity (-s)"); - String keyspace = args.get(0); - String cfname = args.get(1); - Integer duration = Integer.parseInt(args.get(2)); - // generate the list of samplers - List<Sampler> targets = Lists.newArrayList(); - for (String s : samplers.split(",")) - { - try - { - targets.add(Sampler.valueOf(s.toUpperCase())); - } catch (Exception e) - { - throw new IllegalArgumentException(s + " is not a valid sampler, choose one of: " + join(Sampler.values(), ", ")); - } - } - - Map<Sampler, CompositeData> results; - try - { - results = probe.getPartitionSample(keyspace, cfname, size, duration, topCount, targets); - } catch (OpenDataException e) - { - throw new RuntimeException(e); - } - boolean first = true; - for(Entry<Sampler, CompositeData> result : results.entrySet()) - { - CompositeData sampling = result.getValue(); - // weird casting for http://bugs.sun.com/view_bug.do?bug_id=6548436 - List<CompositeData> topk = (List<CompositeData>) (Object) Lists.newArrayList(((TabularDataSupport) sampling.get("partitions")).values()); - Collections.sort(topk, new Ordering<CompositeData>() - { - public int compare(CompositeData left, CompositeData right) - { - return Long.compare((long) right.get("count"), (long) left.get("count")); - } - }); - if(!first) - System.out.println(); - System.out.println(result.getKey().toString()+ " Sampler:"); - System.out.printf(" Cardinality: ~%d (%d capacity)%n", (long) sampling.get("cardinality"), size); - System.out.printf(" Top %d partitions:%n", topCount); - if (topk.size() == 0) - { - System.out.println("\tNothing recorded during sampling period..."); - } else - { - int offset = 0; - for (CompositeData entry : topk) - offset = Math.max(offset, entry.get("string").toString().length()); - System.out.printf("\t%-" + offset + "s%10s%10s%n", "Partition", "Count", "+/-"); - for (CompositeData entry : topk) - System.out.printf("\t%-" + offset + "s%10d%10d%n", entry.get("string").toString(), entry.get("count"), entry.get("error")); - } - first = false; - } + throw new RuntimeException(e); } + return ownershipByDc; } - - @Command(name = "cfhistograms", description = "Print statistic histograms for a given column family") - public static class CfHistograms extends NodeToolCmd - { - @Arguments(usage = "<keyspace> <cfname>", description = "The keyspace and column family name") - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - checkArgument(args.size() == 2, "cfhistograms requires ks and cf args"); - - String keyspace = args.get(0); - String cfname = args.get(1); - - ColumnFamilyStoreMBean store = probe.getCfsProxy(keyspace, cfname); - - long[] estimatedRowSizeHistogram = store.getEstimatedRowSizeHistogram(); - long[] estimatedColumnCountHistogram = store.getEstimatedColumnCountHistogram(); - - if (ArrayUtils.isEmpty(estimatedRowSizeHistogram) || ArrayUtils.isEmpty(estimatedColumnCountHistogram)) - { - System.err.println("No SSTables exists, unable to calculate 'Partition Size' and 'Cell Count' percentiles"); - } - - // calculate percentile of row size and column count - String[] percentiles = new String[]{"50%", "75%", "95%", "98%", "99%", "Min", "Max"}; - double[] readLatency = probe.metricPercentilesAsArray(store.getRecentReadLatencyHistogramMicros()); - double[] writeLatency = probe.metricPercentilesAsArray(store.getRecentWriteLatencyHistogramMicros()); - double[] estimatedRowSizePercentiles = probe.metricPercentilesAsArray(estimatedRowSizeHistogram); - double[] estimatedColumnCountPercentiles = probe.metricPercentilesAsArray(estimatedColumnCountHistogram); - double[] sstablesPerRead = probe.metricPercentilesAsArray(store.getRecentSSTablesPerReadHistogram()); - - System.out.println(format("%s/%s histograms", keyspace, cfname)); - System.out.println(format("%-10s%10s%18s%18s%18s%18s", - "Percentile", "SSTables", "Write Latency", "Read Latency", "Partition Size", "Cell Count")); - System.out.println(format("%-10s%10s%18s%18s%18s%18s", - "", "", "(micros)", "(micros)", "(bytes)", "")); - - for (int i = 0; i < percentiles.length; i++) - { - System.out.println(format("%-10s%10.2f%18.2f%18.2f%18.0f%18.0f", - percentiles[i], - sstablesPerRead[i], - writeLatency[i], - readLatency[i], - estimatedRowSizePercentiles[i], - estimatedColumnCountPercentiles[i])); - } - System.out.println(); - } - } - - @Command(name = "cleanup", description = "Triggers the immediate cleanup of keys no longer belonging to a node. By default, clean all keyspaces") - public static class Cleanup extends NodeToolCmd - { - @Arguments(usage = "[<keyspace> <cfnames>...]", description = "The keyspace followed by one or many column families") - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - List<String> keyspaces = parseOptionalKeyspace(args, probe); - String[] cfnames = parseOptionalColumnFamilies(args); - - for (String keyspace : keyspaces) - { - if (Keyspace.SYSTEM_KS.equals(keyspace)) - continue; - - try - { - probe.forceKeyspaceCleanup(System.out, keyspace, cfnames); - } catch (Exception e) - { - throw new RuntimeException("Error occurred during cleanup", e); - } - } - } - } - - @Command(name = "clearsnapshot", description = "Remove the snapshot with the given name from the given keyspaces. If no snapshotName is specified we will remove all snapshots") - public static class ClearSnapshot extends NodeToolCmd - { - @Arguments(usage = "[<keyspaces>...] ", description = "Remove snapshots from the given keyspaces") - private List<String> keyspaces = new ArrayList<>(); - - @Option(title = "snapshot_name", name = "-t", description = "Remove the snapshot with a given name") - private String snapshotName = EMPTY; - - @Override - public void execute(NodeProbe probe) - { - StringBuilder sb = new StringBuilder(); - - sb.append("Requested clearing snapshot(s) for "); - - if (keyspaces.isEmpty()) - sb.append("[all keyspaces]"); - else - sb.append("[").append(join(keyspaces, ", ")).append("]"); - - if (!snapshotName.isEmpty()) - sb.append(" with snapshot name [").append(snapshotName).append("]"); - - System.out.println(sb.toString()); - - try - { - probe.clearSnapshot(snapshotName, toArray(keyspaces, String.class)); - } catch (IOException e) - { - throw new RuntimeException("Error during clearing snapshots", e); - } - } - } - - @Command(name = "compact", description = "Force a (major) compaction on one or more column families") - public static class Compact extends NodeToolCmd - { - @Arguments(usage = "[<keyspace> <cfnames>...]", description = "The keyspace followed by one or many column families") - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - List<String> keyspaces = parseOptionalKeyspace(args, probe); - String[] cfnames = parseOptionalColumnFamilies(args); - - for (String keyspace : keyspaces) - { - try - { - probe.forceKeyspaceCompaction(keyspace, cfnames); - } catch (Exception e) - { - throw new RuntimeException("Error occurred during compaction", e); - } - } - } - } - - @Command(name = "flush", description = "Flush one or more column families") - public static class Flush extends NodeToolCmd - { - @Arguments(usage = "[<keyspace> <cfnames>...]", description = "The keyspace followed by one or many column families") - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - List<String> keyspaces = parseOptionalKeyspace(args, probe); - String[] cfnames = parseOptionalColumnFamilies(args); - - for (String keyspace : keyspaces) - { - try - { - probe.forceKeyspaceFlush(keyspace, cfnames); - } catch (Exception e) - { - throw new RuntimeException("Error occurred during flushing", e); - } - } - } - } - - @Command(name = "scrub", description = "Scrub (rebuild sstables for) one or more column families") - public static class Scrub extends NodeToolCmd - { - @Arguments(usage = "[<keyspace> <cfnames>...]", description = "The keyspace followed by one or many column families") - private List<String> args = new ArrayList<>(); - - @Option(title = "disable_snapshot", - name = {"-ns", "--no-snapshot"}, - description = "Scrubbed CFs will be snapshotted first, if disableSnapshot is false. (default false)") - private boolean disableSnapshot = false; - - @Option(title = "skip_corrupted", - name = {"-s", "--skip-corrupted"}, - description = "Skip corrupted partitions even when scrubbing counter tables. (default false)") - private boolean skipCorrupted = false; - - @Option(title = "no_validate", - name = {"-n", "--no-validate"}, - description = "Do not validate columns using column validator") - private boolean noValidation = false; - - @Override - public void execute(NodeProbe probe) - { - List<String> keyspaces = parseOptionalKeyspace(args, probe); - String[] cfnames = parseOptionalColumnFamilies(args); - - for (String keyspace : keyspaces) - { - try - { - probe.scrub(System.out, disableSnapshot, skipCorrupted, !noValidation, keyspace, cfnames); - } catch (Exception e) - { - throw new RuntimeException("Error occurred during flushing", e); - } - } - } - } - - @Command(name = "disableautocompaction", description = "Disable autocompaction for the given keyspace and column family") - public static class DisableAutoCompaction extends NodeToolCmd - { - @Arguments(usage = "[<keyspace> <cfnames>...]", description = "The keyspace followed by one or many column families") - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - List<String> keyspaces = parseOptionalKeyspace(args, probe); - String[] cfnames = parseOptionalColumnFamilies(args); - - for (String keyspace : keyspaces) - { - try - { - probe.disableAutoCompaction(keyspace, cfnames); - } catch (IOException e) - { - throw new RuntimeException("Error occurred during disabling auto-compaction", e); - } - } - } - } - - @Command(name = "enableautocompaction", description = "Enable autocompaction for the given keyspace and column family") - public static class EnableAutoCompaction extends NodeToolCmd - { - @Arguments(usage = "[<keyspace> <cfnames>...]", description = "The keyspace followed by one or many column families") - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - List<String> keyspaces = parseOptionalKeyspace(args, probe); - String[] cfnames = parseOptionalColumnFamilies(args); - - for (String keyspace : keyspaces) - { - try - { - probe.enableAutoCompaction(keyspace, cfnames); - } catch (IOException e) - { - throw new RuntimeException("Error occurred during enabling auto-compaction", e); - } - } - } - } - - @Command(name = "upgradesstables", description = "Rewrite sstables (for the requested column families) that are not on the current version (thus upgrading them to said current version)") - public static class UpgradeSSTable extends NodeToolCmd - { - @Arguments(usage = "[<keyspace> <cfnames>...]", description = "The keyspace followed by one or many column families") - private List<String> args = new ArrayList<>(); - - @Option(title = "include_all", name = {"-a", "--include-all-sstables"}, description = "Use -a to include all sstables, even those already on the current version") - private boolean includeAll = false; - - @Override - public void execute(NodeProbe probe) - { - List<String> keyspaces = parseOptionalKeyspace(args, probe); - String[] cfnames = parseOptionalColumnFamilies(args); - - for (String keyspace : keyspaces) - { - try - { - probe.upgradeSSTables(System.out, keyspace, !includeAll, cfnames); - } catch (Exception e) - { - throw new RuntimeException("Error occurred during enabling auto-compaction", e); - } - } - } - } - - @Command(name = "compactionstats", description = "Print statistics on compactions") - public static class CompactionStats extends NodeToolCmd - { - @Option(title = "human_readable", - name = {"-H", "--human-readable"}, - description = "Display bytes in human readable form, i.e. KB, MB, GB, TB") - private boolean humanReadable = false; - - @Override - public void execute(NodeProbe probe) - { - int compactionThroughput = probe.getCompactionThroughput(); - CompactionManagerMBean cm = probe.getCompactionManagerProxy(); - System.out.println("pending tasks: " + probe.getCompactionMetric("PendingTasks")); - long remainingBytes = 0; - List<Map<String, String>> compactions = cm.getCompactions(); - if (!compactions.isEmpty()) - { - List<String[]> lines = new ArrayList<>(); - int[] columnSizes = new int[] { 0, 0, 0, 0, 0, 0, 0 }; - - addLine(lines, columnSizes, "compaction type", "keyspace", "table", "completed", "total", "unit", "progress"); - for (Map<String, String> c : compactions) - { - long total = Long.parseLong(c.get("total")); - long completed = Long.parseLong(c.get("completed")); - String taskType = c.get("taskType"); - String keyspace = c.get("keyspace"); - String columnFamily = c.get("columnfamily"); - String completedStr = humanReadable ? FileUtils.stringifyFileSize(completed) : Long.toString(completed); - String totalStr = humanReadable ? FileUtils.stringifyFileSize(total) : Long.toString(total); - String unit = c.get("unit"); - String percentComplete = total == 0 ? "n/a" : new DecimalFormat("0.00").format((double) completed / total * 100) + "%"; - addLine(lines, columnSizes, taskType, keyspace, columnFamily, completedStr, totalStr, unit, percentComplete); - if (taskType.equals(OperationType.COMPACTION.toString())) - remainingBytes += total - completed; - } - - StringBuilder buffer = new StringBuilder(); - for (int columnSize : columnSizes) { - buffer.append("%"); - buffer.append(columnSize + 3); - buffer.append("s"); - } - buffer.append("%n"); - String format = buffer.toString(); - - for (String[] line : lines) - { - System.out.printf(format, line[0], line[1], line[2], line[3], line[4], line[5], line[6]); - } - - String remainingTime = "n/a"; - if (compactionThroughput != 0) - { - long remainingTimeInSecs = remainingBytes / (1024L * 1024L * compactionThroughput); - remainingTime = format("%dh%02dm%02ds", remainingTimeInSecs / 3600, (remainingTimeInSecs % 3600) / 60, (remainingTimeInSecs % 60)); - } - System.out.printf("%25s%10s%n", "Active compaction remaining time : ", remainingTime); - } - } - - private void addLine(List<String[]> lines, int[] columnSizes, String... columns) { - lines.add(columns); - for (int i = 0; i < columns.length; i++) { - columnSizes[i] = Math.max(columnSizes[i], columns[i].length()); - } - } - } - - @Command(name = "compactionhistory", description = "Print history of compaction") - public static class CompactionHistory extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - System.out.println("Compaction History: "); - - TabularData tabularData = probe.getCompactionHistory(); - if (tabularData.isEmpty()) - { - System.out.printf("There is no compaction history"); - return; - } - - String format = "%-41s%-19s%-29s%-26s%-15s%-15s%s%n"; - List<String> indexNames = tabularData.getTabularType().getIndexNames(); - System.out.printf(format, toArray(indexNames, Object.class)); - - Set<?> values = tabularData.keySet(); - for (Object eachValue : values) - { - List<?> value = (List<?>) eachValue; - System.out.printf(format, toArray(value, Object.class)); - } - } - } - - @Command(name = "decommission", description = "Decommission the *node I am connecting to*") - public static class Decommission extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - try - { - probe.decommission(); - } catch (InterruptedException e) - { - throw new RuntimeException("Error decommissioning node", e); - } - } - } - - @Command(name = "describecluster", description = "Print the name, snitch, partitioner and schema version of a cluster") - public static class DescribeCluster extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - // display cluster name, snitch and partitioner - System.out.println("Cluster Information:"); - System.out.println("\tName: " + probe.getClusterName()); - System.out.println("\tSnitch: " + probe.getEndpointSnitchInfoProxy().getSnitchName()); - System.out.println("\tPartitioner: " + probe.getPartitioner()); - - // display schema version for each node - System.out.println("\tSchema versions:"); - Map<String, List<String>> schemaVersions = probe.getSpProxy().getSchemaVersions(); - for (String version : schemaVersions.keySet()) - { - System.out.println(format("\t\t%s: %s%n", version, schemaVersions.get(version))); - } - } - } - - @Command(name = "disablebinary", description = "Disable native transport (binary protocol)") - public static class DisableBinary extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - probe.stopNativeTransport(); - } - } - - @Command(name = "enablebinary", description = "Reenable native transport (binary protocol)") - public static class EnableBinary extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - probe.startNativeTransport(); - } - } - - @Command(name = "enablegossip", description = "Reenable gossip") - public static class EnableGossip extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - probe.startGossiping(); - } - } - - @Command(name = "disablegossip", description = "Disable gossip (effectively marking the node down)") - public static class DisableGossip extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - probe.stopGossiping(); - } - } - - @Command(name = "enablehandoff", description = "Reenable the future hints storing on the current node") - public static class EnableHandoff extends NodeToolCmd - { - @Arguments(usage = "<dc-name>,<dc-name>", description = "Enable hinted handoff only for these DCs") - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - checkArgument(args.size() <= 1, "enablehandoff does not accept two args"); - if(args.size() == 1) - probe.enableHintedHandoff(args.get(0)); - else - probe.enableHintedHandoff(); - } - } - - @Command(name = "enablethrift", description = "Reenable thrift server") - public static class EnableThrift extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - probe.startThriftServer(); - } - } - - @Command(name = "getcompactionthreshold", description = "Print min and max compaction thresholds for a given column family") - public static class GetCompactionThreshold extends NodeToolCmd - { - @Arguments(usage = "<keyspace> <cfname>", description = "The keyspace with a column family") - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - checkArgument(args.size() == 2, "getcompactionthreshold requires ks and cf args"); - String ks = args.get(0); - String cf = args.get(1); - - ColumnFamilyStoreMBean cfsProxy = probe.getCfsProxy(ks, cf); - System.out.println("Current compaction thresholds for " + ks + "/" + cf + ": \n" + - " min = " + cfsProxy.getMinimumCompactionThreshold() + ", " + - " max = " + cfsProxy.getMaximumCompactionThreshold()); - } - } - - @Command(name = "getcompactionthroughput", description = "Print the MB/s throughput cap for compaction in the system") - public static class GetCompactionThroughput extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - System.out.println("Current compaction throughput: " + probe.getCompactionThroughput() + " MB/s"); - } - } - - @Command(name = "getstreamthroughput", description = "Print the Mb/s throughput cap for streaming in the system") - public static class GetStreamThroughput extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - System.out.println("Current stream throughput: " + probe.getStreamThroughput() + " Mb/s"); - } - } - - @Command(name = "getendpoints", description = "Print the end points that owns the key") - public static class GetEndpoints extends NodeToolCmd - { - @Arguments(usage = "<keyspace> <cfname> <key>", description = "The keyspace, the column family, and the partition key for which we need to find the endpoint") - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - checkArgument(args.size() == 3, "getendpoints requires ks, cf and key args"); - String ks = args.get(0); - String cf = args.get(1); - String key = args.get(2); - - List<InetAddress> endpoints = probe.getEndpoints(ks, cf, key); - for (InetAddress endpoint : endpoints) - { - System.out.println(endpoint.getHostAddress()); - } - } - } - - @Command(name = "getsstables", description = "Print the sstable filenames that own the key") - public static class GetSSTables extends NodeToolCmd - { - @Arguments(usage = "<keyspace> <cfname> <key>", description = "The keyspace, the column family, and the key") - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - checkArgument(args.size() == 3, "getsstables requires ks, cf and key args"); - String ks = args.get(0); - String cf = args.get(1); - String key = args.get(2); - - List<String> sstables = probe.getSSTables(ks, cf, key); - for (String sstable : sstables) - { - System.out.println(sstable); - } - } - } - - @Command(name = "gossipinfo", description = "Shows the gossip information for the cluster") - public static class GossipInfo extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - System.out.println(probe.getGossipInfo()); - } - } - - @Command(name = "invalidatekeycache", description = "Invalidate the key cache") - public static class InvalidateKeyCache extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - probe.invalidateKeyCache(); - } - } - - @Command(name = "invalidaterowcache", description = "Invalidate the row cache") - public static class InvalidateRowCache extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - probe.invalidateRowCache(); - } - } - - @Command(name = "invalidatecountercache", description = "Invalidate the counter cache") - public static class InvalidateCounterCache extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - probe.invalidateCounterCache(); - } - } - - @Command(name = "join", description = "Join the ring") - public static class Join extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - checkState(!probe.isJoined(), "This node has already joined the ring."); - - try - { - probe.joinRing(); - } catch (IOException e) - { - throw new RuntimeException("Error during joining the ring", e); - } - } - } - - @Command(name = "move", description = "Move node on the token ring to a new token") - public static class Move extends NodeToolCmd - { - @Arguments(usage = "<new token>", description = "The new token.", required = true) - private String newToken = EMPTY; - - @Override - public void execute(NodeProbe probe) - { - try - { - probe.move(newToken); - } catch (IOException e) - { - throw new RuntimeException("Error during moving node", e); - } - } - } - - - - @Command(name = "pausehandoff", description = "Pause hints delivery process") - public static class PauseHandoff extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - probe.pauseHintsDelivery(); - } - } - - @Command(name = "resumehandoff", description = "Resume hints delivery process") - public static class ResumeHandoff extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - probe.resumeHintsDelivery(); - } - } - - - @Command(name = "proxyhistograms", description = "Print statistic histograms for network operations") - public static class ProxyHistograms extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - StorageProxyMBean sp = probe.getSpProxy(); - String[] percentiles = new String[]{"50%", "75%", "95%", "98%", "99%", "Min", "Max"}; - double[] readLatency = probe.metricPercentilesAsArray(sp.getRecentReadLatencyHistogramMicros()); - double[] writeLatency = probe.metricPercentilesAsArray(sp.getRecentWriteLatencyHistogramMicros()); - double[] rangeLatency = probe.metricPercentilesAsArray(sp.getRecentRangeLatencyHistogramMicros()); - - System.out.println("proxy histograms"); - System.out.println(format("%-10s%18s%18s%18s", - "Percentile", "Read Latency", "Write Latency", "Range Latency")); - System.out.println(format("%-10s%18s%18s%18s", - "", "(micros)", "(micros)", "(micros)")); - for (int i = 0; i < percentiles.length; i++) - { - System.out.println(format("%-10s%18.2f%18.2f%18.2f", - percentiles[i], - readLatency[i], - writeLatency[i], - rangeLatency[i])); - } - System.out.println(); - } - } - - @Command(name = "rebuild", description = "Rebuild data by streaming from other nodes (similarly to bootstrap)") - public static class Rebuild extends NodeToolCmd - { - @Arguments(usage = "<src-dc-name>", description = "Name of DC from which to select sources for streaming. By default, pick any DC") - private String sourceDataCenterName = null; - - @Override - public void execute(NodeProbe probe) - { - probe.rebuild(sourceDataCenterName); - } - } - - @Command(name = "refresh", description = "Load newly placed SSTables to the system without restart") - public static class Refresh extends NodeToolCmd - { - @Arguments(usage = "<keyspace> <cfname>", description = "The keyspace and column family name") - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - checkArgument(args.size() == 2, "refresh requires ks and cf args"); - probe.loadNewSSTables(args.get(0), args.get(1)); - } - } - - @Deprecated - @Command(name = "removetoken", description = "DEPRECATED (see removenode)", hidden = true) - public static class RemoveToken extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - System.err.println("Warn: removetoken is deprecated, please use removenode instead"); - } - } - - @Command(name = "removenode", description = "Show status of current node removal, force completion of pending removal or remove provided ID") - public static class RemoveNode extends NodeToolCmd - { - @Arguments(title = "remove_operation", usage = "<status>|<force>|<ID>", description = "Show status of current node removal, force completion of pending removal, or remove provided ID", required = true) - private String removeOperation = EMPTY; - - @Override - public void execute(NodeProbe probe) - { - switch (removeOperation) - { - case "status": - System.out.println("RemovalStatus: " + probe.getRemovalStatus()); - break; - case "force": - System.out.println("RemovalStatus: " + probe.getRemovalStatus()); - probe.forceRemoveCompletion(); - break; - default: - probe.removeNode(removeOperation); - break; - } - } - } - - @Command(name = "repair", description = "Repair one or more column families") - public static class Repair extends NodeToolCmd - { - @Arguments(usage = "[<keyspace> <cfnames>...]", description = "The keyspace followed by one or many column families") - private List<String> args = new ArrayList<>(); - - @Option(title = "parallel", name = {"-par", "--parallel"}, description = "Use -par to carry out a parallel repair") - private boolean parallel = false; - - @Option(title = "dc parallel", name = {"-dcpar", "--dc-parallel"}, description = "Use -dcpar to repair data centers in parallel.") - private boolean dcParallel = false; - - @Option(title = "local_dc", name = {"-local", "--in-local-dc"}, description = "Use -local to only repair against nodes in the same datacenter") - private boolean localDC = false; - - @Option(title = "specific_dc", name = {"-dc", "--in-dc"}, description = "Use -dc to repair specific datacenters") - private List<String> specificDataCenters = new ArrayList<>(); - - @Option(title = "specific_host", name = {"-hosts", "--in-hosts"}, description = "Use -hosts to repair specific hosts") - private List<String> specificHosts = new ArrayList<>(); - - @Option(title = "start_token", name = {"-st", "--start-token"}, description = "Use -st to specify a token at which the repair range starts") - private String startToken = EMPTY; - - @Option(title = "end_token", name = {"-et", "--end-token"}, description = "Use -et to specify a token at which repair range ends") - private String endToken = EMPTY; - - @Option(title = "primary_range", name = {"-pr", "--partitioner-range"}, description = "Use -pr to repair only the first range returned by the partitioner") - private boolean primaryRange = false; - - @Option(title = "incremental_repair", name = {"-inc", "--incremental"}, description = "Use -inc to use the new incremental repair") - private boolean incrementalRepair = false; - - @Override - public void execute(NodeProbe probe) - { - List<String> keyspaces = parseOptionalKeyspace(args, probe); - String[] cfnames = parseOptionalColumnFamilies(args); - - if (primaryRange && (!specificDataCenters.isEmpty() || !specificHosts.isEmpty())) - throw new RuntimeException("Primary range repair should be performed on all nodes in the cluster."); - - for (String keyspace : keyspaces) - { - try - { - RepairParallelism parallelismDegree = RepairParallelism.SEQUENTIAL; - if (parallel) - parallelismDegree = RepairParallelism.PARALLEL; - else if (dcParallel) - parallelismDegree = RepairParallelism.DATACENTER_AWARE; - - Collection<String> dataCenters = null; - Collection<String> hosts = null; - if (!specificDataCenters.isEmpty()) - dataCenters = newArrayList(specificDataCenters); - else if (localDC) - dataCenters = newArrayList(probe.getDataCenter()); - else if(!specificHosts.isEmpty()) - hosts = newArrayList(specificHosts); - if (!startToken.isEmpty() || !endToken.isEmpty()) - probe.forceRepairRangeAsync(System.out, keyspace, parallelismDegree, dataCenters,hosts, startToken, endToken, !incrementalRepair); - else - probe.forceRepairAsync(System.out, keyspace, parallelismDegree, dataCenters, hosts, primaryRange, !incrementalRepair, cfnames); - } catch (Exception e) - { - throw new RuntimeException("Error occurred during repair", e); - } - } - } - } - - @Command(name = "setcachecapacity", description = "Set global key, row, and counter cache capacities (in MB units)") - public static class SetCacheCapacity extends NodeToolCmd - { - @Arguments(title = "<key-cache-capacity> <row-cache-capacity> <counter-cache-capacity>", - usage = "<key-cache-capacity> <row-cache-capacity> <counter-cache-capacity>", - description = "Key cache, row cache, and counter cache (in MB)", - required = true) - private List<Integer> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - checkArgument(args.size() == 3, "setcachecapacity requires key-cache-capacity, row-cache-capacity, and counter-cache-capacity args."); - probe.setCacheCapacities(args.get(0), args.get(1), args.get(2)); - } - } - - @Command(name = "setcompactionthreshold", description = "Set min and max compaction thresholds for a given column family") - public static class SetCompactionThreshold extends NodeToolCmd - { - @Arguments(title = "<keyspace> <cfname> <minthreshold> <maxthreshold>", usage = "<keyspace> <cfname> <minthreshold> <maxthreshold>", description = "The keyspace, the column family, min and max threshold", required = true) - private List<String> args = new ArrayList<>(); - - @Override - public void execute(NodeProbe probe) - { - checkArgument(args.size() == 4, "setcompactionthreshold requires ks, cf, min, and max threshold args."); - - int minthreshold = parseInt(args.get(2)); - int maxthreshold = parseInt(args.get(3)); - checkArgument(minthreshold >= 0 && maxthreshold >= 0, "Thresholds must be positive integers"); - checkArgument(minthreshold <= maxthreshold, "Min threshold cannot be greater than max."); - checkArgument(minthreshold >= 2 || maxthreshold == 0, "Min threshold must be at least 2"); - - probe.setCompactionThreshold(args.get(0), args.get(1), minthreshold, maxthreshold); - } - } - - @Command(name = "setcompactionthroughput", description = "Set the MB/s throughput cap for compaction in the system, or 0 to disable throttling") - public static class SetCompactionThroughput extends NodeToolCmd - { - @Arguments(title = "compaction_throughput", usage = "<value_in_mb>", description = "Value in MB, 0 to disable throttling", required = true) - private Integer compactionThroughput = null; - - @Override - public void execute(NodeProbe probe) - { - probe.setCompactionThroughput(compactionThroughput); - } - } - - @Command(name = "sethintedhandoffthrottlekb", description = "Set hinted handoff throttle in kb per second, per delivery thread.") - public static class SetHintedHandoffThrottleInKB extends NodeToolCmd - { - @Arguments(title = "throttle_in_kb", usage = "<value_in_kb_per_sec>", description = "Value in KB per second", required = true) - private Integer throttleInKB = null; - - @Override - public void execute(NodeProbe probe) - { - probe.setHintedHandoffThrottleInKB(throttleInKB); - } - } - - @Command(name = "setstreamthroughput", description = "Set the Mb/s throughput cap for streaming in the system, or 0 to disable throttling") - public static class SetStreamThroughput extends NodeToolCmd - { - @Arguments(title = "stream_throughput", usage = "<value_in_mb>", description = "Value in Mb, 0 to disable throttling", required = true) - private Integer streamThroughput = null; - - @Override - public void execute(NodeProbe probe) - { - probe.setStreamThroughput(streamThroughput); - } - } - - @Command(name = "settraceprobability", description = "Sets the probability for tracing any given request to value. 0 disables, 1 enables for all requests, 0 is the default") - public static class SetTraceProbability extends NodeToolCmd - { - @Arguments(title = "trace_probability", usage = "<value>", description = "Trace probability between 0 and 1 (ex: 0.2)", required = true) - private Double traceProbability = null; - - @Override - public void execute(NodeProbe probe) - { - checkArgument(traceProbability >= 0 && traceProbability <= 1, "Trace probability must be between 0 and 1"); - probe.setTraceProbability(traceProbability); - } - } - - @Command(name = "snapshot", description = "Take a snapshot of specified keyspaces or a snapshot of the specified column family") - public static class Snapshot extends NodeToolCmd - { - @Arguments(usage = "[<keyspaces...>]", description = "List of keyspaces. By default, all keyspaces") - private List<String> keyspaces = new ArrayList<>(); - - @Option(title = "cfname", name = {"-cf", "--column-family"}, description = "The column family name (you must specify one and only one keyspace for using this option)") - private String columnFamily = null; - - @Option(title = "tag", name = {"-t", "--tag"}, description = "The name of the snapshot") - private String snapshotName = Long.toString(System.currentTimeMillis()); - - @Option(title = "kclist", name = { "-kc", "--kc-list" }, description = "The list of Keyspace.Column family to take snapshot.(you must not specify only keyspace)") - private String kcList = null; - - @Override - public void execute(NodeProbe probe) - { - try - { - StringBuilder sb = new StringBuilder(); - - sb.append("Requested creating snapshot(s) for "); - - // Create a separate path for kclist to avoid breaking of already existing scripts - if (null != kcList && !kcList.isEmpty()) - { - kcList = kcList.replace(" ", ""); - if (keyspaces.isEmpty() && null == columnFamily) - sb.append("[").append(kcList).append("]"); - else - { - throw new IOException( - "When specifying the Keyspace columfamily list for a snapshot, you should not specify columnfamily"); - } - if (!snapshotName.isEmpty()) - sb.append(" with snapshot name [").append(snapshotName).append("]"); - System.out.println(sb.toString()); - probe.takeMultipleColumnFamilySnapshot(snapshotName, kcList.split(",")); - System.out.println("Snapshot directory: " + snapshotName); - } - else - { - if (keyspaces.isEmpty()) - sb.append("[all keyspaces]"); - else - sb.append("[").append(join(keyspaces, ", ")).append("]"); - - if (!snapshotName.isEmpty()) - sb.append(" with snapshot name [").append(snapshotName).append("]"); - - System.out.println(sb.toString()); - - probe.takeSnapshot(snapshotName, columnFamily, toArray(keyspaces, String.class)); - System.out.println("Snapshot directory: " + snapshotName); - } - } - catch (IOException e) - { - throw new RuntimeException("Error during taking a snapshot", e); - } - } - } - - @Command(name = "listsnapshots", description = "Lists all the snapshots along with the size on disk and true size.") - public static class ListSnapshots extends NodeToolCmd - { - @Override - public void execute(NodeProbe probe) - { - try - { - System.out.println("Snapshot Details: "); - - final Map<String,TabularData> snapshotDetails = probe.getSnapshotDetails(); - if (snapshotDetails.isEmpty()) - { - System.out.printf("There are no snapshots"); - return; - } - - final long trueSnapshotsSize = probe.trueSnapshotsSize(); - final String format = "%-20s%-29s%-29s%-19s%-19s%n"; - // display column names only once - final List<String> indexNames = snapshotDetails.entrySet().iterator().next().getValue().getTabularType().getIndexNames(); - System.out.printf(format, (Object[]) indexNames.toArray(new String[indexNames.size()])); - - for (final Map.Entry<String, TabularData> snapshotDetail : snapshotDetails.entrySet()) - { - Set<?> values = snapshotDetail.getValue().keySet(); - for (Object eachValue : values) - { - final List<?> value = (List<?>) eachValue; - System.out.printf(format, value.toArray(new Object[value.size()])); - } - } - - System.out.println("\nTotal TrueDiskSpaceUsed: " + FileUtils.stringifyFileSize(trueSnapshotsSize) + "\n"); - } - catch (Exception e) - { - throw new RuntimeException("Error during list snapshot", e); - } - } - } - - @Command(name = "status", description = "Print cluster information (state, load, IDs, ...)") - public static class Status extends NodeToolCmd - { - @Arguments(usage = "[<keyspace>]", description = "The keyspace name") - private String keyspace = null; - - @Option(title = "resolve_ip", name = {"-r", "--resolve-ip"}, description = "Show node domain names instead of IPs") - private boolean resolveIp = false; - - private boolean hasEffectiveOwns = false; - private boolean isTokenPerNode = true; - private int maxAddressLength = 0; - private String format = null; - private Collection<String> joiningNodes, leavingNodes, movingNodes, liveNodes, unreachableNodes; - private Map<String, String> loadMap, hostIDMap, tokensToEndpoints; - private EndpointSnitchInfoMBean epSnitchInfo; - - @Override - public void execute(NodeProbe probe) - { - joiningNodes = probe.getJoiningNodes(); - leavingNodes = probe.getLeavingNodes(); - movingNodes = probe.getMovingNodes(); - loadMap = probe.getLoadMap(); - tokensToEndpoints = probe.getTokenToEndpointMap(); - liveNodes = probe.getLiveNodes(); - unreachableNodes = probe.getUnreachableNodes(); - hostIDMap = probe.getHostIdMap(); - epSnitchInfo = probe.getEndpointSnitchInfoProxy(); - - StringBuffer errors = new StringBuffer(); - - Map<InetAddress, Float> ownerships = null; - try - { - ownerships = probe.effectiveOwnership(keyspace); - hasEffectiveOwns = true; - } catch (IllegalStateException e) - { - ownerships = probe.getOwnership(); - errors.append("Note: " + e.getMessage() + "%n"); - } - catch (IllegalArgumentException ex) - { - System.out.printf("%nError: " + ex.getMessage() + "%n"); - System.exit(1); - } - - SortedMap<String, SetHostStat> dcs = getOwnershipByDc(probe, resolveIp, tokensToEndpoints, ownerships); - - // More tokens than nodes (aka vnodes)? - if (dcs.values().size() < tokensToEndpoints.keySet().size()) - isTokenPerNode = false; - - findMaxAddressLength(dcs); - - // Datacenters - for (Map.Entry<String, SetHostStat> dc : dcs.entrySet()) - { - String dcHeader = String.format("Datacenter: %s%n", dc.getKey()); - System.out.printf(dcHeader); - for (int i = 0; i < (dcHeader.length() - 1); i++) System.out.print('='); - System.out.println(); - - // Legend - System.out.println("Status=Up/Down"); - System.out.println("|/ State=Normal/Leaving/Joining/Moving"); - - printNodesHeader(hasEffectiveOwns, isTokenPerNode); - - ArrayListMultimap<InetAddress, HostStat> hostToTokens = ArrayListMultimap.
<TRUNCATED>