
org.apache.cassandra.tools.NodeTool Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of cassandra-all Show documentation
Show all versions of cassandra-all Show documentation
The Apache Cassandra Project develops a highly scalable second-generation distributed database, bringing together Dynamo's fully distributed design and Bigtable's ColumnFamily-based data model.
The newest version!
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.cassandra.tools;
import java.io.*;
import java.lang.management.MemoryUsage;
import java.net.InetAddress;
import java.net.UnknownHostException;
import java.text.DecimalFormat;
import java.text.SimpleDateFormat;
import java.util.*;
import java.util.Map.Entry;
import java.util.concurrent.ExecutionException;
import javax.management.openmbean.TabularData;
import com.google.common.base.Joiner;
import com.google.common.base.Throwables;
import com.google.common.collect.ArrayListMultimap;
import com.google.common.collect.LinkedHashMultimap;
import com.google.common.collect.Maps;
import com.yammer.metrics.reporting.JmxReporter;
import io.airlift.command.*;
import org.apache.cassandra.concurrent.JMXEnabledThreadPoolExecutorMBean;
import org.apache.cassandra.db.ColumnFamilyStoreMBean;
import org.apache.cassandra.db.Keyspace;
import org.apache.cassandra.db.compaction.CompactionManagerMBean;
import org.apache.cassandra.db.compaction.OperationType;
import org.apache.cassandra.io.util.FileUtils;
import org.apache.cassandra.locator.EndpointSnitchInfoMBean;
import org.apache.cassandra.net.MessagingServiceMBean;
import org.apache.cassandra.service.CacheServiceMBean;
import org.apache.cassandra.streaming.ProgressInfo;
import org.apache.cassandra.streaming.SessionInfo;
import org.apache.cassandra.streaming.StreamState;
import org.apache.cassandra.utils.EstimatedHistogram;
import org.apache.cassandra.utils.FBUtilities;
import static com.google.common.base.Preconditions.checkArgument;
import static com.google.common.base.Preconditions.checkState;
import static com.google.common.base.Throwables.getStackTraceAsString;
import static com.google.common.collect.Iterables.toArray;
import static com.google.common.collect.Lists.newArrayList;
import static java.lang.Integer.parseInt;
import static java.lang.String.format;
import static org.apache.commons.lang3.ArrayUtils.EMPTY_STRING_ARRAY;
import static org.apache.commons.lang3.StringUtils.*;
public class NodeTool
{
private static final String HISTORYFILE = "nodetool.history";
public static void main(String... args)
{
List> commands = newArrayList(
Help.class,
Info.class,
Ring.class,
NetStats.class,
CfStats.class,
CfHistograms.class,
Cleanup.class,
ClearSnapshot.class,
Compact.class,
Scrub.class,
Flush.class,
UpgradeSSTable.class,
DisableAutoCompaction.class,
EnableAutoCompaction.class,
CompactionStats.class,
CompactionHistory.class,
Decommission.class,
DescribeCluster.class,
DisableBinary.class,
EnableBinary.class,
EnableGossip.class,
DisableGossip.class,
EnableHandoff.class,
EnableThrift.class,
GetCompactionThreshold.class,
GetCompactionThroughput.class,
GetStreamThroughput.class,
GetEndpoints.class,
GetSSTables.class,
GossipInfo.class,
InvalidateKeyCache.class,
InvalidateRowCache.class,
InvalidateCounterCache.class,
Join.class,
Move.class,
PauseHandoff.class,
ResumeHandoff.class,
ProxyHistograms.class,
Rebuild.class,
Refresh.class,
RemoveToken.class,
RemoveNode.class,
Repair.class,
SetCacheCapacity.class,
SetHintedHandoffThrottleInKB.class,
SetCompactionThreshold.class,
SetCompactionThroughput.class,
SetStreamThroughput.class,
SetTraceProbability.class,
Snapshot.class,
ListSnapshots.class,
Status.class,
StatusBinary.class,
StatusThrift.class,
Stop.class,
StopDaemon.class,
Version.class,
DescribeRing.class,
RebuildIndex.class,
RangeKeySample.class,
EnableBackup.class,
DisableBackup.class,
ResetLocalSchema.class,
ReloadTriggers.class,
SetCacheKeysToSave.class,
DisableThrift.class,
DisableHandoff.class,
Drain.class,
TruncateHints.class,
TpStats.class,
SetLoggingLevel.class,
GetLoggingLevels.class
);
Cli parser = Cli.builder("nodetool")
.withDescription("Manage your Cassandra cluster")
.withDefaultCommand(Help.class)
.withCommands(commands)
.build();
int status = 0;
try
{
Runnable parse = parser.parse(args);
printHistory(args);
parse.run();
} catch (IllegalArgumentException |
IllegalStateException |
ParseArgumentsMissingException |
ParseArgumentsUnexpectedException |
ParseOptionConversionException |
ParseOptionMissingException |
ParseOptionMissingValueException |
ParseCommandMissingException |
ParseCommandUnrecognizedException e)
{
badUse(e);
status = 1;
} catch (Throwable throwable)
{
err(Throwables.getRootCause(throwable));
status = 2;
}
System.exit(status);
}
private static void printHistory(String... args)
{
//don't bother to print if no args passed (meaning, nodetool is just printing out the sub-commands list)
if (args.length == 0)
return;
String cmdLine = Joiner.on(" ").skipNulls().join(args);
cmdLine = cmdLine.replaceFirst("(?<=(-pw|--password))\\s+\\S+", " ");
try (FileWriter writer = new FileWriter(new File(FBUtilities.getToolsOutputDirectory(), HISTORYFILE), true))
{
SimpleDateFormat sdf = new SimpleDateFormat("yyyy-MM-dd HH:mm:ss,SSS");
writer.append(sdf.format(new Date())).append(": ").append(cmdLine).append(System.lineSeparator());
}
catch (IOException | IOError ioe)
{
//quietly ignore any errors about not being able to write out history
}
}
private static void badUse(Exception e)
{
System.out.println("nodetool: " + e.getMessage());
System.out.println("See 'nodetool help' or 'nodetool help '.");
}
private static void err(Throwable e)
{
System.err.println("error: " + e.getMessage());
System.err.println("-- StackTrace --");
System.err.println(getStackTraceAsString(e));
}
public static abstract class NodeToolCmd implements Runnable
{
@Option(type = OptionType.GLOBAL, name = {"-h", "--host"}, description = "Node hostname or ip address")
private String host = "127.0.0.1";
@Option(type = OptionType.GLOBAL, name = {"-p", "--port"}, description = "Remote jmx agent port number")
private String port = "7199";
@Option(type = OptionType.GLOBAL, name = {"-u", "--username"}, description = "Remote jmx agent username")
private String username = EMPTY;
@Option(type = OptionType.GLOBAL, name = {"-pw", "--password"}, description = "Remote jmx agent password")
private String password = EMPTY;
@Option(type = OptionType.GLOBAL, name = {"-pwf", "--password-file"}, description = "Path to the JMX password file")
private String passwordFilePath = EMPTY;
@Override
public void run()
{
if (isNotEmpty(username)) {
if (isNotEmpty(passwordFilePath))
password = readUserPasswordFromFile(username, passwordFilePath);
if (isEmpty(password))
password = promptAndReadPassword();
}
try (NodeProbe probe = connect())
{
execute(probe);
}
catch (IOException e)
{
throw new RuntimeException("Error while closing JMX connection", e);
}
}
private String readUserPasswordFromFile(String username, String passwordFilePath) {
String password = EMPTY;
File passwordFile = new File(passwordFilePath);
try (Scanner scanner = new Scanner(passwordFile).useDelimiter("\\s+"))
{
while (scanner.hasNextLine())
{
if (scanner.hasNext())
{
String jmxRole = scanner.next();
if (jmxRole.equals(username) && scanner.hasNext())
{
password = scanner.next();
break;
}
}
scanner.nextLine();
}
} catch (FileNotFoundException e)
{
throw new RuntimeException(e);
}
return password;
}
private String promptAndReadPassword()
{
String password = EMPTY;
Console console = System.console();
if (console != null)
password = String.valueOf(console.readPassword("Password:"));
return password;
}
protected abstract void execute(NodeProbe probe);
private NodeProbe connect()
{
NodeProbe nodeClient = null;
try
{
if (username.isEmpty())
nodeClient = new NodeProbe(host, parseInt(port));
else
nodeClient = new NodeProbe(host, parseInt(port), username, password);
} catch (IOException e)
{
Throwable rootCause = Throwables.getRootCause(e);
System.err.println(format("nodetool: Failed to connect to '%s:%s' - %s: '%s'.", host, port, rootCause.getClass().getSimpleName(), rootCause.getMessage()));
System.exit(1);
}
return nodeClient;
}
protected List parseOptionalKeyspace(List cmdArgs, NodeProbe nodeProbe)
{
List keyspaces = new ArrayList<>();
if (cmdArgs == null || cmdArgs.isEmpty())
keyspaces.addAll(nodeProbe.getKeyspaces());
else
keyspaces.add(cmdArgs.get(0));
for (String keyspace : keyspaces)
{
if (!nodeProbe.getKeyspaces().contains(keyspace))
throw new IllegalArgumentException("Keyspace [" + keyspace + "] does not exist.");
}
return Collections.unmodifiableList(keyspaces);
}
protected String[] parseOptionalColumnFamilies(List cmdArgs)
{
return cmdArgs.size() <= 1 ? EMPTY_STRING_ARRAY : toArray(cmdArgs.subList(1, cmdArgs.size()), String.class);
}
}
@Command(name = "info", description = "Print node information (uptime, load, ...)")
public static class Info extends NodeToolCmd
{
@Option(name = {"-T", "--tokens"}, description = "Display all tokens")
private boolean tokens = false;
@Override
public void execute(NodeProbe probe)
{
boolean gossipInitialized = probe.isInitialized();
System.out.printf("%-17s: %s%n", "ID", probe.getLocalHostId());
System.out.printf("%-17s: %s%n", "Gossip active", gossipInitialized);
System.out.printf("%-17s: %s%n", "Thrift active", probe.isThriftServerRunning());
System.out.printf("%-17s: %s%n", "Native Transport active", probe.isNativeTransportRunning());
System.out.printf("%-17s: %s%n", "Load", probe.getLoadString());
if (gossipInitialized)
System.out.printf("%-17s: %s%n", "Generation No", probe.getCurrentGenerationNumber());
else
System.out.printf("%-17s: %s%n", "Generation No", 0);
// Uptime
long secondsUp = probe.getUptime() / 1000;
System.out.printf("%-17s: %d%n", "Uptime (seconds)", secondsUp);
// Memory usage
MemoryUsage heapUsage = probe.getHeapMemoryUsage();
double memUsed = (double) heapUsage.getUsed() / (1024 * 1024);
double memMax = (double) heapUsage.getMax() / (1024 * 1024);
System.out.printf("%-17s: %.2f / %.2f%n", "Heap Memory (MB)", memUsed, memMax);
// Data Center/Rack
System.out.printf("%-17s: %s%n", "Data Center", probe.getDataCenter());
System.out.printf("%-17s: %s%n", "Rack", probe.getRack());
// Exceptions
System.out.printf("%-17s: %s%n", "Exceptions", probe.getStorageMetric("Exceptions"));
CacheServiceMBean cacheService = probe.getCacheServiceMBean();
// Key Cache: Hits, Requests, RecentHitRate, SavePeriodInSeconds
System.out.printf("%-17s: entries %d, size %d (bytes), capacity %d (bytes), %d hits, %d requests, %.3f recent hit rate, %d save period in seconds%n",
"Key Cache",
probe.getCacheMetric("KeyCache", "Entries"),
probe.getCacheMetric("KeyCache", "Size"),
probe.getCacheMetric("KeyCache", "Capacity"),
probe.getCacheMetric("KeyCache", "Hits"),
probe.getCacheMetric("KeyCache", "Requests"),
probe.getCacheMetric("KeyCache", "HitRate"),
cacheService.getKeyCacheSavePeriodInSeconds());
// Row Cache: Hits, Requests, RecentHitRate, SavePeriodInSeconds
System.out.printf("%-17s: entries %d, size %d (bytes), capacity %d (bytes), %d hits, %d requests, %.3f recent hit rate, %d save period in seconds%n",
"Row Cache",
probe.getCacheMetric("RowCache", "Entries"),
probe.getCacheMetric("RowCache", "Size"),
probe.getCacheMetric("RowCache", "Capacity"),
probe.getCacheMetric("RowCache", "Hits"),
probe.getCacheMetric("RowCache", "Requests"),
probe.getCacheMetric("RowCache", "HitRate"),
cacheService.getRowCacheSavePeriodInSeconds());
// Counter Cache: Hits, Requests, RecentHitRate, SavePeriodInSeconds
System.out.printf("%-17s: entries %d, size %d (bytes), capacity %d (bytes), %d hits, %d requests, %.3f recent hit rate, %d save period in seconds%n",
"Counter Cache",
probe.getCacheMetric("CounterCache", "Entries"),
probe.getCacheMetric("CounterCache", "Size"),
probe.getCacheMetric("CounterCache", "Capacity"),
probe.getCacheMetric("CounterCache", "Hits"),
probe.getCacheMetric("CounterCache", "Requests"),
probe.getCacheMetric("CounterCache", "HitRate"),
cacheService.getCounterCacheSavePeriodInSeconds());
// Tokens
List tokens = probe.getTokens();
if (tokens.size() == 1 || this.tokens)
for (String token : tokens)
System.out.printf("%-17s: %s%n", "Token", token);
else
System.out.printf("%-17s: (invoke with -T/--tokens to see all %d tokens)%n", "Token", tokens.size());
}
}
@Command(name = "ring", description = "Print information about the token ring")
public static class Ring extends NodeToolCmd
{
@Arguments(description = "Specify a keyspace for accurate ownership information (topology awareness)")
private String keyspace = null;
@Option(title = "resolve_ip", name = {"-r", "--resolve-ip"}, description = "Show node domain names instead of IPs")
private boolean resolveIp = false;
@Override
public void execute(NodeProbe probe)
{
Map tokensToEndpoints = probe.getTokenToEndpointMap();
LinkedHashMultimap endpointsToTokens = LinkedHashMultimap.create();
boolean haveVnodes = false;
for (Map.Entry entry : tokensToEndpoints.entrySet())
{
haveVnodes |= endpointsToTokens.containsKey(entry.getValue());
endpointsToTokens.put(entry.getValue(), entry.getKey());
}
int maxAddressLength = Collections.max(endpointsToTokens.keys(), new Comparator()
{
@Override
public int compare(String first, String second)
{
return ((Integer) first.length()).compareTo(second.length());
}
}).length();
String formatPlaceholder = "%%-%ds %%-12s%%-7s%%-8s%%-16s%%-20s%%-44s%%n";
String format = format(formatPlaceholder, maxAddressLength);
// Calculate per-token ownership of the ring
Map ownerships;
try
{
ownerships = probe.effectiveOwnership(keyspace);
} catch (IllegalStateException ex)
{
ownerships = probe.getOwnership();
System.out.printf("Note: Ownership information does not include topology; for complete information, specify a keyspace%n");
}
System.out.println();
for (Entry entry : getOwnershipByDc(probe, resolveIp, tokensToEndpoints, ownerships).entrySet())
printDc(probe, format, entry.getKey(), endpointsToTokens, entry.getValue());
if (haveVnodes)
{
System.out.println(" Warning: \"nodetool ring\" is used to output all the tokens of a node.");
System.out.println(" To view status related info of a node use \"nodetool status\" instead.\n");
}
}
private void printDc(NodeProbe probe, String format,
String dc,
LinkedHashMultimap endpointsToTokens,
SetHostStat hoststats)
{
Collection liveNodes = probe.getLiveNodes();
Collection deadNodes = probe.getUnreachableNodes();
Collection joiningNodes = probe.getJoiningNodes();
Collection leavingNodes = probe.getLeavingNodes();
Collection movingNodes = probe.getMovingNodes();
Map loadMap = probe.getLoadMap();
System.out.println("Datacenter: " + dc);
System.out.println("==========");
// get the total amount of replicas for this dc and the last token in this dc's ring
List tokens = new ArrayList<>();
String lastToken = "";
for (HostStat stat : hoststats)
{
tokens.addAll(endpointsToTokens.get(stat.endpoint.getHostAddress()));
lastToken = tokens.get(tokens.size() - 1);
}
System.out.printf(format, "Address", "Rack", "Status", "State", "Load", "Owns", "Token");
if (hoststats.size() > 1)
System.out.printf(format, "", "", "", "", "", "", lastToken);
else
System.out.println();
for (HostStat stat : hoststats)
{
String endpoint = stat.endpoint.getHostAddress();
String rack;
try
{
rack = probe.getEndpointSnitchInfoProxy().getRack(endpoint);
}
catch (UnknownHostException e)
{
rack = "Unknown";
}
String status = liveNodes.contains(endpoint)
? "Up"
: deadNodes.contains(endpoint)
? "Down"
: "?";
String state = "Normal";
if (joiningNodes.contains(endpoint))
state = "Joining";
else if (leavingNodes.contains(endpoint))
state = "Leaving";
else if (movingNodes.contains(endpoint))
state = "Moving";
String load = loadMap.containsKey(endpoint)
? loadMap.get(endpoint)
: "?";
String owns = stat.owns != null ? new DecimalFormat("##0.00%").format(stat.owns) : "?";
System.out.printf(format, stat.ipOrDns(), rack, status, state, load, owns, stat.token);
}
System.out.println();
}
}
@Command(name = "netstats", description = "Print network information on provided host (connecting node by default)")
public static class NetStats extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
System.out.printf("Mode: %s%n", probe.getOperationMode());
Set statuses = probe.getStreamStatus();
if (statuses.isEmpty())
System.out.println("Not sending any streams.");
for (StreamState status : statuses)
{
System.out.printf("%s %s%n", status.description, status.planId.toString());
for (SessionInfo info : status.sessions)
{
System.out.printf(" %s%n", info.peer.toString());
if (!info.receivingSummaries.isEmpty())
{
System.out.printf(" Receiving %d files, %d bytes total%n", info.getTotalFilesToReceive(), info.getTotalSizeToReceive());
for (ProgressInfo progress : info.getReceivingFiles())
{
System.out.printf(" %s%n", progress.toString());
}
}
if (!info.sendingSummaries.isEmpty())
{
System.out.printf(" Sending %d files, %d bytes total%n", info.getTotalFilesToSend(), info.getTotalSizeToSend());
for (ProgressInfo progress : info.getSendingFiles())
{
System.out.printf(" %s%n", progress.toString());
}
}
}
}
System.out.printf("Read Repair Statistics:%nAttempted: %d%nMismatch (Blocking): %d%nMismatch (Background): %d%n", probe.getReadRepairAttempted(), probe.getReadRepairRepairedBlocking(), probe.getReadRepairRepairedBackground());
MessagingServiceMBean ms = probe.msProxy;
System.out.printf("%-25s", "Pool Name");
System.out.printf("%10s", "Active");
System.out.printf("%10s", "Pending");
System.out.printf("%15s%n", "Completed");
int pending;
long completed;
pending = 0;
for (int n : ms.getCommandPendingTasks().values())
pending += n;
completed = 0;
for (long n : ms.getCommandCompletedTasks().values())
completed += n;
System.out.printf("%-25s%10s%10s%15s%n", "Commands", "n/a", pending, completed);
pending = 0;
for (int n : ms.getResponsePendingTasks().values())
pending += n;
completed = 0;
for (long n : ms.getResponseCompletedTasks().values())
completed += n;
System.out.printf("%-25s%10s%10s%15s%n", "Responses", "n/a", pending, completed);
}
}
@Command(name = "cfstats", description = "Print statistics on column families")
public static class CfStats extends NodeToolCmd
{
@Arguments(usage = "[...]", description = "List of column families (or keyspace) names")
private List cfnames = new ArrayList<>();
@Option(name = "-i", description = "Ignore the list of column families and display the remaining cfs")
private boolean ignore = false;
@Override
public void execute(NodeProbe probe)
{
OptionFilter filter = new OptionFilter(ignore, cfnames);
Map> cfstoreMap = new HashMap<>();
// get a list of column family stores
Iterator> cfamilies = probe.getColumnFamilyStoreMBeanProxies();
while (cfamilies.hasNext())
{
Map.Entry entry = cfamilies.next();
String keyspaceName = entry.getKey();
ColumnFamilyStoreMBean cfsProxy = entry.getValue();
if (!cfstoreMap.containsKey(keyspaceName) && filter.isColumnFamilyIncluded(entry.getKey(), cfsProxy.getColumnFamilyName()))
{
List columnFamilies = new ArrayList<>();
columnFamilies.add(cfsProxy);
cfstoreMap.put(keyspaceName, columnFamilies);
} else if (filter.isColumnFamilyIncluded(entry.getKey(), cfsProxy.getColumnFamilyName()))
{
cfstoreMap.get(keyspaceName).add(cfsProxy);
}
}
// make sure all specified kss and cfs exist
filter.verifyKeyspaces(probe.getKeyspaces());
filter.verifyColumnFamilies();
// print out the table statistics
for (Map.Entry> entry : cfstoreMap.entrySet())
{
String keyspaceName = entry.getKey();
List columnFamilies = entry.getValue();
long keyspaceReadCount = 0;
long keyspaceWriteCount = 0;
int keyspacePendingFlushes = 0;
double keyspaceTotalReadTime = 0.0f;
double keyspaceTotalWriteTime = 0.0f;
System.out.println("Keyspace: " + keyspaceName);
for (ColumnFamilyStoreMBean cfstore : columnFamilies)
{
String cfName = cfstore.getColumnFamilyName();
long writeCount = ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "WriteLatency")).getCount();
long readCount = ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "ReadLatency")).getCount();
if (readCount > 0)
{
keyspaceReadCount += readCount;
keyspaceTotalReadTime += (long) probe.getColumnFamilyMetric(keyspaceName, cfName, "ReadTotalLatency");
}
if (writeCount > 0)
{
keyspaceWriteCount += writeCount;
keyspaceTotalWriteTime += (long) probe.getColumnFamilyMetric(keyspaceName, cfName, "WriteTotalLatency");
}
keyspacePendingFlushes += (long) probe.getColumnFamilyMetric(keyspaceName, cfName, "PendingFlushes");
}
double keyspaceReadLatency = keyspaceReadCount > 0
? keyspaceTotalReadTime / keyspaceReadCount / 1000
: Double.NaN;
double keyspaceWriteLatency = keyspaceWriteCount > 0
? keyspaceTotalWriteTime / keyspaceWriteCount / 1000
: Double.NaN;
System.out.println("\tRead Count: " + keyspaceReadCount);
System.out.println("\tRead Latency: " + format("%s", keyspaceReadLatency) + " ms.");
System.out.println("\tWrite Count: " + keyspaceWriteCount);
System.out.println("\tWrite Latency: " + format("%s", keyspaceWriteLatency) + " ms.");
System.out.println("\tPending Flushes: " + keyspacePendingFlushes);
// print out column family statistics for this keyspace
for (ColumnFamilyStoreMBean cfstore : columnFamilies)
{
String cfName = cfstore.getColumnFamilyName();
if (cfName.contains("."))
System.out.println("\t\tTable (index): " + cfName);
else
System.out.println("\t\tTable: " + cfName);
System.out.println("\t\tSSTable count: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "LiveSSTableCount"));
int[] leveledSStables = cfstore.getSSTableCountPerLevel();
if (leveledSStables != null)
{
System.out.print("\t\tSSTables in each level: [");
for (int level = 0; level < leveledSStables.length; level++)
{
int count = leveledSStables[level];
System.out.print(count);
long maxCount = 4L; // for L0
if (level > 0)
maxCount = (long) Math.pow(10, level);
// show max threshold for level when exceeded
if (count > maxCount)
System.out.print("/" + maxCount);
if (level < leveledSStables.length - 1)
System.out.print(", ");
else
System.out.println("]");
}
}
System.out.println("\t\tSpace used (live), bytes: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "LiveDiskSpaceUsed"));
System.out.println("\t\tSpace used (total), bytes: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "TotalDiskSpaceUsed"));
System.out.println("\t\tSpace used by snapshots (total), bytes: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "SnapshotsSize"));
System.out.println("\t\tSSTable Compression Ratio: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "CompressionRatio"));
System.out.println("\t\tMemtable cell count: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "MemtableColumnsCount"));
System.out.println("\t\tMemtable data size, bytes: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "MemtableLiveDataSize"));
System.out.println("\t\tMemtable switch count: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "MemtableSwitchCount"));
System.out.println("\t\tLocal read count: " + ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "ReadLatency")).getCount());
double localReadLatency = ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "ReadLatency")).getMean() / 1000;
double localRLatency = localReadLatency > 0 ? localReadLatency : Double.NaN;
System.out.printf("\t\tLocal read latency: %01.3f ms%n", localRLatency);
System.out.println("\t\tLocal write count: " + ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "WriteLatency")).getCount());
double localWriteLatency = ((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "WriteLatency")).getMean() / 1000;
double localWLatency = localWriteLatency > 0 ? localWriteLatency : Double.NaN;
System.out.printf("\t\tLocal write latency: %01.3f ms%n", localWLatency);
System.out.println("\t\tPending flushes: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "PendingFlushes"));
System.out.println("\t\tBloom filter false positives: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "BloomFilterFalsePositives"));
System.out.println("\t\tBloom filter false ratio: " + format("%01.5f", probe.getColumnFamilyMetric(keyspaceName, cfName, "RecentBloomFilterFalseRatio")));
System.out.println("\t\tBloom filter space used, bytes: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "BloomFilterDiskSpaceUsed"));
System.out.println("\t\tCompacted partition minimum bytes: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "MinRowSize"));
System.out.println("\t\tCompacted partition maximum bytes: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "MaxRowSize"));
System.out.println("\t\tCompacted partition mean bytes: " + probe.getColumnFamilyMetric(keyspaceName, cfName, "MeanRowSize"));
System.out.println("\t\tAverage live cells per slice (last five minutes): " + ((JmxReporter.HistogramMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "LiveScannedHistogram")).getMean());
System.out.println("\t\tAverage tombstones per slice (last five minutes): " + ((JmxReporter.HistogramMBean) probe.getColumnFamilyMetric(keyspaceName, cfName, "TombstoneScannedHistogram")).getMean());
System.out.println("");
}
System.out.println("----------------");
}
}
/**
* Used for filtering keyspaces and columnfamilies to be displayed using the cfstats command.
*/
private static class OptionFilter
{
private Map> filter = new HashMap<>();
private Map> verifier = new HashMap<>();
private List filterList = new ArrayList<>();
private boolean ignoreMode;
public OptionFilter(boolean ignoreMode, List filterList)
{
this.filterList.addAll(filterList);
this.ignoreMode = ignoreMode;
for (String s : filterList)
{
String[] keyValues = s.split("\\.", 2);
// build the map that stores the ks' and cfs to use
if (!filter.containsKey(keyValues[0]))
{
filter.put(keyValues[0], new ArrayList());
verifier.put(keyValues[0], new ArrayList());
if (keyValues.length == 2)
{
filter.get(keyValues[0]).add(keyValues[1]);
verifier.get(keyValues[0]).add(keyValues[1]);
}
} else
{
if (keyValues.length == 2)
{
filter.get(keyValues[0]).add(keyValues[1]);
verifier.get(keyValues[0]).add(keyValues[1]);
}
}
}
}
public boolean isColumnFamilyIncluded(String keyspace, String columnFamily)
{
// supplying empty params list is treated as wanting to display all kss & cfs
if (filterList.isEmpty())
return !ignoreMode;
List cfs = filter.get(keyspace);
// no such keyspace is in the map
if (cfs == null)
return ignoreMode;
// only a keyspace with no cfs was supplied
// so ignore or include (based on the flag) every column family in specified keyspace
else if (cfs.size() == 0)
return !ignoreMode;
// keyspace exists, and it contains specific cfs
verifier.get(keyspace).remove(columnFamily);
return ignoreMode ^ cfs.contains(columnFamily);
}
public void verifyKeyspaces(List keyspaces)
{
for (String ks : verifier.keySet())
if (!keyspaces.contains(ks))
throw new IllegalArgumentException("Unknown keyspace: " + ks);
}
public void verifyColumnFamilies()
{
for (String ks : filter.keySet())
if (verifier.get(ks).size() > 0)
throw new IllegalArgumentException("Unknown column families: " + verifier.get(ks).toString() + " in keyspace: " + ks);
}
}
}
@Command(name = "cfhistograms", description = "Print statistic histograms for a given column family")
public static class CfHistograms extends NodeToolCmd
{
@Arguments(usage = " ", description = "The keyspace and column family name")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
checkArgument(args.size() == 2, "cfhistograms requires ks and cf args");
String keyspace = args.get(0);
String cfname = args.get(1);
// calculate percentile of row size and column count
long[] estimatedRowSize = (long[]) probe.getColumnFamilyMetric(keyspace, cfname, "EstimatedRowSizeHistogram");
long[] estimatedColumnCount = (long[]) probe.getColumnFamilyMetric(keyspace, cfname, "EstimatedColumnCountHistogram");
long[] bucketOffsets = new EstimatedHistogram().getBucketOffsets();
EstimatedHistogram rowSizeHist = new EstimatedHistogram(bucketOffsets, estimatedRowSize);
EstimatedHistogram columnCountHist = new EstimatedHistogram(bucketOffsets, estimatedColumnCount);
// build arrays to store percentile values
double[] estimatedRowSizePercentiles = new double[7];
double[] estimatedColumnCountPercentiles = new double[7];
double[] offsetPercentiles = new double[]{0.5, 0.75, 0.95, 0.98, 0.99};
for (int i = 0; i < offsetPercentiles.length; i++)
{
estimatedRowSizePercentiles[i] = rowSizeHist.percentile(offsetPercentiles[i]);
estimatedColumnCountPercentiles[i] = columnCountHist.percentile(offsetPercentiles[i]);
}
// min value
estimatedRowSizePercentiles[5] = rowSizeHist.min();
estimatedColumnCountPercentiles[5] = columnCountHist.min();
// max value
estimatedRowSizePercentiles[6] = rowSizeHist.max();
estimatedColumnCountPercentiles[6] = columnCountHist.max();
String[] percentiles = new String[]{"50%", "75%", "95%", "98%", "99%", "Min", "Max"};
double[] readLatency = probe.metricPercentilesAsArray((JmxReporter.HistogramMBean) probe.getColumnFamilyMetric(keyspace, cfname, "ReadLatency"));
double[] writeLatency = probe.metricPercentilesAsArray((JmxReporter.TimerMBean) probe.getColumnFamilyMetric(keyspace, cfname, "WriteLatency"));
double[] sstablesPerRead = probe.metricPercentilesAsArray((JmxReporter.HistogramMBean) probe.getColumnFamilyMetric(keyspace, cfname, "SSTablesPerReadHistogram"));
System.out.println(format("%s/%s histograms", keyspace, cfname));
System.out.println(format("%-10s%10s%18s%18s%18s%18s",
"Percentile", "SSTables", "Write Latency", "Read Latency", "Partition Size", "Cell Count"));
System.out.println(format("%-10s%10s%18s%18s%18s%18s",
"", "", "(micros)", "(micros)", "(bytes)", ""));
for (int i = 0; i < percentiles.length; i++)
{
System.out.println(format("%-10s%10.2f%18.2f%18.2f%18.0f%18.0f",
percentiles[i],
sstablesPerRead[i],
writeLatency[i],
readLatency[i],
estimatedRowSizePercentiles[i],
estimatedColumnCountPercentiles[i]));
}
System.out.println();
}
}
@Command(name = "cleanup", description = "Triggers the immediate cleanup of keys no longer belonging to a node. By default, clean all keyspaces")
public static class Cleanup extends NodeToolCmd
{
@Arguments(usage = "[ ...]", description = "The keyspace followed by one or many column families")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
List keyspaces = parseOptionalKeyspace(args, probe);
String[] cfnames = parseOptionalColumnFamilies(args);
for (String keyspace : keyspaces)
{
if (Keyspace.SYSTEM_KS.equals(keyspace))
continue;
try
{
probe.forceKeyspaceCleanup(System.out, keyspace, cfnames);
} catch (Exception e)
{
throw new RuntimeException("Error occurred during cleanup", e);
}
}
}
}
@Command(name = "clearsnapshot", description = "Remove the snapshot with the given name from the given keyspaces. If no snapshotName is specified we will remove all snapshots")
public static class ClearSnapshot extends NodeToolCmd
{
@Arguments(usage = "[...] ", description = "Remove snapshots from the given keyspaces")
private List keyspaces = new ArrayList<>();
@Option(title = "snapshot_name", name = "-t", description = "Remove the snapshot with a given name")
private String snapshotName = EMPTY;
@Override
public void execute(NodeProbe probe)
{
StringBuilder sb = new StringBuilder();
sb.append("Requested clearing snapshot(s) for ");
if (keyspaces.isEmpty())
sb.append("[all keyspaces]");
else
sb.append("[").append(join(keyspaces, ", ")).append("]");
if (!snapshotName.isEmpty())
sb.append(" with snapshot name [").append(snapshotName).append("]");
System.out.println(sb.toString());
try
{
probe.clearSnapshot(snapshotName, toArray(keyspaces, String.class));
} catch (IOException e)
{
throw new RuntimeException("Error during clearing snapshots", e);
}
}
}
@Command(name = "compact", description = "Force a (major) compaction on one or more column families")
public static class Compact extends NodeToolCmd
{
@Arguments(usage = "[ ...]", description = "The keyspace followed by one or many column families")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
List keyspaces = parseOptionalKeyspace(args, probe);
String[] cfnames = parseOptionalColumnFamilies(args);
for (String keyspace : keyspaces)
{
try
{
probe.forceKeyspaceCompaction(keyspace, cfnames);
} catch (Exception e)
{
throw new RuntimeException("Error occurred during compaction", e);
}
}
}
}
@Command(name = "flush", description = "Flush one or more column families")
public static class Flush extends NodeToolCmd
{
@Arguments(usage = "[ ...]", description = "The keyspace followed by one or many column families")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
List keyspaces = parseOptionalKeyspace(args, probe);
String[] cfnames = parseOptionalColumnFamilies(args);
for (String keyspace : keyspaces)
{
try
{
probe.forceKeyspaceFlush(keyspace, cfnames);
} catch (Exception e)
{
throw new RuntimeException("Error occurred during flushing", e);
}
}
}
}
@Command(name = "scrub", description = "Scrub (rebuild sstables for) one or more column families")
public static class Scrub extends NodeToolCmd
{
@Arguments(usage = "[ ...]", description = "The keyspace followed by one or many column families")
private List args = new ArrayList<>();
@Option(title = "disable_snapshot",
name = {"-ns", "--no-snapshot"},
description = "Scrubbed CFs will be snapshotted first, if disableSnapshot is false. (default false)")
private boolean disableSnapshot = false;
@Option(title = "skip_corrupted",
name = {"-s", "--skip-corrupted"},
description = "Skip corrupted partitions even when scrubbing counter tables. (default false)")
private boolean skipCorrupted = false;
@Override
public void execute(NodeProbe probe)
{
List keyspaces = parseOptionalKeyspace(args, probe);
String[] cfnames = parseOptionalColumnFamilies(args);
for (String keyspace : keyspaces)
{
try
{
probe.scrub(System.out, disableSnapshot, skipCorrupted, keyspace, cfnames);
} catch (Exception e)
{
throw new RuntimeException("Error occurred during flushing", e);
}
}
}
}
@Command(name = "disableautocompaction", description = "Disable autocompaction for the given keyspace and column family")
public static class DisableAutoCompaction extends NodeToolCmd
{
@Arguments(usage = "[ ...]", description = "The keyspace followed by one or many column families")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
List keyspaces = parseOptionalKeyspace(args, probe);
String[] cfnames = parseOptionalColumnFamilies(args);
for (String keyspace : keyspaces)
{
try
{
probe.disableAutoCompaction(keyspace, cfnames);
} catch (IOException e)
{
throw new RuntimeException("Error occurred during disabling auto-compaction", e);
}
}
}
}
@Command(name = "enableautocompaction", description = "Enable autocompaction for the given keyspace and column family")
public static class EnableAutoCompaction extends NodeToolCmd
{
@Arguments(usage = "[ ...]", description = "The keyspace followed by one or many column families")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
List keyspaces = parseOptionalKeyspace(args, probe);
String[] cfnames = parseOptionalColumnFamilies(args);
for (String keyspace : keyspaces)
{
try
{
probe.enableAutoCompaction(keyspace, cfnames);
} catch (IOException e)
{
throw new RuntimeException("Error occurred during enabling auto-compaction", e);
}
}
}
}
@Command(name = "upgradesstables", description = "Rewrite sstables (for the requested column families) that are not on the current version (thus upgrading them to said current version)")
public static class UpgradeSSTable extends NodeToolCmd
{
@Arguments(usage = "[ ...]", description = "The keyspace followed by one or many column families")
private List args = new ArrayList<>();
@Option(title = "include_all", name = {"-a", "--include-all-sstables"}, description = "Use -a to include all sstables, even those already on the current version")
private boolean includeAll = false;
@Override
public void execute(NodeProbe probe)
{
List keyspaces = parseOptionalKeyspace(args, probe);
String[] cfnames = parseOptionalColumnFamilies(args);
for (String keyspace : keyspaces)
{
try
{
probe.upgradeSSTables(System.out, keyspace, !includeAll, cfnames);
} catch (Exception e)
{
throw new RuntimeException("Error occurred during enabling auto-compaction", e);
}
}
}
}
@Command(name = "compactionstats", description = "Print statistics on compactions")
public static class CompactionStats extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
int compactionThroughput = probe.getCompactionThroughput();
CompactionManagerMBean cm = probe.getCompactionManagerProxy();
System.out.println("pending tasks: " + probe.getCompactionMetric("PendingTasks"));
if (cm.getCompactions().size() > 0)
System.out.printf("%25s%16s%16s%16s%16s%10s%10s%n", "compaction type", "keyspace", "table", "completed", "total", "unit", "progress");
long remainingBytes = 0;
for (Map c : cm.getCompactions())
{
String percentComplete = new Long(c.get("total")) == 0
? "n/a"
: new DecimalFormat("0.00").format((double) new Long(c.get("completed")) / new Long(c.get("total")) * 100) + "%";
System.out.printf("%25s%16s%16s%16s%16s%10s%10s%n", c.get("taskType"), c.get("keyspace"), c.get("columnfamily"), c.get("completed"), c.get("total"), c.get("unit"), percentComplete);
if (c.get("taskType").equals(OperationType.COMPACTION.toString()))
remainingBytes += (new Long(c.get("total")) - new Long(c.get("completed")));
}
long remainingTimeInSecs = compactionThroughput == 0 || remainingBytes == 0
? -1
: (remainingBytes) / (1024L * 1024L * compactionThroughput);
String remainingTime = remainingTimeInSecs < 0
? "n/a"
: format("%dh%02dm%02ds", remainingTimeInSecs / 3600, (remainingTimeInSecs % 3600) / 60, (remainingTimeInSecs % 60));
System.out.printf("%25s%10s%n", "Active compaction remaining time : ", remainingTime);
}
}
@Command(name = "compactionhistory", description = "Print history of compaction")
public static class CompactionHistory extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
System.out.println("Compaction History: ");
TabularData tabularData = probe.getCompactionHistory();
if (tabularData.isEmpty())
{
System.out.printf("There is no compaction history");
return;
}
String format = "%-41s%-19s%-29s%-26s%-15s%-15s%s%n";
List indexNames = tabularData.getTabularType().getIndexNames();
System.out.printf(format, toArray(indexNames, Object.class));
Set> values = tabularData.keySet();
for (Object eachValue : values)
{
List> value = (List>) eachValue;
System.out.printf(format, toArray(value, Object.class));
}
}
}
@Command(name = "decommission", description = "Decommission the *node I am connecting to*")
public static class Decommission extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
try
{
probe.decommission();
} catch (InterruptedException e)
{
throw new RuntimeException("Error decommissioning node", e);
}
}
}
@Command(name = "describecluster", description = "Print the name, snitch, partitioner and schema version of a cluster")
public static class DescribeCluster extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
// display cluster name, snitch and partitioner
System.out.println("Cluster Information:");
System.out.println("\tName: " + probe.getClusterName());
System.out.println("\tSnitch: " + probe.getEndpointSnitchInfoProxy().getSnitchName());
System.out.println("\tPartitioner: " + probe.getPartitioner());
// display schema version for each node
System.out.println("\tSchema versions:");
Map> schemaVersions = probe.getSpProxy().getSchemaVersions();
for (String version : schemaVersions.keySet())
{
System.out.println(format("\t\t%s: %s%n", version, schemaVersions.get(version)));
}
}
}
@Command(name = "disablebinary", description = "Disable native transport (binary protocol)")
public static class DisableBinary extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.stopNativeTransport();
}
}
@Command(name = "enablebinary", description = "Reenable native transport (binary protocol)")
public static class EnableBinary extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.startNativeTransport();
}
}
@Command(name = "enablegossip", description = "Reenable gossip")
public static class EnableGossip extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.startGossiping();
}
}
@Command(name = "disablegossip", description = "Disable gossip (effectively marking the node down)")
public static class DisableGossip extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.stopGossiping();
}
}
@Command(name = "enablehandoff", description = "Reenable the future hints storing on the current node")
public static class EnableHandoff extends NodeToolCmd
{
@Arguments(usage = ",", description = "Enable hinted handoff only for these DCs")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
checkArgument(args.size() <= 1, "enablehandoff does not accept two args");
if(args.size() == 1)
probe.enableHintedHandoff(args.get(0));
else
probe.enableHintedHandoff();
}
}
@Command(name = "enablethrift", description = "Reenable thrift server")
public static class EnableThrift extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.startThriftServer();
}
}
@Command(name = "getcompactionthreshold", description = "Print min and max compaction thresholds for a given column family")
public static class GetCompactionThreshold extends NodeToolCmd
{
@Arguments(usage = " ", description = "The keyspace with a column family")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
checkArgument(args.size() == 2, "getcompactionthreshold requires ks and cf args");
String ks = args.get(0);
String cf = args.get(1);
ColumnFamilyStoreMBean cfsProxy = probe.getCfsProxy(ks, cf);
System.out.println("Current compaction thresholds for " + ks + "/" + cf + ": \n" +
" min = " + cfsProxy.getMinimumCompactionThreshold() + ", " +
" max = " + cfsProxy.getMaximumCompactionThreshold());
}
}
@Command(name = "getcompactionthroughput", description = "Print the MB/s throughput cap for compaction in the system")
public static class GetCompactionThroughput extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
System.out.println("Current compaction throughput: " + probe.getCompactionThroughput() + " MB/s");
}
}
@Command(name = "getstreamthroughput", description = "Print the MB/s throughput cap for streaming in the system")
public static class GetStreamThroughput extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
System.out.println("Current stream throughput: " + probe.getStreamThroughput() + " MB/s");
}
}
@Command(name = "getendpoints", description = "Print the end points that owns the key")
public static class GetEndpoints extends NodeToolCmd
{
@Arguments(usage = " ", description = "The keyspace, the column family, and the key for which we need to find the endpoint")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
checkArgument(args.size() == 3, "getendpoints requires ks, cf and key args");
String ks = args.get(0);
String cf = args.get(1);
String key = args.get(2);
List endpoints = probe.getEndpoints(ks, cf, key);
for (InetAddress endpoint : endpoints)
{
System.out.println(endpoint.getHostAddress());
}
}
}
@Command(name = "getsstables", description = "Print the sstable filenames that own the key")
public static class GetSSTables extends NodeToolCmd
{
@Arguments(usage = " ", description = "The keyspace, the column family, and the key")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
checkArgument(args.size() == 3, "getsstables requires ks, cf and key args");
String ks = args.get(0);
String cf = args.get(1);
String key = args.get(2);
List sstables = probe.getSSTables(ks, cf, key);
for (String sstable : sstables)
{
System.out.println(sstable);
}
}
}
@Command(name = "gossipinfo", description = "Shows the gossip information for the cluster")
public static class GossipInfo extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
System.out.println(probe.getGossipInfo());
}
}
@Command(name = "invalidatekeycache", description = "Invalidate the key cache")
public static class InvalidateKeyCache extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.invalidateKeyCache();
}
}
@Command(name = "invalidaterowcache", description = "Invalidate the row cache")
public static class InvalidateRowCache extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.invalidateRowCache();
}
}
@Command(name = "invalidatecountercache", description = "Invalidate the counter cache")
public static class InvalidateCounterCache extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.invalidateCounterCache();
}
}
@Command(name = "join", description = "Join the ring")
public static class Join extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
checkState(!probe.isJoined(), "This node has already joined the ring.");
try
{
probe.joinRing();
} catch (IOException e)
{
throw new RuntimeException("Error during joining the ring", e);
}
}
}
@Command(name = "move", description = "Move node on the token ring to a new token")
public static class Move extends NodeToolCmd
{
@Arguments(usage = "", description = "The new token.", required = true)
private String newToken = EMPTY;
@Override
public void execute(NodeProbe probe)
{
try
{
probe.move(newToken);
} catch (IOException e)
{
throw new RuntimeException("Error during moving node", e);
}
}
}
@Command(name = "pausehandoff", description = "Pause hints delivery process")
public static class PauseHandoff extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.pauseHintsDelivery();
}
}
@Command(name = "resumehandoff", description = "Resume hints delivery process")
public static class ResumeHandoff extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.resumeHintsDelivery();
}
}
@Command(name = "proxyhistograms", description = "Print statistic histograms for network operations")
public static class ProxyHistograms extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
String[] percentiles = new String[]{"50%", "75%", "95%", "98%", "99%", "Min", "Max"};
double[] readLatency = probe.metricPercentilesAsArray(probe.getProxyMetric("Read"));
double[] writeLatency = probe.metricPercentilesAsArray(probe.getProxyMetric("Write"));
double[] rangeLatency = probe.metricPercentilesAsArray(probe.getProxyMetric("RangeSlice"));
System.out.println("proxy histograms");
System.out.println(format("%-10s%18s%18s%18s",
"Percentile", "Read Latency", "Write Latency", "Range Latency"));
System.out.println(format("%-10s%18s%18s%18s",
"", "(micros)", "(micros)", "(micros)"));
for (int i = 0; i < percentiles.length; i++)
{
System.out.println(format("%-10s%18.2f%18.2f%18.2f",
percentiles[i],
readLatency[i],
writeLatency[i],
rangeLatency[i]));
}
System.out.println();
}
}
@Command(name = "rebuild", description = "Rebuild data by streaming from other nodes (similarly to bootstrap)")
public static class Rebuild extends NodeToolCmd
{
@Arguments(usage = "", description = "Name of DC from which to select sources for streaming. By default, pick any DC")
private String sourceDataCenterName = null;
@Override
public void execute(NodeProbe probe)
{
probe.rebuild(sourceDataCenterName);
}
}
@Command(name = "refresh", description = "Load newly placed SSTables to the system without restart")
public static class Refresh extends NodeToolCmd
{
@Arguments(usage = " ", description = "The keyspace and column family name")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
checkArgument(args.size() == 2, "refresh requires ks and cf args");
probe.loadNewSSTables(args.get(0), args.get(1));
}
}
@Deprecated
@Command(name = "removetoken", description = "DEPRECATED (see removenode)", hidden = true)
public static class RemoveToken extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
System.err.println("Warn: removetoken is deprecated, please use removenode instead");
}
}
@Command(name = "removenode", description = "Show status of current node removal, force completion of pending removal or remove provided ID")
public static class RemoveNode extends NodeToolCmd
{
@Arguments(title = "remove_operation", usage = "||", description = "Show status of current node removal, force completion of pending removal, or remove provided ID", required = true)
private String removeOperation = EMPTY;
@Override
public void execute(NodeProbe probe)
{
switch (removeOperation)
{
case "status":
System.out.println("RemovalStatus: " + probe.getRemovalStatus());
break;
case "force":
System.out.println("RemovalStatus: " + probe.getRemovalStatus());
probe.forceRemoveCompletion();
break;
default:
probe.removeNode(removeOperation);
break;
}
}
}
@Command(name = "repair", description = "Repair one or more column families")
public static class Repair extends NodeToolCmd
{
@Arguments(usage = "[ ...]", description = "The keyspace followed by one or many column families")
private List args = new ArrayList<>();
@Option(title = "parallel", name = {"-par", "--parallel"}, description = "Use -par to carry out a parallel repair")
private boolean parallel = false;
@Option(title = "local_dc", name = {"-local", "--in-local-dc"}, description = "Use -local to only repair against nodes in the same datacenter")
private boolean localDC = false;
@Option(title = "specific_dc", name = {"-dc", "--in-dc"}, description = "Use -dc to repair specific datacenters")
private List specificDataCenters = new ArrayList<>();
@Option(title = "specific_host", name = {"-hosts", "--in-hosts"}, description = "Use -hosts to repair specific hosts")
private List specificHosts = new ArrayList<>();
@Option(title = "start_token", name = {"-st", "--start-token"}, description = "Use -st to specify a token at which the repair range starts")
private String startToken = EMPTY;
@Option(title = "end_token", name = {"-et", "--end-token"}, description = "Use -et to specify a token at which repair range ends")
private String endToken = EMPTY;
@Option(title = "primary_range", name = {"-pr", "--partitioner-range"}, description = "Use -pr to repair only the first range returned by the partitioner")
private boolean primaryRange = false;
@Option(title = "incremental_repair", name = {"-inc", "--incremental"}, description = "Use -inc to use the new incremental repair")
private boolean incrementalRepair = false;
@Override
public void execute(NodeProbe probe)
{
List keyspaces = parseOptionalKeyspace(args, probe);
String[] cfnames = parseOptionalColumnFamilies(args);
if (primaryRange && (localDC || !specificHosts.isEmpty() || !specificHosts.isEmpty()))
throw new RuntimeException("Primary range repair should be performed on all nodes in the cluster.");
for (String keyspace : keyspaces)
{
try
{
Collection dataCenters = null;
Collection hosts = null;
if (!specificDataCenters.isEmpty())
dataCenters = newArrayList(specificDataCenters);
else if (localDC)
dataCenters = newArrayList(probe.getDataCenter());
else if(!specificHosts.isEmpty())
hosts = newArrayList(specificHosts);
if (!startToken.isEmpty() || !endToken.isEmpty())
probe.forceRepairRangeAsync(System.out, keyspace, !parallel, dataCenters,hosts, startToken, endToken, !incrementalRepair);
else
probe.forceRepairAsync(System.out, keyspace, !parallel, dataCenters, hosts, primaryRange, !incrementalRepair, cfnames);
} catch (Exception e)
{
throw new RuntimeException("Error occurred during repair", e);
}
}
}
}
@Command(name = "setcachecapacity", description = "Set global key, row, and counter cache capacities (in MB units)")
public static class SetCacheCapacity extends NodeToolCmd
{
@Arguments(title = " ",
usage = " ",
description = "Key cache, row cache, and counter cache (in MB)",
required = true)
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
checkArgument(args.size() == 3, "setcachecapacity requires key-cache-capacity, row-cache-capacity, and counter-cache-capacity args.");
probe.setCacheCapacities(args.get(0), args.get(1), args.get(2));
}
}
@Command(name = "setcompactionthreshold", description = "Set min and max compaction thresholds for a given column family")
public static class SetCompactionThreshold extends NodeToolCmd
{
@Arguments(title = " ", usage = " ", description = "The keyspace, the column family, min and max threshold", required = true)
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
checkArgument(args.size() == 4, "setcompactionthreshold requires ks, cf, min, and max threshold args.");
int minthreshold = parseInt(args.get(2));
int maxthreshold = parseInt(args.get(3));
checkArgument(minthreshold >= 0 && maxthreshold >= 0, "Thresholds must be positive integers");
checkArgument(minthreshold <= maxthreshold, "Min threshold cannot be greater than max.");
checkArgument(minthreshold >= 2 || maxthreshold == 0, "Min threshold must be at least 2");
probe.setCompactionThreshold(args.get(0), args.get(1), minthreshold, maxthreshold);
}
}
@Command(name = "setcompactionthroughput", description = "Set the MB/s throughput cap for compaction in the system, or 0 to disable throttling")
public static class SetCompactionThroughput extends NodeToolCmd
{
@Arguments(title = "compaction_throughput", usage = "", description = "Value in MB, 0 to disable throttling", required = true)
private Integer compactionThroughput = null;
@Override
public void execute(NodeProbe probe)
{
probe.setCompactionThroughput(compactionThroughput);
}
}
@Command(name = "sethintedhandoffthrottlekb", description = "Set hinted handoff throttle in kb per second, per delivery thread.")
public static class SetHintedHandoffThrottleInKB extends NodeToolCmd
{
@Arguments(title = "throttle_in_kb", usage = "", description = "Value in KB per second", required = true)
private Integer throttleInKB = null;
@Override
public void execute(NodeProbe probe)
{
probe.setHintedHandoffThrottleInKB(throttleInKB);
}
}
@Command(name = "setstreamthroughput", description = "Set the MB/s throughput cap for streaming in the system, or 0 to disable throttling")
public static class SetStreamThroughput extends NodeToolCmd
{
@Arguments(title = "stream_throughput", usage = "", description = "Value in MB, 0 to disable throttling", required = true)
private Integer streamThroughput = null;
@Override
public void execute(NodeProbe probe)
{
probe.setStreamThroughput(streamThroughput);
}
}
@Command(name = "settraceprobability", description = "Sets the probability for tracing any given request to value. 0 disables, 1 enables for all requests, 0 is the default")
public static class SetTraceProbability extends NodeToolCmd
{
@Arguments(title = "trace_probability", usage = "", description = "Trace probability between 0 and 1 (ex: 0.2)", required = true)
private Double traceProbability = null;
@Override
public void execute(NodeProbe probe)
{
checkArgument(traceProbability >= 0 && traceProbability <= 1, "Trace probability must be between 0 and 1");
probe.setTraceProbability(traceProbability);
}
}
@Command(name = "snapshot", description = "Take a snapshot of specified keyspaces or a snapshot of the specified column family")
public static class Snapshot extends NodeToolCmd
{
@Arguments(usage = "[]", description = "List of keyspaces. By default, all keyspaces")
private List keyspaces = new ArrayList<>();
@Option(title = "cfname", name = {"-cf", "--column-family"}, description = "The column family name (you must specify one and only one keyspace for using this option)")
private String columnFamily = null;
@Option(title = "tag", name = {"-t", "--tag"}, description = "The name of the snapshot")
private String snapshotName = Long.toString(System.currentTimeMillis());
@Override
public void execute(NodeProbe probe)
{
try
{
StringBuilder sb = new StringBuilder();
sb.append("Requested creating snapshot(s) for ");
if (keyspaces.isEmpty())
sb.append("[all keyspaces]");
else
sb.append("[").append(join(keyspaces, ", ")).append("]");
if (!snapshotName.isEmpty())
sb.append(" with snapshot name [").append(snapshotName).append("]");
System.out.println(sb.toString());
probe.takeSnapshot(snapshotName, columnFamily, toArray(keyspaces, String.class));
System.out.println("Snapshot directory: " + snapshotName);
} catch (IOException e)
{
throw new RuntimeException("Error during taking a snapshot", e);
}
}
}
@Command(name = "listsnapshots", description = "Lists all the snapshots along with the size on disk and true size.")
public static class ListSnapshots extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
try
{
System.out.println("Snapshot Details: ");
final Map snapshotDetails = probe.getSnapshotDetails();
if (snapshotDetails.isEmpty())
{
System.out.printf("There are no snapshots");
return;
}
final long trueSnapshotsSize = probe.trueSnapshotsSize();
final String format = "%-20s%-29s%-29s%-19s%-19s%n";
// display column names only once
final List indexNames = snapshotDetails.entrySet().iterator().next().getValue().getTabularType().getIndexNames();
System.out.printf(format, (Object[]) indexNames.toArray(new String[indexNames.size()]));
for (final Map.Entry snapshotDetail : snapshotDetails.entrySet())
{
Set> values = snapshotDetail.getValue().keySet();
for (Object eachValue : values)
{
final List> value = (List>) eachValue;
System.out.printf(format, value.toArray(new Object[value.size()]));
}
}
System.out.println("\nTotal TrueDiskSpaceUsed: " + FileUtils.stringifyFileSize(trueSnapshotsSize) + "\n");
}
catch (Exception e)
{
throw new RuntimeException("Error during list snapshot", e);
}
}
}
@Command(name = "status", description = "Print cluster information (state, load, IDs, ...)")
public static class Status extends NodeToolCmd
{
@Arguments(usage = "[]", description = "The keyspace name")
private String keyspace = null;
@Option(title = "resolve_ip", name = {"-r", "--resolve-ip"}, description = "Show node domain names instead of IPs")
private boolean resolveIp = false;
private boolean hasEffectiveOwns = false;
private boolean isTokenPerNode = true;
private int maxAddressLength = 0;
private String format = null;
private Collection joiningNodes, leavingNodes, movingNodes, liveNodes, unreachableNodes;
private Map loadMap, hostIDMap, tokensToEndpoints;
private EndpointSnitchInfoMBean epSnitchInfo;
@Override
public void execute(NodeProbe probe)
{
joiningNodes = probe.getJoiningNodes();
leavingNodes = probe.getLeavingNodes();
movingNodes = probe.getMovingNodes();
loadMap = probe.getLoadMap();
tokensToEndpoints = probe.getTokenToEndpointMap();
liveNodes = probe.getLiveNodes();
unreachableNodes = probe.getUnreachableNodes();
hostIDMap = probe.getHostIdMap();
epSnitchInfo = probe.getEndpointSnitchInfoProxy();
Map ownerships;
try
{
ownerships = probe.effectiveOwnership(keyspace);
hasEffectiveOwns = true;
} catch (IllegalStateException e)
{
ownerships = probe.getOwnership();
System.out.printf("Note: Ownership information does not include topology; for complete information, specify a keyspace%n");
}
Map dcs = getOwnershipByDc(probe, resolveIp, tokensToEndpoints, ownerships);
// More tokens than nodes (aka vnodes)?
if (dcs.values().size() < tokensToEndpoints.keySet().size())
isTokenPerNode = false;
findMaxAddressLength(dcs);
// Datacenters
for (Map.Entry dc : dcs.entrySet())
{
String dcHeader = String.format("Datacenter: %s%n", dc.getKey());
System.out.printf(dcHeader);
for (int i = 0; i < (dcHeader.length() - 1); i++) System.out.print('=');
System.out.println();
// Legend
System.out.println("Status=Up/Down");
System.out.println("|/ State=Normal/Leaving/Joining/Moving");
printNodesHeader(hasEffectiveOwns, isTokenPerNode);
ArrayListMultimap hostToTokens = ArrayListMultimap.create();
for (HostStat stat : dc.getValue())
hostToTokens.put(stat.endpoint, stat);
for (InetAddress endpoint : hostToTokens.keySet())
{
Float owns = ownerships.get(endpoint);
List tokens = hostToTokens.get(endpoint);
printNode(endpoint.getHostAddress(), owns, tokens, hasEffectiveOwns, isTokenPerNode);
}
}
}
private void findMaxAddressLength(Map dcs)
{
maxAddressLength = 0;
for (Map.Entry dc : dcs.entrySet())
{
for (HostStat stat : dc.getValue())
{
maxAddressLength = Math.max(maxAddressLength, stat.ipOrDns().length());
}
}
}
private void printNodesHeader(boolean hasEffectiveOwns, boolean isTokenPerNode)
{
String fmt = getFormat(hasEffectiveOwns, isTokenPerNode);
String owns = hasEffectiveOwns ? "Owns (effective)" : "Owns";
if (isTokenPerNode)
System.out.printf(fmt, "-", "-", "Address", "Load", owns, "Host ID", "Token", "Rack");
else
System.out.printf(fmt, "-", "-", "Address", "Load", "Tokens", owns, "Host ID", "Rack");
}
private void printNode(String endpoint, Float owns, List tokens, boolean hasEffectiveOwns, boolean isTokenPerNode)
{
String status, state, load, strOwns, hostID, rack, fmt;
fmt = getFormat(hasEffectiveOwns, isTokenPerNode);
if (liveNodes.contains(endpoint)) status = "U";
else if (unreachableNodes.contains(endpoint)) status = "D";
else status = "?";
if (joiningNodes.contains(endpoint)) state = "J";
else if (leavingNodes.contains(endpoint)) state = "L";
else if (movingNodes.contains(endpoint)) state = "M";
else state = "N";
load = loadMap.containsKey(endpoint) ? loadMap.get(endpoint) : "?";
strOwns = owns != null ? new DecimalFormat("##0.0%").format(owns) : "?";
hostID = hostIDMap.get(endpoint);
try
{
rack = epSnitchInfo.getRack(endpoint);
} catch (UnknownHostException e)
{
throw new RuntimeException(e);
}
String endpointDns = tokens.get(0).ipOrDns();
if (isTokenPerNode)
System.out.printf(fmt, status, state, endpointDns, load, strOwns, hostID, tokens.get(0).token, rack);
else
System.out.printf(fmt, status, state, endpointDns, load, tokens.size(), strOwns, hostID, rack);
}
private String getFormat(
boolean hasEffectiveOwns,
boolean isTokenPerNode)
{
if (format == null)
{
StringBuilder buf = new StringBuilder();
String addressPlaceholder = String.format("%%-%ds ", maxAddressLength);
buf.append("%s%s "); // status
buf.append(addressPlaceholder); // address
buf.append("%-9s "); // load
if (!isTokenPerNode)
buf.append("%-6s "); // "Tokens"
if (hasEffectiveOwns)
buf.append("%-16s "); // "Owns (effective)"
else
buf.append("%-6s "); // "Owns
buf.append("%-36s "); // Host ID
if (isTokenPerNode)
buf.append("%-39s "); // token
buf.append("%s%n"); // "Rack"
format = buf.toString();
}
return format;
}
}
private static Map getOwnershipByDc(NodeProbe probe, boolean resolveIp,
Map tokenToEndpoint,
Map ownerships)
{
Map ownershipByDc = Maps.newLinkedHashMap();
EndpointSnitchInfoMBean epSnitchInfo = probe.getEndpointSnitchInfoProxy();
try
{
for (Entry tokenAndEndPoint : tokenToEndpoint.entrySet())
{
String dc = epSnitchInfo.getDatacenter(tokenAndEndPoint.getValue());
if (!ownershipByDc.containsKey(dc))
ownershipByDc.put(dc, new SetHostStat(resolveIp));
ownershipByDc.get(dc).add(tokenAndEndPoint.getKey(), tokenAndEndPoint.getValue(), ownerships);
}
}
catch (UnknownHostException e)
{
throw new RuntimeException(e);
}
return ownershipByDc;
}
static class SetHostStat implements Iterable
{
final List hostStats = new ArrayList();
final boolean resolveIp;
public SetHostStat(boolean resolveIp)
{
this.resolveIp = resolveIp;
}
public int size()
{
return hostStats.size();
}
@Override
public Iterator iterator()
{
return hostStats.iterator();
}
public void add(String token, String host, Map ownerships) throws UnknownHostException
{
InetAddress endpoint = InetAddress.getByName(host);
Float owns = ownerships.get(endpoint);
hostStats.add(new HostStat(token, endpoint, resolveIp, owns));
}
}
static class HostStat
{
public final InetAddress endpoint;
public final boolean resolveIp;
public final Float owns;
public final String token;
public HostStat(String token, InetAddress endpoint, boolean resolveIp, Float owns)
{
this.token = token;
this.endpoint = endpoint;
this.resolveIp = resolveIp;
this.owns = owns;
}
public String ipOrDns()
{
return resolveIp ? endpoint.getHostName() : endpoint.getHostAddress();
}
}
@Command(name = "statusbinary", description = "Status of native transport (binary protocol)")
public static class StatusBinary extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
System.out.println(
probe.isNativeTransportRunning()
? "running"
: "not running");
}
}
@Command(name = "statusthrift", description = "Status of thrift server")
public static class StatusThrift extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
System.out.println(
probe.isThriftServerRunning()
? "running"
: "not running");
}
}
@Command(name = "stop", description = "Stop compaction")
public static class Stop extends NodeToolCmd
{
@Arguments(title = "compaction_type", usage = "", description = "Supported types are COMPACTION, VALIDATION, CLEANUP, SCRUB, INDEX_BUILD", required = true)
private OperationType compactionType = OperationType.UNKNOWN;
@Override
public void execute(NodeProbe probe)
{
probe.stop(compactionType.name());
}
}
@Command(name = "stopdaemon", description = "Stop cassandra daemon")
public static class StopDaemon extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
try
{
probe.stopCassandraDaemon();
} catch (Exception ignored)
{
// ignored
}
System.out.println("Cassandra has shutdown.");
}
}
@Command(name = "version", description = "Print cassandra version")
public static class Version extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
System.out.println("ReleaseVersion: " + probe.getReleaseVersion());
}
}
@Command(name = "describering", description = "Shows the token ranges info of a given keyspace")
public static class DescribeRing extends NodeToolCmd
{
@Arguments(description = "The keyspace name", required = true)
String keyspace = EMPTY;
@Override
public void execute(NodeProbe probe)
{
System.out.println("Schema Version:" + probe.getSchemaVersion());
System.out.println("TokenRange: ");
try
{
for (String tokenRangeString : probe.describeRing(keyspace))
{
System.out.println("\t" + tokenRangeString);
}
} catch (IOException e)
{
throw new RuntimeException(e);
}
}
}
@Command(name = "rangekeysample", description = "Shows the sampled keys held across all keyspaces")
public static class RangeKeySample extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
System.out.println("RangeKeySample: ");
List tokenStrings = probe.sampleKeyRange();
for (String tokenString : tokenStrings)
{
System.out.println("\t" + tokenString);
}
}
}
@Command(name = "rebuild_index", description = "A full rebuild of native secondary indexes for a given column family")
public static class RebuildIndex extends NodeToolCmd
{
@Arguments(usage = " ", description = "The keyspace and column family name followed by a list of index names (IndexNameExample: Standard3.IdxName Standard3.IdxName1)")
List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
checkArgument(args.size() >= 3, "rebuild_index requires ks, cf and idx args");
probe.rebuildIndex(args.get(0), args.get(1), toArray(args.subList(2, args.size()), String.class));
}
}
@Command(name = "resetlocalschema", description = "Reset node's local schema and resync")
public static class ResetLocalSchema extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
try
{
probe.resetLocalSchema();
} catch (IOException e)
{
throw new RuntimeException(e);
}
}
}
@Command(name = "enablebackup", description = "Enable incremental backup")
public static class EnableBackup extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.setIncrementalBackupsEnabled(true);
}
}
@Command(name = "disablebackup", description = "Disable incremental backup")
public static class DisableBackup extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.setIncrementalBackupsEnabled(false);
}
}
@Command(name = "setcachekeystosave", description = "Set number of keys saved by each cache for faster post-restart warmup. 0 to disable")
public static class SetCacheKeysToSave extends NodeToolCmd
{
@Arguments(title = " ",
usage = " ",
description = "The number of keys saved by each cache. 0 to disable",
required = true)
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
checkArgument(args.size() == 3, "setcachekeystosave requires key-cache-keys-to-save, row-cache-keys-to-save, and counter-cache-keys-to-save args.");
probe.setCacheKeysToSave(args.get(0), args.get(1), args.get(2));
}
}
@Command(name = "reloadtriggers", description = "Reload trigger classes")
public static class ReloadTriggers extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.reloadTriggers();
}
}
@Command(name = "disablehandoff", description = "Disable gossip (effectively marking the node down)")
public static class DisableHandoff extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.disableHintedHandoff();
}
}
@Command(name = "disablethrift", description = "Disable thrift server")
public static class DisableThrift extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
probe.stopThriftServer();
}
}
@Command(name = "drain", description = "Drain the node (stop accepting writes and flush all column families)")
public static class Drain extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
try
{
probe.drain();
} catch (IOException | InterruptedException | ExecutionException e)
{
throw new RuntimeException("Error occured during flushing", e);
}
}
}
@Command(name = "tpstats", description = "Print usage statistics of thread pools")
public static class TpStats extends NodeTool.NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
System.out.printf("%-25s%10s%10s%15s%10s%18s%n", "Pool Name", "Active", "Pending", "Completed", "Blocked", "All time blocked");
Iterator> threads = probe.getThreadPoolMBeanProxies();
while (threads.hasNext())
{
Map.Entry thread = threads.next();
String poolName = thread.getKey();
JMXEnabledThreadPoolExecutorMBean threadPoolProxy = thread.getValue();
System.out.printf("%-25s%10s%10s%15s%10s%18s%n",
poolName,
threadPoolProxy.getActiveCount(),
threadPoolProxy.getPendingTasks(),
threadPoolProxy.getCompletedTasks(),
threadPoolProxy.getCurrentlyBlockedTasks(),
threadPoolProxy.getTotalBlockedTasks());
}
System.out.printf("%n%-20s%10s%n", "Message type", "Dropped");
for (Map.Entry entry : probe.getDroppedMessages().entrySet())
System.out.printf("%-20s%10s%n", entry.getKey(), entry.getValue());
}
}
@Command(name = "truncatehints", description = "Truncate all hints on the local node, or truncate hints for the endpoint(s) specified.")
public static class TruncateHints extends NodeToolCmd
{
@Arguments(usage = "[endpoint ... ]", description = "Endpoint address(es) to delete hints for, either ip address (\"127.0.0.1\") or hostname")
private String endpoint = EMPTY;
@Override
public void execute(NodeProbe probe)
{
if (endpoint.isEmpty())
probe.truncateHints();
else
probe.truncateHints(endpoint);
}
}
@Command(name = "setlogginglevel", description = "Set the log level threshold for a given class. If both class and level are empty/null, it will reset to the initial configuration")
public static class SetLoggingLevel extends NodeToolCmd
{
@Arguments(usage = " ", description = "The class to change the level for and the log level threshold to set (can be empty)")
private List args = new ArrayList<>();
@Override
public void execute(NodeProbe probe)
{
String classQualifier = args.size() >= 1 ? args.get(0) : EMPTY;
String level = args.size() == 2 ? args.get(1) : EMPTY;
probe.setLoggingLevel(classQualifier, level);
}
}
@Command(name = "getlogginglevels", description = "Get the runtime logging levels")
public static class GetLoggingLevels extends NodeToolCmd
{
@Override
public void execute(NodeProbe probe)
{
// what if some one set a very long logger name? 50 space may not be enough...
System.out.printf("%n%-50s%10s%n", "Logger Name", "Log Level");
for (Map.Entry entry : probe.getLoggingLevels().entrySet())
System.out.printf("%-50s%10s%n", entry.getKey(), entry.getValue());
}
}
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy