All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.datastax.driver.core.Cluster Maven / Gradle / Ivy

Go to download

A driver for Apache Cassandra 1.2+ that works exclusively with the Cassandra Query Language version 3 (CQL3) and Cassandra's binary protocol.

There is a newer version: 4.0.0
Show newest version
/*
 * Copyright DataStax, Inc.
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 * http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package com.datastax.driver.core;

import static com.datastax.driver.core.SchemaElement.KEYSPACE;

import com.datastax.driver.core.exceptions.AuthenticationException;
import com.datastax.driver.core.exceptions.BusyConnectionException;
import com.datastax.driver.core.exceptions.ConnectionException;
import com.datastax.driver.core.exceptions.InvalidQueryException;
import com.datastax.driver.core.exceptions.NoHostAvailableException;
import com.datastax.driver.core.exceptions.SyntaxError;
import com.datastax.driver.core.exceptions.UnsupportedProtocolVersionException;
import com.datastax.driver.core.policies.AddressTranslator;
import com.datastax.driver.core.policies.IdentityTranslator;
import com.datastax.driver.core.policies.LatencyAwarePolicy;
import com.datastax.driver.core.policies.LoadBalancingPolicy;
import com.datastax.driver.core.policies.Policies;
import com.datastax.driver.core.policies.ReconnectionPolicy;
import com.datastax.driver.core.policies.RetryPolicy;
import com.datastax.driver.core.policies.SpeculativeExecutionPolicy;
import com.datastax.driver.core.utils.MoreFutures;
import com.google.common.annotations.VisibleForTesting;
import com.google.common.base.Functions;
import com.google.common.base.Predicates;
import com.google.common.base.Strings;
import com.google.common.base.Throwables;
import com.google.common.collect.HashMultimap;
import com.google.common.collect.Iterables;
import com.google.common.collect.Lists;
import com.google.common.collect.MapMaker;
import com.google.common.collect.SetMultimap;
import com.google.common.collect.Sets;
import com.google.common.util.concurrent.AsyncFunction;
import com.google.common.util.concurrent.FutureCallback;
import com.google.common.util.concurrent.Futures;
import com.google.common.util.concurrent.ListenableFuture;
import com.google.common.util.concurrent.ListeningExecutorService;
import com.google.common.util.concurrent.MoreExecutors;
import com.google.common.util.concurrent.SettableFuture;
import com.google.common.util.concurrent.Uninterruptibles;
import java.io.Closeable;
import java.net.InetAddress;
import java.net.InetSocketAddress;
import java.net.UnknownHostException;
import java.util.ArrayList;
import java.util.Collection;
import java.util.Collections;
import java.util.HashMap;
import java.util.Iterator;
import java.util.List;
import java.util.Map;
import java.util.Map.Entry;
import java.util.ResourceBundle;
import java.util.Set;
import java.util.concurrent.BlockingQueue;
import java.util.concurrent.CancellationException;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.ConcurrentMap;
import java.util.concurrent.CopyOnWriteArraySet;
import java.util.concurrent.ExecutionException;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Future;
import java.util.concurrent.FutureTask;
import java.util.concurrent.ScheduledExecutorService;
import java.util.concurrent.ThreadPoolExecutor;
import java.util.concurrent.TimeUnit;
import java.util.concurrent.atomic.AtomicInteger;
import java.util.concurrent.atomic.AtomicReference;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

/**
 * Information and known state of a Cassandra cluster.
 *
 * 

This is the main entry point of the driver. A simple example of access to a Cassandra cluster * would be: * *

 *   Cluster cluster = Cluster.builder().addContactPoint("192.168.0.1").build();
 *   Session session = cluster.connect("db1");
 *
 *   for (Row row : session.execute("SELECT * FROM table1"))
 *       // do something ...
 * 
* *

A cluster object maintains a permanent connection to one of the cluster nodes which it uses * solely to maintain information on the state and current topology of the cluster. Using the * connection, the driver will discover all the nodes currently in the cluster as well as new nodes * joining the cluster subsequently. */ public class Cluster implements Closeable { private static final Logger logger = LoggerFactory.getLogger(Cluster.class); private static final ResourceBundle driverProperties = ResourceBundle.getBundle("com.datastax.driver.core.Driver"); static { logDriverVersion(); // Force initialization to fail fast if there is an issue detecting the version GuavaCompatibility.init(); } @VisibleForTesting static final int NEW_NODE_DELAY_SECONDS = SystemProperties.getInt("com.datastax.driver.NEW_NODE_DELAY_SECONDS", 1); // Some per-JVM number that allows to generate unique cluster names when // multiple Cluster instance are created in the same JVM. private static final AtomicInteger CLUSTER_ID = new AtomicInteger(0); private static final int NOTIF_LOCK_TIMEOUT_SECONDS = SystemProperties.getInt("com.datastax.driver.NOTIF_LOCK_TIMEOUT_SECONDS", 60); final Manager manager; /** * Constructs a new Cluster instance. * *

This constructor is mainly exposed so Cluster can be sub-classed as a means to make * testing/mocking easier or to "intercept" its method call. Most users shouldn't extend this * class however and should prefer either using the {@link #builder} or calling {@link #buildFrom} * with a custom Initializer. * * @param name the name to use for the cluster (this is not the Cassandra cluster name, see {@link * #getClusterName}). * @param contactPoints the list of contact points to use for the new cluster. * @param configuration the configuration for the new cluster. */ protected Cluster( String name, List contactPoints, Configuration configuration) { this(name, contactPoints, configuration, Collections.emptySet()); } /** * Constructs a new Cluster instance. * *

This constructor is mainly exposed so Cluster can be sub-classed as a means to make * testing/mocking easier or to "intercept" its method call. Most users shouldn't extend this * class however and should prefer using the {@link #builder}. * * @param initializer the initializer to use. * @see #buildFrom */ protected Cluster(Initializer initializer) { this( initializer.getClusterName(), checkNotEmpty(initializer.getContactPoints()), initializer.getConfiguration(), initializer.getInitialListeners()); } private static List checkNotEmpty(List contactPoints) { if (contactPoints.isEmpty()) throw new IllegalArgumentException("Cannot build a cluster without contact points"); return contactPoints; } private Cluster( String name, List contactPoints, Configuration configuration, Collection listeners) { this.manager = new Manager(name, contactPoints, configuration, listeners); } /** * Initialize this Cluster instance. * *

This method creates an initial connection to one of the contact points used to construct the * {@code Cluster} instance. That connection is then used to populate the cluster {@link * Metadata}. * *

Calling this method is optional in the sense that any call to one of the {@code connect} * methods of this object will automatically trigger a call to this method beforehand. It is thus * only useful to call this method if for some reason you want to populate the metadata (or test * that at least one contact point can be reached) without creating a first {@code Session}. * *

Please note that this method only creates one control connection for gathering cluster * metadata. In particular, it doesn't create any connection pools. Those are created when a new * {@code Session} is created through {@code connect}. * *

This method has no effect if the cluster is already initialized. * * @return this {@code Cluster} object. * @throws NoHostAvailableException if no host amongst the contact points can be reached. * @throws AuthenticationException if an authentication error occurs while contacting the initial * contact points. * @throws IllegalStateException if the Cluster was closed prior to calling this method. This can * occur either directly (through {@link #close()} or {@link #closeAsync()}), or as a result * of an error while initializing the Cluster. */ public Cluster init() { this.manager.init(); return this; } /** * Build a new cluster based on the provided initializer. * *

Note that for building a cluster pragmatically, Cluster.Builder provides a slightly less * verbose shortcut with {@link Builder#build}. * *

Also note that that all the contact points provided by {@code initializer} must share the * same port. * * @param initializer the Cluster.Initializer to use * @return the newly created Cluster instance * @throws IllegalArgumentException if the list of contact points provided by {@code initializer} * is empty or if not all those contact points have the same port. */ public static Cluster buildFrom(Initializer initializer) { return new Cluster(initializer); } /** * Creates a new {@link Cluster.Builder} instance. * *

This is a convenience method for {@code new Cluster.Builder()}. * * @return the new cluster builder. */ public static Cluster.Builder builder() { return new Cluster.Builder(); } /** * Returns the current version of the driver. * *

This is intended for products that wrap or extend the driver, as a way to check * compatibility if end-users override the driver version in their application. * * @return the version. */ public static String getDriverVersion() { return driverProperties.getString("driver.version"); } /** * Logs the driver version to the console. * *

This method logs the version using the logger {@code com.datastax.driver.core} and level * {@code INFO}. */ public static void logDriverVersion() { Logger core = LoggerFactory.getLogger("com.datastax.driver.core"); core.info("DataStax Java driver {} for Apache Cassandra", getDriverVersion()); } /** * Creates a new session on this cluster but does not initialize it. * *

Because this method does not perform any initialization, it cannot fail. The initialization * of the session (the connection of the Session to the Cassandra nodes) will occur if either the * {@link Session#init} method is called explicitly, or whenever the returned session object is * used. * *

Once a session returned by this method gets initialized (see above), it will be set to no * keyspace. If you want to set such session to a keyspace, you will have to explicitly execute a * 'USE mykeyspace' query. * *

Note that if you do not particularly need to defer initialization, it is simpler to use one * of the {@code connect()} method of this class. * * @return a new, non-initialized session on this cluster. */ public Session newSession() { checkNotClosed(manager); return manager.newSession(); } /** * Creates a new session on this cluster and initialize it. * *

Note that this method will initialize the newly created session, trying to connect to the * Cassandra nodes before returning. If you only want to create a Session object without * initializing it right away, see {@link #newSession}. * * @return a new session on this cluster sets to no keyspace. * @throws NoHostAvailableException if the Cluster has not been initialized yet ({@link #init} has * not be called and this is the first connect call) and no host amongst the contact points * can be reached. * @throws AuthenticationException if an authentication error occurs while contacting the initial * contact points. * @throws IllegalStateException if the Cluster was closed prior to calling this method. This can * occur either directly (through {@link #close()} or {@link #closeAsync()}), or as a result * of an error while initializing the Cluster. */ public Session connect() { try { return Uninterruptibles.getUninterruptibly(connectAsync()); } catch (ExecutionException e) { throw DriverThrowables.propagateCause(e); } } /** * Creates a new session on this cluster, initialize it and sets the keyspace to the provided one. * *

Note that this method will initialize the newly created session, trying to connect to the * Cassandra nodes before returning. If you only want to create a Session object without * initializing it right away, see {@link #newSession}. * * @param keyspace The name of the keyspace to use for the created {@code Session}. * @return a new session on this cluster sets to keyspace {@code keyspaceName}. * @throws NoHostAvailableException if the Cluster has not been initialized yet ({@link #init} has * not be called and this is the first connect call) and no host amongst the contact points * can be reached, or if no host can be contacted to set the {@code keyspace}. * @throws AuthenticationException if an authentication error occurs while contacting the initial * contact points. * @throws InvalidQueryException if the keyspace does not exist. * @throws IllegalStateException if the Cluster was closed prior to calling this method. This can * occur either directly (through {@link #close()} or {@link #closeAsync()}), or as a result * of an error while initializing the Cluster. */ public Session connect(String keyspace) { try { return Uninterruptibles.getUninterruptibly(connectAsync(keyspace)); } catch (ExecutionException e) { throw DriverThrowables.propagateCause(e); } } /** * Creates a new session on this cluster and initializes it asynchronously. * *

This will also initialize the {@code Cluster} if needed; note that cluster initialization * happens synchronously on the thread that called this method. Therefore it is recommended to * initialize the cluster at application startup, and not rely on this method to do it. * * @return a future that will complete when the session is fully initialized. * @throws NoHostAvailableException if the Cluster has not been initialized yet ({@link #init} has * not been called and this is the first connect call) and no host amongst the contact points * can be reached. * @throws IllegalStateException if the Cluster was closed prior to calling this method. This can * occur either directly (through {@link #close()} or {@link #closeAsync()}), or as a result * of an error while initializing the Cluster. * @see #connect() */ public ListenableFuture connectAsync() { return connectAsync(null); } /** * Creates a new session on this cluster, and initializes it to the given keyspace asynchronously. * *

This will also initialize the {@code Cluster} if needed; note that cluster initialization * happens synchronously on the thread that called this method. Therefore it is recommended to * initialize the cluster at application startup, and not rely on this method to do it. * * @param keyspace The name of the keyspace to use for the created {@code Session}. * @return a future that will complete when the session is fully initialized. * @throws NoHostAvailableException if the Cluster has not been initialized yet ({@link #init} has * not been called and this is the first connect call) and no host amongst the contact points * can be reached. * @throws IllegalStateException if the Cluster was closed prior to calling this method. This can * occur either directly (through {@link #close()} or {@link #closeAsync()}), or as a result * of an error while initializing the Cluster. */ public ListenableFuture connectAsync(final String keyspace) { checkNotClosed(manager); init(); final Session session = manager.newSession(); ListenableFuture sessionInitialized = session.initAsync(); if (keyspace == null) { return sessionInitialized; } else { final String useQuery = "USE " + keyspace; ListenableFuture keyspaceSet = GuavaCompatibility.INSTANCE.transformAsync( sessionInitialized, new AsyncFunction() { @Override public ListenableFuture apply(Session session) throws Exception { return session.executeAsync(useQuery); } }); ListenableFuture withErrorHandling = GuavaCompatibility.INSTANCE.withFallback( keyspaceSet, new AsyncFunction() { @Override public ListenableFuture apply(Throwable t) throws Exception { session.closeAsync(); if (t instanceof SyntaxError) { // Give a more explicit message, because it's probably caused by a bad keyspace // name SyntaxError e = (SyntaxError) t; t = new SyntaxError( e.getAddress(), String.format( "Error executing \"%s\" (%s). Check that your keyspace name is valid", useQuery, e.getMessage())); } throw Throwables.propagate(t); } }); return GuavaCompatibility.INSTANCE.transform(withErrorHandling, Functions.constant(session)); } } /** * The name of this cluster object. * *

Note that this is not the Cassandra cluster name, but rather a name assigned to this Cluster * object. Currently, that name is only used for one purpose: to distinguish exposed JMX metrics * when multiple Cluster instances live in the same JVM (which should be rare in the first place). * That name can be set at Cluster building time (through {@link Builder#withClusterName} for * instance) but will default to a name like {@code cluster1} where each Cluster instance in the * same JVM will have a different number. * * @return the name for this cluster instance. */ public String getClusterName() { return manager.clusterName; } /** * Returns read-only metadata on the connected cluster. * *

This includes the known nodes with their status as seen by the driver, as well as the schema * definitions. Since this return metadata on the connected cluster, this method may trigger the * creation of a connection if none has been established yet (neither {@code init()} nor {@code * connect()} has been called yet). * * @return the cluster metadata. * @throws NoHostAvailableException if the Cluster has not been initialized yet and no host * amongst the contact points can be reached. * @throws AuthenticationException if an authentication error occurs while contacting the initial * contact points. * @throws IllegalStateException if the Cluster was closed prior to calling this method. This can * occur either directly (through {@link #close()} or {@link #closeAsync()}), or as a result * of an error while initializing the Cluster. */ public Metadata getMetadata() { manager.init(); return manager.metadata; } /** * The cluster configuration. * * @return the cluster configuration. */ public Configuration getConfiguration() { return manager.configuration; } /** * The cluster metrics. * * @return the cluster metrics, or {@code null} if this cluster has not yet been {@link #init() * initialized}, or if metrics collection has been disabled (that is if {@link * Configuration#getMetricsOptions} returns {@code null}). */ public Metrics getMetrics() { checkNotClosed(manager); return manager.metrics; } /** * Registers the provided listener to be notified on hosts up/down/added/removed events. * *

Registering the same listener multiple times is a no-op. * *

This method should be used to register additional listeners on an already-initialized * cluster. To add listeners to a cluster object prior to its initialization, use {@link * Builder#withInitialListeners(Collection)}. Calling this method on a non-initialized cluster * will result in the listener being {@link * com.datastax.driver.core.Host.StateListener#onRegister(Cluster) notified} twice of cluster * registration: once inside this method, and once at cluster initialization. * * @param listener the new {@link Host.StateListener} to register. * @return this {@code Cluster} object; */ public Cluster register(Host.StateListener listener) { checkNotClosed(manager); boolean added = manager.listeners.add(listener); if (added) listener.onRegister(this); return this; } /** * Unregisters the provided listener from being notified on hosts events. * *

This method is a no-op if {@code listener} hasn't previously been registered against this * Cluster. * * @param listener the {@link Host.StateListener} to unregister. * @return this {@code Cluster} object; */ public Cluster unregister(Host.StateListener listener) { checkNotClosed(manager); boolean removed = manager.listeners.remove(listener); if (removed) listener.onUnregister(this); return this; } /** * Registers the provided tracker to be updated with hosts read latencies. * *

Registering the same tracker multiple times is a no-op. * *

Beware that the registered tracker's {@link LatencyTracker#update(Host, Statement, * Exception, long) update} method will be called very frequently (at the end of every query to a * Cassandra host) and should thus not be costly. * *

The main use case for a {@link LatencyTracker} is to allow load balancing policies to * implement latency awareness. For example, {@link LatencyAwarePolicy} registers it's own * internal {@code LatencyTracker} (automatically, you don't have to call this method directly). * * @param tracker the new {@link LatencyTracker} to register. * @return this {@code Cluster} object; */ public Cluster register(LatencyTracker tracker) { checkNotClosed(manager); boolean added = manager.latencyTrackers.add(tracker); if (added) tracker.onRegister(this); return this; } /** * Unregisters the provided latency tracking from being updated with host read latencies. * *

This method is a no-op if {@code tracker} hasn't previously been registered against this * Cluster. * * @param tracker the {@link LatencyTracker} to unregister. * @return this {@code Cluster} object; */ public Cluster unregister(LatencyTracker tracker) { checkNotClosed(manager); boolean removed = manager.latencyTrackers.remove(tracker); if (removed) tracker.onUnregister(this); return this; } /** * Registers the provided listener to be updated with schema change events. * *

Registering the same listener multiple times is a no-op. * * @param listener the new {@link SchemaChangeListener} to register. * @return this {@code Cluster} object; */ public Cluster register(SchemaChangeListener listener) { checkNotClosed(manager); boolean added = manager.schemaChangeListeners.add(listener); if (added) listener.onRegister(this); return this; } /** * Unregisters the provided schema change listener from being updated with schema change events. * *

This method is a no-op if {@code listener} hasn't previously been registered against this * Cluster. * * @param listener the {@link SchemaChangeListener} to unregister. * @return this {@code Cluster} object; */ public Cluster unregister(SchemaChangeListener listener) { checkNotClosed(manager); boolean removed = manager.schemaChangeListeners.remove(listener); if (removed) listener.onUnregister(this); return this; } /** * Initiates a shutdown of this cluster instance. * *

This method is asynchronous and return a future on the completion of the shutdown process. * As soon a the cluster is shutdown, no new request will be accepted, but already submitted * queries are allowed to complete. This method closes all connections from all sessions and * reclaims all resources used by this Cluster instance. * *

If for some reason you wish to expedite this process, the {@link CloseFuture#force} can be * called on the result future. * *

This method has no particular effect if the cluster was already closed (in which case the * returned future will return immediately). * * @return a future on the completion of the shutdown process. */ public CloseFuture closeAsync() { return manager.close(); } /** * Initiates a shutdown of this cluster instance and blocks until that shutdown completes. * *

This method is a shortcut for {@code closeAsync().get()}. */ @Override public void close() { try { closeAsync().get(); } catch (ExecutionException e) { throw DriverThrowables.propagateCause(e); } catch (InterruptedException e) { Thread.currentThread().interrupt(); } } /** * Whether this Cluster instance has been closed. * *

Note that this method returns true as soon as one of the close methods ({@link #closeAsync} * or {@link #close}) has been called, it does not guarantee that the closing is done. If you want * to guarantee that the closing is done, you can call {@code close()} and wait until it returns * (or call the get method on {@code closeAsync()} with a very short timeout and check this * doesn't timeout). * * @return {@code true} if this Cluster instance has been closed, {@code false} otherwise. */ public boolean isClosed() { return manager.closeFuture.get() != null; } private static void checkNotClosed(Manager manager) { if (manager.errorDuringInit()) { throw new IllegalStateException( "Can't use this cluster instance because it encountered an error in its initialization", manager.getInitException()); } else if (manager.isClosed()) { throw new IllegalStateException( "Can't use this cluster instance because it was previously closed"); } } /** * Initializer for {@link Cluster} instances. * *

If you want to create a new {@code Cluster} instance programmatically, then it is advised to * use {@link Cluster.Builder} which can be obtained from the {@link Cluster#builder} method. * *

But it is also possible to implement a custom {@code Initializer} that retrieves * initialization from a web-service or from a configuration file. */ public interface Initializer { /** * An optional name for the created cluster. * *

Such name is optional (a default name will be created otherwise) and is currently only use * for JMX reporting of metrics. See {@link Cluster#getClusterName} for more information. * * @return the name for the created cluster or {@code null} to use an automatically generated * name. */ public String getClusterName(); /** * Returns the initial Cassandra hosts to connect to. * * @return the initial Cassandra contact points. See {@link Builder#addContactPoint} for more * details on contact points. */ public List getContactPoints(); /** * The configuration to use for the new cluster. * *

Note that some configuration can be modified after the cluster initialization but some * others cannot. In particular, the ones that cannot be changed afterwards includes: * *

    *
  • the port use to connect to Cassandra nodes (see {@link ProtocolOptions}). *
  • the policies used (see {@link Policies}). *
  • the authentication info provided (see {@link Configuration}). *
  • whether metrics are enabled (see {@link Configuration}). *
* * @return the configuration to use for the new cluster. */ public Configuration getConfiguration(); /** * Optional listeners to register against the newly created cluster. * *

Note that contrary to listeners registered post Cluster creation, the listeners returned * by this method will see {@link Host.StateListener#onAdd} events for the initial contact * points. * * @return a possibly empty collection of {@code Host.StateListener} to register against the * newly created cluster. */ public Collection getInitialListeners(); } /** Helper class to build {@link Cluster} instances. */ public static class Builder implements Initializer { private String clusterName; private final List addresses = new ArrayList(); private final List rawAddresses = new ArrayList(); private int port = ProtocolOptions.DEFAULT_PORT; private int maxSchemaAgreementWaitSeconds = ProtocolOptions.DEFAULT_MAX_SCHEMA_AGREEMENT_WAIT_SECONDS; private ProtocolVersion protocolVersion; private AuthProvider authProvider = AuthProvider.NONE; private final Policies.Builder policiesBuilder = Policies.builder(); private final Configuration.Builder configurationBuilder = Configuration.builder(); private ProtocolOptions.Compression compression = ProtocolOptions.Compression.NONE; private SSLOptions sslOptions = null; private boolean metricsEnabled = true; private boolean jmxEnabled = true; private boolean allowBetaProtocolVersion = false; private boolean noCompact = false; private Collection listeners; @Override public String getClusterName() { return clusterName; } @Override public List getContactPoints() { if (rawAddresses.isEmpty()) return addresses; List allAddresses = new ArrayList(addresses); for (InetAddress address : rawAddresses) allAddresses.add(new InetSocketAddress(address, port)); return allAddresses; } /** * An optional name for the create cluster. * *

Note: this is not related to the Cassandra cluster name (though you are free to provide * the same name). See {@link Cluster#getClusterName} for details. * *

If you use this method and create more than one Cluster instance in the same JVM (which * should be avoided unless you need to connect to multiple Cassandra clusters), you should make * sure each Cluster instance get a unique name or you may have a problem with JMX reporting. * * @param name the cluster name to use for the created Cluster instance. * @return this Builder. */ public Builder withClusterName(String name) { this.clusterName = name; return this; } /** * The port to use to connect to the Cassandra host. * *

If not set through this method, the default port (9042) will be used instead. * * @param port the port to set. * @return this Builder. */ public Builder withPort(int port) { this.port = port; return this; } /** * Create cluster connection using latest development protocol version, which is currently in * beta. Calling this method will result into setting USE_BETA flag in all outgoing messages, * which allows server to negotiate the supported protocol version even if it is currently in * beta. * *

This feature is only available starting with version {@link ProtocolVersion#V5 V5}. * *

Use with caution, refer to the server and protocol documentation for the details on latest * protocol version. * * @return this Builder. */ public Builder allowBetaProtocolVersion() { if (protocolVersion != null) throw new IllegalArgumentException( "Can't use beta flag with initial protocol version of " + protocolVersion); this.allowBetaProtocolVersion = true; this.protocolVersion = ProtocolVersion.NEWEST_BETA; return this; } /** * Sets the maximum time to wait for schema agreement before returning from a DDL query. * *

If not set through this method, the default value (10 seconds) will be used. * * @param maxSchemaAgreementWaitSeconds the new value to set. * @return this Builder. * @throws IllegalStateException if the provided value is zero or less. */ public Builder withMaxSchemaAgreementWaitSeconds(int maxSchemaAgreementWaitSeconds) { if (maxSchemaAgreementWaitSeconds <= 0) throw new IllegalArgumentException("Max schema agreement wait must be greater than zero"); this.maxSchemaAgreementWaitSeconds = maxSchemaAgreementWaitSeconds; return this; } /** * The native protocol version to use. * *

The driver supports versions 1 to 5 of the native protocol. Higher versions of the * protocol have more features and should be preferred, but this also depends on the Cassandra * version: * *

* *

* * * * * * * *
Native protocol version to Cassandra version correspondence
Protocol versionMinimum Cassandra version
11.2
22.0
32.1
42.2
53.10
* *

By default, the driver will "auto-detect" which protocol version it can use when * connecting to the first node. More precisely, it will try first with {@link * ProtocolVersion#NEWEST_SUPPORTED}, and if not supported fallback to the highest version * supported by the first node it connects to. Please note that once the version is * "auto-detected", it won't change: if the first node the driver connects to is a Cassandra 1.2 * node and auto-detection is used (the default), then the native protocol version 1 will be use * for the lifetime of the Cluster instance. * *

By using {@link Builder#allowBetaProtocolVersion()}, it is possible to force driver to * connect to Cassandra node that supports the latest protocol beta version. Leaving this flag * out will let client to connect with latest released version. * *

This method allows to force the use of a particular protocol version. Forcing version 1 is * always fine since all Cassandra version (at least all those supporting the native protocol in * the first place) so far support it. However, please note that a number of features of the * driver won't be available if that version of the protocol is in use, including result set * paging, {@link BatchStatement}, executing a non-prepared query with binary values ({@link * Session#execute(String, Object...)}), ... (those methods will throw an * UnsupportedFeatureException). Using the protocol version 1 should thus only be considered * when using Cassandra 1.2, until nodes have been upgraded to Cassandra 2.0. * *

If version 2 of the protocol is used, then Cassandra 1.2 nodes will be ignored (the driver * won't connect to them). * *

The default behavior (auto-detection) is fine in almost all case, but you may want to * force a particular version if you have a Cassandra cluster with mixed 1.2/2.0 nodes (i.e. * during a Cassandra upgrade). * * @param version the native protocol version to use. {@code null} is also supported to trigger * auto-detection (see above) but this is the default (so you don't have to call this method * for that behavior). * @return this Builder. */ public Builder withProtocolVersion(ProtocolVersion version) { if (allowBetaProtocolVersion) throw new IllegalStateException( "Can not set the version explicitly if `allowBetaProtocolVersion` was used."); if (version.compareTo(ProtocolVersion.NEWEST_SUPPORTED) > 0) throw new IllegalArgumentException( "Can not use " + version + " protocol version. " + "Newest supported protocol version is: " + ProtocolVersion.NEWEST_SUPPORTED + ". " + "For beta versions, use `allowBetaProtocolVersion` instead"); this.protocolVersion = version; return this; } /** * Adds a contact point - or many if the given address resolves to multiple InetAddress * s (A records). * *

Contact points are addresses of Cassandra nodes that the driver uses to discover the * cluster topology. Only one contact point is required (the driver will retrieve the address of * the other nodes automatically), but it is usually a good idea to provide more than one * contact point, because if that single contact point is unavailable, the driver cannot * initialize itself correctly. * *

Note that by default (that is, unless you use the {@link #withLoadBalancingPolicy}) method * of this builder), the first successfully contacted host will be used to define the local * data-center for the client. If follows that if you are running Cassandra in a multiple * data-center setting, it is a good idea to only provide contact points that are in the same * datacenter than the client, or to provide manually the load balancing policy that suits your * need. * *

If the host name points to a DNS record with multiple a-records, all InetAddresses * returned will be used. Make sure that all resulting InetAddresss returned point * to the same cluster and datacenter. * * @param address the address of the node(s) to connect to. * @return this Builder. * @throws IllegalArgumentException if the given {@code address} could not be resolved. * @throws SecurityException if a security manager is present and permission to resolve the host * name is denied. */ public Builder addContactPoint(String address) { // We explicitly check for nulls because InetAdress.getByName() will happily // accept it and use localhost (while a null here almost likely mean a user error, // not "connect to localhost") if (address == null) throw new NullPointerException(); try { addContactPoints(InetAddress.getAllByName(address)); return this; } catch (UnknownHostException e) { throw new IllegalArgumentException("Failed to add contact point: " + address, e); } } /** * Adds contact points. * *

See {@link Builder#addContactPoint} for more details on contact points. * *

Note that all contact points must be resolvable; if any of them cannot be * resolved, this method will fail. * * @param addresses addresses of the nodes to add as contact points. * @return this Builder. * @throws IllegalArgumentException if any of the given {@code addresses} could not be resolved. * @throws SecurityException if a security manager is present and permission to resolve the host * name is denied. * @see Builder#addContactPoint */ public Builder addContactPoints(String... addresses) { for (String address : addresses) addContactPoint(address); return this; } /** * Adds contact points. * *

See {@link Builder#addContactPoint} for more details on contact points. * *

Note that all contact points must be resolvable; if any of them cannot be * resolved, this method will fail. * * @param addresses addresses of the nodes to add as contact points. * @return this Builder. * @throws IllegalArgumentException if any of the given {@code addresses} could not be resolved. * @throws SecurityException if a security manager is present and permission to resolve the host * name is denied. * @see Builder#addContactPoint */ public Builder addContactPoints(InetAddress... addresses) { Collections.addAll(this.rawAddresses, addresses); return this; } /** * Adds contact points. * *

See {@link Builder#addContactPoint} for more details on contact points. * * @param addresses addresses of the nodes to add as contact points. * @return this Builder * @see Builder#addContactPoint */ public Builder addContactPoints(Collection addresses) { this.rawAddresses.addAll(addresses); return this; } /** * Adds contact points. * *

See {@link Builder#addContactPoint} for more details on contact points. Contrarily to * other {@code addContactPoints} methods, this method allows to provide a different port for * each contact point. Since Cassandra nodes must always all listen on the same port, this is * rarely what you want and most users should prefer other {@code addContactPoints} methods to * this one. However, this can be useful if the Cassandra nodes are behind a router and are not * accessed directly. Note that if you are in this situation (Cassandra nodes are behind a * router, not directly accessible), you almost surely want to provide a specific {@link * AddressTranslator} (through {@link #withAddressTranslator}) to translate actual Cassandra * node addresses to the addresses the driver should use, otherwise the driver will not be able * to auto-detect new nodes (and will generally not function optimally). * * @param addresses addresses of the nodes to add as contact points. * @return this Builder * @see Builder#addContactPoint */ public Builder addContactPointsWithPorts(InetSocketAddress... addresses) { Collections.addAll(this.addresses, addresses); return this; } /** * Adds contact points. * *

See {@link Builder#addContactPoint} for more details on contact points. Contrarily to * other {@code addContactPoints} methods, this method allows to provide a different port for * each contact point. Since Cassandra nodes must always all listen on the same port, this is * rarely what you want and most users should prefer other {@code addContactPoints} methods to * this one. However, this can be useful if the Cassandra nodes are behind a router and are not * accessed directly. Note that if you are in this situation (Cassandra nodes are behind a * router, not directly accessible), you almost surely want to provide a specific {@link * AddressTranslator} (through {@link #withAddressTranslator}) to translate actual Cassandra * node addresses to the addresses the driver should use, otherwise the driver will not be able * to auto-detect new nodes (and will generally not function optimally). * * @param addresses addresses of the nodes to add as contact points. * @return this Builder * @see Builder#addContactPoint */ public Builder addContactPointsWithPorts(Collection addresses) { this.addresses.addAll(addresses); return this; } /** * Configures the load balancing policy to use for the new cluster. * *

If no load balancing policy is set through this method, {@link * Policies#defaultLoadBalancingPolicy} will be used instead. * * @param policy the load balancing policy to use. * @return this Builder. */ public Builder withLoadBalancingPolicy(LoadBalancingPolicy policy) { policiesBuilder.withLoadBalancingPolicy(policy); return this; } /** * Configures the reconnection policy to use for the new cluster. * *

If no reconnection policy is set through this method, {@link * Policies#DEFAULT_RECONNECTION_POLICY} will be used instead. * * @param policy the reconnection policy to use. * @return this Builder. */ public Builder withReconnectionPolicy(ReconnectionPolicy policy) { policiesBuilder.withReconnectionPolicy(policy); return this; } /** * Configures the retry policy to use for the new cluster. * *

If no retry policy is set through this method, {@link Policies#DEFAULT_RETRY_POLICY} will * be used instead. * * @param policy the retry policy to use. * @return this Builder. */ public Builder withRetryPolicy(RetryPolicy policy) { policiesBuilder.withRetryPolicy(policy); return this; } /** * Configures the address translator to use for the new cluster. * *

See {@link AddressTranslator} for more detail on address translation, but the default * translator, {@link IdentityTranslator}, should be correct in most cases. If unsure, stick to * the default. * * @param translator the translator to use. * @return this Builder. */ public Builder withAddressTranslator(AddressTranslator translator) { policiesBuilder.withAddressTranslator(translator); return this; } /** * Configures the generator that will produce the client-side timestamp sent with each query. * *

This feature is only available with version {@link ProtocolVersion#V3 V3} or above of the * native protocol. With earlier versions, timestamps are always generated server-side, and * setting a generator through this method will have no effect. * *

If no generator is set through this method, the driver will default to client-side * timestamps by using {@link AtomicMonotonicTimestampGenerator}. * * @param timestampGenerator the generator to use. * @return this Builder. */ public Builder withTimestampGenerator(TimestampGenerator timestampGenerator) { policiesBuilder.withTimestampGenerator(timestampGenerator); return this; } /** * Configures the speculative execution policy to use for the new cluster. * *

If no policy is set through this method, {@link * Policies#defaultSpeculativeExecutionPolicy()} will be used instead. * * @param policy the policy to use. * @return this Builder. */ public Builder withSpeculativeExecutionPolicy(SpeculativeExecutionPolicy policy) { policiesBuilder.withSpeculativeExecutionPolicy(policy); return this; } /** * Configures the {@link CodecRegistry} instance to use for the new cluster. * *

If no codec registry is set through this method, {@link CodecRegistry#DEFAULT_INSTANCE} * will be used instead. * *

Note that if two or more {@link Cluster} instances are configured to use the default codec * registry, they are going to share the same instance. In this case, care should be taken when * registering new codecs on it as any codec registered by one cluster would be immediately * available to others sharing the same default instance. * * @param codecRegistry the codec registry to use. * @return this Builder. */ public Builder withCodecRegistry(CodecRegistry codecRegistry) { configurationBuilder.withCodecRegistry(codecRegistry); return this; } /** * Uses the provided credentials when connecting to Cassandra hosts. * *

This should be used if the Cassandra cluster has been configured to use the {@code * PasswordAuthenticator}. If the the default {@code AllowAllAuthenticator} is used instead, * using this method has no effect. * * @param username the username to use to login to Cassandra hosts. * @param password the password corresponding to {@code username}. * @return this Builder. */ public Builder withCredentials(String username, String password) { this.authProvider = new PlainTextAuthProvider(username, password); return this; } /** * Use the specified AuthProvider when connecting to Cassandra hosts. * *

Use this method when a custom authentication scheme is in place. You shouldn't call both * this method and {@code withCredentials} on the same {@code Builder} instance as one will * supersede the other * * @param authProvider the {@link AuthProvider} to use to login to Cassandra hosts. * @return this Builder */ public Builder withAuthProvider(AuthProvider authProvider) { this.authProvider = authProvider; return this; } /** * Sets the compression to use for the transport. * * @param compression the compression to set. * @return this Builder. * @see ProtocolOptions.Compression */ public Builder withCompression(ProtocolOptions.Compression compression) { this.compression = compression; return this; } /** * Disables metrics collection for the created cluster (metrics are enabled by default * otherwise). * * @return this builder. */ public Builder withoutMetrics() { this.metricsEnabled = false; return this; } /** * Enables the use of SSL for the created {@code Cluster}. * *

Calling this method will use the JDK-based implementation with the default options (see * {@link RemoteEndpointAwareJdkSSLOptions.Builder}). This is thus a shortcut for {@code * withSSL(JdkSSLOptions.builder().build())}. * *

Note that if SSL is enabled, the driver will not connect to any Cassandra nodes that * doesn't have SSL enabled and it is strongly advised to enable SSL on every Cassandra node if * you plan on using SSL in the driver. * * @return this builder. */ public Builder withSSL() { this.sslOptions = RemoteEndpointAwareJdkSSLOptions.builder().build(); return this; } /** * Enable the use of SSL for the created {@code Cluster} using the provided options. * * @param sslOptions the SSL options to use. * @return this builder. */ public Builder withSSL(SSLOptions sslOptions) { this.sslOptions = sslOptions; return this; } /** * Register the provided listeners in the newly created cluster. * *

Note: repeated calls to this method will override the previous ones. * * @param listeners the listeners to register. * @return this builder. */ public Builder withInitialListeners(Collection listeners) { this.listeners = listeners; return this; } /** * Disables JMX reporting of the metrics. * *

JMX reporting is enabled by default (see {@link Metrics}) but can be disabled using this * option. If metrics are disabled, this is a no-op. * * @return this builder. */ public Builder withoutJMXReporting() { this.jmxEnabled = false; return this; } /** * Sets the PoolingOptions to use for the newly created Cluster. * *

If no pooling options are set through this method, default pooling options will be used. * * @param options the pooling options to use. * @return this builder. */ public Builder withPoolingOptions(PoolingOptions options) { configurationBuilder.withPoolingOptions(options); return this; } /** * Sets the SocketOptions to use for the newly created Cluster. * *

If no socket options are set through this method, default socket options will be used. * * @param options the socket options to use. * @return this builder. */ public Builder withSocketOptions(SocketOptions options) { configurationBuilder.withSocketOptions(options); return this; } /** * Sets the QueryOptions to use for the newly created Cluster. * *

If no query options are set through this method, default query options will be used. * * @param options the query options to use. * @return this builder. */ public Builder withQueryOptions(QueryOptions options) { configurationBuilder.withQueryOptions(options); return this; } /** * Sets the threading options to use for the newly created Cluster. * *

If no options are set through this method, a new instance of {@link ThreadingOptions} will * be used. * * @param options the options. * @return this builder. */ public Builder withThreadingOptions(ThreadingOptions options) { configurationBuilder.withThreadingOptions(options); return this; } /** * Set the {@link NettyOptions} to use for the newly created Cluster. * *

If no Netty options are set through this method, {@link NettyOptions#DEFAULT_INSTANCE} * will be used as a default value, which means that no customization will be applied. * * @param nettyOptions the {@link NettyOptions} to use. * @return this builder. */ public Builder withNettyOptions(NettyOptions nettyOptions) { configurationBuilder.withNettyOptions(nettyOptions); return this; } /** * Enables the NO_COMPACT startup option. *

* When this option is supplied, SELECT, UPDATE, DELETE and * BATCH statements on COMPACT STORAGE tables function in "compatibility" mode which * allows seeing these tables as if they were "regular" CQL tables. *

* This option only effects interactions with tables using COMPACT STORAGE and is only supported by * C* 4.0+ and DSE 6.0+. * * @return this builder. * @see CASSANDRA-10857 */ public Builder withNoCompact() { this.noCompact = true; return this; } /** * The configuration that will be used for the new cluster. * *

You should not modify this object directly because changes made to the returned * object may not be used by the cluster build. Instead, you should use the other methods of * this {@code Builder}. * * @return the configuration to use for the new cluster. */ @Override public Configuration getConfiguration() { ProtocolOptions protocolOptions = new ProtocolOptions( port, protocolVersion, maxSchemaAgreementWaitSeconds, sslOptions, authProvider, noCompact) .setCompression(compression); MetricsOptions metricsOptions = new MetricsOptions(metricsEnabled, jmxEnabled); return configurationBuilder .withProtocolOptions(protocolOptions) .withMetricsOptions(metricsOptions) .withPolicies(policiesBuilder.build()) .build(); } @Override public Collection getInitialListeners() { return listeners == null ? Collections.emptySet() : listeners; } /** * Builds the cluster with the configured set of initial contact points and policies. * *

This is a convenience method for {@code Cluster.buildFrom(this)}. * * @return the newly built Cluster instance. */ public Cluster build() { return Cluster.buildFrom(this); } } static long timeSince(long startNanos, TimeUnit destUnit) { return destUnit.convert(System.nanoTime() - startNanos, TimeUnit.NANOSECONDS); } private static String generateClusterName() { return "cluster" + CLUSTER_ID.incrementAndGet(); } /** * The sessions and hosts managed by this a Cluster instance. * *

Note: the reason we create a Manager object separate from Cluster is that Manager is not * publicly visible. For instance, we wouldn't want user to be able to call the {@link #onUp} and * {@link #onDown} methods. */ class Manager implements Connection.DefaultResponseHandler { final String clusterName; private volatile boolean isInit; private volatile boolean isFullyInit; private Exception initException; // Initial contacts point final List contactPoints; final Set sessions = new CopyOnWriteArraySet(); Metadata metadata; final Configuration configuration; Metrics metrics; Connection.Factory connectionFactory; ControlConnection controlConnection; final ConvictionPolicy.Factory convictionPolicyFactory = new ConvictionPolicy.DefaultConvictionPolicy.Factory(); ListeningExecutorService executor; ListeningExecutorService blockingExecutor; ScheduledExecutorService reconnectionExecutor; ScheduledExecutorService scheduledTasksExecutor; BlockingQueue executorQueue; BlockingQueue blockingExecutorQueue; BlockingQueue reconnectionExecutorQueue; BlockingQueue scheduledTasksExecutorQueue; ConnectionReaper reaper; final AtomicReference closeFuture = new AtomicReference(); // All the queries that have been prepared (we keep them so we can re-prepared them when a node // fail or a // new one join the cluster). // Note: we could move this down to the session level, but since prepared statement are global // to a node, // this would yield a slightly less clear behavior. ConcurrentMap preparedQueries; final Set listeners; final Set latencyTrackers = new CopyOnWriteArraySet(); final Set schemaChangeListeners = new CopyOnWriteArraySet(); EventDebouncer nodeListRefreshRequestDebouncer; EventDebouncer nodeRefreshRequestDebouncer; EventDebouncer schemaRefreshRequestDebouncer; private Manager( String clusterName, List contactPoints, Configuration configuration, Collection listeners) { this.clusterName = clusterName == null ? generateClusterName() : clusterName; this.configuration = configuration; this.contactPoints = contactPoints; this.listeners = new CopyOnWriteArraySet(listeners); } // Initialization is not too performance intensive and in practice there shouldn't be contention // on it so synchronized is good enough. synchronized void init() { checkNotClosed(this); if (isInit) { return; } isInit = true; try { logger.debug("Starting new cluster with contact points " + contactPoints); this.configuration.register(this); ThreadingOptions threadingOptions = this.configuration.getThreadingOptions(); // executor ExecutorService tmpExecutor = threadingOptions.createExecutor(clusterName); this.executorQueue = (tmpExecutor instanceof ThreadPoolExecutor) ? ((ThreadPoolExecutor) tmpExecutor).getQueue() : null; this.executor = MoreExecutors.listeningDecorator(tmpExecutor); // blocking executor ExecutorService tmpBlockingExecutor = threadingOptions.createBlockingExecutor(clusterName); this.blockingExecutorQueue = (tmpBlockingExecutor instanceof ThreadPoolExecutor) ? ((ThreadPoolExecutor) tmpBlockingExecutor).getQueue() : null; this.blockingExecutor = MoreExecutors.listeningDecorator(tmpBlockingExecutor); // reconnection executor this.reconnectionExecutor = threadingOptions.createReconnectionExecutor(clusterName); this.reconnectionExecutorQueue = (reconnectionExecutor instanceof ThreadPoolExecutor) ? ((ThreadPoolExecutor) reconnectionExecutor).getQueue() : null; // scheduled tasks executor this.scheduledTasksExecutor = threadingOptions.createScheduledTasksExecutor(clusterName); this.scheduledTasksExecutorQueue = (scheduledTasksExecutor instanceof ThreadPoolExecutor) ? ((ThreadPoolExecutor) scheduledTasksExecutor).getQueue() : null; this.reaper = new ConnectionReaper(threadingOptions.createReaperExecutor(clusterName)); this.metadata = new Metadata(this); this.connectionFactory = new Connection.Factory(this, configuration); this.controlConnection = new ControlConnection(this); this.metrics = configuration.getMetricsOptions().isEnabled() ? new Metrics(this) : null; this.preparedQueries = new MapMaker().weakValues().makeMap(); // create debouncers - at this stage, they are not running yet final QueryOptions queryOptions = configuration.getQueryOptions(); this.nodeListRefreshRequestDebouncer = new EventDebouncer( "Node list refresh", scheduledTasksExecutor, new NodeListRefreshRequestDeliveryCallback()) { @Override int maxPendingEvents() { return configuration.getQueryOptions().getMaxPendingRefreshNodeListRequests(); } @Override long delayMs() { return configuration.getQueryOptions().getRefreshNodeListIntervalMillis(); } }; this.nodeRefreshRequestDebouncer = new EventDebouncer( "Node refresh", scheduledTasksExecutor, new NodeRefreshRequestDeliveryCallback()) { @Override int maxPendingEvents() { return configuration.getQueryOptions().getMaxPendingRefreshNodeRequests(); } @Override long delayMs() { return configuration.getQueryOptions().getRefreshNodeIntervalMillis(); } }; this.schemaRefreshRequestDebouncer = new EventDebouncer( "Schema refresh", scheduledTasksExecutor, new SchemaRefreshRequestDeliveryCallback()) { @Override int maxPendingEvents() { return configuration.getQueryOptions().getMaxPendingRefreshSchemaRequests(); } @Override long delayMs() { return configuration.getQueryOptions().getRefreshSchemaIntervalMillis(); } }; this.scheduledTasksExecutor.scheduleWithFixedDelay( new CleanupIdleConnectionsTask(), 10, 10, TimeUnit.SECONDS); for (InetSocketAddress address : contactPoints) { // We don't want to signal -- call onAdd() -- because nothing is ready // yet (loadbalancing policy, control connection, ...). All we want is // create the Host object so we can initialize the control connection. metadata.addIfAbsent(metadata.newHost(address)); } Collection allHosts = metadata.allHosts(); // At this stage, metadata.allHosts() only contains the contact points, that's what we want // to // pass to LBP.init(). // But the control connection will initialize first and discover more hosts, so make a copy. Set contactPointHosts = Sets.newHashSet(allHosts); negotiateProtocolVersionAndConnect(); // The control connection can mark hosts down if it failed to connect to them, or remove // them if they weren't found // in the control host's system.peers. Separate them: Set downContactPointHosts = Sets.newHashSet(); Set removedContactPointHosts = Sets.newHashSet(); for (Host host : contactPointHosts) { if (!allHosts.contains(host)) removedContactPointHosts.add(host); else if (host.state == Host.State.DOWN) downContactPointHosts.add(host); } contactPointHosts.removeAll(removedContactPointHosts); contactPointHosts.removeAll(downContactPointHosts); // Now that the control connection is ready, we have all the information we need about the // nodes (datacenter, // rack...) to initialize the load balancing policy loadBalancingPolicy().init(Cluster.this, contactPointHosts); speculativeExecutionPolicy().init(Cluster.this); configuration.getPolicies().getRetryPolicy().init(Cluster.this); reconnectionPolicy().init(Cluster.this); configuration.getPolicies().getAddressTranslator().init(Cluster.this); for (LatencyTracker tracker : latencyTrackers) tracker.onRegister(Cluster.this); for (Host.StateListener listener : listeners) listener.onRegister(Cluster.this); for (Host host : removedContactPointHosts) { loadBalancingPolicy().onRemove(host); for (Host.StateListener listener : listeners) listener.onRemove(host); } for (Host host : downContactPointHosts) { loadBalancingPolicy().onDown(host); for (Host.StateListener listener : listeners) listener.onDown(host); startPeriodicReconnectionAttempt(host, true); } configuration.getPoolingOptions().setProtocolVersion(protocolVersion()); for (Host host : allHosts) { // If the host is down at this stage, it's a contact point that the control connection // failed to reach. // Reconnection attempts are already scheduled, and the LBP and listeners have been // notified above. if (host.state == Host.State.DOWN) continue; // Otherwise, we want to do the equivalent of onAdd(). But since we know for sure that no // sessions or prepared // statements exist at this point, we can skip some of the steps (plus this avoids // scheduling concurrent pool // creations if a session is created right after this method returns). logger.info("New Cassandra host {} added", host); if (!host.supports(connectionFactory.protocolVersion)) { logUnsupportedVersionProtocol(host, connectionFactory.protocolVersion); continue; } if (!contactPointHosts.contains(host)) loadBalancingPolicy().onAdd(host); host.setUp(); for (Host.StateListener listener : listeners) listener.onAdd(host); } // start debouncers this.nodeListRefreshRequestDebouncer.start(); this.schemaRefreshRequestDebouncer.start(); this.nodeRefreshRequestDebouncer.start(); isFullyInit = true; } catch (RuntimeException e) { initException = e; close(); throw e; } } private void negotiateProtocolVersionAndConnect() { boolean shouldNegotiate = (configuration.getProtocolOptions().initialProtocolVersion == null); while (true) { try { controlConnection.connect(); return; } catch (UnsupportedProtocolVersionException e) { if (!shouldNegotiate) { throw e; } // Do not trust version of server's response, as C* behavior in case of protocol // negotiation is not // properly documented, and varies over time (specially after CASSANDRA-11464). Instead, // always // retry at attempted version - 1, if such a version exists; and otherwise, stop and fail. ProtocolVersion attemptedVersion = e.getUnsupportedVersion(); ProtocolVersion retryVersion = attemptedVersion.getLowerSupported(); if (retryVersion == null) { throw e; } logger.info( "Cannot connect with protocol version {}, trying with {}", attemptedVersion, retryVersion); connectionFactory.protocolVersion = retryVersion; } } } ProtocolVersion protocolVersion() { return connectionFactory.protocolVersion; } Cluster getCluster() { return Cluster.this; } LoadBalancingPolicy loadBalancingPolicy() { return configuration.getPolicies().getLoadBalancingPolicy(); } SpeculativeExecutionPolicy speculativeExecutionPolicy() { return configuration.getPolicies().getSpeculativeExecutionPolicy(); } ReconnectionPolicy reconnectionPolicy() { return configuration.getPolicies().getReconnectionPolicy(); } InetSocketAddress translateAddress(InetSocketAddress address) { InetSocketAddress translated = configuration.getPolicies().getAddressTranslator().translate(address); return translated == null ? address : translated; } InetSocketAddress translateAddress(InetAddress address) { InetSocketAddress sa = new InetSocketAddress(address, connectionFactory.getPort()); return translateAddress(sa); } private Session newSession() { SessionManager session = new SessionManager(Cluster.this); sessions.add(session); return session; } boolean removeSession(Session session) { return sessions.remove(session); } void reportQuery(Host host, Statement statement, Exception exception, long latencyNanos) { for (LatencyTracker tracker : latencyTrackers) { tracker.update(host, statement, exception, latencyNanos); } } boolean isClosed() { return closeFuture.get() != null; } boolean errorDuringInit() { return (isInit && initException != null); } Exception getInitException() { return initException; } private CloseFuture close() { CloseFuture future = closeFuture.get(); if (future != null) return future; if (isInit) { logger.debug("Shutting down"); // stop debouncers if (nodeListRefreshRequestDebouncer != null) { nodeListRefreshRequestDebouncer.stop(); } if (nodeRefreshRequestDebouncer != null) { nodeRefreshRequestDebouncer.stop(); } if (schemaRefreshRequestDebouncer != null) { schemaRefreshRequestDebouncer.stop(); } // If we're shutting down, there is no point in waiting on scheduled reconnections, nor on // notifications // delivery or blocking tasks so we use shutdownNow shutdownNow(reconnectionExecutor); shutdownNow(scheduledTasksExecutor); shutdownNow(blockingExecutor); // but for the worker executor, we want to let submitted tasks finish unless the shutdown is // forced. if (executor != null) { executor.shutdown(); } // We also close the metrics if (metrics != null) metrics.shutdown(); loadBalancingPolicy().close(); speculativeExecutionPolicy().close(); configuration.getPolicies().getRetryPolicy().close(); reconnectionPolicy().close(); configuration.getPolicies().getAddressTranslator().close(); for (LatencyTracker tracker : latencyTrackers) tracker.onUnregister(Cluster.this); for (Host.StateListener listener : listeners) listener.onUnregister(Cluster.this); for (SchemaChangeListener listener : schemaChangeListeners) listener.onUnregister(Cluster.this); // Then we shutdown all connections List futures = new ArrayList(sessions.size() + 1); if (controlConnection != null) { futures.add(controlConnection.closeAsync()); } for (Session session : sessions) futures.add(session.closeAsync()); future = new ClusterCloseFuture(futures); // The rest will happen asynchronously, when all connections are successfully closed } else { future = CloseFuture.immediateFuture(); } return closeFuture.compareAndSet(null, future) ? future : closeFuture.get(); // We raced, it's ok, return the future that was actually set } private void shutdownNow(ExecutorService executor) { if (executor != null) { List pendingTasks = executor.shutdownNow(); // If some tasks were submitted to this executor but not yet commenced, make sure the // corresponding futures complete for (Runnable pendingTask : pendingTasks) { if (pendingTask instanceof FutureTask) ((FutureTask) pendingTask).cancel(false); } } } void logUnsupportedVersionProtocol(Host host, ProtocolVersion version) { logger.warn( "Detected added or restarted Cassandra host {} but ignoring it since it does not support the version {} of the native " + "protocol which is currently in use. If you want to force the use of a particular version of the native protocol, use " + "Cluster.Builder#usingProtocolVersion() when creating the Cluster instance.", host, version); } void logClusterNameMismatch(Host host, String expectedClusterName, String actualClusterName) { logger.warn( "Detected added or restarted Cassandra host {} but ignoring it since its cluster name '{}' does not match the one " + "currently known ({})", host, actualClusterName, expectedClusterName); } public ListenableFuture triggerOnUp(final Host host) { if (!isClosed()) { return executor.submit( new ExceptionCatchingRunnable() { @Override public void runMayThrow() throws InterruptedException, ExecutionException { onUp(host, null); } }); } else { return MoreFutures.VOID_SUCCESS; } } // Use triggerOnUp unless you're sure you want to run this on the current thread. private void onUp(final Host host, Connection reusedConnection) throws InterruptedException, ExecutionException { if (isClosed()) return; if (!host.supports(connectionFactory.protocolVersion)) { logUnsupportedVersionProtocol(host, connectionFactory.protocolVersion); return; } try { boolean locked = host.notificationsLock.tryLock(NOTIF_LOCK_TIMEOUT_SECONDS, TimeUnit.SECONDS); if (!locked) { logger.warn( "Could not acquire notifications lock within {} seconds, ignoring UP notification for {}", NOTIF_LOCK_TIMEOUT_SECONDS, host); return; } try { // We don't want to use the public Host.isUp() as this would make us skip the rest for // suspected hosts if (host.state == Host.State.UP) return; Host.statesLogger.debug("[{}] marking host UP", host); // If there is a reconnection attempt scheduled for that node, cancel it Future scheduledAttempt = host.reconnectionAttempt.getAndSet(null); if (scheduledAttempt != null) { logger.debug("Cancelling reconnection attempt since node is UP"); scheduledAttempt.cancel(false); } try { if (getCluster().getConfiguration().getQueryOptions().isReprepareOnUp()) reusedConnection = prepareAllQueries(host, reusedConnection); } catch (InterruptedException e) { Thread.currentThread().interrupt(); // Don't propagate because we don't want to prevent other listener to run } catch (UnsupportedProtocolVersionException e) { logUnsupportedVersionProtocol(host, e.getUnsupportedVersion()); return; } catch (ClusterNameMismatchException e) { logClusterNameMismatch(host, e.expectedClusterName, e.actualClusterName); return; } // Session#onUp() expects the load balancing policy to have been updated first, so that // Host distances are up to date. This mean the policy could return the node before the // new pool have been created. This is harmless if there is no prior pool since // RequestHandler // will ignore the node, but we do want to make sure there is no prior pool so we don't // query from a pool we will shutdown right away. for (SessionManager s : sessions) s.removePool(host); loadBalancingPolicy().onUp(host); controlConnection.onUp(host); logger.trace("Adding/renewing host pools for newly UP host {}", host); List> futures = Lists.newArrayListWithCapacity(sessions.size()); for (SessionManager s : sessions) futures.add(s.forceRenewPool(host, reusedConnection)); try { // Only mark the node up once all session have re-added their pool (if the // load-balancing // policy says it should), so that Host.isUp() don't return true before we're // reconnected // to the node. List poolCreationResults = Futures.allAsList(futures).get(); // If any of the creation failed, they will have signaled a connection failure // which will trigger a reconnection to the node. So don't bother marking UP. if (Iterables.any(poolCreationResults, Predicates.equalTo(false))) { logger.debug("Connection pool cannot be created, not marking {} UP", host); return; } host.setUp(); for (Host.StateListener listener : listeners) listener.onUp(host); } catch (ExecutionException e) { Throwable t = e.getCause(); // That future is not really supposed to throw unexpected exceptions if (!(t instanceof InterruptedException) && !(t instanceof CancellationException)) logger.error( "Unexpected error while marking node UP: while this shouldn't happen, this shouldn't be critical", t); } // Now, check if there isn't pools to create/remove following the addition. // We do that now only so that it's not called before we've set the node up. for (SessionManager s : sessions) s.updateCreatedPools().get(); } finally { host.notificationsLock.unlock(); } } finally { if (reusedConnection != null && !reusedConnection.hasOwner()) reusedConnection.closeAsync(); } } public ListenableFuture triggerOnDown(final Host host, boolean startReconnection) { return triggerOnDown(host, false, startReconnection); } public ListenableFuture triggerOnDown( final Host host, final boolean isHostAddition, final boolean startReconnection) { if (!isClosed()) { return executor.submit( new ExceptionCatchingRunnable() { @Override public void runMayThrow() throws InterruptedException, ExecutionException { onDown(host, isHostAddition, startReconnection); } }); } else { return MoreFutures.VOID_SUCCESS; } } // Use triggerOnDown unless you're sure you want to run this on the current thread. private void onDown(final Host host, final boolean isHostAddition, boolean startReconnection) throws InterruptedException, ExecutionException { if (isClosed()) return; boolean locked = host.notificationsLock.tryLock(NOTIF_LOCK_TIMEOUT_SECONDS, TimeUnit.SECONDS); if (!locked) { logger.warn( "Could not acquire notifications lock within {} seconds, ignoring DOWN notification for {}", NOTIF_LOCK_TIMEOUT_SECONDS, host); return; } try { // Note: we don't want to skip that method if !host.isUp() because we set isUp // late in onUp, and so we can rely on isUp if there is an error during onUp. // But if there is a reconnection attempt in progress already, then we know // we've already gone through that method since the last successful onUp(), so // we're good skipping it. if (host.reconnectionAttempt.get() != null) { logger.debug("Aborting onDown because a reconnection is running on DOWN host {}", host); return; } Host.statesLogger.debug("[{}] marking host DOWN", host); // Remember if we care about this node at all. We must call this before // we've signalled the load balancing policy, since most policy will always // IGNORE down nodes anyway. HostDistance distance = loadBalancingPolicy().distance(host); boolean wasUp = host.isUp(); host.setDown(); loadBalancingPolicy().onDown(host); controlConnection.onDown(host); for (SessionManager s : sessions) s.onDown(host); // Contrarily to other actions of that method, there is no reason to notify listeners // unless the host was UP at the beginning of this function since even if a onUp fail // mid-method, listeners won't have been notified of the UP. if (wasUp) { for (Host.StateListener listener : listeners) listener.onDown(host); } // Don't start a reconnection if we ignore the node anyway (JAVA-314) if (distance == HostDistance.IGNORED || !startReconnection) return; startPeriodicReconnectionAttempt(host, isHostAddition); } finally { host.notificationsLock.unlock(); } } void startPeriodicReconnectionAttempt(final Host host, final boolean isHostAddition) { new AbstractReconnectionHandler( host.toString(), reconnectionExecutor, reconnectionPolicy().newSchedule(), host.reconnectionAttempt) { @Override protected Connection tryReconnect() throws ConnectionException, InterruptedException, UnsupportedProtocolVersionException, ClusterNameMismatchException { return connectionFactory.open(host); } @Override protected void onReconnection(Connection connection) { // Make sure we have up-to-date infos on that host before adding it (so we typically // catch that an upgraded node uses a new cassandra version). if (controlConnection.refreshNodeInfo(host)) { logger.debug("Successful reconnection to {}, setting host UP", host); try { if (isHostAddition) { onAdd(host, connection); submitNodeListRefresh(); } else onUp(host, connection); } catch (InterruptedException e) { Thread.currentThread().interrupt(); } catch (Exception e) { logger.error("Unexpected error while setting node up", e); } } else { logger.debug("Not enough info for {}, ignoring host", host); connection.closeAsync(); } } @Override protected boolean onConnectionException(ConnectionException e, long nextDelayMs) { if (logger.isDebugEnabled()) logger.debug( "Failed reconnection to {} ({}), scheduling retry in {} milliseconds", host, e.getMessage(), nextDelayMs); return true; } @Override protected boolean onUnknownException(Exception e, long nextDelayMs) { logger.error( String.format( "Unknown error during reconnection to %s, scheduling retry in %d milliseconds", host, nextDelayMs), e); return true; } @Override protected boolean onAuthenticationException(AuthenticationException e, long nextDelayMs) { logger.error( String.format( "Authentication error during reconnection to %s, scheduling retry in %d milliseconds", host, nextDelayMs), e); return true; } }.start(); } void startSingleReconnectionAttempt(final Host host) { if (isClosed() || host.isUp()) return; logger.debug("Scheduling one-time reconnection to {}", host); // Setting an initial delay of 0 to start immediately, and all the exception handlers return // false to prevent further attempts new AbstractReconnectionHandler( host.toString(), reconnectionExecutor, reconnectionPolicy().newSchedule(), host.reconnectionAttempt, 0) { @Override protected Connection tryReconnect() throws ConnectionException, InterruptedException, UnsupportedProtocolVersionException, ClusterNameMismatchException { return connectionFactory.open(host); } @Override protected void onReconnection(Connection connection) { // Make sure we have up-to-date infos on that host before adding it (so we typically // catch that an upgraded node uses a new cassandra version). if (controlConnection.refreshNodeInfo(host)) { logger.debug("Successful reconnection to {}, setting host UP", host); try { onUp(host, connection); } catch (InterruptedException e) { Thread.currentThread().interrupt(); } catch (Exception e) { logger.error("Unexpected error while setting node up", e); } } else { logger.debug("Not enough info for {}, ignoring host", host); connection.closeAsync(); } } @Override protected boolean onConnectionException(ConnectionException e, long nextDelayMs) { if (logger.isDebugEnabled()) logger.debug("Failed one-time reconnection to {} ({})", host, e.getMessage()); return false; } @Override protected boolean onUnknownException(Exception e, long nextDelayMs) { logger.error(String.format("Unknown error during one-time reconnection to %s", host), e); return false; } @Override protected boolean onAuthenticationException(AuthenticationException e, long nextDelayMs) { logger.error( String.format("Authentication error during one-time reconnection to %s", host), e); return false; } }.start(); } public ListenableFuture triggerOnAdd(final Host host) { if (!isClosed()) { return executor.submit( new ExceptionCatchingRunnable() { @Override public void runMayThrow() throws InterruptedException, ExecutionException { onAdd(host, null); } }); } else { return MoreFutures.VOID_SUCCESS; } } // Use triggerOnAdd unless you're sure you want to run this on the current thread. private void onAdd(final Host host, Connection reusedConnection) throws InterruptedException, ExecutionException { if (isClosed()) return; if (!host.supports(connectionFactory.protocolVersion)) { logUnsupportedVersionProtocol(host, connectionFactory.protocolVersion); return; } try { boolean locked = host.notificationsLock.tryLock(NOTIF_LOCK_TIMEOUT_SECONDS, TimeUnit.SECONDS); if (!locked) { logger.warn( "Could not acquire notifications lock within {} seconds, ignoring ADD notification for {}", NOTIF_LOCK_TIMEOUT_SECONDS, host); return; } try { Host.statesLogger.debug("[{}] adding host", host); // Adds to the load balancing first and foremost, as doing so might change the decision // it will make for distance() on that node (not likely but we leave that possibility). // This does mean the policy may start returning that node for query plan, but as long // as no pools have been created (below) this will be ignored by RequestHandler so it's // fine. loadBalancingPolicy().onAdd(host); // Next, if the host should be ignored, well, ignore it. if (loadBalancingPolicy().distance(host) == HostDistance.IGNORED) { // We still mark the node UP though as it should be (and notifiy the listeners). // We'll mark it down if we have a notification anyway and we've documented that // especially // for IGNORED hosts, the isUp() method was a best effort guess host.setUp(); for (Host.StateListener listener : listeners) listener.onAdd(host); return; } try { reusedConnection = prepareAllQueries(host, reusedConnection); } catch (InterruptedException e) { Thread.currentThread().interrupt(); // Don't propagate because we don't want to prevent other listener to run } catch (UnsupportedProtocolVersionException e) { logUnsupportedVersionProtocol(host, e.getUnsupportedVersion()); return; } catch (ClusterNameMismatchException e) { logClusterNameMismatch(host, e.expectedClusterName, e.actualClusterName); return; } controlConnection.onAdd(host); List> futures = Lists.newArrayListWithCapacity(sessions.size()); for (SessionManager s : sessions) futures.add(s.maybeAddPool(host, reusedConnection)); try { // Only mark the node up once all session have added their pool (if the load-balancing // policy says it should), so that Host.isUp() don't return true before we're // reconnected // to the node. List poolCreationResults = Futures.allAsList(futures).get(); // If any of the creation failed, they will have signaled a connection failure // which will trigger a reconnection to the node. So don't bother marking UP. if (Iterables.any(poolCreationResults, Predicates.equalTo(false))) { logger.debug("Connection pool cannot be created, not marking {} UP", host); return; } host.setUp(); for (Host.StateListener listener : listeners) listener.onAdd(host); } catch (ExecutionException e) { Throwable t = e.getCause(); // That future is not really supposed to throw unexpected exceptions if (!(t instanceof InterruptedException) && !(t instanceof CancellationException)) logger.error( "Unexpected error while adding node: while this shouldn't happen, this shouldn't be critical", t); } // Now, check if there isn't pools to create/remove following the addition. // We do that now only so that it's not called before we've set the node up. for (SessionManager s : sessions) s.updateCreatedPools().get(); } finally { host.notificationsLock.unlock(); } } finally { if (reusedConnection != null && !reusedConnection.hasOwner()) reusedConnection.closeAsync(); } } public ListenableFuture triggerOnRemove(final Host host) { if (!isClosed()) { return executor.submit( new ExceptionCatchingRunnable() { @Override public void runMayThrow() throws InterruptedException, ExecutionException { onRemove(host); } }); } else { return MoreFutures.VOID_SUCCESS; } } // Use triggerOnRemove unless you're sure you want to run this on the current thread. private void onRemove(Host host) throws InterruptedException, ExecutionException { if (isClosed()) return; boolean locked = host.notificationsLock.tryLock(NOTIF_LOCK_TIMEOUT_SECONDS, TimeUnit.SECONDS); if (!locked) { logger.warn( "Could not acquire notifications lock within {} seconds, ignoring REMOVE notification for {}", NOTIF_LOCK_TIMEOUT_SECONDS, host); return; } try { host.setDown(); Host.statesLogger.debug("[{}] removing host", host); loadBalancingPolicy().onRemove(host); controlConnection.onRemove(host); for (SessionManager s : sessions) s.onRemove(host); for (Host.StateListener listener : listeners) listener.onRemove(host); } finally { host.notificationsLock.unlock(); } } public void signalHostDown(Host host, boolean isHostAddition) { // Don't mark the node down until we've fully initialized the controlConnection as this might // mess up with // the protocol detection if (!isFullyInit || isClosed()) return; triggerOnDown(host, isHostAddition, true); } public void removeHost(Host host, boolean isInitialConnection) { if (host == null) return; if (metadata.remove(host)) { if (isInitialConnection) { logger.warn( "You listed {} in your contact points, but it wasn't found in the control host's system.peers at startup", host); } else { logger.info("Cassandra host {} removed", host); triggerOnRemove(host); } } } public void ensurePoolsSizing() { if (protocolVersion().compareTo(ProtocolVersion.V3) >= 0) return; for (SessionManager session : sessions) { for (HostConnectionPool pool : session.pools.values()) pool.ensureCoreConnections(); } } public PreparedStatement addPrepared(PreparedStatement stmt) { PreparedStatement previous = preparedQueries.putIfAbsent(stmt.getPreparedId().boundValuesMetadata.id, stmt); if (previous != null) { logger.warn( "Re-preparing already prepared query is generally an anti-pattern and will likely affect performance. " + "Consider preparing the statement only once. Query='{}'", stmt.getQueryString()); // The one object in the cache will get GCed once it's not referenced by the client anymore // since we use a weak reference. // So we need to make sure that the instance we do return to the user is the one that is in // the cache. // However if the result metadata changed since the last PREPARE call, this also needs to be // updated. previous.getPreparedId().resultSetMetadata = stmt.getPreparedId().resultSetMetadata; return previous; } return stmt; } /** * @param reusedConnection an existing connection (from a reconnection attempt) that we want to * reuse to prepare the statements (might be null). * @return a connection that the rest of the initialization process can use (it will be made * part of a connection pool). Can be reusedConnection, or one that was open in the method. */ private Connection prepareAllQueries(Host host, Connection reusedConnection) throws InterruptedException, UnsupportedProtocolVersionException, ClusterNameMismatchException { if (preparedQueries.isEmpty()) return reusedConnection; logger.debug( "Preparing {} prepared queries on newly up node {}", preparedQueries.size(), host); Connection connection = null; try { connection = (reusedConnection == null) ? connectionFactory.open(host) : reusedConnection; // Furthermore, along with each prepared query we keep the current keyspace at the time of // preparation // as we need to make it is the same when we re-prepare on new/restarted nodes. Most query // will use the // same keyspace so keeping it each time is slightly wasteful, but this doesn't really // matter and is // simpler. Besides, we do avoid in prepareAllQueries to not set the current keyspace more // than needed. // We need to make sure we prepared every query with the right current keyspace, i.e. the // one originally // used for preparing it. However, since we are likely that all prepared query belong to // only a handful // of different keyspace (possibly only one), and to avoid setting the current keyspace more // than needed, // we first sort the query per keyspace. SetMultimap perKeyspace = HashMultimap.create(); for (PreparedStatement ps : preparedQueries.values()) { // It's possible for a query to not have a current keyspace. But since null doesn't work // well as // map keys, we use the empty string instead (that is not a valid keyspace name). String keyspace = ps.getQueryKeyspace() == null ? "" : ps.getQueryKeyspace(); perKeyspace.put(keyspace, ps.getQueryString()); } for (String keyspace : perKeyspace.keySet()) { // Empty string mean no particular keyspace to set if (!keyspace.isEmpty()) connection.setKeyspace(keyspace); List futures = new ArrayList(preparedQueries.size()); for (String query : perKeyspace.get(keyspace)) { futures.add(connection.write(new Requests.Prepare(query))); } for (Connection.Future future : futures) { try { future.get(); } catch (ExecutionException e) { // This "might" happen if we drop a CF but haven't removed it's prepared queries // (which we don't do // currently). It's not a big deal however as if it's a more serious problem it'll // show up later when // the query is tried for execution. logger.debug("Unexpected error while preparing queries on new/newly up host", e); } } } return connection; } catch (ConnectionException e) { // Ignore, not a big deal if (connection != null) connection.closeAsync(); return null; } catch (AuthenticationException e) { // That's a bad news, but ignore at this point if (connection != null) connection.closeAsync(); return null; } catch (BusyConnectionException e) { // Ignore, not a big deal // In theory the problem is transient so the connection could be reused later, but if the // core pool size is 1 // it's better to close this one so that we start with a fresh connection. if (connection != null) connection.closeAsync(); return null; } } ListenableFuture submitSchemaRefresh( final SchemaElement targetType, final String targetKeyspace, final String targetName, final List targetSignature) { SchemaRefreshRequest request = new SchemaRefreshRequest(targetType, targetKeyspace, targetName, targetSignature); logger.trace("Submitting schema refresh: {}", request); return schemaRefreshRequestDebouncer.eventReceived(request); } ListenableFuture submitNodeListRefresh() { logger.trace("Submitting node list and token map refresh"); return nodeListRefreshRequestDebouncer.eventReceived(new NodeListRefreshRequest()); } ListenableFuture submitNodeRefresh(InetSocketAddress address, HostEvent eventType) { NodeRefreshRequest request = new NodeRefreshRequest(address, eventType); logger.trace("Submitting node refresh: {}", request); return nodeRefreshRequestDebouncer.eventReceived(request); } // refresh the schema using the provided connection, and notice the future with the provided // resultset once done public void refreshSchemaAndSignal( final Connection connection, final DefaultResultSetFuture future, final ResultSet rs, final SchemaElement targetType, final String targetKeyspace, final String targetName, final List targetSignature) { if (logger.isDebugEnabled()) logger.debug( "Refreshing schema for {}{}", targetType == null ? "everything" : targetKeyspace, (targetType == KEYSPACE) ? "" : "." + targetName + " (" + targetType + ")"); maybeRefreshSchemaAndSignal( connection, future, rs, targetType, targetKeyspace, targetName, targetSignature); } public void waitForSchemaAgreementAndSignal( final Connection connection, final DefaultResultSetFuture future, final ResultSet rs) { maybeRefreshSchemaAndSignal(connection, future, rs, null, null, null, null); } private void maybeRefreshSchemaAndSignal( final Connection connection, final DefaultResultSetFuture future, final ResultSet rs, final SchemaElement targetType, final String targetKeyspace, final String targetName, final List targetSignature) { final boolean refreshSchema = (targetKeyspace != null); // if false, only wait for schema agreement executor.submit( new Runnable() { @Override public void run() { boolean schemaInAgreement = false; try { // Before refreshing the schema, wait for schema agreement so // that querying a table just after having created it don't fail. schemaInAgreement = ControlConnection.waitForSchemaAgreement(connection, Cluster.Manager.this); if (!schemaInAgreement) logger.warn( "No schema agreement from live replicas after {} s. The schema may not be up to date on some nodes.", configuration.getProtocolOptions().getMaxSchemaAgreementWaitSeconds()); ListenableFuture schemaReady; if (refreshSchema) { schemaReady = submitSchemaRefresh(targetType, targetKeyspace, targetName, targetSignature); // JAVA-1120: skip debouncing delay and force immediate delivery if (!schemaReady.isDone()) schemaRefreshRequestDebouncer.scheduleImmediateDelivery(); } else { schemaReady = MoreFutures.VOID_SUCCESS; } final boolean finalSchemaInAgreement = schemaInAgreement; schemaReady.addListener( new Runnable() { @Override public void run() { rs.getExecutionInfo().setSchemaInAgreement(finalSchemaInAgreement); future.setResult(rs); } }, GuavaCompatibility.INSTANCE.sameThreadExecutor()); } catch (Exception e) { logger.warn("Error while waiting for schema agreement", e); // This is not fatal, complete the future anyway rs.getExecutionInfo().setSchemaInAgreement(schemaInAgreement); future.setResult(rs); } } }); } // Called when some message has been received but has been initiated from the server (streamId < // 0). // This is called on an I/O thread, so all blocking operation must be done on an executor. @Override public void handle(Message.Response response) { if (!(response instanceof Responses.Event)) { logger.error("Received an unexpected message from the server: {}", response); return; } final ProtocolEvent event = ((Responses.Event) response).event; logger.debug("Received event {}, scheduling delivery", response); switch (event.type) { case TOPOLOGY_CHANGE: ProtocolEvent.TopologyChange tpc = (ProtocolEvent.TopologyChange) event; InetSocketAddress tpAddr = translateAddress(tpc.node.getAddress()); Host.statesLogger.debug("[{}] received event {}", tpAddr, tpc.change); switch (tpc.change) { case NEW_NODE: submitNodeRefresh(tpAddr, HostEvent.ADDED); break; case REMOVED_NODE: submitNodeRefresh(tpAddr, HostEvent.REMOVED); break; case MOVED_NODE: submitNodeListRefresh(); break; } break; case STATUS_CHANGE: ProtocolEvent.StatusChange stc = (ProtocolEvent.StatusChange) event; InetSocketAddress stAddr = translateAddress(stc.node.getAddress()); Host.statesLogger.debug("[{}] received event {}", stAddr, stc.status); switch (stc.status) { case UP: submitNodeRefresh(stAddr, HostEvent.UP); break; case DOWN: submitNodeRefresh(stAddr, HostEvent.DOWN); break; } break; case SCHEMA_CHANGE: if (!configuration.getQueryOptions().isMetadataEnabled()) return; ProtocolEvent.SchemaChange scc = (ProtocolEvent.SchemaChange) event; switch (scc.change) { case CREATED: case UPDATED: submitSchemaRefresh( scc.targetType, scc.targetKeyspace, scc.targetName, scc.targetSignature); break; case DROPPED: if (scc.targetType == KEYSPACE) { final KeyspaceMetadata removedKeyspace = manager.metadata.removeKeyspace(scc.targetKeyspace); if (removedKeyspace != null) { executor.submit( new Runnable() { @Override public void run() { manager.metadata.triggerOnKeyspaceRemoved(removedKeyspace); } }); } } else { KeyspaceMetadata keyspace = manager.metadata.keyspaces.get(scc.targetKeyspace); if (keyspace == null) { logger.warn( "Received a DROPPED notification for {} {}.{}, but this keyspace is unknown in our metadata", scc.targetType, scc.targetKeyspace, scc.targetName); } else { switch (scc.targetType) { case TABLE: // we can't tell whether it's a table or a view, // but since two objects cannot have the same name, // try removing both final TableMetadata removedTable = keyspace.removeTable(scc.targetName); if (removedTable != null) { executor.submit( new Runnable() { @Override public void run() { manager.metadata.triggerOnTableRemoved(removedTable); } }); } else { final MaterializedViewMetadata removedView = keyspace.removeMaterializedView(scc.targetName); if (removedView != null) { executor.submit( new Runnable() { @Override public void run() { manager.metadata.triggerOnMaterializedViewRemoved(removedView); } }); } } break; case TYPE: final UserType removedType = keyspace.removeUserType(scc.targetName); if (removedType != null) { executor.submit( new Runnable() { @Override public void run() { manager.metadata.triggerOnUserTypeRemoved(removedType); } }); } break; case FUNCTION: final FunctionMetadata removedFunction = keyspace.removeFunction( Metadata.fullFunctionName(scc.targetName, scc.targetSignature)); if (removedFunction != null) { executor.submit( new Runnable() { @Override public void run() { manager.metadata.triggerOnFunctionRemoved(removedFunction); } }); } break; case AGGREGATE: final AggregateMetadata removedAggregate = keyspace.removeAggregate( Metadata.fullFunctionName(scc.targetName, scc.targetSignature)); if (removedAggregate != null) { executor.submit( new Runnable() { @Override public void run() { manager.metadata.triggerOnAggregateRemoved(removedAggregate); } }); } break; } } } break; } break; } } void refreshConnectedHosts() { // Deal first with the control connection: if it's connected to a node that is not LOCAL, try // reconnecting (thus letting the loadBalancingPolicy pick a better node) Host ccHost = controlConnection.connectedHost(); if (ccHost == null || loadBalancingPolicy().distance(ccHost) != HostDistance.LOCAL) controlConnection.triggerReconnect(); try { for (SessionManager s : sessions) Uninterruptibles.getUninterruptibly(s.updateCreatedPools()); } catch (ExecutionException e) { throw DriverThrowables.propagateCause(e); } } void refreshConnectedHost(Host host) { // Deal with the control connection if it was using this host Host ccHost = controlConnection.connectedHost(); if (ccHost == null || ccHost.equals(host) && loadBalancingPolicy().distance(ccHost) != HostDistance.LOCAL) controlConnection.triggerReconnect(); for (SessionManager s : sessions) s.updateCreatedPools(host); } private class ClusterCloseFuture extends CloseFuture.Forwarding { ClusterCloseFuture(List futures) { super(futures); } @Override public CloseFuture force() { // The only ExecutorService we haven't forced yet is executor shutdownNow(executor); return super.force(); } @Override protected void onFuturesDone() { /* * When we reach this, all sessions should be shutdown. We've also started a shutdown * of the thread pools used by this object. Remains 2 things before marking the shutdown * as done: * 1) we need to wait for the completion of the shutdown of the Cluster threads pools. * 2) we need to shutdown the Connection.Factory, i.e. the executors used by Netty. * But at least for 2), we must not do it on the current thread because that could be * a netty worker, which we're going to shutdown. So creates some thread for that. */ (new Thread("Shutdown-checker") { @Override public void run() { // Just wait indefinitely on the the completion of the thread pools. Provided the // user // call force(), we'll never really block forever. try { if (reconnectionExecutor != null) { reconnectionExecutor.awaitTermination(Long.MAX_VALUE, TimeUnit.SECONDS); } if (scheduledTasksExecutor != null) { scheduledTasksExecutor.awaitTermination(Long.MAX_VALUE, TimeUnit.SECONDS); } if (executor != null) { executor.awaitTermination(Long.MAX_VALUE, TimeUnit.SECONDS); } if (blockingExecutor != null) { blockingExecutor.awaitTermination(Long.MAX_VALUE, TimeUnit.SECONDS); } // Some of the jobs on the executors can be doing query stuff, so close the // connectionFactory at the very last if (connectionFactory != null) { connectionFactory.shutdown(); } if (reaper != null) { reaper.shutdown(); } set(null); } catch (InterruptedException e) { Thread.currentThread().interrupt(); setException(e); } } }) .start(); } } private class CleanupIdleConnectionsTask implements Runnable { @Override public void run() { try { long now = System.currentTimeMillis(); for (SessionManager session : sessions) { session.cleanupIdleConnections(now); } } catch (Exception e) { logger.warn("Error while trashing idle connections", e); } } } private class SchemaRefreshRequest { private final SchemaElement targetType; private final String targetKeyspace; private final String targetName; private final List targetSignature; public SchemaRefreshRequest( SchemaElement targetType, String targetKeyspace, String targetName, List targetSignature) { this.targetType = targetType; this.targetKeyspace = Strings.emptyToNull(targetKeyspace); this.targetName = Strings.emptyToNull(targetName); this.targetSignature = targetSignature; } /** * Coalesce schema refresh requests. The algorithm is simple: if more than 2 keyspaces need * refresh, then refresh the entire schema; otherwise if more than 2 elements in the same * keyspace need refresh, then refresh the entire keyspace. * * @param that the other request to merge with the current one. * @return A coalesced request */ SchemaRefreshRequest coalesce(SchemaRefreshRequest that) { if (this.targetType == null || that.targetType == null) return new SchemaRefreshRequest(null, null, null, null); if (!this.targetKeyspace.equals(that.targetKeyspace)) return new SchemaRefreshRequest(null, null, null, null); if (this.targetName == null || that.targetName == null) return new SchemaRefreshRequest(KEYSPACE, targetKeyspace, null, null); if (!this.targetName.equals(that.targetName)) return new SchemaRefreshRequest(KEYSPACE, targetKeyspace, null, null); return this; } @Override public String toString() { if (this.targetType == null) return "Refresh ALL"; if (this.targetName == null) return "Refresh keyspace " + targetKeyspace; return String.format("Refresh %s %s.%s", targetType, targetKeyspace, targetName); } } private class SchemaRefreshRequestDeliveryCallback implements EventDebouncer.DeliveryCallback { @Override public ListenableFuture deliver(final List events) { return executor.submit( new ExceptionCatchingRunnable() { @Override public void runMayThrow() throws InterruptedException, ExecutionException { SchemaRefreshRequest coalesced = null; for (SchemaRefreshRequest request : events) { coalesced = coalesced == null ? request : coalesced.coalesce(request); } assert coalesced != null; logger.trace("Coalesced schema refresh request: {}", coalesced); controlConnection.refreshSchema( coalesced.targetType, coalesced.targetKeyspace, coalesced.targetName, coalesced.targetSignature); } }); } } private class NodeRefreshRequest { private final InetSocketAddress address; private final HostEvent eventType; private NodeRefreshRequest(InetSocketAddress address, HostEvent eventType) { this.address = address; this.eventType = eventType; } @Override public String toString() { return address + " " + eventType; } } private class NodeRefreshRequestDeliveryCallback implements EventDebouncer.DeliveryCallback { @Override public ListenableFuture deliver(List events) { Map hosts = new HashMap(); // only keep the last event for each host for (NodeRefreshRequest req : events) { hosts.put(req.address, req.eventType); } List> futures = new ArrayList>(hosts.size()); for (final Entry entry : hosts.entrySet()) { InetSocketAddress address = entry.getKey(); HostEvent eventType = entry.getValue(); switch (eventType) { case UP: Host upHost = metadata.getHost(address); if (upHost == null) { upHost = metadata.newHost(address); Host previous = metadata.addIfAbsent(upHost); if (previous != null) { // We got beat by another thread at adding the host. Let it win and ignore the // notification here. continue; } futures.add(schedule(hostAdded(upHost))); } else { futures.add(schedule(hostUp(upHost))); } break; case ADDED: Host newHost = metadata.newHost(address); Host previous = metadata.addIfAbsent(newHost); if (previous == null) { futures.add(schedule(hostAdded(newHost))); } else if (!previous.isUp()) { futures.add(schedule(hostUp(previous))); } break; case DOWN: // Note that there is a slight risk we can receive the event late and thus // mark the host down even though we already had reconnected successfully. // But it is unlikely, and don't have too much consequence since we'll try // reconnecting // right away, so we favor the detection to make the Host.isUp method more reliable. Host downHost = metadata.getHost(address); if (downHost != null) { // Only process DOWN events if we have no active connections to the host . // Otherwise, we // wait for the connections to fail. This is to prevent against a bad control host // aggressively marking DOWN all of its peers. if (downHost.convictionPolicy.hasActiveConnections()) { logger.debug( "Ignoring down event on {} because it still has active connections", downHost); } else { futures.add(execute(hostDown(downHost))); } } break; case REMOVED: Host removedHost = metadata.getHost(address); if (removedHost != null) futures.add(execute(hostRemoved(removedHost))); break; } } return Futures.allAsList(futures); } private ListenableFuture execute(ExceptionCatchingRunnable task) { return executor.submit(task); } private ListenableFuture schedule(final ExceptionCatchingRunnable task) { // Cassandra tends to send notifications for new/up nodes a bit early (it is triggered once // gossip is up, but that is before the client-side server is up), so we add a delay // (otherwise the connection will likely fail and have to be retry which is wasteful). // This has been fixed by CASSANDRA-8236 and does not apply to protocol versions >= 4 // and C* versions >= 2.2.0 if (protocolVersion().compareTo(ProtocolVersion.V4) < 0) { final SettableFuture future = SettableFuture.create(); scheduledTasksExecutor.schedule( new ExceptionCatchingRunnable() { @Override public void runMayThrow() throws Exception { ListenableFuture f = execute(task); GuavaCompatibility.INSTANCE.addCallback( f, new FutureCallback() { @Override public void onSuccess(Object result) { future.set(null); } @Override public void onFailure(Throwable t) { future.setException(t); } }); } }, NEW_NODE_DELAY_SECONDS, TimeUnit.SECONDS); return future; } else { return execute(task); } } // Make sure we call controlConnection.refreshNodeInfo(host) // so that we have up-to-date infos on that host before adding it (so we typically // catch that an upgraded node uses a new cassandra version). private ExceptionCatchingRunnable hostAdded(final Host host) { return new ExceptionCatchingRunnable() { @Override public void runMayThrow() throws Exception { if (controlConnection.refreshNodeInfo(host)) { onAdd(host, null); submitNodeListRefresh(); } else { logger.debug("Not enough info for {}, ignoring host", host); } } }; } private ExceptionCatchingRunnable hostUp(final Host host) { return new ExceptionCatchingRunnable() { @Override public void runMayThrow() throws Exception { if (controlConnection.refreshNodeInfo(host)) { onUp(host, null); } else { logger.debug("Not enough info for {}, ignoring host", host); } } }; } private ExceptionCatchingRunnable hostDown(final Host host) { return new ExceptionCatchingRunnable() { @Override public void runMayThrow() throws Exception { onDown(host, false, true); } }; } private ExceptionCatchingRunnable hostRemoved(final Host host) { return new ExceptionCatchingRunnable() { @Override public void runMayThrow() throws Exception { if (metadata.remove(host)) { logger.info("Cassandra host {} removed", host); onRemove(host); submitNodeListRefresh(); } } }; } } private class NodeListRefreshRequest { @Override public String toString() { return "Refresh node list and token map"; } } private class NodeListRefreshRequestDeliveryCallback implements EventDebouncer.DeliveryCallback { @Override public ListenableFuture deliver(List events) { // The number of received requests does not matter // as long as one request is made, refresh the entire node list return executor.submit( new ExceptionCatchingRunnable() { @Override public void runMayThrow() throws InterruptedException, ExecutionException { controlConnection.refreshNodeListAndTokenMap(); } }); } } } private enum HostEvent { UP, DOWN, ADDED, REMOVED } /** * Periodically ensures that closed connections are properly terminated once they have no more * pending requests. * *

This is normally done when the connection errors out, or when the last request is processed; * this class acts as a last-effort protection since unterminated connections can lead to * deadlocks. If it terminates a connection, this indicates a bug; warnings are logged so that * this can be reported. * * @see Connection#tryTerminate(boolean) */ static class ConnectionReaper { private static final int INTERVAL_MS = 15000; private final ScheduledExecutorService executor; @VisibleForTesting final Map connections = new ConcurrentHashMap(); private volatile boolean shutdown; private final Runnable reaperTask = new Runnable() { @Override public void run() { long now = System.currentTimeMillis(); Iterator> iterator = connections.entrySet().iterator(); while (iterator.hasNext()) { Entry entry = iterator.next(); Connection connection = entry.getKey(); Long terminateTime = entry.getValue(); if (terminateTime <= now) { boolean terminated = connection.tryTerminate(true); if (terminated) iterator.remove(); } } } }; ConnectionReaper(ScheduledExecutorService executor) { this.executor = executor; this.executor.scheduleWithFixedDelay( reaperTask, INTERVAL_MS, INTERVAL_MS, TimeUnit.MILLISECONDS); } void register(Connection connection, long terminateTime) { if (shutdown) { // This should not happen since the reaper is shut down after all sessions. logger.warn("Connection registered after reaper shutdown: {}", connection); connection.tryTerminate(true); } else { connections.put(connection, terminateTime); } } void shutdown() { shutdown = true; // Force shutdown to avoid waiting for the interval, and run the task manually one last time executor.shutdownNow(); reaperTask.run(); } } }