All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.datastax.driver.core.NettyOptions Maven / Gradle / Ivy

Go to download

A driver for Scylla and Apache Cassandra 1.2+ that works exclusively with the Cassandra Query Language version 3 (CQL3) and Cassandra's binary protocol.

The newest version!
/*
 * Copyright DataStax, Inc.
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 * http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package com.datastax.driver.core;

import io.netty.bootstrap.Bootstrap;
import io.netty.buffer.PooledByteBufAllocator;
import io.netty.channel.ChannelOption;
import io.netty.channel.EventLoopGroup;
import io.netty.channel.socket.SocketChannel;
import io.netty.util.HashedWheelTimer;
import io.netty.util.Timer;
import java.util.concurrent.ThreadFactory;

/**
 * A set of hooks that allow clients to customize the driver's underlying Netty layer.
 *
 * 

Clients that need to hook into the driver's underlying Netty layer can subclass this class and * provide the necessary customization by overriding its methods. * *

Typically, clients would register this class with {@link Cluster#builder()}: * *

* *

 *     NettyOptions nettyOptions = ...
 *     Cluster cluster = Cluster.builder()
 *          .addContactPoint(...)
 *          .withNettyOptions(nettyOptions)
 *          .build();
 * 
* *

Extending the NettyOptions API * *

Contrary to other driver options, the options available in this class should be considered as * advanced features and as such, they should only be modified by expert users. * *

A misconfiguration introduced by the means of this API can have unexpected results and * cause the driver to completely fail to connect. * *

Moreover, since versions 2.0.9 and 2.1.4 (see JAVA-538), the driver is available in two * different flavors: with a standard Maven dependency on Netty, or with a "shaded" (internalized) * Netty dependency. * *

Given that NettyOptions API exposes Netty classes ({@link SocketChannel}, etc.), it should * only be extended by clients using the non-shaded version of driver. * *

Extending this API with shaded Netty classes is not supported, and in particular for * OSGi applications, it is likely that such a configuration would lead to compile and/or runtime * errors. * * @since 2.0.10 */ public class NettyOptions { /** The default instance of {@link NettyOptions} to use. */ public static final NettyOptions DEFAULT_INSTANCE = new NettyOptions(); /** * Return the {@code EventLoopGroup} instance to use. * *

This hook is invoked only once at {@link Cluster} initialization; the returned instance will * be kept in use throughout the cluster lifecycle. * *

Typically, implementors would return a newly-created instance; it is however possible to * re-use a shared instance, but in this case implementors should also override {@link * #onClusterClose(EventLoopGroup)} to prevent the shared instance to be closed when the cluster * is closed. * *

The default implementation returns a new instance of {@code * io.netty.channel.epoll.EpollEventLoopGroup} if {@link NettyUtil#isEpollAvailable() epoll is * available}, or {@code io.netty.channel.nio.NioEventLoopGroup} otherwise. * * @param threadFactory The {@link ThreadFactory} to use when creating a new {@code * EventLoopGroup} instance; The driver will provide its own internal thread factory here. It * is safe to ignore it and use another thread factory. Note however that for optimal * performance it is recommended to use a factory that returns {@link * io.netty.util.concurrent.FastThreadLocalThread} instances (such as Netty's {@link * java.util.concurrent.Executors.DefaultThreadFactory}). * @return the {@code EventLoopGroup} instance to use. */ public EventLoopGroup eventLoopGroup(ThreadFactory threadFactory) { return NettyUtil.newEventLoopGroupInstance(threadFactory); } /** * Return the specific {@code SocketChannel} subclass to use. * *

This hook is invoked only once at {@link Cluster} initialization; the returned instance will * then be used each time the driver creates a new {@link Connection} and configures a new * instance of {@link Bootstrap} for it. * *

The default implementation returns {@code io.netty.channel.epoll.EpollSocketChannel} if * {@link NettyUtil#isEpollAvailable() epoll is available}, or {@code * io.netty.channel.socket.nio.NioSocketChannel} otherwise. * * @return The {@code SocketChannel} subclass to use. */ public Class channelClass() { return NettyUtil.channelClass(); } /** * Hook invoked each time the driver creates a new {@link Connection} and configures a new * instance of {@link Bootstrap} for it. * *

This hook is guaranteed to be called after the driver has applied all {@link * SocketOptions}s. * *

This is a good place to add extra {@link io.netty.channel.ChannelHandler ChannelOption}s to * the boostrap; e.g. plug a custom {@link io.netty.buffer.ByteBufAllocator ByteBufAllocator} * implementation: * *

* *

   * ByteBufAllocator myCustomByteBufAllocator = ...
   *
   * public void afterBootstrapInitialized(Bootstrap bootstrap) {
   *     bootstrap.option(ChannelOption.ALLOCATOR, myCustomByteBufAllocator);
   * }
   * 
* *

Note that the default implementation of this method configures a pooled {@code * ByteBufAllocator} (Netty 4.0 defaults to unpooled). If you override this method to set * unrelated options, make sure you call {@code super.afterBootstrapInitialized(bootstrap)}. * * @param bootstrap the {@link Bootstrap} being initialized. */ public void afterBootstrapInitialized(Bootstrap bootstrap) { // In Netty 4.1.x, pooled will be the default, so this won't be necessary anymore bootstrap.option(ChannelOption.ALLOCATOR, PooledByteBufAllocator.DEFAULT); } /** * Hook invoked each time the driver creates a new {@link Connection} and initializes the {@link * SocketChannel channel}. * *

This hook is guaranteed to be called after the driver has registered all its * internal channel handlers, and applied the configured {@link SSLOptions}, if any. * *

This is a good place to add extra {@link io.netty.channel.ChannelHandler ChannelHandler}s to * the channel's pipeline; e.g. to add a custom SSL handler to the beginning of the handler chain, * do the following: * *

* *

   * ChannelPipeline pipeline = channel.pipeline();
   * SSLEngine myCustomSSLEngine = ...
   * SslHandler myCustomSSLHandler = new SslHandler(myCustomSSLEngine);
   * pipeline.addFirst("ssl", myCustomSSLHandler);
   * 
* *

Note: if you intend to provide your own SSL implementation, do not enable the driver's * built-in {@link SSLOptions} at the same time. * * @param channel the {@link SocketChannel} instance, after being initialized by the driver. * @throws Exception if this methods encounters any errors. */ public void afterChannelInitialized(SocketChannel channel) throws Exception { // noop } /** * Hook invoked when the cluster is shutting down after a call to {@link Cluster#close()}. * *

This is guaranteed to be called only after all connections have been individually closed, * and their channels closed, and only once per {@link EventLoopGroup} instance. * *

This gives the implementor a chance to close the {@link EventLoopGroup} properly, if * required. * *

The default implementation initiates a {@link EventLoopGroup#shutdownGracefully() graceful * shutdown} of the passed {@link EventLoopGroup}, then waits uninterruptibly for the shutdown to * complete or timeout. * *

Implementation note: if the {@link EventLoopGroup} instance is being shared, or used for * other purposes than to coordinate Netty events for the current cluster, then it should not be * shut down here; subclasses would have to override this method accordingly to take the * appropriate action. * * @param eventLoopGroup the event loop group used by the cluster being closed */ public void onClusterClose(EventLoopGroup eventLoopGroup) { eventLoopGroup.shutdownGracefully().syncUninterruptibly(); } /** * Return the {@link Timer} instance used by Read Timeouts and Speculative Execution. * *

This hook is invoked only once at {@link Cluster} initialization; the returned instance will * be kept in use throughout the cluster lifecycle. * *

Typically, implementors would return a newly-created instance; it is however possible to * re-use a shared instance, but in this case implementors should also override {@link * #onClusterClose(Timer)} to prevent the shared instance to be closed when the cluster is closed. * *

The default implementation returns a new instance created by {@link * HashedWheelTimer#HashedWheelTimer(ThreadFactory)}. * * @param threadFactory The {@link ThreadFactory} to use when creating a new {@link * HashedWheelTimer} instance; The driver will provide its own internal thread factory here. * It is safe to ignore it and use another thread factory. Note however that for optimal * performance it is recommended to use a factory that returns {@link * io.netty.util.concurrent.FastThreadLocalThread} instances (such as Netty's {@link * java.util.concurrent.Executors.DefaultThreadFactory}). * @return the {@link Timer} instance to use. */ public Timer timer(ThreadFactory threadFactory) { return new HashedWheelTimer(threadFactory); } /** * Hook invoked when the cluster is shutting down after a call to {@link Cluster#close()}. * *

This is guaranteed to be called only after all connections have been individually closed, * and their channels closed, and only once per {@link Timer} instance. * *

This gives the implementor a chance to close the {@link Timer} properly, if required. * *

The default implementation calls a {@link Timer#stop()} of the passed {@link Timer} * instance. * *

Implementation note: if the {@link Timer} instance is being shared, or used for other * purposes than to schedule actions for the current cluster, than it should not be stopped here; * subclasses would have to override this method accordingly to take the appropriate action. * * @param timer the timer used by the cluster being closed */ public void onClusterClose(Timer timer) { timer.stop(); } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy