All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.hadoop.hbase.client.HTablePool Maven / Gradle / Ivy

There is a newer version: 4.15.0-HBase-1.5
Show newest version
/**
 *
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.hadoop.hbase.client;

import java.io.Closeable;
import java.io.IOException;
import java.util.Collection;
import java.util.List;
import java.util.Map;

import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.client.coprocessor.Batch;
import org.apache.hadoop.hbase.client.coprocessor.Batch.Callback;
import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.PoolMap;
import org.apache.hadoop.hbase.util.PoolMap.PoolType;

import com.google.protobuf.Descriptors;
import com.google.protobuf.Message;
import com.google.protobuf.Service;
import com.google.protobuf.ServiceException;

/**
 * A simple pool of HTable instances.
 *
 * Each HTablePool acts as a pool for all tables. To use, instantiate an
 * HTablePool and use {@link #getTable(String)} to get an HTable from the pool.
 *
   * This method is not needed anymore, clients should call
   * HTableInterface.close() rather than returning the tables to the pool
   *
 * Once you are done with it, close your instance of {@link HTableInterface}
 * by calling {@link HTableInterface#close()} rather than returning the tables
 * to the pool with (deprecated) {@link #putTable(HTableInterface)}.
 *
 * 

* A pool can be created with a maxSize which defines the most HTable * references that will ever be retained for each table. Otherwise the default * is {@link Integer#MAX_VALUE}. * *

* Pool will manage its own connections to the cluster. See * {@link HConnectionManager}. * @deprecated as of 0.98.1. See {@link HConnection#getTable(String)}. */ @InterfaceAudience.Private @Deprecated public class HTablePool implements Closeable { private final PoolMap tables; private final int maxSize; private final PoolType poolType; private final Configuration config; private final HTableInterfaceFactory tableFactory; /** * Default Constructor. Default HBaseConfiguration and no limit on pool size. */ public HTablePool() { this(HBaseConfiguration.create(), Integer.MAX_VALUE); } /** * Constructor to set maximum versions and use the specified configuration. * * @param config * configuration * @param maxSize * maximum number of references to keep for each table */ public HTablePool(final Configuration config, final int maxSize) { this(config, maxSize, null, null); } /** * Constructor to set maximum versions and use the specified configuration and * table factory. * * @param config * configuration * @param maxSize * maximum number of references to keep for each table * @param tableFactory * table factory */ public HTablePool(final Configuration config, final int maxSize, final HTableInterfaceFactory tableFactory) { this(config, maxSize, tableFactory, PoolType.Reusable); } /** * Constructor to set maximum versions and use the specified configuration and * pool type. * * @param config * configuration * @param maxSize * maximum number of references to keep for each table * @param poolType * pool type which is one of {@link PoolType#Reusable} or * {@link PoolType#ThreadLocal} */ public HTablePool(final Configuration config, final int maxSize, final PoolType poolType) { this(config, maxSize, null, poolType); } /** * Constructor to set maximum versions and use the specified configuration, * table factory and pool type. The HTablePool supports the * {@link PoolType#Reusable} and {@link PoolType#ThreadLocal}. If the pool * type is null or not one of those two values, then it will default to * {@link PoolType#Reusable}. * * @param config * configuration * @param maxSize * maximum number of references to keep for each table * @param tableFactory * table factory * @param poolType * pool type which is one of {@link PoolType#Reusable} or * {@link PoolType#ThreadLocal} */ public HTablePool(final Configuration config, final int maxSize, final HTableInterfaceFactory tableFactory, PoolType poolType) { // Make a new configuration instance so I can safely cleanup when // done with the pool. this.config = config == null ? HBaseConfiguration.create() : config; this.maxSize = maxSize; this.tableFactory = tableFactory == null ? new HTableFactory() : tableFactory; if (poolType == null) { this.poolType = PoolType.Reusable; } else { switch (poolType) { case Reusable: case ThreadLocal: this.poolType = poolType; break; default: this.poolType = PoolType.Reusable; break; } } this.tables = new PoolMap(this.poolType, this.maxSize); } /** * Get a reference to the specified table from the pool. * * @param tableName * table name * @return a reference to the specified table * @throws RuntimeException * if there is a problem instantiating the HTable */ public HTableInterface getTable(String tableName) { // call the old getTable implementation renamed to findOrCreateTable HTableInterface table = findOrCreateTable(tableName); // return a proxy table so when user closes the proxy, the actual table // will be returned to the pool return new PooledHTable(table); } /** * Get a reference to the specified table from the pool. * * Create a new one if one is not available. * * @param tableName * table name * @return a reference to the specified table * @throws RuntimeException * if there is a problem instantiating the HTable */ private HTableInterface findOrCreateTable(String tableName) { HTableInterface table = tables.get(tableName); if (table == null) { table = createHTable(tableName); } return table; } /** * Get a reference to the specified table from the pool. *

* * Create a new one if one is not available. * * @param tableName * table name * @return a reference to the specified table * @throws RuntimeException * if there is a problem instantiating the HTable */ public HTableInterface getTable(byte[] tableName) { return getTable(Bytes.toString(tableName)); } /** * This method is not needed anymore, clients should call * HTableInterface.close() rather than returning the tables to the pool * * @param table * the proxy table user got from pool * @deprecated */ public void putTable(HTableInterface table) throws IOException { // we need to be sure nobody puts a proxy implementation in the pool // but if the client code is not updated // and it will continue to call putTable() instead of calling close() // then we need to return the wrapped table to the pool instead of the // proxy // table if (table instanceof PooledHTable) { returnTable(((PooledHTable) table).getWrappedTable()); } else { // normally this should not happen if clients pass back the same // table // object they got from the pool // but if it happens then it's better to reject it throw new IllegalArgumentException("not a pooled table: " + table); } } /** * Puts the specified HTable back into the pool. *

* * If the pool already contains maxSize references to the table, then * the table instance gets closed after flushing buffered edits. * * @param table * table */ private void returnTable(HTableInterface table) throws IOException { // this is the old putTable method renamed and made private String tableName = Bytes.toString(table.getTableName()); if (tables.size(tableName) >= maxSize) { // release table instance since we're not reusing it this.tables.removeValue(tableName, table); this.tableFactory.releaseHTableInterface(table); return; } tables.put(tableName, table); } protected HTableInterface createHTable(String tableName) { return this.tableFactory.createHTableInterface(config, Bytes.toBytes(tableName)); } /** * Closes all the HTable instances , belonging to the given table, in the * table pool. *

* Note: this is a 'shutdown' of the given table pool and different from * {@link #putTable(HTableInterface)}, that is used to return the table * instance to the pool for future re-use. * * @param tableName */ public void closeTablePool(final String tableName) throws IOException { Collection tables = this.tables.values(tableName); if (tables != null) { for (HTableInterface table : tables) { this.tableFactory.releaseHTableInterface(table); } } this.tables.remove(tableName); } /** * See {@link #closeTablePool(String)}. * * @param tableName */ public void closeTablePool(final byte[] tableName) throws IOException { closeTablePool(Bytes.toString(tableName)); } /** * Closes all the HTable instances , belonging to all tables in the table * pool. *

* Note: this is a 'shutdown' of all the table pools. */ @Override public void close() throws IOException { for (String tableName : tables.keySet()) { closeTablePool(tableName); } this.tables.clear(); } public int getCurrentPoolSize(String tableName) { return tables.size(tableName); } /** * A proxy class that implements HTableInterface.close method to return the * wrapped table back to the table pool * */ class PooledHTable implements HTableInterface { private boolean open = false; private HTableInterface table; // actual table implementation public PooledHTable(HTableInterface table) { this.table = table; this.open = true; } @Override public byte[] getTableName() { checkState(); return table.getTableName(); } @Override public TableName getName() { return table.getName(); } @Override public Configuration getConfiguration() { checkState(); return table.getConfiguration(); } @Override public HTableDescriptor getTableDescriptor() throws IOException { checkState(); return table.getTableDescriptor(); } @Override public boolean exists(Get get) throws IOException { checkState(); return table.exists(get); } @Override public boolean[] existsAll(List gets) throws IOException { checkState(); return table.existsAll(gets); } @Override public Boolean[] exists(List gets) throws IOException { checkState(); return table.exists(gets); } @Override public void batch(List actions, Object[] results) throws IOException, InterruptedException { checkState(); table.batch(actions, results); } /** * {@inheritDoc} * @deprecated If any exception is thrown by one of the actions, there is no way to * retrieve the partially executed results. Use {@link #batch(List, Object[])} instead. */ @Override public Object[] batch(List actions) throws IOException, InterruptedException { checkState(); return table.batch(actions); } @Override public Result get(Get get) throws IOException { checkState(); return table.get(get); } @Override public Result[] get(List gets) throws IOException { checkState(); return table.get(gets); } @Override @SuppressWarnings("deprecation") @Deprecated public Result getRowOrBefore(byte[] row, byte[] family) throws IOException { checkState(); return table.getRowOrBefore(row, family); } @Override public ResultScanner getScanner(Scan scan) throws IOException { checkState(); return table.getScanner(scan); } @Override public ResultScanner getScanner(byte[] family) throws IOException { checkState(); return table.getScanner(family); } @Override public ResultScanner getScanner(byte[] family, byte[] qualifier) throws IOException { checkState(); return table.getScanner(family, qualifier); } @Override public void put(Put put) throws IOException { checkState(); table.put(put); } @Override public void put(List puts) throws IOException { checkState(); table.put(puts); } @Override public boolean checkAndPut(byte[] row, byte[] family, byte[] qualifier, byte[] value, Put put) throws IOException { checkState(); return table.checkAndPut(row, family, qualifier, value, put); } @Override public boolean checkAndPut(byte[] row, byte[] family, byte[] qualifier, CompareOp compareOp, byte[] value, Put put) throws IOException { checkState(); return table.checkAndPut(row, family, qualifier, compareOp, value, put); } @Override public void delete(Delete delete) throws IOException { checkState(); table.delete(delete); } @Override public void delete(List deletes) throws IOException { checkState(); table.delete(deletes); } @Override public boolean checkAndDelete(byte[] row, byte[] family, byte[] qualifier, byte[] value, Delete delete) throws IOException { checkState(); return table.checkAndDelete(row, family, qualifier, value, delete); } @Override public boolean checkAndDelete(byte[] row, byte[] family, byte[] qualifier, CompareOp compareOp, byte[] value, Delete delete) throws IOException { checkState(); return table.checkAndDelete(row, family, qualifier, compareOp, value, delete); } @Override public Result increment(Increment increment) throws IOException { checkState(); return table.increment(increment); } @Override public long incrementColumnValue(byte[] row, byte[] family, byte[] qualifier, long amount) throws IOException { checkState(); return table.incrementColumnValue(row, family, qualifier, amount); } @Override public long incrementColumnValue(byte[] row, byte[] family, byte[] qualifier, long amount, Durability durability) throws IOException { checkState(); return table.incrementColumnValue(row, family, qualifier, amount, durability); } @Override public boolean isAutoFlush() { checkState(); return table.isAutoFlush(); } @Override public void flushCommits() throws IOException { checkState(); table.flushCommits(); } /** * Returns the actual table back to the pool * * @throws IOException */ @Override public void close() throws IOException { checkState(); open = false; returnTable(table); } @Override public CoprocessorRpcChannel coprocessorService(byte[] row) { checkState(); return table.coprocessorService(row); } @Override public Map coprocessorService(Class service, byte[] startKey, byte[] endKey, Batch.Call callable) throws ServiceException, Throwable { checkState(); return table.coprocessorService(service, startKey, endKey, callable); } @Override public void coprocessorService(Class service, byte[] startKey, byte[] endKey, Batch.Call callable, Callback callback) throws ServiceException, Throwable { checkState(); table.coprocessorService(service, startKey, endKey, callable, callback); } @Override public String toString() { return "PooledHTable{" + ", table=" + table + '}'; } /** * Expose the wrapped HTable to tests in the same package * * @return wrapped htable */ HTableInterface getWrappedTable() { return table; } @Override public void batchCallback(List actions, Object[] results, Callback callback) throws IOException, InterruptedException { checkState(); table.batchCallback(actions, results, callback); } /** * {@inheritDoc} * @deprecated If any exception is thrown by one of the actions, there is no way to * retrieve the partially executed results. Use * {@link #batchCallback(List, Object[], org.apache.hadoop.hbase.client.coprocessor.Batch.Callback)} * instead. */ @Override public Object[] batchCallback(List actions, Callback callback) throws IOException, InterruptedException { checkState(); return table.batchCallback(actions, callback); } @Override public void mutateRow(RowMutations rm) throws IOException { checkState(); table.mutateRow(rm); } @Override public Result append(Append append) throws IOException { checkState(); return table.append(append); } @Override public void setAutoFlush(boolean autoFlush) { checkState(); table.setAutoFlush(autoFlush, autoFlush); } @Override public void setAutoFlush(boolean autoFlush, boolean clearBufferOnFail) { checkState(); table.setAutoFlush(autoFlush, clearBufferOnFail); } @Override public void setAutoFlushTo(boolean autoFlush) { table.setAutoFlushTo(autoFlush); } @Override public long getWriteBufferSize() { checkState(); return table.getWriteBufferSize(); } @Override public void setWriteBufferSize(long writeBufferSize) throws IOException { checkState(); table.setWriteBufferSize(writeBufferSize); } boolean isOpen() { return open; } private void checkState() { if (!isOpen()) { throw new IllegalStateException("Table=" + table.getName().getNameAsString() + " already closed"); } } @Override public long incrementColumnValue(byte[] row, byte[] family, byte[] qualifier, long amount, boolean writeToWAL) throws IOException { return table.incrementColumnValue(row, family, qualifier, amount, writeToWAL); } @Override public Map batchCoprocessorService( Descriptors.MethodDescriptor method, Message request, byte[] startKey, byte[] endKey, R responsePrototype) throws ServiceException, Throwable { checkState(); return table.batchCoprocessorService(method, request, startKey, endKey, responsePrototype); } @Override public void batchCoprocessorService( Descriptors.MethodDescriptor method, Message request, byte[] startKey, byte[] endKey, R responsePrototype, Callback callback) throws ServiceException, Throwable { checkState(); table.batchCoprocessorService(method, request, startKey, endKey, responsePrototype, callback); } @Override public boolean checkAndMutate(byte[] row, byte[] family, byte[] qualifier, CompareOp compareOp, byte[] value, RowMutations mutation) throws IOException { checkState(); return table.checkAndMutate(row, family, qualifier, compareOp, value, mutation); } @Override public void setOperationTimeout(int operationTimeout) { table.setOperationTimeout(operationTimeout); } @Override public int getOperationTimeout() { return table.getOperationTimeout(); } @Override @Deprecated public void setRpcTimeout(int rpcTimeout) { table.setRpcTimeout(rpcTimeout); } @Override @Deprecated public int getRpcTimeout() { return table.getRpcTimeout(); } @Override public int getReadRpcTimeout() { return table.getReadRpcTimeout(); } @Override public void setReadRpcTimeout(int readRpcTimeout) { table.setReadRpcTimeout(readRpcTimeout); } @Override public int getWriteRpcTimeout() { return table.getWriteRpcTimeout(); } @Override public void setWriteRpcTimeout(int writeRpcTimeout) { table.setWriteRpcTimeout(writeRpcTimeout); } } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy