All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.hadoop.hbase.client.example.MultiThreadedClientExample Maven / Gradle / Ivy

There is a newer version: 3.0.0-beta-1
Show newest version
/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.hadoop.hbase.client.example;

import java.io.IOException;
import java.util.ArrayList;
import java.util.List;
import java.util.concurrent.Callable;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Executors;
import java.util.concurrent.ForkJoinPool;
import java.util.concurrent.Future;
import java.util.concurrent.ThreadFactory;
import java.util.concurrent.ThreadLocalRandom;
import java.util.concurrent.TimeUnit;
import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.hbase.Cell;
import org.apache.hadoop.hbase.CellBuilderFactory;
import org.apache.hadoop.hbase.CellBuilderType;
import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.client.Connection;
import org.apache.hadoop.hbase.client.ConnectionFactory;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.client.RegionLocator;
import org.apache.hadoop.hbase.client.Result;
import org.apache.hadoop.hbase.client.ResultScanner;
import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.client.Table;
import org.apache.hadoop.hbase.filter.KeyOnlyFilter;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;
import org.apache.yetus.audience.InterfaceAudience;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;

/**
 * Example on how to use HBase's {@link Connection} and {@link Table} in a multi-threaded
 * environment. Each table is a light weight object that is created and thrown away. Connections are
 * heavy weight objects that hold on to zookeeper connections, async processes, and other state.
 *
 * 
 * Usage:
 * bin/hbase org.apache.hadoop.hbase.client.example.MultiThreadedClientExample testTableName 500000
 * 
*

* The table should already be created before running the command. This example expects one column * family named d. *

*

* This is meant to show different operations that are likely to be done in a real world * application. These operations are: *

*
    *
  • 30% of all operations performed are batch writes. 30 puts are created and sent out at a time. * The response for all puts is waited on.
  • *
  • 20% of all operations are single writes. A single put is sent out and the response is waited * for.
  • *
  • 50% of all operations are scans. These scans start at a random place and scan up to 100 rows. *
  • *
*/ @InterfaceAudience.Private public class MultiThreadedClientExample extends Configured implements Tool { private static final Logger LOG = LoggerFactory.getLogger(MultiThreadedClientExample.class); private static final int DEFAULT_NUM_OPERATIONS = 500000; /** * The name of the column family. d for default. */ private static final byte[] FAMILY = Bytes.toBytes("d"); /** * For the example we're just using one qualifier. */ private static final byte[] QUAL = Bytes.toBytes("test"); private final ExecutorService internalPool; private final int threads; public MultiThreadedClientExample() throws IOException { // Base number of threads. // This represents the number of threads you application has // that can be interacting with an hbase client. this.threads = Runtime.getRuntime().availableProcessors() * 4; // Daemon threads are great for things that get shut down. ThreadFactory threadFactory = new ThreadFactoryBuilder().setDaemon(true).setNameFormat("internal-pol-%d").build(); this.internalPool = Executors.newFixedThreadPool(threads, threadFactory); } @Override public int run(String[] args) throws Exception { if (args.length < 1 || args.length > 2) { System.out.println("Usage: " + this.getClass().getName() + " tableName [num_operations]"); return -1; } final TableName tableName = TableName.valueOf(args[0]); int numOperations = DEFAULT_NUM_OPERATIONS; // the second arg is the number of operations to send. if (args.length == 2) { numOperations = Integer.parseInt(args[1]); } // Threads for the client only. // // We don't want to mix hbase and business logic. // ExecutorService service = new ForkJoinPool(threads * 2); // Create two different connections showing how it's possible to // separate different types of requests onto different connections final Connection writeConnection = ConnectionFactory.createConnection(getConf(), service); final Connection readConnection = ConnectionFactory.createConnection(getConf(), service); // At this point the entire cache for the region locations is full. // Only do this if the number of regions in a table is easy to fit into memory. // // If you are interacting with more than 25k regions on a client then it's probably not good // to do this at all. warmUpConnectionCache(readConnection, tableName); warmUpConnectionCache(writeConnection, tableName); List> futures = new ArrayList<>(numOperations); for (int i = 0; i < numOperations; i++) { double r = ThreadLocalRandom.current().nextDouble(); Future f; // For the sake of generating some synthetic load this queues // some different callables. // These callables are meant to represent real work done by your application. if (r < .30) { f = internalPool.submit(new WriteExampleCallable(writeConnection, tableName)); } else if (r < .50) { f = internalPool.submit(new SingleWriteExampleCallable(writeConnection, tableName)); } else { f = internalPool.submit(new ReadExampleCallable(writeConnection, tableName)); } futures.add(f); } // Wait a long time for all the reads/writes to complete for (Future f : futures) { f.get(10, TimeUnit.MINUTES); } // Clean up after our selves for cleanliness internalPool.shutdownNow(); service.shutdownNow(); return 0; } private void warmUpConnectionCache(Connection connection, TableName tn) throws IOException { try (RegionLocator locator = connection.getRegionLocator(tn)) { LOG.info("Warmed up region location cache for " + tn + " got " + locator.getAllRegionLocations().size()); } } /** * Class that will show how to send batches of puts at the same time. */ public static class WriteExampleCallable implements Callable { private final Connection connection; private final TableName tableName; public WriteExampleCallable(Connection connection, TableName tableName) { this.connection = connection; this.tableName = tableName; } @Override public Boolean call() throws Exception { // Table implements Closable so we use the try with resource structure here. // https://docs.oracle.com/javase/tutorial/essential/exceptions/tryResourceClose.html try (Table t = connection.getTable(tableName)) { byte[] value = Bytes.toBytes(Double.toString(ThreadLocalRandom.current().nextDouble())); int rows = 30; // Array to put the batch ArrayList puts = new ArrayList<>(rows); for (int i = 0; i < 30; i++) { byte[] rk = Bytes.toBytes(ThreadLocalRandom.current().nextLong()); Put p = new Put(rk); p.add(CellBuilderFactory.create(CellBuilderType.SHALLOW_COPY).setRow(rk).setFamily(FAMILY) .setQualifier(QUAL).setTimestamp(p.getTimestamp()).setType(Cell.Type.Put) .setValue(value).build()); puts.add(p); } // now that we've assembled the batch it's time to push it to hbase. t.put(puts); } return true; } } /** * Class to show how to send a single put. */ public static class SingleWriteExampleCallable implements Callable { private final Connection connection; private final TableName tableName; public SingleWriteExampleCallable(Connection connection, TableName tableName) { this.connection = connection; this.tableName = tableName; } @Override public Boolean call() throws Exception { try (Table t = connection.getTable(tableName)) { byte[] value = Bytes.toBytes(Double.toString(ThreadLocalRandom.current().nextDouble())); byte[] rk = Bytes.toBytes(ThreadLocalRandom.current().nextLong()); Put p = new Put(rk); p.add(CellBuilderFactory.create(CellBuilderType.SHALLOW_COPY).setRow(rk).setFamily(FAMILY) .setQualifier(QUAL).setTimestamp(p.getTimestamp()).setType(Cell.Type.Put).setValue(value) .build()); t.put(p); } return true; } } /** * Class to show how to scan some rows starting at a random location. */ public static class ReadExampleCallable implements Callable { private final Connection connection; private final TableName tableName; public ReadExampleCallable(Connection connection, TableName tableName) { this.connection = connection; this.tableName = tableName; } @Override public Boolean call() throws Exception { // total length in bytes of all read rows. int result = 0; // Number of rows the scan will read before being considered done. int toRead = 100; try (Table t = connection.getTable(tableName)) { byte[] rk = Bytes.toBytes(ThreadLocalRandom.current().nextLong()); Scan s = new Scan(rk); // This filter will keep the values from being sent accross the wire. // This is good for counting or other scans that are checking for // existence and don't rely on the value. s.setFilter(new KeyOnlyFilter()); // Don't go back to the server for every single row. // We know these rows are small. So ask for 20 at a time. // This would be application specific. // // The goal is to reduce round trips but asking for too // many rows can lead to GC problems on client and server sides. s.setCaching(20); // Don't use the cache. While this is a silly test program it's still good to be // explicit that scans normally don't use the block cache. s.setCacheBlocks(false); // Open up the scanner and close it automatically when done. try (ResultScanner rs = t.getScanner(s)) { // Now go through rows. for (Result r : rs) { // Keep track of things size to simulate doing some real work. result += r.getRow().length; toRead -= 1; // Most online applications won't be // reading the entire table so this break // simulates small to medium size scans, // without needing to know an end row. if (toRead <= 0) { break; } } } } return result > 0; } } public static void main(String[] args) throws Exception { ToolRunner.run(new MultiThreadedClientExample(), args); } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy