All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.aerospike.client.command.BatchNode Maven / Gradle / Ivy

/*
 * Copyright 2012-2019 Aerospike, Inc.
 *
 * Portions may be licensed to Aerospike, Inc. under one or more contributor
 * license agreements WHICH ARE COMPATIBLE WITH THE APACHE LICENSE, VERSION 2.0.
 *
 * Licensed under the Apache License, Version 2.0 (the "License"); you may not
 * use this file except in compliance with the License. You may obtain a copy of
 * the License at http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
 * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
 * License for the specific language governing permissions and limitations under
 * the License.
 */
package com.aerospike.client.command;

import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
import java.util.concurrent.atomic.AtomicReferenceArray;

import com.aerospike.client.AerospikeException;
import com.aerospike.client.BatchRead;
import com.aerospike.client.Key;
import com.aerospike.client.ResultCode;
import com.aerospike.client.cluster.Cluster;
import com.aerospike.client.cluster.Node;
import com.aerospike.client.cluster.Partition;
import com.aerospike.client.cluster.Partitions;
import com.aerospike.client.policy.BatchPolicy;

public final class BatchNode {

	public static List generateList(Cluster cluster, BatchPolicy policy, Key[] keys) {
		Node[] nodes = cluster.getNodes();

		if (nodes.length == 0) {
			throw new AerospikeException(ResultCode.SERVER_NOT_AVAILABLE, "Command failed because cluster is empty.");
		}

		// Create initial key capacity for each node as average + 25%.
		int keysPerNode = keys.length / nodes.length;
		keysPerNode += keysPerNode >>> 2;

		// The minimum key capacity is 10.
		if (keysPerNode < 10) {
			keysPerNode = 10;
		}

		// Split keys by server node.
		List batchNodes = new ArrayList(nodes.length);

		for (int i = 0; i < keys.length; i++) {
			Partition partition = new Partition(keys[i]);
			Node node = getNode(cluster, policy, partition, 0);
			BatchNode batchNode = findBatchNode(batchNodes, node);

			if (batchNode == null) {
				batchNodes.add(new BatchNode(node, keysPerNode, i));
			}
			else {
				batchNode.addKey(i);
			}
		}
		return batchNodes;
	}

	public static List generateList(Cluster cluster, BatchPolicy policy, Key[] keys, int sequence, BatchNode batchSeed) {
		Node[] nodes = cluster.getNodes();

		if (nodes.length == 0) {
			throw new AerospikeException(ResultCode.SERVER_NOT_AVAILABLE, "Command failed because cluster is empty.");
		}

		// Create initial key capacity for each node as average + 25%.
		int keysPerNode = batchSeed.offsetsSize / nodes.length;
		keysPerNode += keysPerNode >>> 2;

		// The minimum key capacity is 10.
		if (keysPerNode < 10) {
			keysPerNode = 10;
		}

		// Split keys by server node.
		List batchNodes = new ArrayList(nodes.length);

		for (int i = 0; i < batchSeed.offsetsSize; i++) {
			int offset = batchSeed.offsets[i];
			Partition partition = new Partition(keys[offset]);
			Node node = getNode(cluster, policy, partition, sequence);
			BatchNode batchNode = findBatchNode(batchNodes, node);

			if (batchNode == null) {
				batchNodes.add(new BatchNode(node, keysPerNode, offset));
			}
			else {
				batchNode.addKey(offset);
			}
		}
		return batchNodes;
	}

	public static List generateList(Cluster cluster, BatchPolicy policy, List records) {
		Node[] nodes = cluster.getNodes();

		if (nodes.length == 0) {
			throw new AerospikeException(ResultCode.SERVER_NOT_AVAILABLE, "Command failed because cluster is empty.");
		}

		// Create initial key capacity for each node as average + 25%.
		int max = records.size();
		int keysPerNode = max / nodes.length;
		keysPerNode += keysPerNode >>> 2;

		// The minimum key capacity is 10.
		if (keysPerNode < 10) {
			keysPerNode = 10;
		}

		// Split keys by server node.
		List batchNodes = new ArrayList(nodes.length);

		for (int i = 0; i < max; i++) {
			Partition partition = new Partition(records.get(i).key);
			Node node = getNode(cluster, policy, partition, 0);
			BatchNode batchNode = findBatchNode(batchNodes, node);

			if (batchNode == null) {
				batchNodes.add(new BatchNode(node, keysPerNode, i));
			}
			else {
				batchNode.addKey(i);
			}
		}
		return batchNodes;
	}

	public static List generateList(Cluster cluster, BatchPolicy policy, List records, int sequence, BatchNode batchSeed) {
		Node[] nodes = cluster.getNodes();

		if (nodes.length == 0) {
			throw new AerospikeException(ResultCode.SERVER_NOT_AVAILABLE, "Command failed because cluster is empty.");
		}

		// Create initial key capacity for each node as average + 25%.
		int keysPerNode = batchSeed.offsetsSize / nodes.length;
		keysPerNode += keysPerNode >>> 2;

		// The minimum key capacity is 10.
		if (keysPerNode < 10) {
			keysPerNode = 10;
		}

		// Split keys by server node.
		List batchNodes = new ArrayList(nodes.length);

		for (int i = 0; i < batchSeed.offsetsSize; i++) {
			int offset = batchSeed.offsets[i];
			Partition partition = new Partition(records.get(offset).key);
			Node node = getNode(cluster, policy, partition, sequence);
			BatchNode batchNode = findBatchNode(batchNodes, node);

			if (batchNode == null) {
				batchNodes.add(new BatchNode(node, keysPerNode, offset));
			}
			else {
				batchNode.addKey(offset);
			}
		}
		return batchNodes;
	}

	private static Node getNode(Cluster cluster, BatchPolicy policy, Partition partition, int sequence) {
		// Must copy hashmap reference for copy on write semantics to work.
		HashMap map = cluster.partitionMap;
		Partitions partitions = map.get(partition.namespace);

		if (partitions == null) {
			throw new AerospikeException.InvalidNamespace(partition.namespace, map.size());
		}

		if (partitions.cpMode && ! policy.linearizeRead) {
			// Strong Consistency namespaces always use master node when read policy is sequential.
			return cluster.getMasterNode(partitions, partition);
		}

		switch (policy.replica) {
		case SEQUENCE:
			return getSequenceNode(cluster, partitions, partition, sequence);

		case PREFER_RACK:
			return getRackNode(cluster, partitions, partition, sequence);

		default:
		case MASTER:
			return cluster.getMasterNode(partitions, partition);

		case MASTER_PROLES:
			return cluster.getMasterProlesNode(partitions, partition);

		case RANDOM:
			return cluster.getRandomNode();
		}
	}

	private static final Node getSequenceNode(Cluster cluster, Partitions partitions, Partition partition, int sequence) {
		AtomicReferenceArray[] replicas = partitions.replicas;

		for (int i = 0; i < replicas.length; i++) {
			int index = Math.abs(sequence % replicas.length);
			Node node = replicas[index].get(partition.partitionId);

			if (node != null && node.isActive()) {
				return node;
			}
			sequence++;
		}
		Node[] nodeArray = cluster.getNodes();
		throw new AerospikeException.InvalidNode(nodeArray.length, partition);
	}

	private static Node getRackNode(Cluster cluster, Partitions partitions, Partition partition, int sequence) {
		AtomicReferenceArray[] replicas = partitions.replicas;
		Node fallback = null;

		for (int i = 0; i < replicas.length; i++) {
			int index = Math.abs(sequence % replicas.length);
			Node node = replicas[index].get(partition.partitionId);

			if (node != null && node.isActive()) {
				if (node.hasRack(partition.namespace, cluster.rackId)) {
					return node;
				}

				if (fallback == null) {
					fallback = node;
				}
			}
			sequence++;
		}

		if (fallback != null) {
			return fallback;
		}

		Node[] nodeArray = cluster.getNodes();
		throw new AerospikeException.InvalidNode(nodeArray.length, partition);
	}

	private static BatchNode findBatchNode(List nodes, Node node) {
		for (BatchNode batchNode : nodes) {
			// Note: using pointer equality for performance.
			if (batchNode.node == node) {
				return batchNode;
			}
		}
		return null;
	}

	public final Node node;
	public int[] offsets;
	public int offsetsSize;

	public BatchNode(Node node, int capacity, int offset) {
		this.node = node;
		this.offsets = new int[capacity];
		this.offsets[0] = offset;
		this.offsetsSize = 1;
	}

	public void addKey(int offset) {
		if (offsetsSize >= offsets.length) {
			int[] copy = new int[offsetsSize * 2];
			System.arraycopy(offsets, 0, copy, 0, offsetsSize);
			offsets = copy;
		}
		offsets[offsetsSize++] = offset;
	}
}




© 2015 - 2025 Weber Informatics LLC | Privacy Policy