All Downloads are FREE. Search and download functionalities are using the official Maven repository.

gwtrpc.shaded.com.google.common.util.concurrent.Striped Maven / Gradle / Ivy

There is a newer version: 1.0-alpha-8
Show newest version
/*
 * Copyright (C) 2011 The Guava Authors
 *
 * Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except
 * in compliance with the License. You may obtain a copy of the License at
 *
 * http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software distributed under the License
 * is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express
 * or implied. See the License for the specific language governing permissions and limitations under
 * the License.
 */

package com.google.common.util.concurrent;

import com.google.common.annotations.Beta;
import com.google.common.annotations.GwtIncompatible;
import com.google.common.annotations.VisibleForTesting;
import com.google.common.base.MoreObjects;
import com.google.common.base.Preconditions;
import com.google.common.base.Supplier;
import com.google.common.collect.ImmutableList;
import com.google.common.collect.Iterables;
import com.google.common.collect.MapMaker;
import com.google.common.math.IntMath;
import com.google.common.primitives.Ints;
import java.lang.ref.Reference;
import java.lang.ref.ReferenceQueue;
import java.lang.ref.WeakReference;
import java.math.RoundingMode;
import java.util.Arrays;
import java.util.Collections;
import java.util.List;
import java.util.concurrent.ConcurrentMap;
import java.util.concurrent.Semaphore;
import java.util.concurrent.atomic.AtomicReferenceArray;
import java.util.concurrent.locks.Condition;
import java.util.concurrent.locks.Lock;
import java.util.concurrent.locks.ReadWriteLock;
import java.util.concurrent.locks.ReentrantLock;
import java.util.concurrent.locks.ReentrantReadWriteLock;

/**
 * A striped {@code Lock/Semaphore/ReadWriteLock}. This offers the underlying lock striping similar
 * to that of {@code ConcurrentHashMap} in a reusable form, and extends it for semaphores and
 * read-write locks. Conceptually, lock striping is the technique of dividing a lock into many
 * stripes, increasing the granularity of a single lock and allowing independent operations
 * to lock different stripes and proceed concurrently, instead of creating contention for a single
 * lock.
 *
 * 

The guarantee provided by this class is that equal keys lead to the same lock (or semaphore), * i.e. {@code if (key1.equals(key2))} then {@code striped.get(key1) == striped.get(key2)} (assuming * {@link Object#hashCode()} is correctly implemented for the keys). Note that if {@code key1} is * not equal to {@code key2}, it is not guaranteed that {@code * striped.get(key1) != striped.get(key2)}; the elements might nevertheless be mapped to the same * lock. The lower the number of stripes, the higher the probability of this happening. * *

There are three flavors of this class: {@code Striped}, {@code Striped}, and * {@code Striped}. For each type, two implementations are offered: {@linkplain * #lock(int) strong} and {@linkplain #lazyWeakLock(int) weak} {@code Striped}, {@linkplain * #semaphore(int, int) strong} and {@linkplain #lazyWeakSemaphore(int, int) weak} {@code * Striped}, and {@linkplain #readWriteLock(int) strong} and {@linkplain * #lazyWeakReadWriteLock(int) weak} {@code Striped}. Strong means that all * stripes (locks/semaphores) are initialized eagerly, and are not reclaimed unless {@code Striped} * itself is reclaimable. Weak means that locks/semaphores are created lazily, and they are * allowed to be reclaimed if nobody is holding on to them. This is useful, for example, if one * wants to create a {@code Striped} of many locks, but worries that in most cases only a * small portion of these would be in use. * *

Prior to this class, one might be tempted to use {@code Map}, where {@code K} * represents the task. This maximizes concurrency by having each unique key mapped to a unique * lock, but also maximizes memory footprint. On the other extreme, one could use a single lock for * all tasks, which minimizes memory footprint but also minimizes concurrency. Instead of choosing * either of these extremes, {@code Striped} allows the user to trade between required concurrency * and memory footprint. For example, if a set of tasks are CPU-bound, one could easily create a * very compact {@code Striped} of {@code availableProcessors() * 4} stripes, instead of * possibly thousands of locks which could be created in a {@code Map} structure. * * @author Dimitris Andreou * @since 13.0 */ @Beta @GwtIncompatible public abstract class Striped { /** * If there are at least this many stripes, we assume the memory usage of a ConcurrentMap will be * smaller than a large array. (This assumes that in the lazy case, most stripes are unused. As * always, if many stripes are in use, a non-lazy striped makes more sense.) */ private static final int LARGE_LAZY_CUTOFF = 1024; private Striped() {} /** * Returns the stripe that corresponds to the passed key. It is always guaranteed that if {@code * key1.equals(key2)}, then {@code get(key1) == get(key2)}. * * @param key an arbitrary, non-null key * @return the stripe that the passed key corresponds to */ public abstract L get(Object key); /** * Returns the stripe at the specified index. Valid indexes are 0, inclusively, to {@code size()}, * exclusively. * * @param index the index of the stripe to return; must be in {@code [0...size())} * @return the stripe at the specified index */ public abstract L getAt(int index); /** * Returns the index to which the given key is mapped, so that getAt(indexFor(key)) == get(key). */ abstract int indexFor(Object key); /** Returns the total number of stripes in this instance. */ public abstract int size(); /** * Returns the stripes that correspond to the passed objects, in ascending (as per {@link * #getAt(int)}) order. Thus, threads that use the stripes in the order returned by this method * are guaranteed to not deadlock each other. * *

It should be noted that using a {@code Striped} with relatively few stripes, and {@code * bulkGet(keys)} with a relative large number of keys can cause an excessive number of shared * stripes (much like the birthday paradox, where much fewer than anticipated birthdays are needed * for a pair of them to match). Please consider carefully the implications of the number of * stripes, the intended concurrency level, and the typical number of keys used in a {@code * bulkGet(keys)} operation. See Balls in * Bins model for mathematical formulas that can be used to estimate the probability of * collisions. * * @param keys arbitrary non-null keys * @return the stripes corresponding to the objects (one per each object, derived by delegating to * {@link #get(Object)}; may contain duplicates), in an increasing index order. */ public Iterable bulkGet(Iterable keys) { // Initially using the array to store the keys, then reusing it to store the respective L's final Object[] array = Iterables.toArray(keys, Object.class); if (array.length == 0) { return ImmutableList.of(); } int[] stripes = new int[array.length]; for (int i = 0; i < array.length; i++) { stripes[i] = indexFor(array[i]); } Arrays.sort(stripes); // optimize for runs of identical stripes int previousStripe = stripes[0]; array[0] = getAt(previousStripe); for (int i = 1; i < array.length; i++) { int currentStripe = stripes[i]; if (currentStripe == previousStripe) { array[i] = array[i - 1]; } else { array[i] = getAt(currentStripe); previousStripe = currentStripe; } } /* * Note that the returned Iterable holds references to the returned stripes, to avoid * error-prone code like: * * Striped stripedLock = Striped.lazyWeakXXX(...)' * Iterable locks = stripedLock.bulkGet(keys); * for (Lock lock : locks) { * lock.lock(); * } * operation(); * for (Lock lock : locks) { * lock.unlock(); * } * * If we only held the int[] stripes, translating it on the fly to L's, the original locks might * be garbage collected after locking them, ending up in a huge mess. */ @SuppressWarnings("unchecked") // we carefully replaced all keys with their respective L's List asList = (List) Arrays.asList(array); return Collections.unmodifiableList(asList); } // Static factories /** * Creates a {@code Striped} with eagerly initialized, strongly referenced locks. Every lock * is reentrant. * * @param stripes the minimum number of stripes (locks) required * @return a new {@code Striped} */ public static Striped lock(int stripes) { return new CompactStriped<>( stripes, new Supplier() { @Override public Lock get() { return new PaddedLock(); } }); } /** * Creates a {@code Striped} with lazily initialized, weakly referenced locks. Every lock is * reentrant. * * @param stripes the minimum number of stripes (locks) required * @return a new {@code Striped} */ public static Striped lazyWeakLock(int stripes) { return lazy( stripes, new Supplier() { @Override public Lock get() { return new ReentrantLock(false); } }); } private static Striped lazy(int stripes, Supplier supplier) { return stripes < LARGE_LAZY_CUTOFF ? new SmallLazyStriped(stripes, supplier) : new LargeLazyStriped(stripes, supplier); } /** * Creates a {@code Striped} with eagerly initialized, strongly referenced semaphores, * with the specified number of permits. * * @param stripes the minimum number of stripes (semaphores) required * @param permits the number of permits in each semaphore * @return a new {@code Striped} */ public static Striped semaphore(int stripes, final int permits) { return new CompactStriped<>( stripes, new Supplier() { @Override public Semaphore get() { return new PaddedSemaphore(permits); } }); } /** * Creates a {@code Striped} with lazily initialized, weakly referenced semaphores, * with the specified number of permits. * * @param stripes the minimum number of stripes (semaphores) required * @param permits the number of permits in each semaphore * @return a new {@code Striped} */ public static Striped lazyWeakSemaphore(int stripes, final int permits) { return lazy( stripes, new Supplier() { @Override public Semaphore get() { return new Semaphore(permits, false); } }); } /** * Creates a {@code Striped} with eagerly initialized, strongly referenced * read-write locks. Every lock is reentrant. * * @param stripes the minimum number of stripes (locks) required * @return a new {@code Striped} */ public static Striped readWriteLock(int stripes) { return new CompactStriped<>(stripes, READ_WRITE_LOCK_SUPPLIER); } /** * Creates a {@code Striped} with lazily initialized, weakly referenced read-write * locks. Every lock is reentrant. * * @param stripes the minimum number of stripes (locks) required * @return a new {@code Striped} */ public static Striped lazyWeakReadWriteLock(int stripes) { return lazy(stripes, WEAK_SAFE_READ_WRITE_LOCK_SUPPLIER); } private static final Supplier READ_WRITE_LOCK_SUPPLIER = new Supplier() { @Override public ReadWriteLock get() { return new ReentrantReadWriteLock(); } }; private static final Supplier WEAK_SAFE_READ_WRITE_LOCK_SUPPLIER = new Supplier() { @Override public ReadWriteLock get() { return new WeakSafeReadWriteLock(); } }; /** * ReadWriteLock implementation whose read and write locks retain a reference back to this lock. * Otherwise, a reference to just the read lock or just the write lock would not suffice to ensure * the {@code ReadWriteLock} is retained. */ private static final class WeakSafeReadWriteLock implements ReadWriteLock { private final ReadWriteLock delegate; WeakSafeReadWriteLock() { this.delegate = new ReentrantReadWriteLock(); } @Override public Lock readLock() { return new WeakSafeLock(delegate.readLock(), this); } @Override public Lock writeLock() { return new WeakSafeLock(delegate.writeLock(), this); } } /** Lock object that ensures a strong reference is retained to a specified object. */ private static final class WeakSafeLock extends ForwardingLock { private final Lock delegate; @SuppressWarnings("unused") private final WeakSafeReadWriteLock strongReference; WeakSafeLock(Lock delegate, WeakSafeReadWriteLock strongReference) { this.delegate = delegate; this.strongReference = strongReference; } @Override Lock delegate() { return delegate; } @Override public Condition newCondition() { return new WeakSafeCondition(delegate.newCondition(), strongReference); } } /** Condition object that ensures a strong reference is retained to a specified object. */ private static final class WeakSafeCondition extends ForwardingCondition { private final Condition delegate; @SuppressWarnings("unused") private final WeakSafeReadWriteLock strongReference; WeakSafeCondition(Condition delegate, WeakSafeReadWriteLock strongReference) { this.delegate = delegate; this.strongReference = strongReference; } @Override Condition delegate() { return delegate; } } private abstract static class PowerOfTwoStriped extends Striped { /** Capacity (power of two) minus one, for fast mod evaluation */ final int mask; PowerOfTwoStriped(int stripes) { Preconditions.checkArgument(stripes > 0, "Stripes must be positive"); this.mask = stripes > Ints.MAX_POWER_OF_TWO ? ALL_SET : ceilToPowerOfTwo(stripes) - 1; } @Override final int indexFor(Object key) { int hash = smear(key.hashCode()); return hash & mask; } @Override public final L get(Object key) { return getAt(indexFor(key)); } } /** * Implementation of Striped where 2^k stripes are represented as an array of the same length, * eagerly initialized. */ private static class CompactStriped extends PowerOfTwoStriped { /** Size is a power of two. */ private final Object[] array; private CompactStriped(int stripes, Supplier supplier) { super(stripes); Preconditions.checkArgument(stripes <= Ints.MAX_POWER_OF_TWO, "Stripes must be <= 2^30)"); this.array = new Object[mask + 1]; for (int i = 0; i < array.length; i++) { array[i] = supplier.get(); } } @SuppressWarnings("unchecked") // we only put L's in the array @Override public L getAt(int index) { return (L) array[index]; } @Override public int size() { return array.length; } } /** * Implementation of Striped where up to 2^k stripes can be represented, using an * AtomicReferenceArray of size 2^k. To map a user key into a stripe, we take a k-bit slice of the * user key's (smeared) hashCode(). The stripes are lazily initialized and are weakly referenced. */ @VisibleForTesting static class SmallLazyStriped extends PowerOfTwoStriped { final AtomicReferenceArray> locks; final Supplier supplier; final int size; final ReferenceQueue queue = new ReferenceQueue(); SmallLazyStriped(int stripes, Supplier supplier) { super(stripes); this.size = (mask == ALL_SET) ? Integer.MAX_VALUE : mask + 1; this.locks = new AtomicReferenceArray<>(size); this.supplier = supplier; } @Override public L getAt(int index) { if (size != Integer.MAX_VALUE) { Preconditions.checkElementIndex(index, size()); } // else no check necessary, all index values are valid ArrayReference existingRef = locks.get(index); L existing = existingRef == null ? null : existingRef.get(); if (existing != null) { return existing; } L created = supplier.get(); ArrayReference newRef = new ArrayReference(created, index, queue); while (!locks.compareAndSet(index, existingRef, newRef)) { // we raced, we need to re-read and try again existingRef = locks.get(index); existing = existingRef == null ? null : existingRef.get(); if (existing != null) { return existing; } } drainQueue(); return created; } // N.B. Draining the queue is only necessary to ensure that we don't accumulate empty references // in the array. We could skip this if we decide we don't care about holding on to Reference // objects indefinitely. private void drainQueue() { Reference ref; while ((ref = queue.poll()) != null) { // We only ever register ArrayReferences with the queue so this is always safe. ArrayReference arrayRef = (ArrayReference) ref; // Try to clear out the array slot, n.b. if we fail that is fine, in either case the // arrayRef will be out of the array after this step. locks.compareAndSet(arrayRef.index, arrayRef, null); } } @Override public int size() { return size; } private static final class ArrayReference extends WeakReference { final int index; ArrayReference(L referent, int index, ReferenceQueue queue) { super(referent, queue); this.index = index; } } } /** * Implementation of Striped where up to 2^k stripes can be represented, using a ConcurrentMap * where the key domain is [0..2^k). To map a user key into a stripe, we take a k-bit slice of the * user key's (smeared) hashCode(). The stripes are lazily initialized and are weakly referenced. */ @VisibleForTesting static class LargeLazyStriped extends PowerOfTwoStriped { final ConcurrentMap locks; final Supplier supplier; final int size; LargeLazyStriped(int stripes, Supplier supplier) { super(stripes); this.size = (mask == ALL_SET) ? Integer.MAX_VALUE : mask + 1; this.supplier = supplier; this.locks = new MapMaker().weakValues().makeMap(); } @Override public L getAt(int index) { if (size != Integer.MAX_VALUE) { Preconditions.checkElementIndex(index, size()); } // else no check necessary, all index values are valid L existing = locks.get(index); if (existing != null) { return existing; } L created = supplier.get(); existing = locks.putIfAbsent(index, created); return MoreObjects.firstNonNull(existing, created); } @Override public int size() { return size; } } /** A bit mask were all bits are set. */ private static final int ALL_SET = ~0; private static int ceilToPowerOfTwo(int x) { return 1 << IntMath.log2(x, RoundingMode.CEILING); } /* * This method was written by Doug Lea with assistance from members of JCP JSR-166 Expert Group * and released to the public domain, as explained at * http://creativecommons.org/licenses/publicdomain * * As of 2010/06/11, this method is identical to the (package private) hash method in OpenJDK 7's * java.util.HashMap class. */ // Copied from java/com/google/common/collect/Hashing.java private static int smear(int hashCode) { hashCode ^= (hashCode >>> 20) ^ (hashCode >>> 12); return hashCode ^ (hashCode >>> 7) ^ (hashCode >>> 4); } private static class PaddedLock extends ReentrantLock { /* * Padding from 40 into 64 bytes, same size as cache line. Might be beneficial to add a fourth * long here, to minimize chance of interference between consecutive locks, but I couldn't * observe any benefit from that. */ long unused1; long unused2; long unused3; PaddedLock() { super(false); } } private static class PaddedSemaphore extends Semaphore { // See PaddedReentrantLock comment long unused1; long unused2; long unused3; PaddedSemaphore(int permits) { super(permits, false); } } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy