com.sun.xml.ws.rx.ha.HighlyAvailableMap Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of webservices-rt Show documentation
Show all versions of webservices-rt Show documentation
This module contains the Metro runtime code.
/*
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS HEADER.
*
* Copyright (c) 1997-2017 Oracle and/or its affiliates. All rights reserved.
*
* The contents of this file are subject to the terms of either the GNU
* General Public License Version 2 only ("GPL") or the Common Development
* and Distribution License("CDDL") (collectively, the "License"). You
* may not use this file except in compliance with the License. You can
* obtain a copy of the License at
* https://oss.oracle.com/licenses/CDDL+GPL-1.1
* or LICENSE.txt. See the License for the specific
* language governing permissions and limitations under the License.
*
* When distributing the software, include this License Header Notice in each
* file and include the License file at LICENSE.txt.
*
* GPL Classpath Exception:
* Oracle designates this particular file as subject to the "Classpath"
* exception as provided by Oracle in the GPL Version 2 section of the License
* file that accompanied this code.
*
* Modifications:
* If applicable, add the following below the License Header, with the fields
* enclosed by brackets [] replaced by your own identifying information:
* "Portions Copyright [year] [name of copyright owner]"
*
* Contributor(s):
* If you wish your version of this file to be governed by only the CDDL or
* only the GPL Version 2, indicate your decision by adding "[Contributor]
* elects to include this software in this distribution under the [CDDL or GPL
* Version 2] license." If you don't indicate a single choice of license, a
* recipient has the option to distribute your version of this file under
* either the CDDL, the GPL Version 2 or to extend the choice of license to
* its licensees as provided above. However, if you add GPL Version 2 code
* and therefore, elected the GPL Version 2 license, then the option applies
* only if the new code is made subject to such option by the copyright
* holder.
*/
package com.sun.xml.ws.rx.ha;
import com.sun.istack.logging.Logger;
import java.io.Serializable;
import java.util.Collection;
import java.util.HashMap;
import java.util.Map;
import java.util.Map.Entry;
import java.util.Set;
import com.sun.xml.ws.api.ha.HaInfo;
import com.sun.xml.ws.api.ha.HighAvailabilityProvider;
import com.sun.xml.ws.commons.ha.HaContext;
import com.sun.xml.ws.commons.ha.StickyKey;
import java.util.concurrent.locks.ReadWriteLock;
import java.util.concurrent.locks.ReentrantReadWriteLock;
import java.util.logging.Level;
import org.glassfish.ha.store.api.BackingStore;
/**
*
* @author Marek Potociar (marek.potociar at sun.com)
*/
public final class HighlyAvailableMap implements Map {
private static final Logger LOGGER = Logger.getLogger(HighlyAvailableMap.class);
public static final class NoopReplicationManager implements ReplicationManager {
private final String loggerProlog;
public NoopReplicationManager(String name) {
this.loggerProlog = "[" + name + "]: ";
}
public V load(K key) {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "load() method invoked for key: " + key);
}
return null;
}
public void save(K key, V value, boolean isNew) {
// noop
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "save() method invoked for [key=" + key + ", value=" + value + ", isNew=" + isNew + "]");
}
}
public void remove(K key) {
// noop
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "remove() method invoked for key: " + key);
}
}
public void close() {
// noop
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "close() invoked");
}
}
public void destroy() {
// noop
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "destroy() invoked");
}
}
}
public static final class SimpleReplicationManager implements ReplicationManager {
private final BackingStore backingStore;
private final String loggerProlog;
public SimpleReplicationManager(String name, BackingStore backingStore) {
this.backingStore = backingStore;
this.loggerProlog = "[" + name + "]: ";
}
public V load(K key) {
final V data = HighAvailabilityProvider.loadFrom(backingStore, key, null);
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "loaded data for key [" + key + "]: " + data);
}
return data;
}
public void save(K key, V value, boolean isNew) {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "sending for replication [key=" + key + ", value=" + value + ", isNew=" + isNew + "]");
}
HighAvailabilityProvider.saveTo(backingStore, key, value, isNew);
}
public void remove(K key) {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "removing data for key: " + key);
}
HighAvailabilityProvider.removeFrom(backingStore, key);
}
public void close() {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "closing backing store");
}
HighAvailabilityProvider.close(backingStore);
}
public void destroy() {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "destroying backing store");
}
HighAvailabilityProvider.destroy(backingStore);
}
}
public static final class StickyReplicationManager implements ReplicationManager {
private final BackingStore backingStore;
private final String loggerProlog;
public StickyReplicationManager(String name, BackingStore backingStore) {
this.backingStore = backingStore;
this.loggerProlog = "[" + name + "]: ";
}
public V load(K key) {
final V data = HighAvailabilityProvider.loadFrom(backingStore, new StickyKey(key), null);
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "loaded data for key [" + key + "]: " + data);
}
return data;
}
public void save(final K key, final V value, final boolean isNew) {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "sending for replication [key=" + key + ", value=" + value + ", isNew=" + isNew + "]");
}
HaInfo haInfo = HaContext.currentHaInfo();
if (haInfo != null) {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "Existing HaInfo found, using it for data replication: " + HaContext.asString(haInfo));
}
HaContext.udpateReplicaInstance(HighAvailabilityProvider.saveTo(backingStore, new StickyKey(key, haInfo.getKey()), value, isNew));
} else {
final StickyKey stickyKey = new StickyKey(key);
final String replicaId = HighAvailabilityProvider.saveTo(backingStore, stickyKey, value, isNew);
haInfo = new HaInfo(stickyKey.getHashKey(), replicaId, false);
HaContext.updateHaInfo(haInfo);
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "No HaInfo found, created new after data replication: " + HaContext.asString(haInfo));
}
}
}
public void remove(K key) {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "removing data for key: " + key);
}
HighAvailabilityProvider.removeFrom(backingStore, new StickyKey(key));
}
public void close() {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "closing backing store");
}
HighAvailabilityProvider.close(backingStore);
}
public void destroy() {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "destroying backing store");
}
HighAvailabilityProvider.destroy(backingStore);
}
}
private final Map localMap;
private final ReadWriteLock dataLock = new ReentrantReadWriteLock();
private final ReplicationManager replicationManager;
private final String loggerProlog;
public static HighlyAvailableMap create(final String name, BackingStore backingStore) {
return new HighlyAvailableMap(name, new HashMap(), new SimpleReplicationManager(name + "_MANAGER", backingStore));
}
public static HighlyAvailableMap createSticky(final String name, BackingStore backingStore) {
return new HighlyAvailableMap(name, new HashMap(), new StickyReplicationManager(name + "_MANAGER", backingStore));
}
public static HighlyAvailableMap create(final String name, ReplicationManager replicationManager) {
if (replicationManager == null) {
replicationManager = new NoopReplicationManager(name + "_MANAGER");
}
return new HighlyAvailableMap(name, new HashMap(), replicationManager);
}
private HighlyAvailableMap(final String name, Map wrappedMap, ReplicationManager replicationManager) {
this.loggerProlog = "[" + name + "]: ";
this.localMap = wrappedMap;
this.replicationManager = replicationManager;
}
public int size() {
dataLock.readLock().lock();
try {
return localMap.size();
} finally {
dataLock.readLock().unlock();
}
}
public boolean isEmpty() {
dataLock.readLock().lock();
try {
return localMap.isEmpty();
} finally {
dataLock.readLock().unlock();
}
}
public boolean containsKey(Object key) {
@SuppressWarnings("unchecked")
K _key = (K) key;
dataLock.readLock().lock();
try {
if (localMap.containsKey(_key)) {
return true;
}
return tryLoad(_key) != null;
} finally {
dataLock.readLock().unlock();
}
}
@SuppressWarnings("unchecked")
public boolean containsValue(Object value) {
throw new UnsupportedOperationException();
}
public V get(Object key) {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "Retrieving data for key ["+ key + "]");
}
@SuppressWarnings("unchecked")
K _key = (K) key;
dataLock.readLock().lock();
try {
V value = localMap.get(_key);
if (value != null) {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "Data for key ["+ key + "] found in a local cache: " + value);
}
return value;
}
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "Data for key ["+ key + "] not found in the local cache - consulting replication manager");
}
return tryLoad(_key);
} finally {
dataLock.readLock().unlock();
}
}
public V put(K key, V value) {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "Storing data for key ["+ key + "]: " + value);
}
dataLock.writeLock().lock();
try {
V oldValue = localMap.put(key, value);
replicationManager.save(key, value, oldValue == null);
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "Old data replaced for key ["+ key + "]: " + oldValue);
}
return oldValue;
} finally {
dataLock.writeLock().unlock();
}
}
public V remove(Object key) {
@SuppressWarnings("unchecked")
K _key = (K) key;
V oldValue = get(_key);
dataLock.writeLock().lock();
try {
localMap.remove(_key);
replicationManager.remove(_key);
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "Removing data for key ["+ key + "]: " + oldValue);
}
return oldValue;
} finally {
dataLock.writeLock().unlock();
}
}
public void putAll(Map extends K, ? extends V> m) {
dataLock.writeLock().lock();
try {
for (Entry extends K, ? extends V> e : m.entrySet()) {
put(e.getKey(), e.getValue());
}
} finally {
dataLock.writeLock().unlock();
}
}
private V tryLoad(K key) {
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "Using replication manager to load data for key ["+ key + "]");
}
dataLock.readLock().unlock();
dataLock.writeLock().lock();
try {
V value = localMap.get(key);
if (value != null) {
return value;
}
value = replicationManager.load(key);
if (value != null) {
localMap.put(key, value);
}
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "Replication manager returned data for key ["+ key + "]: " + value);
}
return value;
} finally {
dataLock.readLock().lock();
dataLock.writeLock().unlock();
}
}
public void clear() {
dataLock.writeLock().lock();
try {
for (K key : localMap.keySet()) {
replicationManager.remove(key);
}
localMap.clear();
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "HA map cleared");
}
} finally {
dataLock.writeLock().unlock();
}
}
public Set keySet() {
dataLock.readLock().lock();
try {
return localMap.keySet();
} finally {
dataLock.readLock().unlock();
}
}
public Collection values() {
dataLock.readLock().lock();
try {
return localMap.values();
} finally {
dataLock.readLock().unlock();
}
}
public Set> entrySet() {
dataLock.readLock().lock();
try {
return localMap.entrySet();
} finally {
dataLock.readLock().unlock();
}
}
public Map getLocalMapCopy() {
dataLock.readLock().lock();
try {
return new HashMap(localMap);
} finally {
dataLock.readLock().unlock();
}
}
public void invalidateCache() {
dataLock.writeLock().lock();
try {
localMap.clear();
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "local cache invalidated");
}
} finally {
dataLock.writeLock().unlock();
}
}
public ReplicationManager getReplicationManager() {
return replicationManager;
}
public void close() {
replicationManager.close();
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "HA map closed");
}
}
public void destroy() {
replicationManager.destroy();
if (LOGGER.isLoggable(Level.FINER)) {
LOGGER.finer(loggerProlog + "HA map destroyed");
}
}
}