org.apache.hudi.common.util.collection.RocksDbDiskMap Maven / Gradle / Ivy
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hudi.common.util.collection;
import org.apache.hudi.exception.HoodieException;
import org.apache.hudi.exception.HoodieNotSupportedException;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.io.IOException;
import java.io.Serializable;
import java.util.AbstractMap;
import java.util.Collection;
import java.util.HashSet;
import java.util.Iterator;
import java.util.Map;
import java.util.Set;
import java.util.Spliterators;
import java.util.stream.Stream;
import java.util.stream.StreamSupport;
/**
* This class provides a disk spillable only map implementation.
* All of the data is stored using the RocksDB implementation.
*/
public final class RocksDbDiskMap extends DiskMap {
// ColumnFamily allows partitioning data within RockDB, which allows
// independent configuration and faster deletes across partitions
// https://github.com/facebook/rocksdb/wiki/Column-Families
// For this use case, we use a single static column family/ partition
//
private static final String ROCKSDB_COL_FAMILY = "rocksdb-diskmap";
private static final Logger LOG = LoggerFactory.getLogger(RocksDbDiskMap.class);
// Stores the key and corresponding value's latest metadata spilled to disk
private final Set keySet;
private RocksDBDAO rocksDb;
public RocksDbDiskMap(String rocksDbStoragePath) throws IOException {
super(rocksDbStoragePath, ExternalSpillableMap.DiskMapType.ROCKS_DB.name());
this.keySet = new HashSet<>();
}
@Override
public int size() {
return keySet.size();
}
@Override
public boolean isEmpty() {
return keySet.isEmpty();
}
@Override
public boolean containsKey(Object key) {
return keySet.contains((T) key);
}
@Override
public boolean containsValue(Object value) {
throw new HoodieNotSupportedException("unable to compare values in map");
}
@Override
public R get(Object key) {
if (!containsKey(key)) {
return null;
}
return getRocksDb().get(ROCKSDB_COL_FAMILY, (T) key);
}
@Override
public R put(T key, R value) {
getRocksDb().put(ROCKSDB_COL_FAMILY, key, value);
keySet.add(key);
return value;
}
@Override
public R remove(Object key) {
R value = get(key);
if (value != null) {
keySet.remove((T) key);
getRocksDb().delete(ROCKSDB_COL_FAMILY, (T) key);
}
return value;
}
@Override
public void putAll(Map extends T, ? extends R> keyValues) {
getRocksDb().writeBatch(batch -> keyValues.forEach((key, value) -> getRocksDb().putInBatch(batch, ROCKSDB_COL_FAMILY, key, value)));
keySet.addAll(keyValues.keySet());
}
@Override
public void clear() {
close();
}
@Override
public Set keySet() {
return keySet;
}
@Override
public Collection values() {
throw new HoodieException("Unsupported Operation Exception");
}
@Override
public Set> entrySet() {
Set> entrySet = new HashSet<>();
for (T key : keySet) {
entrySet.add(new AbstractMap.SimpleEntry<>(key, get(key)));
}
return entrySet;
}
/**
* Custom iterator to iterate over values written to disk.
*/
@Override
public Iterator iterator() {
return getRocksDb().iterator(ROCKSDB_COL_FAMILY);
}
@Override
public Stream valueStream() {
return StreamSupport.stream(Spliterators.spliteratorUnknownSize(iterator(), 0), false);
}
@Override
public long sizeOfFileOnDiskInBytes() {
return getRocksDb().getTotalBytesWritten();
}
@Override
public void close() {
keySet.clear();
if (null != rocksDb) {
rocksDb.close();
}
rocksDb = null;
super.close();
}
private RocksDBDAO getRocksDb() {
if (null == rocksDb) {
synchronized (this) {
if (null == rocksDb) {
rocksDb = new RocksDBDAO(ROCKSDB_COL_FAMILY, diskMapPath);
rocksDb.addColumnFamily(ROCKSDB_COL_FAMILY);
}
}
}
return rocksDb;
}
}