Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.kafka.streams.state.internals;
import org.apache.kafka.clients.consumer.ConsumerRecord;
import org.apache.kafka.common.metrics.Sensor;
import org.apache.kafka.common.utils.Bytes;
import org.apache.kafka.streams.KeyValue;
import org.apache.kafka.streams.errors.ProcessorStateException;
import org.apache.kafka.streams.processor.BatchingStateRestoreCallback;
import org.apache.kafka.streams.processor.ProcessorContext;
import org.apache.kafka.streams.processor.StateStore;
import org.apache.kafka.streams.processor.internals.ProcessorContextUtils;
import org.apache.kafka.streams.processor.internals.metrics.StreamsMetricsImpl;
import org.apache.kafka.streams.processor.internals.metrics.TaskMetrics;
import org.apache.kafka.streams.state.KeyValueIterator;
import org.rocksdb.RocksDBException;
import org.rocksdb.WriteBatch;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.util.Collection;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
public class AbstractRocksDBSegmentedBytesStore implements SegmentedBytesStore {
private static final Logger LOG = LoggerFactory.getLogger(AbstractRocksDBSegmentedBytesStore.class);
private final String name;
private final AbstractSegments segments;
private final String metricScope;
private final KeySchema keySchema;
private ProcessorContext context;
private Sensor expiredRecordSensor;
private long observedStreamTime = ConsumerRecord.NO_TIMESTAMP;
private volatile boolean open;
AbstractRocksDBSegmentedBytesStore(final String name,
final String metricScope,
final KeySchema keySchema,
final AbstractSegments segments) {
this.name = name;
this.metricScope = metricScope;
this.keySchema = keySchema;
this.segments = segments;
}
@Override
public KeyValueIterator fetch(final Bytes key,
final long from,
final long to) {
return fetch(key, from, to, true);
}
@Override
public KeyValueIterator backwardFetch(final Bytes key,
final long from,
final long to) {
return fetch(key, from, to, false);
}
KeyValueIterator fetch(final Bytes key,
final long from,
final long to,
final boolean forward) {
final List searchSpace = keySchema.segmentsToSearch(segments, from, to, forward);
final Bytes binaryFrom = keySchema.lowerRangeFixedSize(key, from);
final Bytes binaryTo = keySchema.upperRangeFixedSize(key, to);
return new SegmentIterator<>(
searchSpace.iterator(),
keySchema.hasNextCondition(key, key, from, to),
binaryFrom,
binaryTo,
forward);
}
@Override
public KeyValueIterator fetch(final Bytes keyFrom,
final Bytes keyTo,
final long from,
final long to) {
return fetch(keyFrom, keyTo, from, to, true);
}
@Override
public KeyValueIterator backwardFetch(final Bytes keyFrom,
final Bytes keyTo,
final long from,
final long to) {
return fetch(keyFrom, keyTo, from, to, false);
}
KeyValueIterator fetch(final Bytes keyFrom,
final Bytes keyTo,
final long from,
final long to,
final boolean forward) {
if (keyFrom.compareTo(keyTo) > 0) {
LOG.warn("Returning empty iterator for fetch with invalid key range: from > to. " +
"This may be due to range arguments set in the wrong order, " +
"or serdes that don't preserve ordering when lexicographically comparing the serialized bytes. " +
"Note that the built-in numerical serdes do not follow this for negative numbers");
return KeyValueIterators.emptyIterator();
}
final List searchSpace = keySchema.segmentsToSearch(segments, from, to, forward);
final Bytes binaryFrom = keySchema.lowerRange(keyFrom, from);
final Bytes binaryTo = keySchema.upperRange(keyTo, to);
return new SegmentIterator<>(
searchSpace.iterator(),
keySchema.hasNextCondition(keyFrom, keyTo, from, to),
binaryFrom,
binaryTo,
forward);
}
@Override
public KeyValueIterator all() {
final List searchSpace = segments.allSegments(true);
return new SegmentIterator<>(
searchSpace.iterator(),
keySchema.hasNextCondition(null, null, 0, Long.MAX_VALUE),
null,
null,
true);
}
@Override
public KeyValueIterator backwardAll() {
final List searchSpace = segments.allSegments(false);
return new SegmentIterator<>(
searchSpace.iterator(),
keySchema.hasNextCondition(null, null, 0, Long.MAX_VALUE),
null,
null,
false);
}
@Override
public KeyValueIterator fetchAll(final long timeFrom,
final long timeTo) {
final List searchSpace = segments.segments(timeFrom, timeTo, true);
return new SegmentIterator<>(
searchSpace.iterator(),
keySchema.hasNextCondition(null, null, timeFrom, timeTo),
null,
null,
true);
}
@Override
public KeyValueIterator backwardFetchAll(final long timeFrom,
final long timeTo) {
final List searchSpace = segments.segments(timeFrom, timeTo, false);
return new SegmentIterator<>(
searchSpace.iterator(),
keySchema.hasNextCondition(null, null, timeFrom, timeTo),
null,
null,
false);
}
@Override
public void remove(final Bytes key) {
final long timestamp = keySchema.segmentTimestamp(key);
observedStreamTime = Math.max(observedStreamTime, timestamp);
final S segment = segments.getSegmentForTimestamp(timestamp);
if (segment == null) {
return;
}
segment.delete(key);
}
@Override
public void put(final Bytes key,
final byte[] value) {
final long timestamp = keySchema.segmentTimestamp(key);
observedStreamTime = Math.max(observedStreamTime, timestamp);
final long segmentId = segments.segmentId(timestamp);
final S segment = segments.getOrCreateSegmentIfLive(segmentId, context, observedStreamTime);
if (segment == null) {
expiredRecordSensor.record(1.0d, ProcessorContextUtils.getCurrentSystemTime(context));
LOG.warn("Skipping record for expired segment.");
} else {
segment.put(key, value);
}
}
@Override
public byte[] get(final Bytes key) {
final S segment = segments.getSegmentForTimestamp(keySchema.segmentTimestamp(key));
if (segment == null) {
return null;
}
return segment.get(key);
}
@Override
public String name() {
return name;
}
@Deprecated
@Override
public void init(final ProcessorContext context,
final StateStore root) {
this.context = context;
final StreamsMetricsImpl metrics = ProcessorContextUtils.getMetricsImpl(context);
final String threadId = Thread.currentThread().getName();
final String taskName = context.taskId().toString();
expiredRecordSensor = TaskMetrics.droppedRecordsSensorOrExpiredWindowRecordDropSensor(
threadId,
taskName,
metricScope,
name(),
metrics
);
segments.openExisting(this.context, observedStreamTime);
// register and possibly restore the state from the logs
context.register(root, new RocksDBSegmentsBatchingRestoreCallback());
open = true;
}
@Override
public void flush() {
segments.flush();
}
@Override
public void close() {
open = false;
segments.close();
}
@Override
public boolean persistent() {
return true;
}
@Override
public boolean isOpen() {
return open;
}
// Visible for testing
List getSegments() {
return segments.allSegments(false);
}
// Visible for testing
void restoreAllInternal(final Collection> records) {
try {
final Map writeBatchMap = getWriteBatches(records);
for (final Map.Entry entry : writeBatchMap.entrySet()) {
final S segment = entry.getKey();
final WriteBatch batch = entry.getValue();
segment.write(batch);
batch.close();
}
} catch (final RocksDBException e) {
throw new ProcessorStateException("Error restoring batch to store " + this.name, e);
}
}
// Visible for testing
Map getWriteBatches(final Collection> records) {
// advance stream time to the max timestamp in the batch
for (final KeyValue record : records) {
final long timestamp = keySchema.segmentTimestamp(Bytes.wrap(record.key));
observedStreamTime = Math.max(observedStreamTime, timestamp);
}
final Map writeBatchMap = new HashMap<>();
for (final KeyValue record : records) {
final long timestamp = keySchema.segmentTimestamp(Bytes.wrap(record.key));
final long segmentId = segments.segmentId(timestamp);
final S segment = segments.getOrCreateSegmentIfLive(segmentId, context, observedStreamTime);
if (segment != null) {
try {
final WriteBatch batch = writeBatchMap.computeIfAbsent(segment, s -> new WriteBatch());
segment.addToBatch(record, batch);
} catch (final RocksDBException e) {
throw new ProcessorStateException("Error restoring batch to store " + this.name, e);
}
}
}
return writeBatchMap;
}
private class RocksDBSegmentsBatchingRestoreCallback implements BatchingStateRestoreCallback {
@Override
public void restoreAll(final Collection> records) {
restoreAllInternal(records);
}
}
}