
com.hivemq.persistence.local.xodus.RetainedMessageXodusLocalPersistence Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of hivemq-community-edition-embedded Show documentation
Show all versions of hivemq-community-edition-embedded Show documentation
HiveMQ CE is a Java-based open source MQTT broker that fully supports MQTT 3.x and MQTT 5
The newest version!
/*
* Copyright 2019-present HiveMQ GmbH
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.hivemq.persistence.local.xodus;
import com.google.common.annotations.VisibleForTesting;
import com.google.common.collect.ImmutableMap;
import com.hivemq.bootstrap.ioc.lazysingleton.LazySingleton;
import com.hivemq.configuration.service.InternalConfigurations;
import com.hivemq.exceptions.UnrecoverableException;
import com.hivemq.extension.sdk.api.annotations.NotNull;
import com.hivemq.extension.sdk.api.annotations.Nullable;
import com.hivemq.extensions.iteration.BucketChunkResult;
import com.hivemq.migration.meta.PersistenceType;
import com.hivemq.persistence.PersistenceStartup;
import com.hivemq.persistence.RetainedMessage;
import com.hivemq.persistence.local.DeltaCounter;
import com.hivemq.persistence.local.xodus.bucket.Bucket;
import com.hivemq.persistence.payload.PublishPayloadPersistence;
import com.hivemq.persistence.retained.RetainedMessageLocalPersistence;
import com.hivemq.util.LocalPersistenceFileUtil;
import com.hivemq.util.ThreadPreConditions;
import jetbrains.exodus.ByteIterable;
import jetbrains.exodus.ExodusException;
import jetbrains.exodus.env.Cursor;
import jetbrains.exodus.env.StoreConfig;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import javax.annotation.PostConstruct;
import javax.inject.Inject;
import java.util.Map;
import java.util.Set;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.atomic.AtomicLong;
import static com.google.common.base.Preconditions.checkArgument;
import static com.google.common.base.Preconditions.checkNotNull;
import static com.hivemq.persistence.local.xodus.RetainedMessageSerializer.deserializeKey;
import static com.hivemq.persistence.local.xodus.RetainedMessageSerializer.deserializeValue;
import static com.hivemq.persistence.local.xodus.RetainedMessageSerializer.serializeKey;
import static com.hivemq.persistence.local.xodus.RetainedMessageSerializer.serializeValue;
import static com.hivemq.persistence.local.xodus.XodusUtils.byteIterableToBytes;
import static com.hivemq.persistence.local.xodus.XodusUtils.byteIterableToString;
import static com.hivemq.persistence.local.xodus.XodusUtils.bytesToByteIterable;
import static com.hivemq.persistence.local.xodus.XodusUtils.stringToByteIterable;
import static com.hivemq.util.ThreadPreConditions.SINGLE_WRITER_THREAD_PREFIX;
@LazySingleton
public class RetainedMessageXodusLocalPersistence extends XodusLocalPersistence
implements RetainedMessageLocalPersistence {
private static final Logger log = LoggerFactory.getLogger(RetainedMessageXodusLocalPersistence.class);
public static final String PERSISTENCE_VERSION = "040500";
private final @NotNull PublishPayloadPersistence payloadPersistence;
private final @NotNull AtomicLong retainMessageCounter = new AtomicLong(0);
@VisibleForTesting
final @NotNull ConcurrentHashMap topicTrees = new ConcurrentHashMap<>();
@Inject
public RetainedMessageXodusLocalPersistence(
final @NotNull LocalPersistenceFileUtil localPersistenceFileUtil,
final @NotNull PublishPayloadPersistence payloadPersistence,
final @NotNull EnvironmentUtil environmentUtil,
final @NotNull PersistenceStartup persistenceStartup) {
super(environmentUtil,
localPersistenceFileUtil,
persistenceStartup,
InternalConfigurations.PERSISTENCE_BUCKET_COUNT.get(),
//check if enabled
InternalConfigurations.RETAINED_MESSAGE_PERSISTENCE_TYPE.get().equals(PersistenceType.FILE));
this.payloadPersistence = payloadPersistence;
for (int i = 0; i < bucketCount; i++) {
topicTrees.put(i, new PublishTopicTree());
}
}
@Override
protected @NotNull String getName() {
return PERSISTENCE_NAME;
}
@Override
protected @NotNull String getVersion() {
return PERSISTENCE_VERSION;
}
@Override
protected @NotNull StoreConfig getStoreConfig() {
return StoreConfig.WITHOUT_DUPLICATES;
}
@Override
protected @NotNull Logger getLogger() {
return log;
}
@PostConstruct
protected void postConstruct() {
super.postConstruct();
}
@Override
public void init() {
try {
final DeltaCounter retainMessageDelta = DeltaCounter.finishWith(retainMessageCounter::addAndGet);
for (int i = 0; i < buckets.length; i++) {
final Bucket bucket = buckets[i];
final PublishTopicTree publishTopicTree = topicTrees.get(i);
bucket.getEnvironment().executeInReadonlyTransaction(txn -> {
try (final Cursor cursor = bucket.getStore().openCursor(txn)) {
while (cursor.getNext()) {
final RetainedMessage message = deserializeValue(byteIterableToBytes(cursor.getValue()));
payloadPersistence.incrementReferenceCounterOnBootstrap(message.getPublishId());
final String topic = deserializeKey(byteIterableToBytes(cursor.getKey()));
publishTopicTree.add(topic);
retainMessageDelta.increment();
}
}
});
}
// We can't attach the DeltaRunner as a commit hook because we do not flush here. Need to run it manually.
retainMessageDelta.run();
} catch (final ExodusException e) {
log.error("An error occurred while preparing the Retained Message persistence.");
log.debug("Original Exception:", e);
throw new UnrecoverableException(false);
}
}
@Override
public void clear(final int bucketIndex) {
ThreadPreConditions.startsWith(SINGLE_WRITER_THREAD_PREFIX);
topicTrees.put(bucketIndex, new PublishTopicTree());
final Bucket bucket = buckets[bucketIndex];
bucket.getEnvironment().executeInExclusiveTransaction(txn -> {
final DeltaCounter retainMessageDelta = DeltaCounter.finishWith(retainMessageCounter::addAndGet);
txn.setCommitHook(retainMessageDelta);
try (final Cursor cursor = bucket.getStore().openCursor(txn)) {
while (cursor.getNext()) {
final RetainedMessage message = deserializeValue(byteIterableToBytes(cursor.getValue()));
payloadPersistence.decrementReferenceCounter(message.getPublishId());
retainMessageDelta.decrement();
cursor.deleteCurrent();
}
}
});
}
@Override
public long size() {
return retainMessageCounter.get();
}
@Override
public void remove(final @NotNull String topic, final int bucketIndex) {
checkNotNull(topic, "Topic must not be null");
ThreadPreConditions.startsWith(SINGLE_WRITER_THREAD_PREFIX);
final Bucket bucket = buckets[bucketIndex];
bucket.getEnvironment().executeInExclusiveTransaction(txn -> {
final ByteIterable key = stringToByteIterable(topic);
final ByteIterable byteIterable = bucket.getStore().get(txn, bytesToByteIterable(serializeKey(topic)));
if (byteIterable == null) {
log.trace("Removing retained message for topic {} (no message was stored previously)", topic);
return;
}
final RetainedMessage message = deserializeValue(byteIterableToBytes(byteIterable));
log.trace("Removing retained message for topic {}", topic);
bucket.getStore().delete(txn, key);
topicTrees.get(bucketIndex).remove(topic);
payloadPersistence.decrementReferenceCounter(message.getPublishId());
retainMessageCounter.decrementAndGet();
});
}
@Override
public @Nullable RetainedMessage get(final @NotNull String topic, final int bucketIndex) {
checkNotNull(topic, "Topic must not be null");
ThreadPreConditions.startsWith(SINGLE_WRITER_THREAD_PREFIX);
final Bucket bucket = buckets[bucketIndex];
return bucket.getEnvironment().computeInReadonlyTransaction(txn -> {
final ByteIterable byteIterable = bucket.getStore().get(txn, bytesToByteIterable(serializeKey(topic)));
if (byteIterable != null) {
final RetainedMessage message = deserializeValue(byteIterableToBytes(byteIterable));
if (message.hasExpired()) {
return null;
}
final byte[] payload = payloadPersistence.get(message.getPublishId());
if (payload == null) {
log.warn("No payload was found for the retained message on topic {}.", topic);
return null;
}
message.setMessage(payload);
return message;
}
//Not found :(
return null;
});
}
@Override
public void put(
final @NotNull RetainedMessage retainedMessage, final @NotNull String topic, final int bucketIndex) {
checkNotNull(topic, "Topic must not be null");
checkNotNull(retainedMessage, "Retained message must not be null");
ThreadPreConditions.startsWith(SINGLE_WRITER_THREAD_PREFIX);
final Bucket bucket = buckets[bucketIndex];
bucket.getEnvironment().executeInExclusiveTransaction(txn -> {
try (final Cursor cursor = bucket.getStore().openCursor(txn)) {
final ByteIterable byteIterable = cursor.getSearchKey(bytesToByteIterable(serializeKey(topic)));
if (byteIterable != null) {
final RetainedMessage retainedMessageFromStore =
deserializeValue(byteIterableToBytes(cursor.getValue()));
log.trace("Replacing retained message for topic {}", topic);
txn.setCommitHook(() -> {
// The previous retained message is replaced, so we have to decrement the reference count.
payloadPersistence.decrementReferenceCounter(retainedMessageFromStore.getPublishId());
// And add the new payload
payloadPersistence.add(retainedMessage.getMessage(), retainedMessage.getPublishId());
});
bucket.getStore()
.put(txn,
bytesToByteIterable(serializeKey(topic)),
bytesToByteIterable(serializeValue(retainedMessage)));
} else {
txn.setCommitHook(() -> {
//persist needs increment.
retainMessageCounter.incrementAndGet();
topicTrees.get(bucketIndex).add(topic);
payloadPersistence.add(retainedMessage.getMessage(), retainedMessage.getPublishId());
});
bucket.getStore()
.put(txn,
bytesToByteIterable(serializeKey(topic)),
bytesToByteIterable(serializeValue(retainedMessage)));
log.trace("Creating new retained message for topic {}", topic);
}
}
});
}
@Override
public @NotNull Set getAllTopics(final @NotNull String subscription, final int bucketId) {
checkArgument(bucketId >= 0 && bucketId < bucketCount, "Bucket index out of range");
ThreadPreConditions.startsWith(SINGLE_WRITER_THREAD_PREFIX);
return topicTrees.get(bucketId).get(subscription);
}
@Override
public void cleanUp(final int bucketId) {
checkArgument(bucketId >= 0 && bucketId < bucketCount, "Bucket index out of range");
ThreadPreConditions.startsWith(SINGLE_WRITER_THREAD_PREFIX);
if (stopped.get()) {
return;
}
final Bucket bucket = buckets[bucketId];
bucket.getEnvironment().executeInExclusiveTransaction(txn -> {
try (final Cursor cursor = bucket.getStore().openCursor(txn)) {
while (cursor.getNext()) {
final RetainedMessage message = deserializeValue(byteIterableToBytes(cursor.getValue()));
if (message.hasExpired()) {
cursor.deleteCurrent();
payloadPersistence.decrementReferenceCounter(message.getPublishId());
retainMessageCounter.decrementAndGet();
topicTrees.get(bucketId).remove(deserializeKey(byteIterableToBytes(cursor.getKey())));
}
}
}
});
}
@Override
public @NotNull BucketChunkResult
© 2015 - 2025 Weber Informatics LLC | Privacy Policy