All Downloads are FREE. Search and download functionalities are using the official Maven repository.

io.streamnative.pulsar.handlers.kop.migration.metadata.ManagedLedgerPropertiesMigrationMetadataManager Maven / Gradle / Ivy

There is a newer version: 4.0.0.4
Show newest version
/**
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package io.streamnative.pulsar.handlers.kop.migration.metadata;

import com.google.common.annotations.VisibleForTesting;
import com.google.common.collect.ImmutableMap;
import io.netty.channel.Channel;
import io.streamnative.pulsar.handlers.kop.AdminManager;
import io.streamnative.pulsar.handlers.kop.KafkaTopicLookupService;
import io.streamnative.pulsar.handlers.kop.utils.KopTopic;
import java.nio.ByteBuffer;
import java.util.ArrayList;
import java.util.Collections;
import java.util.Map;
import java.util.Properties;
import java.util.Set;
import java.util.concurrent.CompletableFuture;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.ConcurrentSkipListSet;
import lombok.extern.slf4j.Slf4j;
import org.apache.bookkeeper.mledger.ManagedLedger;
import org.apache.bookkeeper.mledger.ManagedLedgerException;
import org.apache.kafka.clients.admin.AdminClient;
import org.apache.kafka.clients.admin.AdminClientConfig;
import org.apache.kafka.clients.admin.TopicDescription;
import org.apache.kafka.clients.consumer.Consumer;
import org.apache.kafka.clients.consumer.KafkaConsumer;
import org.apache.kafka.clients.producer.KafkaProducer;
import org.apache.kafka.clients.producer.Producer;
import org.apache.kafka.common.KafkaFuture;
import org.apache.kafka.common.message.CreateTopicsRequestData;
import org.apache.kafka.common.requests.ApiError;
import org.apache.kafka.common.serialization.ByteBufferDeserializer;
import org.apache.kafka.common.serialization.ByteBufferSerializer;
import org.apache.kafka.common.serialization.StringDeserializer;
import org.apache.kafka.common.serialization.StringSerializer;

/**
 * A MigrationMetadata Manager that uses Managed Ledger properties for metadata storage.
 */
@Slf4j
public class ManagedLedgerPropertiesMigrationMetadataManager implements MigrationMetadataManager {

    @VisibleForTesting
    final Map numOutstandingRequests = new ConcurrentHashMap<>();
    @VisibleForTesting
    final Map> kafkaProducers = new ConcurrentHashMap<>();
    @VisibleForTesting
    final Map> kafkaConsumers = new ConcurrentHashMap<>();
    @VisibleForTesting
    final Map adminClients = new ConcurrentHashMap<>();

    // Cache topics that are not configured with migration so save lookup costs
    private final Set nonMigratoryTopics = new ConcurrentSkipListSet<>();

    private final AdminManager adminManager;
    private final KafkaTopicLookupService topicLookupService;

    public ManagedLedgerPropertiesMigrationMetadataManager(KafkaTopicLookupService topicLookupService,
                                                           AdminManager adminManager) {
        this.adminManager = adminManager;
        this.topicLookupService = topicLookupService;
    }

    @Override
    public KafkaProducer getKafkaProducerForTopic(String topic, String namespacePrefix,
                                                                      String kafkaClusterAddress) {
        String fullTopicName = new KopTopic(topic, namespacePrefix).getFullName();
        return kafkaProducers.computeIfAbsent(fullTopicName, key -> {
            Properties props = new Properties();
            props.put("bootstrap.servers", kafkaClusterAddress);
            props.put("key.serializer", StringSerializer.class);
            props.put("value.serializer", ByteBufferSerializer.class);
            return new KafkaProducer<>(props);
        });
    }

    @Override
    public KafkaConsumer getKafkaConsumerForTopic(String topic, String namespacePrefix,
                                                                      String kafkaClusterAddress) {
        String fullTopicName = new KopTopic(topic, namespacePrefix).getFullName();
        return kafkaConsumers.computeIfAbsent(fullTopicName, key -> {
            Properties props = new Properties();
            props.put("bootstrap.servers", kafkaClusterAddress);
            props.put("key.deserializer", StringDeserializer.class);
            props.put("value.deserializer", ByteBufferDeserializer.class);
            return new KafkaConsumer<>(props);
        });
    }

    @Override
    public AdminClient getAdminClientForKafka(String kafkaClusterAddress) {
        return adminClients.computeIfAbsent(kafkaClusterAddress, key -> {
            Properties props = new Properties();
            props.put(AdminClientConfig.BOOTSTRAP_SERVERS_CONFIG, kafkaClusterAddress);
            return AdminClient.create(props);
        });
    }

    private CompletableFuture getManagedLedger(String topic, String namespacePrefix,
                                                              Channel channel) {
        String fullPartitionName = KopTopic.toString(topic, 0, namespacePrefix);
        return topicLookupService.getTopic(fullPartitionName, channel).thenApply(persistentTopic -> {
            if (!persistentTopic.isPresent()) {
                throw new IllegalArgumentException("Cannot get topic " + fullPartitionName);
            }
            return persistentTopic.get().getManagedLedger();
        });
    }

    @Override
    public CompletableFuture getMigrationMetadata(String topic, String namespacePrefix,
                                                                     Channel channel) {
        if (nonMigratoryTopics.contains(topic)) {
            return CompletableFuture.completedFuture(null);
        }
        return getManagedLedger(topic, namespacePrefix, channel).thenApply(managedLedger -> {
            MigrationMetadata migrationMetadata = MigrationMetadata.fromProperties(managedLedger.getProperties());
            if (migrationMetadata == null) {
                nonMigratoryTopics.add(topic);
            }
            return migrationMetadata;
        });
    }

    @Override
    public CompletableFuture getKafkaClusterAddress(String topic, String namespacePrefix, Channel channel) {
        return getMigrationMetadata(topic, namespacePrefix, channel).thenApply(migrationMetadata -> {
            if (migrationMetadata == null) {
                return null;
            }
            return migrationMetadata.getKafkaClusterAddress();
        });
    }

    @Override
    public CompletableFuture getMigrationStatus(String topic, String namespacePrefix,
                                                                 Channel channel) {
        return getMigrationMetadata(topic, namespacePrefix, channel).thenApply(migrationMetadata -> {
            if (migrationMetadata == null) {
                return null;
            }
            return migrationMetadata.getMigrationStatus();
        });
    }

    @Override
    public void startProxyRequest(String topic, String namespacePrefix) {
        String fullTopicName = new KopTopic(topic, namespacePrefix).getFullName();
        numOutstandingRequests.put(fullTopicName, numOutstandingRequests.getOrDefault(topic, 0) + 1);
    }

    @Override
    public void finishProxyRequest(String topic, String namespacePrefix) {
        String fullTopicName = new KopTopic(topic, namespacePrefix).getFullName();
        numOutstandingRequests.compute(fullTopicName, (key, value) -> {
            if (value == null) {
                log.error("Cannot finish request for topic {}; no request has been proxied", topic);
                return null;
            }
            if (value == 0) {
                log.error("Cannot finish more requests than started for topic {}", topic);
                return 0;
            }
            return value - 1;
        });
    }

    private CompletableFuture setMigrationMetadata(String topic, String namespacePrefix,
                                                         MigrationMetadata migrationMetadata,
                                                         Channel channel) {
        return getManagedLedger(topic, namespacePrefix, channel).thenAccept(
                managedLedger -> {
                    try {
                        managedLedger.setProperties(migrationMetadata.asProperties());
                    } catch (InterruptedException | ManagedLedgerException e) {
                        throw new IllegalStateException(e);
                    }
                });
    }

    private CompletableFuture setMigrationStatus(String topic, String namespacePrefix, MigrationStatus status,
                                                       Channel channel) {
        return setMigrationMetadata(topic, namespacePrefix, MigrationMetadata.builder().migrationStatus(status).build(),
                channel);
    }

    @Override
    public CompletableFuture createWithMigration(String topic, String namespacePrefix, String kafkaClusterAddress,
                                                       Channel channel) {
        // TODO: Check authorization

        AdminClient adminClient = getAdminClientForKafka(kafkaClusterAddress);
        KafkaFuture future =
                new ArrayList<>(adminClient.describeTopics(Collections.singleton(topic)).values().values()).get(0);

        // https://gist.github.com/bmaggi/8e42a16a02f18d3bff9b0b742a75bfe7
        CompletableFuture wrappingFuture = new CompletableFuture<>();

        future.thenApply(topicDescription -> {
            log.error(topicDescription.toString());
            int numPartitions = topicDescription.partitions().size();
            int replicationFactor = topicDescription.partitions().get(0).replicas().size();
            adminManager.createTopicsAsync(ImmutableMap.of(topic,
                                    new CreateTopicsRequestData.CreatableTopic()
                                            .setNumPartitions(numPartitions)
                                            .setReplicationFactor((short) replicationFactor)),
                            1000,
                            namespacePrefix)
                    .thenCompose(validResult -> CompletableFuture.allOf(validResult.entrySet().stream().map(entry -> {
                        String key = entry.getKey();
                        ApiError value = entry.getValue();
                        if (!value.equals(ApiError.NONE)) {
                            throw value.exception();
                        }
                        log.info("Created topic partition: " + key + " with result " + value);
                        return setMigrationMetadata(topic, namespacePrefix,
                                MigrationMetadata.builder().kafkaClusterAddress(kafkaClusterAddress)
                                        .migrationStatus(MigrationStatus.NOT_STARTED).build(),
                                channel);
                    }).toArray(CompletableFuture[]::new))).join();
            return null;
        }).whenComplete((value, throwable) -> {
            if (throwable != null) {
                wrappingFuture.completeExceptionally(throwable);
            } else {
                wrappingFuture.complete(null);
            }
        });
        return wrappingFuture;
    }

    @Override
    public CompletableFuture migrate(String topic, String namespacePrefix, Channel channel) {
        // TODO: actually start the migration
        setMigrationStatus(topic, namespacePrefix, MigrationStatus.STARTED, channel);
        return CompletableFuture.completedFuture(null);
    }

    @Override
    public void close() {
        kafkaProducers.values().forEach(Producer::close);
        kafkaConsumers.values().forEach(Consumer::close);
        adminClients.values().forEach(AdminClient::close);
    }
}




© 2015 - 2024 Weber Informatics LLC | Privacy Policy