Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
// Generated by delombok at Thu Oct 10 18:56:32 UTC 2024
/* SPDX-License-Identifier: Apache-2.0
Copyright 2022 Atlan Pte. Ltd. */
package com.atlan.model.packages;
import com.atlan.AtlanClient;
import com.atlan.exception.AtlanException;
import com.atlan.exception.InvalidRequestException;
import com.atlan.model.enums.AtlanConnectorType;
import com.atlan.model.enums.AtlanPackageType;
import java.util.List;
import java.util.Map;
public class ConfluentKafkaCrawler extends AbstractCrawler {
public static final String PREFIX = AtlanPackageType.KAFKA_CONFLUENT_CLOUD.getValue();
/**
* Create the base configuration for a new Confluent Cloud Kafka crawler. Sets all admins as connection admins.
*
* @param client connectivity to an Atlan tenant
* @param connectionName name of the connection to create when running the crawler for the first time
* @return the builder for the base configuration of a Confluent Cloud Kafka crawler
* @throws AtlanException if there is not at least one connection admin specified, or any specified are invalid
*/
public static ConfluentKafkaCrawlerBuilder, ?> creator(AtlanClient client, String connectionName) throws AtlanException {
return creator(client, connectionName, List.of(client.getRoleCache().getIdForName("$admin")), null, null);
}
/**
* Create the base configuration for a new Confluent Cloud Kafka crawler.
*
* @param client connectivity to an Atlan tenant
* @param connectionName name of the connection to create when running the crawler for the first time
* @param adminRoles unique identifiers (GUIDs) of roles who will be connection admins on the connection
* @param adminGroups internal names of groups who will be connection admins on the connection
* @param adminUsers usernames of users who will be connection admins on the connection
* @return the builder for the base configuration of a Confluent Cloud Kafka crawler
* @throws AtlanException if there is not at least one connection admin specified, or any specified are invalid
*/
public static ConfluentKafkaCrawlerBuilder, ?> creator(AtlanClient client, String connectionName, List adminRoles, List adminGroups, List adminUsers) throws AtlanException {
return _internal().setup(PREFIX, "@atlan/kafka-confluent-cloud", client, getConnection(client, connectionName, AtlanConnectorType.CONFLUENT_KAFKA, adminRoles, adminGroups, adminUsers, false, false, 0L, "https://cdn.confluent.io/wp-content/uploads/apache-kafka-icon-2021-e1638496305992.jpg")).include(null).exclude(null).skipInternal(true);
}
public static abstract class ConfluentKafkaCrawlerBuilder> extends AbstractCrawlerBuilder {
/**
* Set up the crawler to extract directly from Kafka.
*
* @param bootstrap hostname and port number (host.example.com:9092) for the Kafka bootstrap server
* @param encrypted whether to use encrypted SSL connection (true), or plaintext (false)
* @return the builder, set up to extract directly from Kafka
*/
public B direct(String bootstrap, boolean encrypted) {
localCreds.name("default-confluent-kafka-" + epoch + "-0").host(bootstrap).port(9092).extra("security_protocol", encrypted ? "SASL_SSL" : "SASL_PLAINTEXT").connectorConfigName("atlan-connectors-kafka-confluent-cloud");
return this._parameter("extraction-method", "direct")._credential(localCreds);
}
/**
* Set up the crawler to use API token-based authentication.
*
* @param apiKey through which to access Kafka
* @param apiSecret through which to access Kafka
* @return the builder, set up to use API token-based authentication
*/
public B apiToken(String apiKey, String apiSecret) {
localCreds.authType("basic").username(apiKey).password(apiSecret);
return this._credential(localCreds);
}
/**
* Defines the filter for topics to include when crawling.
*
* @param regex any topic names that match this regular expression will be included in crawling
* @return the builder, set to include only those topics specified
* @throws InvalidRequestException in the unlikely event the provided filter cannot be translated
*/
public B include(String regex) throws InvalidRequestException {
return this._parameter("include-filter", regex);
}
/**
* Defines a regular expression to use for excluding topics when crawling.
*
* @param regex any topic names that match this regular expression will be excluded from crawling
* @return the builder, set to exclude any topics that match the provided regular expression
*/
public B exclude(String regex) {
return this._parameter("exclude-filter", regex);
}
/**
* Whether to skip internal topics when crawling (true) or include them.
*
* @param enabled if true, internal topics will be skipped when crawling
* @return the builder, set to include or exclude internal topics
*/
public B skipInternal(boolean enabled) {
return this._parameter("skip-internal-topics", "" + enabled);
}
/**
* Set all the metadata for the package (labels, annotations, etc).
*
* @return the builder, with metadata set
*/
@Override
protected B metadata() {
return this._label("orchestration.atlan.com/certified", "true")._label("orchestration.atlan.com/source", "confluent-kafka")._label("orchestration.atlan.com/sourceCategory", "eventbus")._label("orchestration.atlan.com/type", "connector")._label("orchestration.atlan.com/verified", "true")._label("package.argoproj.io/installer", "argopm")._label("package.argoproj.io/name", "a-t-ratlans-l-a-s-hkafka-confluent-cloud")._label("package.argoproj.io/registry", "httpsc-o-l-o-ns-l-a-s-hs-l-a-s-hpackages.atlan.com")._label("orchestration.atlan.com/default-confluent-kafka-" + epoch, "true")._label("orchestration.atlan.com/atlan-ui", "true")._annotation("orchestration.atlan.com/allowSchedule", "true")._annotation("orchestration.atlan.com/dependentPackage", "")._annotation("orchestration.atlan.com/docsUrl", "https://ask.atlan.com/hc/en-us/articles/6778924963599")._annotation("orchestration.atlan.com/emoji", "\ud83d\ude80")._annotation("orchestration.atlan.com/icon", "https://cdn.confluent.io/wp-content/uploads/apache-kafka-icon-2021-e1638496305992.jpg")._annotation("orchestration.atlan.com/logo", "https://cdn.confluent.io/wp-content/uploads/apache-kafka-icon-2021-e1638496305992.jpg")._annotation("orchestration.atlan.com/marketplaceLink", "https://packages.atlan.com/-/web/detail/@atlan/kafka-confluent-cloud")._annotation("orchestration.atlan.com/name", "Confluent Kafka Assets")._annotation("orchestration.atlan.com/usecase", "crawling,discovery")._annotation("package.argoproj.io/author", "Atlan")._annotation("package.argoproj.io/description", "Package to crawl Confluent Kafka assets and publish to Atlan for discovery.")._annotation("package.argoproj.io/homepage", "https://packages.atlan.com/-/web/detail/@atlan/kafka-confluent-cloud")._annotation("package.argoproj.io/keywords", "[\"kafka-confluent-cloud\",\"confluent-kafka\",\"eventbus\",\"connector\",\"kafka\"]")._annotation("package.argoproj.io/name", "@atlan/kafka-confluent-cloud")._annotation("package.argoproj.io/registry", "https://packages.atlan.com")._annotation("package.argoproj.io/repository", "git+https://github.com/atlanhq/marketplace-packages.git")._annotation("package.argoproj.io/support", "[email protected]")._annotation("orchestration.atlan.com/atlanName", PREFIX + "-default-confluent-kafka-" + epoch)._parameters(Map.ofEntries(Map.entry("credential-guid", "{{credentialGuid}}"), Map.entry("connection", connection.toJson(client)), Map.entry("publish-mode", "production"), Map.entry("atlas-auth-type", "internal")));
}
@java.lang.Override
@java.lang.SuppressWarnings("all")
@lombok.Generated
protected B $fillValuesFrom(final C instance) {
super.$fillValuesFrom(instance);
ConfluentKafkaCrawler.ConfluentKafkaCrawlerBuilder.$fillValuesFromInstanceIntoBuilder(instance, this);
return self();
}
@java.lang.SuppressWarnings("all")
@lombok.Generated
private static void $fillValuesFromInstanceIntoBuilder(final ConfluentKafkaCrawler instance, final ConfluentKafkaCrawler.ConfluentKafkaCrawlerBuilder, ?> b) {
}
@java.lang.Override
@java.lang.SuppressWarnings("all")
@lombok.Generated
protected abstract B self();
@java.lang.Override
@java.lang.SuppressWarnings("all")
@lombok.Generated
public abstract C build();
@java.lang.Override
@java.lang.SuppressWarnings("all")
@lombok.Generated
public java.lang.String toString() {
return "ConfluentKafkaCrawler.ConfluentKafkaCrawlerBuilder(super=" + super.toString() + ")";
}
}
@java.lang.SuppressWarnings("all")
@lombok.Generated
private static final class ConfluentKafkaCrawlerBuilderImpl extends ConfluentKafkaCrawler.ConfluentKafkaCrawlerBuilder {
@java.lang.SuppressWarnings("all")
@lombok.Generated
private ConfluentKafkaCrawlerBuilderImpl() {
}
@java.lang.Override
@java.lang.SuppressWarnings("all")
@lombok.Generated
protected ConfluentKafkaCrawler.ConfluentKafkaCrawlerBuilderImpl self() {
return this;
}
@java.lang.Override
@java.lang.SuppressWarnings("all")
@lombok.Generated
public ConfluentKafkaCrawler build() {
return new ConfluentKafkaCrawler(this);
}
}
@java.lang.SuppressWarnings("all")
@lombok.Generated
protected ConfluentKafkaCrawler(final ConfluentKafkaCrawler.ConfluentKafkaCrawlerBuilder, ?> b) {
super(b);
}
@java.lang.SuppressWarnings("all")
@lombok.Generated
public static ConfluentKafkaCrawler.ConfluentKafkaCrawlerBuilder, ?> _internal() {
return new ConfluentKafkaCrawler.ConfluentKafkaCrawlerBuilderImpl();
}
@java.lang.SuppressWarnings("all")
@lombok.Generated
public ConfluentKafkaCrawler.ConfluentKafkaCrawlerBuilder, ?> toBuilder() {
return new ConfluentKafkaCrawler.ConfluentKafkaCrawlerBuilderImpl().$fillValuesFrom(this);
}
@java.lang.Override
@java.lang.SuppressWarnings("all")
@lombok.Generated
public boolean equals(final java.lang.Object o) {
if (o == this) return true;
if (!(o instanceof ConfluentKafkaCrawler)) return false;
final ConfluentKafkaCrawler other = (ConfluentKafkaCrawler) o;
if (!other.canEqual((java.lang.Object) this)) return false;
if (!super.equals(o)) return false;
return true;
}
@java.lang.SuppressWarnings("all")
@lombok.Generated
protected boolean canEqual(final java.lang.Object other) {
return other instanceof ConfluentKafkaCrawler;
}
@java.lang.Override
@java.lang.SuppressWarnings("all")
@lombok.Generated
public int hashCode() {
final int result = super.hashCode();
return result;
}
@java.lang.Override
@java.lang.SuppressWarnings("all")
@lombok.Generated
public java.lang.String toString() {
return "ConfluentKafkaCrawler(super=" + super.toString() + ")";
}
}