org.openmetadata.schema.services.connections.database.HiveConnection Maven / Gradle / Ivy
package org.openmetadata.schema.services.connections.database;
import java.util.HashMap;
import java.util.Map;
import javax.annotation.processing.Generated;
import javax.validation.Valid;
import javax.validation.constraints.NotNull;
import com.fasterxml.jackson.annotation.JsonCreator;
import com.fasterxml.jackson.annotation.JsonInclude;
import com.fasterxml.jackson.annotation.JsonProperty;
import com.fasterxml.jackson.annotation.JsonPropertyDescription;
import com.fasterxml.jackson.annotation.JsonPropertyOrder;
import com.fasterxml.jackson.annotation.JsonValue;
import org.openmetadata.annotations.PasswordField;
/**
* HiveConnection
*
* Hive SQL Connection Config
*
*/
@JsonInclude(JsonInclude.Include.NON_NULL)
@JsonPropertyOrder({
"type",
"scheme",
"username",
"password",
"hostPort",
"auth",
"kerberosServiceName",
"databaseSchema",
"databaseName",
"authOptions",
"metastoreConnection",
"connectionOptions",
"connectionArguments",
"supportsMetadataExtraction",
"supportsDBTExtraction",
"supportsProfiler",
"sampleDataStorageConfig"
})
@Generated("jsonschema2pojo")
public class HiveConnection {
/**
* Service type.
*
*/
@JsonProperty("type")
@JsonPropertyDescription("Service type.")
private HiveConnection.HiveType type = HiveConnection.HiveType.fromValue("Hive");
/**
* SQLAlchemy driver scheme options.
*
*/
@JsonProperty("scheme")
@JsonPropertyDescription("SQLAlchemy driver scheme options.")
private HiveConnection.HiveScheme scheme = HiveConnection.HiveScheme.fromValue("hive");
/**
* Username
*
* Username to connect to Hive. This user should have privileges to read all the metadata in Hive.
*
*/
@JsonProperty("username")
@JsonPropertyDescription("Username to connect to Hive. This user should have privileges to read all the metadata in Hive.")
private String username;
/**
* Password
*
* Password to connect to Hive.
*
*/
@JsonProperty("password")
@JsonPropertyDescription("Password to connect to Hive.")
@PasswordField
private String password;
/**
* Host and Port
*
* Host and port of the Hive service.
* (Required)
*
*/
@JsonProperty("hostPort")
@JsonPropertyDescription("Host and port of the Hive service.")
@NotNull
private String hostPort;
/**
* Authentication Mode
*
* Authentication mode to connect to hive.
*
*/
@JsonProperty("auth")
@JsonPropertyDescription("Authentication mode to connect to hive.")
private HiveConnection.Auth auth = HiveConnection.Auth.fromValue("NONE");
/**
* Kerberos Service Name
*
* If authenticating with Kerberos specify the Kerberos service name
*
*/
@JsonProperty("kerberosServiceName")
@JsonPropertyDescription("If authenticating with Kerberos specify the Kerberos service name")
private String kerberosServiceName;
/**
* Database Schema
*
* Database Schema of the data source. This is optional parameter, if you would like to restrict the metadata reading to a single schema. When left blank, OpenMetadata Ingestion attempts to scan all the schemas.
*
*/
@JsonProperty("databaseSchema")
@JsonPropertyDescription("Database Schema of the data source. This is optional parameter, if you would like to restrict the metadata reading to a single schema. When left blank, OpenMetadata Ingestion attempts to scan all the schemas.")
private String databaseSchema;
/**
* Database Name
*
* Optional name to give to the database in OpenMetadata. If left blank, we will use default as the database name.
*
*/
@JsonProperty("databaseName")
@JsonPropertyDescription("Optional name to give to the database in OpenMetadata. If left blank, we will use default as the database name.")
private String databaseName;
/**
* URL Authentication Options
*
* Authentication options to pass to Hive connector. These options are based on SQLAlchemy.
*
*/
@JsonProperty("authOptions")
@JsonPropertyDescription("Authentication options to pass to Hive connector. These options are based on SQLAlchemy.")
private String authOptions;
/**
* Hive Metastore Connection Details
*
* Hive Metastore Connection Details
*
*/
@JsonProperty("metastoreConnection")
@JsonPropertyDescription("Hive Metastore Connection Details")
private Object metastoreConnection;
/**
* Additional connection options to build the URL that can be sent to service during the connection.
*
*/
@JsonProperty("connectionOptions")
@JsonPropertyDescription("Additional connection options to build the URL that can be sent to service during the connection.")
@Valid
private ConnectionOptions connectionOptions;
/**
* Additional connection arguments such as security or protocol configs that can be sent to service during connection.
*
*/
@JsonProperty("connectionArguments")
@JsonPropertyDescription("Additional connection arguments such as security or protocol configs that can be sent to service during connection.")
@Valid
private ConnectionArguments connectionArguments;
/**
* Supports Metadata Extraction.
*
*/
@JsonProperty("supportsMetadataExtraction")
@JsonPropertyDescription("Supports Metadata Extraction.")
private Boolean supportsMetadataExtraction = true;
/**
* Supports DBT Extraction.
*
*/
@JsonProperty("supportsDBTExtraction")
@JsonPropertyDescription("Supports DBT Extraction.")
private Boolean supportsDBTExtraction = true;
/**
* Supports Profiler
*
*/
@JsonProperty("supportsProfiler")
@JsonPropertyDescription("Supports Profiler")
private Boolean supportsProfiler = true;
/**
* Storage config to store sample data
*
*/
@JsonProperty("sampleDataStorageConfig")
@JsonPropertyDescription("Storage config to store sample data")
@Valid
private SampleDataStorageConfig sampleDataStorageConfig;
/**
* Service type.
*
*/
@JsonProperty("type")
public HiveConnection.HiveType getType() {
return type;
}
/**
* Service type.
*
*/
@JsonProperty("type")
public void setType(HiveConnection.HiveType type) {
this.type = type;
}
public HiveConnection withType(HiveConnection.HiveType type) {
this.type = type;
return this;
}
/**
* SQLAlchemy driver scheme options.
*
*/
@JsonProperty("scheme")
public HiveConnection.HiveScheme getScheme() {
return scheme;
}
/**
* SQLAlchemy driver scheme options.
*
*/
@JsonProperty("scheme")
public void setScheme(HiveConnection.HiveScheme scheme) {
this.scheme = scheme;
}
public HiveConnection withScheme(HiveConnection.HiveScheme scheme) {
this.scheme = scheme;
return this;
}
/**
* Username
*
* Username to connect to Hive. This user should have privileges to read all the metadata in Hive.
*
*/
@JsonProperty("username")
public String getUsername() {
return username;
}
/**
* Username
*
* Username to connect to Hive. This user should have privileges to read all the metadata in Hive.
*
*/
@JsonProperty("username")
public void setUsername(String username) {
this.username = username;
}
public HiveConnection withUsername(String username) {
this.username = username;
return this;
}
/**
* Password
*
* Password to connect to Hive.
*
*/
@JsonProperty("password")
@PasswordField
public String getPassword() {
return password;
}
/**
* Password
*
* Password to connect to Hive.
*
*/
@JsonProperty("password")
@PasswordField
public void setPassword(String password) {
this.password = password;
}
public HiveConnection withPassword(String password) {
this.password = password;
return this;
}
/**
* Host and Port
*
* Host and port of the Hive service.
* (Required)
*
*/
@JsonProperty("hostPort")
public String getHostPort() {
return hostPort;
}
/**
* Host and Port
*
* Host and port of the Hive service.
* (Required)
*
*/
@JsonProperty("hostPort")
public void setHostPort(String hostPort) {
this.hostPort = hostPort;
}
public HiveConnection withHostPort(String hostPort) {
this.hostPort = hostPort;
return this;
}
/**
* Authentication Mode
*
* Authentication mode to connect to hive.
*
*/
@JsonProperty("auth")
public HiveConnection.Auth getAuth() {
return auth;
}
/**
* Authentication Mode
*
* Authentication mode to connect to hive.
*
*/
@JsonProperty("auth")
public void setAuth(HiveConnection.Auth auth) {
this.auth = auth;
}
public HiveConnection withAuth(HiveConnection.Auth auth) {
this.auth = auth;
return this;
}
/**
* Kerberos Service Name
*
* If authenticating with Kerberos specify the Kerberos service name
*
*/
@JsonProperty("kerberosServiceName")
public String getKerberosServiceName() {
return kerberosServiceName;
}
/**
* Kerberos Service Name
*
* If authenticating with Kerberos specify the Kerberos service name
*
*/
@JsonProperty("kerberosServiceName")
public void setKerberosServiceName(String kerberosServiceName) {
this.kerberosServiceName = kerberosServiceName;
}
public HiveConnection withKerberosServiceName(String kerberosServiceName) {
this.kerberosServiceName = kerberosServiceName;
return this;
}
/**
* Database Schema
*
* Database Schema of the data source. This is optional parameter, if you would like to restrict the metadata reading to a single schema. When left blank, OpenMetadata Ingestion attempts to scan all the schemas.
*
*/
@JsonProperty("databaseSchema")
public String getDatabaseSchema() {
return databaseSchema;
}
/**
* Database Schema
*
* Database Schema of the data source. This is optional parameter, if you would like to restrict the metadata reading to a single schema. When left blank, OpenMetadata Ingestion attempts to scan all the schemas.
*
*/
@JsonProperty("databaseSchema")
public void setDatabaseSchema(String databaseSchema) {
this.databaseSchema = databaseSchema;
}
public HiveConnection withDatabaseSchema(String databaseSchema) {
this.databaseSchema = databaseSchema;
return this;
}
/**
* Database Name
*
* Optional name to give to the database in OpenMetadata. If left blank, we will use default as the database name.
*
*/
@JsonProperty("databaseName")
public String getDatabaseName() {
return databaseName;
}
/**
* Database Name
*
* Optional name to give to the database in OpenMetadata. If left blank, we will use default as the database name.
*
*/
@JsonProperty("databaseName")
public void setDatabaseName(String databaseName) {
this.databaseName = databaseName;
}
public HiveConnection withDatabaseName(String databaseName) {
this.databaseName = databaseName;
return this;
}
/**
* URL Authentication Options
*
* Authentication options to pass to Hive connector. These options are based on SQLAlchemy.
*
*/
@JsonProperty("authOptions")
public String getAuthOptions() {
return authOptions;
}
/**
* URL Authentication Options
*
* Authentication options to pass to Hive connector. These options are based on SQLAlchemy.
*
*/
@JsonProperty("authOptions")
public void setAuthOptions(String authOptions) {
this.authOptions = authOptions;
}
public HiveConnection withAuthOptions(String authOptions) {
this.authOptions = authOptions;
return this;
}
/**
* Hive Metastore Connection Details
*
* Hive Metastore Connection Details
*
*/
@JsonProperty("metastoreConnection")
public Object getMetastoreConnection() {
return metastoreConnection;
}
/**
* Hive Metastore Connection Details
*
* Hive Metastore Connection Details
*
*/
@JsonProperty("metastoreConnection")
public void setMetastoreConnection(Object metastoreConnection) {
this.metastoreConnection = metastoreConnection;
}
public HiveConnection withMetastoreConnection(Object metastoreConnection) {
this.metastoreConnection = metastoreConnection;
return this;
}
/**
* Additional connection options to build the URL that can be sent to service during the connection.
*
*/
@JsonProperty("connectionOptions")
public ConnectionOptions getConnectionOptions() {
return connectionOptions;
}
/**
* Additional connection options to build the URL that can be sent to service during the connection.
*
*/
@JsonProperty("connectionOptions")
public void setConnectionOptions(ConnectionOptions connectionOptions) {
this.connectionOptions = connectionOptions;
}
public HiveConnection withConnectionOptions(ConnectionOptions connectionOptions) {
this.connectionOptions = connectionOptions;
return this;
}
/**
* Additional connection arguments such as security or protocol configs that can be sent to service during connection.
*
*/
@JsonProperty("connectionArguments")
public ConnectionArguments getConnectionArguments() {
return connectionArguments;
}
/**
* Additional connection arguments such as security or protocol configs that can be sent to service during connection.
*
*/
@JsonProperty("connectionArguments")
public void setConnectionArguments(ConnectionArguments connectionArguments) {
this.connectionArguments = connectionArguments;
}
public HiveConnection withConnectionArguments(ConnectionArguments connectionArguments) {
this.connectionArguments = connectionArguments;
return this;
}
/**
* Supports Metadata Extraction.
*
*/
@JsonProperty("supportsMetadataExtraction")
public Boolean getSupportsMetadataExtraction() {
return supportsMetadataExtraction;
}
/**
* Supports Metadata Extraction.
*
*/
@JsonProperty("supportsMetadataExtraction")
public void setSupportsMetadataExtraction(Boolean supportsMetadataExtraction) {
this.supportsMetadataExtraction = supportsMetadataExtraction;
}
public HiveConnection withSupportsMetadataExtraction(Boolean supportsMetadataExtraction) {
this.supportsMetadataExtraction = supportsMetadataExtraction;
return this;
}
/**
* Supports DBT Extraction.
*
*/
@JsonProperty("supportsDBTExtraction")
public Boolean getSupportsDBTExtraction() {
return supportsDBTExtraction;
}
/**
* Supports DBT Extraction.
*
*/
@JsonProperty("supportsDBTExtraction")
public void setSupportsDBTExtraction(Boolean supportsDBTExtraction) {
this.supportsDBTExtraction = supportsDBTExtraction;
}
public HiveConnection withSupportsDBTExtraction(Boolean supportsDBTExtraction) {
this.supportsDBTExtraction = supportsDBTExtraction;
return this;
}
/**
* Supports Profiler
*
*/
@JsonProperty("supportsProfiler")
public Boolean getSupportsProfiler() {
return supportsProfiler;
}
/**
* Supports Profiler
*
*/
@JsonProperty("supportsProfiler")
public void setSupportsProfiler(Boolean supportsProfiler) {
this.supportsProfiler = supportsProfiler;
}
public HiveConnection withSupportsProfiler(Boolean supportsProfiler) {
this.supportsProfiler = supportsProfiler;
return this;
}
/**
* Storage config to store sample data
*
*/
@JsonProperty("sampleDataStorageConfig")
public SampleDataStorageConfig getSampleDataStorageConfig() {
return sampleDataStorageConfig;
}
/**
* Storage config to store sample data
*
*/
@JsonProperty("sampleDataStorageConfig")
public void setSampleDataStorageConfig(SampleDataStorageConfig sampleDataStorageConfig) {
this.sampleDataStorageConfig = sampleDataStorageConfig;
}
public HiveConnection withSampleDataStorageConfig(SampleDataStorageConfig sampleDataStorageConfig) {
this.sampleDataStorageConfig = sampleDataStorageConfig;
return this;
}
@Override
public String toString() {
StringBuilder sb = new StringBuilder();
sb.append(HiveConnection.class.getName()).append('@').append(Integer.toHexString(System.identityHashCode(this))).append('[');
sb.append("type");
sb.append('=');
sb.append(((this.type == null)?"":this.type));
sb.append(',');
sb.append("scheme");
sb.append('=');
sb.append(((this.scheme == null)?"":this.scheme));
sb.append(',');
sb.append("username");
sb.append('=');
sb.append(((this.username == null)?"":this.username));
sb.append(',');
sb.append("password");
sb.append('=');
sb.append(((this.password == null)?"":this.password));
sb.append(',');
sb.append("hostPort");
sb.append('=');
sb.append(((this.hostPort == null)?"":this.hostPort));
sb.append(',');
sb.append("auth");
sb.append('=');
sb.append(((this.auth == null)?"":this.auth));
sb.append(',');
sb.append("kerberosServiceName");
sb.append('=');
sb.append(((this.kerberosServiceName == null)?"":this.kerberosServiceName));
sb.append(',');
sb.append("databaseSchema");
sb.append('=');
sb.append(((this.databaseSchema == null)?"":this.databaseSchema));
sb.append(',');
sb.append("databaseName");
sb.append('=');
sb.append(((this.databaseName == null)?"":this.databaseName));
sb.append(',');
sb.append("authOptions");
sb.append('=');
sb.append(((this.authOptions == null)?"":this.authOptions));
sb.append(',');
sb.append("metastoreConnection");
sb.append('=');
sb.append(((this.metastoreConnection == null)?"":this.metastoreConnection));
sb.append(',');
sb.append("connectionOptions");
sb.append('=');
sb.append(((this.connectionOptions == null)?"":this.connectionOptions));
sb.append(',');
sb.append("connectionArguments");
sb.append('=');
sb.append(((this.connectionArguments == null)?"":this.connectionArguments));
sb.append(',');
sb.append("supportsMetadataExtraction");
sb.append('=');
sb.append(((this.supportsMetadataExtraction == null)?"":this.supportsMetadataExtraction));
sb.append(',');
sb.append("supportsDBTExtraction");
sb.append('=');
sb.append(((this.supportsDBTExtraction == null)?"":this.supportsDBTExtraction));
sb.append(',');
sb.append("supportsProfiler");
sb.append('=');
sb.append(((this.supportsProfiler == null)?"":this.supportsProfiler));
sb.append(',');
sb.append("sampleDataStorageConfig");
sb.append('=');
sb.append(((this.sampleDataStorageConfig == null)?"":this.sampleDataStorageConfig));
sb.append(',');
if (sb.charAt((sb.length()- 1)) == ',') {
sb.setCharAt((sb.length()- 1), ']');
} else {
sb.append(']');
}
return sb.toString();
}
@Override
public int hashCode() {
int result = 1;
result = ((result* 31)+((this.databaseSchema == null)? 0 :this.databaseSchema.hashCode()));
result = ((result* 31)+((this.supportsMetadataExtraction == null)? 0 :this.supportsMetadataExtraction.hashCode()));
result = ((result* 31)+((this.authOptions == null)? 0 :this.authOptions.hashCode()));
result = ((result* 31)+((this.scheme == null)? 0 :this.scheme.hashCode()));
result = ((result* 31)+((this.databaseName == null)? 0 :this.databaseName.hashCode()));
result = ((result* 31)+((this.supportsProfiler == null)? 0 :this.supportsProfiler.hashCode()));
result = ((result* 31)+((this.auth == null)? 0 :this.auth.hashCode()));
result = ((result* 31)+((this.sampleDataStorageConfig == null)? 0 :this.sampleDataStorageConfig.hashCode()));
result = ((result* 31)+((this.type == null)? 0 :this.type.hashCode()));
result = ((result* 31)+((this.kerberosServiceName == null)? 0 :this.kerberosServiceName.hashCode()));
result = ((result* 31)+((this.supportsDBTExtraction == null)? 0 :this.supportsDBTExtraction.hashCode()));
result = ((result* 31)+((this.connectionArguments == null)? 0 :this.connectionArguments.hashCode()));
result = ((result* 31)+((this.password == null)? 0 :this.password.hashCode()));
result = ((result* 31)+((this.connectionOptions == null)? 0 :this.connectionOptions.hashCode()));
result = ((result* 31)+((this.metastoreConnection == null)? 0 :this.metastoreConnection.hashCode()));
result = ((result* 31)+((this.hostPort == null)? 0 :this.hostPort.hashCode()));
result = ((result* 31)+((this.username == null)? 0 :this.username.hashCode()));
return result;
}
@Override
public boolean equals(Object other) {
if (other == this) {
return true;
}
if ((other instanceof HiveConnection) == false) {
return false;
}
HiveConnection rhs = ((HiveConnection) other);
return ((((((((((((((((((this.databaseSchema == rhs.databaseSchema)||((this.databaseSchema!= null)&&this.databaseSchema.equals(rhs.databaseSchema)))&&((this.supportsMetadataExtraction == rhs.supportsMetadataExtraction)||((this.supportsMetadataExtraction!= null)&&this.supportsMetadataExtraction.equals(rhs.supportsMetadataExtraction))))&&((this.authOptions == rhs.authOptions)||((this.authOptions!= null)&&this.authOptions.equals(rhs.authOptions))))&&((this.scheme == rhs.scheme)||((this.scheme!= null)&&this.scheme.equals(rhs.scheme))))&&((this.databaseName == rhs.databaseName)||((this.databaseName!= null)&&this.databaseName.equals(rhs.databaseName))))&&((this.supportsProfiler == rhs.supportsProfiler)||((this.supportsProfiler!= null)&&this.supportsProfiler.equals(rhs.supportsProfiler))))&&((this.auth == rhs.auth)||((this.auth!= null)&&this.auth.equals(rhs.auth))))&&((this.sampleDataStorageConfig == rhs.sampleDataStorageConfig)||((this.sampleDataStorageConfig!= null)&&this.sampleDataStorageConfig.equals(rhs.sampleDataStorageConfig))))&&((this.type == rhs.type)||((this.type!= null)&&this.type.equals(rhs.type))))&&((this.kerberosServiceName == rhs.kerberosServiceName)||((this.kerberosServiceName!= null)&&this.kerberosServiceName.equals(rhs.kerberosServiceName))))&&((this.supportsDBTExtraction == rhs.supportsDBTExtraction)||((this.supportsDBTExtraction!= null)&&this.supportsDBTExtraction.equals(rhs.supportsDBTExtraction))))&&((this.connectionArguments == rhs.connectionArguments)||((this.connectionArguments!= null)&&this.connectionArguments.equals(rhs.connectionArguments))))&&((this.password == rhs.password)||((this.password!= null)&&this.password.equals(rhs.password))))&&((this.connectionOptions == rhs.connectionOptions)||((this.connectionOptions!= null)&&this.connectionOptions.equals(rhs.connectionOptions))))&&((this.metastoreConnection == rhs.metastoreConnection)||((this.metastoreConnection!= null)&&this.metastoreConnection.equals(rhs.metastoreConnection))))&&((this.hostPort == rhs.hostPort)||((this.hostPort!= null)&&this.hostPort.equals(rhs.hostPort))))&&((this.username == rhs.username)||((this.username!= null)&&this.username.equals(rhs.username))));
}
/**
* Authentication Mode
*
* Authentication mode to connect to hive.
*
*/
@Generated("jsonschema2pojo")
public enum Auth {
NONE("NONE"),
LDAP("LDAP"),
KERBEROS("KERBEROS"),
CUSTOM("CUSTOM"),
NOSASL("NOSASL"),
BASIC("BASIC"),
GSSAPI("GSSAPI"),
JWT("JWT"),
PLAIN("PLAIN");
private final String value;
private final static Map CONSTANTS = new HashMap();
static {
for (HiveConnection.Auth c: values()) {
CONSTANTS.put(c.value, c);
}
}
Auth(String value) {
this.value = value;
}
@Override
public String toString() {
return this.value;
}
@JsonValue
public String value() {
return this.value;
}
@JsonCreator
public static HiveConnection.Auth fromValue(String value) {
HiveConnection.Auth constant = CONSTANTS.get(value);
if (constant == null) {
throw new IllegalArgumentException(value);
} else {
return constant;
}
}
}
/**
* SQLAlchemy driver scheme options.
*
*/
@Generated("jsonschema2pojo")
public enum HiveScheme {
HIVE("hive"),
HIVE_HTTP("hive+http"),
HIVE_HTTPS("hive+https");
private final String value;
private final static Map CONSTANTS = new HashMap();
static {
for (HiveConnection.HiveScheme c: values()) {
CONSTANTS.put(c.value, c);
}
}
HiveScheme(String value) {
this.value = value;
}
@Override
public String toString() {
return this.value;
}
@JsonValue
public String value() {
return this.value;
}
@JsonCreator
public static HiveConnection.HiveScheme fromValue(String value) {
HiveConnection.HiveScheme constant = CONSTANTS.get(value);
if (constant == null) {
throw new IllegalArgumentException(value);
} else {
return constant;
}
}
}
/**
* Service type.
*
*/
@Generated("jsonschema2pojo")
public enum HiveType {
HIVE("Hive");
private final String value;
private final static Map CONSTANTS = new HashMap();
static {
for (HiveConnection.HiveType c: values()) {
CONSTANTS.put(c.value, c);
}
}
HiveType(String value) {
this.value = value;
}
@Override
public String toString() {
return this.value;
}
@JsonValue
public String value() {
return this.value;
}
@JsonCreator
public static HiveConnection.HiveType fromValue(String value) {
HiveConnection.HiveType constant = CONSTANTS.get(value);
if (constant == null) {
throw new IllegalArgumentException(value);
} else {
return constant;
}
}
}
}