org.apache.hadoop.hive.conf.Constants Maven / Gradle / Ivy
The newest version!
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hive.conf;
import java.util.regex.Pattern;
public class Constants {
/* Constants for Hive stats */
public static final String HIVE_ENGINE = "hive";
/* Constants for LLAP */
public static final String LLAP_LOGGER_NAME_QUERY_ROUTING = "query-routing";
public static final String LLAP_LOGGER_NAME_CONSOLE = "console";
public static final String LLAP_LOGGER_NAME_RFA = "RFA";
public static final String LLAP_NUM_BUCKETS = "llap.num.buckets";
public static final String LLAP_BUCKET_ID = "llap.bucket.id";
/* Constants for Druid storage handler */
public static final String DRUID_HIVE_STORAGE_HANDLER_ID =
"org.apache.hadoop.hive.druid.DruidStorageHandler";
public static final String DRUID_HIVE_OUTPUT_FORMAT =
"org.apache.hadoop.hive.druid.io.DruidOutputFormat";
public static final String DRUID_DATA_SOURCE = "druid.datasource";
public static final String DRUID_SEGMENT_GRANULARITY = "druid.segment.granularity";
public static final String DRUID_TARGET_SHARDS_PER_GRANULARITY =
"druid.segment.targetShardsPerGranularity";
public static final String DRUID_TIMESTAMP_GRANULARITY_COL_NAME = "__time_granularity";
public static final String DRUID_SHARD_KEY_COL_NAME = "__druid_extra_partition_key";
public static final String DRUID_QUERY_JSON = "druid.query.json";
public static final String DRUID_QUERY_FIELD_NAMES = "druid.fieldNames";
public static final String DRUID_QUERY_FIELD_TYPES = "druid.fieldTypes";
public static final String DRUID_QUERY_TYPE = "druid.query.type";
public static final String JDBC_HIVE_STORAGE_HANDLER_ID =
"org.apache.hive.storage.jdbc.JdbcStorageHandler";
public static final String JDBC_CONFIG_PREFIX = "hive.sql";
public static final String JDBC_CATALOG = JDBC_CONFIG_PREFIX + ".catalog";
public static final String JDBC_SCHEMA = JDBC_CONFIG_PREFIX + ".schema";
public static final String JDBC_TABLE = JDBC_CONFIG_PREFIX + ".table";
public static final String JDBC_DATABASE_TYPE = JDBC_CONFIG_PREFIX + ".database.type";
public static final String JDBC_URL = JDBC_CONFIG_PREFIX + ".jdbc.url";
public static final String JDBC_DRIVER = JDBC_CONFIG_PREFIX + ".jdbc.driver";
public static final String JDBC_USERNAME = JDBC_CONFIG_PREFIX + ".dbcp.username";
public static final String JDBC_PASSWORD = JDBC_CONFIG_PREFIX + ".dbcp.password";
public static final String JDBC_PASSWORD_URI = JDBC_CONFIG_PREFIX + ".dbcp.password.uri";
public static final String JDBC_KEYSTORE = JDBC_CONFIG_PREFIX + ".dbcp.password.keystore";
public static final String JDBC_KEY = JDBC_CONFIG_PREFIX + ".dbcp.password.key";
public static final String JDBC_QUERY = JDBC_CONFIG_PREFIX + ".query";
public static final String JDBC_QUERY_FIELD_NAMES = JDBC_CONFIG_PREFIX + ".query.fieldNames";
public static final String JDBC_QUERY_FIELD_TYPES = JDBC_CONFIG_PREFIX + ".query.fieldTypes";
public static final String JDBC_SPLIT_QUERY = JDBC_CONFIG_PREFIX + ".query.split";
public static final String JDBC_PARTITION_COLUMN = JDBC_CONFIG_PREFIX + ".partitionColumn";
public static final String JDBC_NUM_PARTITIONS = JDBC_CONFIG_PREFIX + ".numPartitions";
public static final String JDBC_LOW_BOUND = JDBC_CONFIG_PREFIX + ".lowerBound";
public static final String JDBC_UPPER_BOUND = JDBC_CONFIG_PREFIX + ".upperBound";
public static final String HIVE_SERVER2_JOB_CREDSTORE_PASSWORD_ENVVAR = "HIVE_JOB_CREDSTORE_PASSWORD";
public static final String HADOOP_CREDENTIAL_PASSWORD_ENVVAR = "HADOOP_CREDSTORE_PASSWORD";
public static final String HADOOP_CREDENTIAL_PROVIDER_PATH_CONFIG = "hadoop.security.credential.provider.path";
public static final String MATERIALIZED_VIEW_REWRITING_TIME_WINDOW = "rewriting.time.window";
public static final String MATERIALIZED_VIEW_SORT_COLUMNS = "materializedview.sort.columns";
public static final String MATERIALIZED_VIEW_DISTRIBUTE_COLUMNS = "materializedview.distribute.columns";
/** A named lock is acquired prior to executing the query; enabling to run queries in parallel which might interfere with eachother. */
public static final String HIVE_QUERY_EXCLUSIVE_LOCK = "hive.query.exclusive.lock";
public static final String SCHEDULED_QUERY_NAMESPACE = "scheduled.query.namespace";
public static final String SCHEDULED_QUERY_SCHEDULENAME = "scheduled.query.schedulename";
public static final String SCHEDULED_QUERY_EXECUTIONID = "scheduled.query.executionid";
public static final String SCHEDULED_QUERY_USER = "scheduled.query.user";
public static final String COMPACTOR_INTIATOR_THREAD_NAME_FORMAT = "Initiator-executor-thread-%d";
public static final String COMPACTOR_CLEANER_THREAD_NAME_FORMAT = "Cleaner-executor-thread-%d";
public static final String MODE = "mode";
public static final String ACID_FETCH_DELETED_ROWS = "acid.fetch.deleted.rows";
public static final String INSERT_ONLY_FETCH_BUCKET_ID = "insertonly.fetch.bucketid";
public static final String ERROR_MESSAGE_NO_DETAILS_AVAILABLE = "No detailed message available";
public static final String ORC_INPUT_FORMAT = "org.apache.hadoop.hive.ql.io.orc.OrcInputFormat";
public static final String ORC_OUTPUT_FORMAT = "org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat";
public static final Pattern COMPACTION_POOLS_PATTERN = Pattern.compile("hive\\.compactor\\.worker\\.(.*)\\.threads");
public static final String HIVE_COMPACTOR_WORKER_POOL = "hive.compactor.worker.pool";
public static final String HTTP_HEADER_REQUEST_TRACK = "X-Request-ID";
public static final String TIME_POSTFIX_REQUEST_TRACK = "_TIME";
public static final String ICEBERG_PARTITION_TABLE_SCHEMA = "partition,spec_id,record_count,file_count," +
"position_delete_record_count,position_delete_file_count,equality_delete_record_count," +
"equality_delete_file_count,last_updated_at,total_data_file_size_in_bytes,last_updated_snapshot_id";
public static final String DELIMITED_JSON_SERDE = "org.apache.hadoop.hive.serde2.DelimitedJSONSerDe";
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy