Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
/*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package io.trino.plugin.hive;
import io.trino.hive.thrift.metastore.DataOperationType;
import io.trino.plugin.hive.acid.AcidOperation;
import io.trino.plugin.hive.acid.AcidTransaction;
import io.trino.plugin.hive.metastore.AcidTransactionOwner;
import io.trino.plugin.hive.metastore.Database;
import io.trino.plugin.hive.metastore.HiveColumnStatistics;
import io.trino.plugin.hive.metastore.HiveMetastore;
import io.trino.plugin.hive.metastore.HivePrincipal;
import io.trino.plugin.hive.metastore.HivePrivilegeInfo;
import io.trino.plugin.hive.metastore.HivePrivilegeInfo.HivePrivilege;
import io.trino.plugin.hive.metastore.Partition;
import io.trino.plugin.hive.metastore.PartitionWithStatistics;
import io.trino.plugin.hive.metastore.PrincipalPrivileges;
import io.trino.plugin.hive.metastore.StatisticsUpdateMode;
import io.trino.plugin.hive.metastore.Table;
import io.trino.plugin.hive.metastore.TableInfo;
import io.trino.plugin.hive.projection.PartitionProjection;
import io.trino.spi.TrinoException;
import io.trino.spi.connector.SchemaTableName;
import io.trino.spi.connector.TableNotFoundException;
import io.trino.spi.function.LanguageFunction;
import io.trino.spi.function.SchemaFunctionName;
import io.trino.spi.predicate.TupleDomain;
import io.trino.spi.security.RoleGrant;
import io.trino.spi.type.TypeManager;
import java.util.Collection;
import java.util.List;
import java.util.Map;
import java.util.Optional;
import java.util.OptionalLong;
import java.util.Set;
import static com.google.common.collect.ImmutableMap.toImmutableMap;
import static io.trino.plugin.hive.HiveErrorCode.HIVE_TABLE_DROPPED_DURING_QUERY;
import static io.trino.plugin.hive.projection.PartitionProjectionProperties.getPartitionProjectionFromTable;
import static java.util.Objects.requireNonNull;
public class HiveMetastoreClosure
{
private final HiveMetastore delegate;
private final TypeManager typeManager;
private final boolean partitionProjectionEnabled;
/**
* Do not use this directly. Instead, the closure should be fetched from the current SemiTransactionalHiveMetastore,
* which can be fetched from the current HiveMetadata.
*/
public HiveMetastoreClosure(HiveMetastore delegate, TypeManager typeManager, boolean partitionProjectionEnabled)
{
this.delegate = requireNonNull(delegate, "delegate is null");
this.typeManager = requireNonNull(typeManager, "typeManager is null");
this.partitionProjectionEnabled = partitionProjectionEnabled;
}
public Optional getDatabase(String databaseName)
{
return delegate.getDatabase(databaseName);
}
public List getAllDatabases()
{
return delegate.getAllDatabases();
}
private Table getExistingTable(String databaseName, String tableName)
{
return delegate.getTable(databaseName, tableName)
.orElseThrow(() -> new TableNotFoundException(new SchemaTableName(databaseName, tableName)));
}
public Optional
getTable(String databaseName, String tableName)
{
return delegate.getTable(databaseName, tableName);
}
public Map getTableColumnStatistics(String databaseName, String tableName, Set columnNames)
{
return delegate.getTableColumnStatistics(databaseName, tableName, columnNames);
}
public Map> getPartitionColumnStatistics(
String databaseName,
String tableName,
Set partitionNames,
Set columnNames)
{
return delegate.getPartitionColumnStatistics(databaseName, tableName, partitionNames, columnNames);
}
public boolean useSparkTableStatistics()
{
return delegate.useSparkTableStatistics();
}
public void updateTableStatistics(String databaseName, String tableName, AcidTransaction transaction, StatisticsUpdateMode mode, PartitionStatistics statisticsUpdate)
{
delegate.updateTableStatistics(databaseName, tableName, transaction, mode, statisticsUpdate);
}
public void updatePartitionStatistics(String databaseName, String tableName, StatisticsUpdateMode mode, Map partitionUpdates)
{
Table table = getExistingTable(databaseName, tableName);
delegate.updatePartitionStatistics(table, mode, partitionUpdates);
}
public List getTables(String databaseName)
{
return delegate.getTables(databaseName);
}
public void createDatabase(Database database)
{
delegate.createDatabase(database);
}
public void dropDatabase(String databaseName, boolean deleteData)
{
delegate.dropDatabase(databaseName, deleteData);
}
public void renameDatabase(String databaseName, String newDatabaseName)
{
delegate.renameDatabase(databaseName, newDatabaseName);
}
public void setDatabaseOwner(String databaseName, HivePrincipal principal)
{
delegate.setDatabaseOwner(databaseName, principal);
}
public void setTableOwner(String databaseName, String tableName, HivePrincipal principal)
{
delegate.setTableOwner(databaseName, tableName, principal);
}
public void createTable(Table table, PrincipalPrivileges principalPrivileges)
{
delegate.createTable(table, principalPrivileges);
}
public void dropTable(String databaseName, String tableName, boolean deleteData)
{
delegate.dropTable(databaseName, tableName, deleteData);
}
public void replaceTable(String databaseName, String tableName, Table newTable, PrincipalPrivileges principalPrivileges)
{
delegate.replaceTable(databaseName, tableName, newTable, principalPrivileges);
}
public void renameTable(String databaseName, String tableName, String newDatabaseName, String newTableName)
{
delegate.renameTable(databaseName, tableName, newDatabaseName, newTableName);
}
public void commentTable(String databaseName, String tableName, Optional comment)
{
delegate.commentTable(databaseName, tableName, comment);
}
public void commentColumn(String databaseName, String tableName, String columnName, Optional comment)
{
delegate.commentColumn(databaseName, tableName, columnName, comment);
}
public void addColumn(String databaseName, String tableName, String columnName, HiveType columnType, String columnComment)
{
delegate.addColumn(databaseName, tableName, columnName, columnType, columnComment);
}
public void renameColumn(String databaseName, String tableName, String oldColumnName, String newColumnName)
{
delegate.renameColumn(databaseName, tableName, oldColumnName, newColumnName);
}
public void dropColumn(String databaseName, String tableName, String columnName)
{
delegate.dropColumn(databaseName, tableName, columnName);
}
public Optional getPartition(String databaseName, String tableName, List partitionValues)
{
return delegate.getTable(databaseName, tableName)
.flatMap(table -> delegate.getPartition(table, partitionValues));
}
public Optional> getPartitionNamesByFilter(
String databaseName,
String tableName,
List columnNames,
TupleDomain partitionKeysFilter)
{
if (partitionProjectionEnabled) {
Table table = getTable(databaseName, tableName)
.orElseThrow(() -> new TrinoException(HIVE_TABLE_DROPPED_DURING_QUERY, "Table does not exists: " + tableName));
Optional projection = getPartitionProjectionFromTable(table, typeManager);
if (projection.isPresent()) {
return projection.get().getProjectedPartitionNamesByFilter(columnNames, partitionKeysFilter);
}
}
return delegate.getPartitionNamesByFilter(databaseName, tableName, columnNames, partitionKeysFilter);
}
public Map> getPartitionsByNames(String databaseName, String tableName, List partitionNames)
{
return delegate.getTable(databaseName, tableName)
.map(table -> getPartitionsByNames(table, partitionNames))
.orElseGet(() -> partitionNames.stream()
.collect(toImmutableMap(name -> name, name -> Optional.empty())));
}
private Map> getPartitionsByNames(Table table, List partitionNames)
{
if (partitionProjectionEnabled) {
Optional projection = getPartitionProjectionFromTable(table, typeManager);
if (projection.isPresent()) {
return projection.get().getProjectedPartitionsByNames(table, partitionNames);
}
}
return delegate.getPartitionsByNames(table, partitionNames);
}
public void addPartitions(String databaseName, String tableName, List partitions)
{
delegate.addPartitions(databaseName, tableName, partitions);
}
public void dropPartition(String databaseName, String tableName, List parts, boolean deleteData)
{
delegate.dropPartition(databaseName, tableName, parts, deleteData);
}
public void alterPartition(String databaseName, String tableName, PartitionWithStatistics partition)
{
delegate.alterPartition(databaseName, tableName, partition);
}
public void createRole(String role, String grantor)
{
delegate.createRole(role, grantor);
}
public void dropRole(String role)
{
delegate.dropRole(role);
}
public Set listRoles()
{
return delegate.listRoles();
}
public void grantRoles(Set roles, Set grantees, boolean adminOption, HivePrincipal grantor)
{
delegate.grantRoles(roles, grantees, adminOption, grantor);
}
public void revokeRoles(Set roles, Set grantees, boolean adminOption, HivePrincipal grantor)
{
delegate.revokeRoles(roles, grantees, adminOption, grantor);
}
public Set listRoleGrants(HivePrincipal principal)
{
return delegate.listRoleGrants(principal);
}
public void grantTablePrivileges(String databaseName, String tableName, String tableOwner, HivePrincipal grantee, HivePrincipal grantor, Set privileges, boolean grantOption)
{
delegate.grantTablePrivileges(databaseName, tableName, tableOwner, grantee, grantor, privileges, grantOption);
}
public void revokeTablePrivileges(String databaseName, String tableName, String tableOwner, HivePrincipal grantee, HivePrincipal grantor, Set privileges, boolean grantOption)
{
delegate.revokeTablePrivileges(databaseName, tableName, tableOwner, grantee, grantor, privileges, grantOption);
}
public Set listTablePrivileges(String databaseName, String tableName, Optional tableOwner, Optional principal)
{
return delegate.listTablePrivileges(databaseName, tableName, tableOwner, principal);
}
public void checkSupportsTransactions()
{
delegate.checkSupportsTransactions();
}
public long openTransaction(AcidTransactionOwner transactionOwner)
{
checkSupportsTransactions();
return delegate.openTransaction(transactionOwner);
}
public void commitTransaction(long transactionId)
{
delegate.commitTransaction(transactionId);
}
public void abortTransaction(long transactionId)
{
delegate.abortTransaction(transactionId);
}
public void sendTransactionHeartbeat(long transactionId)
{
delegate.sendTransactionHeartbeat(transactionId);
}
public void acquireSharedReadLock(
AcidTransactionOwner transactionOwner,
String queryId,
long transactionId,
List fullTables,
List partitions)
{
delegate.acquireSharedReadLock(transactionOwner, queryId, transactionId, fullTables, partitions);
}
public String getValidWriteIds(List tables, long currentTransactionId)
{
return delegate.getValidWriteIds(tables, currentTransactionId);
}
public Optional getConfigValue(String name)
{
return delegate.getConfigValue(name);
}
public long allocateWriteId(String dbName, String tableName, long transactionId)
{
return delegate.allocateWriteId(dbName, tableName, transactionId);
}
public void acquireTableWriteLock(
AcidTransactionOwner transactionOwner,
String queryId,
long transactionId,
String dbName,
String tableName,
DataOperationType operation,
boolean isPartitioned)
{
delegate.acquireTableWriteLock(transactionOwner, queryId, transactionId, dbName, tableName, operation, isPartitioned);
}
public void updateTableWriteId(String dbName, String tableName, long transactionId, long writeId, OptionalLong rowCountChange)
{
delegate.updateTableWriteId(dbName, tableName, transactionId, writeId, rowCountChange);
}
public void addDynamicPartitions(String dbName, String tableName, List partitionNames, long transactionId, long writeId, AcidOperation operation)
{
delegate.addDynamicPartitions(dbName, tableName, partitionNames, transactionId, writeId, operation);
}
public void alterTransactionalTable(Table table, long transactionId, long writeId, PrincipalPrivileges principalPrivileges)
{
delegate.alterTransactionalTable(table, transactionId, writeId, principalPrivileges);
}
public boolean functionExists(SchemaFunctionName name, String signatureToken)
{
return delegate.functionExists(name.getSchemaName(), name.getFunctionName(), signatureToken);
}
public Collection getAllFunctions(String schemaName)
{
return delegate.getAllFunctions(schemaName);
}
public Collection getFunctions(SchemaFunctionName name)
{
return delegate.getFunctions(name.getSchemaName(), name.getFunctionName());
}
public void createFunction(SchemaFunctionName name, LanguageFunction function)
{
delegate.createFunction(name.getSchemaName(), name.getFunctionName(), function);
}
public void replaceFunction(SchemaFunctionName name, LanguageFunction function)
{
delegate.replaceFunction(name.getSchemaName(), name.getFunctionName(), function);
}
public void dropFunction(SchemaFunctionName name, String signatureToken)
{
delegate.dropFunction(name.getSchemaName(), name.getFunctionName(), signatureToken);
}
}