Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hive.ql.session;
import static org.apache.hadoop.hive.metastore.MetaStoreUtils.DEFAULT_DATABASE_NAME;
import java.io.File;
import java.io.IOException;
import java.io.InputStream;
import java.io.PrintStream;
import java.lang.management.ManagementFactory;
import java.net.URI;
import java.net.URISyntaxException;
import java.net.URLClassLoader;
import java.sql.Timestamp;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.Collection;
import java.util.HashMap;
import java.util.HashSet;
import java.util.LinkedHashMap;
import java.util.LinkedList;
import java.util.List;
import java.util.Map;
import java.util.Set;
import java.util.UUID;
import java.util.concurrent.CancellationException;
import java.util.concurrent.locks.ReentrantLock;
import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.ArrayUtils;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.hive.common.FileUtils;
import org.apache.hadoop.hive.common.JavaUtils;
import org.apache.hadoop.hive.common.log.ProgressMonitor;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.conf.HiveConf.ConfVars;
import org.apache.hadoop.hive.metastore.ObjectStore;
import org.apache.hadoop.hive.metastore.api.ColumnStatisticsObj;
import org.apache.hadoop.hive.ql.MapRedStats;
import org.apache.hadoop.hive.ql.exec.Registry;
import org.apache.hadoop.hive.ql.exec.Utilities;
import org.apache.hadoop.hive.ql.exec.spark.session.SparkSession;
import org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionManagerImpl;
import org.apache.hadoop.hive.ql.exec.tez.TezSessionPoolManager;
import org.apache.hadoop.hive.ql.exec.tez.TezSessionState;
import org.apache.hadoop.hive.ql.history.HiveHistory;
import org.apache.hadoop.hive.ql.history.HiveHistoryImpl;
import org.apache.hadoop.hive.ql.history.HiveHistoryProxyHandler;
import org.apache.hadoop.hive.ql.lockmgr.HiveTxnManager;
import org.apache.hadoop.hive.ql.lockmgr.LockException;
import org.apache.hadoop.hive.ql.lockmgr.TxnManagerFactory;
import org.apache.hadoop.hive.ql.log.PerfLogger;
import org.apache.hadoop.hive.ql.metadata.Hive;
import org.apache.hadoop.hive.ql.metadata.HiveException;
import org.apache.hadoop.hive.ql.metadata.HiveUtils;
import org.apache.hadoop.hive.ql.metadata.Table;
import org.apache.hadoop.hive.ql.security.HiveAuthenticationProvider;
import org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider;
import org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook;
import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveAuthorizer;
import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveAuthorizerFactory;
import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveAuthzSessionContext;
import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveAuthzSessionContext.CLIENT_TYPE;
import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveMetastoreClientFactoryImpl;
import org.apache.hadoop.hive.ql.util.ResourceDownloader;
import org.apache.hadoop.hive.shims.HadoopShims;
import org.apache.hadoop.hive.shims.ShimLoader;
import org.apache.hadoop.hive.shims.Utils;
import org.apache.hadoop.security.UserGroupInformation;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import com.google.common.annotations.VisibleForTesting;
import com.google.common.base.Preconditions;
import com.google.common.collect.Maps;
/**
* SessionState encapsulates common data associated with a session.
*
* Also provides support for a thread static session object that can be accessed
* from any point in the code to interact with the user and to retrieve
* configuration information
*/
public class SessionState {
private static final Logger LOG = LoggerFactory.getLogger(SessionState.class);
private static final String TMP_PREFIX = "_tmp_space.db";
private static final String LOCAL_SESSION_PATH_KEY = "_hive.local.session.path";
private static final String HDFS_SESSION_PATH_KEY = "_hive.hdfs.session.path";
private static final String TMP_TABLE_SPACE_KEY = "_hive.tmp_table_space";
static final String LOCK_FILE_NAME = "inuse.lck";
static final String INFO_FILE_NAME = "inuse.info";
private final Map> tempTables = new HashMap>();
private final Map> tempTableColStats =
new HashMap>();
protected ClassLoader parentLoader;
// Session-scope compile lock.
private final ReentrantLock compileLock = new ReentrantLock();
/**
* current configuration.
*/
private final HiveConf sessionConf;
/**
* silent mode.
*/
protected boolean isSilent;
/**
* verbose mode
*/
protected boolean isVerbose;
/**
* The flag to indicate if the session serves the queries from HiveServer2 or not.
*/
private boolean isHiveServerQuery = false;
/**
* The flag to indicate if the session using thrift jdbc binary serde or not.
*/
private boolean isUsingThriftJDBCBinarySerDe = false;
/**
* The flag to indicate if the session already started so we can skip the init
*/
private boolean isStarted = false;
/*
* HiveHistory Object
*/
protected HiveHistory hiveHist;
/**
* Streams to read/write from.
*/
public InputStream in;
public PrintStream out;
public PrintStream info;
public PrintStream err;
/**
* Standard output from any child process(es).
*/
public PrintStream childOut;
/**
* Error output from any child process(es).
*/
public PrintStream childErr;
/**
* Temporary file name used to store results of non-Hive commands (e.g., set, dfs)
* and HiveServer.fetch*() function will read results from this file
*/
protected File tmpOutputFile;
/**
* Temporary file name used to store error output of executing non-Hive commands (e.g., set, dfs)
*/
protected File tmpErrOutputFile;
private String lastCommand;
private HiveAuthorizationProvider authorizer;
private HiveAuthorizer authorizerV2;
private volatile ProgressMonitor progressMonitor;
public enum AuthorizationMode{V1, V2};
private HiveAuthenticationProvider authenticator;
private CreateTableAutomaticGrant createTableGrants;
private Map mapRedStats;
private Map hiveVariables;
// A mapping from a hadoop job ID to the stack traces collected from the map reduce task logs
private Map>> stackTraces;
// This mapping collects all the configuration variables which have been set by the user
// explicitly, either via SET in the CLI, the hiveconf option, or a System property.
// It is a mapping from the variable name to its value. Note that if a user repeatedly
// changes the value of a variable, the corresponding change will be made in this mapping.
private Map overriddenConfigurations;
private Map> localMapRedErrors;
private TezSessionState tezSessionState;
private String currentDatabase;
private final String CONFIG_AUTHZ_SETTINGS_APPLIED_MARKER =
"hive.internal.ss.authz.settings.applied.marker";
private String userIpAddress;
private SparkSession sparkSession;
/**
* Gets information about HDFS encryption
*/
private Map hdfsEncryptionShims = Maps.newHashMap();
/**
* Lineage state.
*/
LineageState ls;
private final String userName;
/**
* scratch path to use for all non-local (ie. hdfs) file system tmp folders
* @return Path for Scratch path for the current session
*/
private Path hdfsSessionPath;
private FSDataOutputStream hdfsSessionPathLockFile = null;
/**
* sub dir of hdfs session path. used to keep tmp tables
* @return Path for temporary tables created by the current session
*/
private Path hdfsTmpTableSpace;
/**
* scratch directory to use for local file system tmp folders
* @return Path for local scratch directory for current session
*/
private Path localSessionPath;
private String hdfsScratchDirURIString;
/**
* Next value to use in naming a temporary table created by an insert...values statement
*/
private int nextValueTempTableSuffix = 1;
/**
* Transaction manager to use for this session. This is instantiated lazily by
* {@link #initTxnMgr(org.apache.hadoop.hive.conf.HiveConf)}
*/
private HiveTxnManager txnMgr = null;
/**
* store the jars loaded last time
*/
private final Set preReloadableAuxJars = new HashSet();
private final Registry registry;
/**
* CURRENT_TIMESTAMP value for query
*/
private Timestamp queryCurrentTimestamp;
private final ResourceMaps resourceMaps;
private final ResourceDownloader resourceDownloader;
private List forwardedAddresses;
private String atsDomainId;
/**
* Get the lineage state stored in this session.
*
* @return LineageState
*/
public LineageState getLineageState() {
return ls;
}
public HiveConf getConf() {
return sessionConf;
}
public File getTmpOutputFile() {
return tmpOutputFile;
}
public void setTmpOutputFile(File f) {
tmpOutputFile = f;
}
public File getTmpErrOutputFile() {
return tmpErrOutputFile;
}
public void setTmpErrOutputFile(File tmpErrOutputFile) {
this.tmpErrOutputFile = tmpErrOutputFile;
}
public void deleteTmpOutputFile() {
FileUtils.deleteTmpFile(tmpOutputFile);
}
public void deleteTmpErrOutputFile() {
FileUtils.deleteTmpFile(tmpErrOutputFile);
}
public boolean getIsSilent() {
if(sessionConf != null) {
return sessionConf.getBoolVar(HiveConf.ConfVars.HIVESESSIONSILENT);
} else {
return isSilent;
}
}
public boolean isHiveServerQuery() {
return this.isHiveServerQuery;
}
public void setIsSilent(boolean isSilent) {
if(sessionConf != null) {
sessionConf.setBoolVar(HiveConf.ConfVars.HIVESESSIONSILENT, isSilent);
}
this.isSilent = isSilent;
}
public ReentrantLock getCompileLock() {
return compileLock;
}
public boolean getIsVerbose() {
return isVerbose;
}
public void setIsVerbose(boolean isVerbose) {
this.isVerbose = isVerbose;
}
public void setIsUsingThriftJDBCBinarySerDe(boolean isUsingThriftJDBCBinarySerDe) {
this.isUsingThriftJDBCBinarySerDe = isUsingThriftJDBCBinarySerDe;
}
public boolean getIsUsingThriftJDBCBinarySerDe() {
return isUsingThriftJDBCBinarySerDe;
}
public void setIsHiveServerQuery(boolean isHiveServerQuery) {
this.isHiveServerQuery = isHiveServerQuery;
}
public SessionState(HiveConf conf) {
this(conf, null);
}
public SessionState(HiveConf conf, String userName) {
this.sessionConf = conf;
this.userName = userName;
this.registry = new Registry(false);
if (LOG.isDebugEnabled()) {
LOG.debug("SessionState user: " + userName);
}
isSilent = conf.getBoolVar(HiveConf.ConfVars.HIVESESSIONSILENT);
ls = new LineageState();
resourceMaps = new ResourceMaps();
// Must be deterministic order map for consistent q-test output across Java versions
overriddenConfigurations = new LinkedHashMap();
overriddenConfigurations.putAll(HiveConf.getConfSystemProperties());
// if there isn't already a session name, go ahead and create it.
if (StringUtils.isEmpty(conf.getVar(HiveConf.ConfVars.HIVESESSIONID))) {
conf.setVar(HiveConf.ConfVars.HIVESESSIONID, makeSessionId());
}
// Using system classloader as the parent. Using thread context
// classloader as parent can pollute the session. See HIVE-11878
parentLoader = SessionState.class.getClassLoader();
// Make sure that each session has its own UDFClassloader. For details see {@link UDFClassLoader}
final ClassLoader currentLoader = Utilities.createUDFClassLoader((URLClassLoader) parentLoader, new String[]{});
this.sessionConf.setClassLoader(currentLoader);
resourceDownloader = new ResourceDownloader(conf,
HiveConf.getVar(conf, ConfVars.DOWNLOADED_RESOURCES_DIR));
}
public Map getHiveVariables() {
if (hiveVariables == null) {
hiveVariables = new HashMap();
}
return hiveVariables;
}
public void setHiveVariables(Map hiveVariables) {
this.hiveVariables = hiveVariables;
}
public String getSessionId() {
return (sessionConf.getVar(HiveConf.ConfVars.HIVESESSIONID));
}
public void updateThreadName() {
final String sessionId = getSessionId();
final String logPrefix = getConf().getLogIdVar(sessionId);
final String currThreadName = Thread.currentThread().getName();
if (!currThreadName.contains(logPrefix)) {
final String newThreadName = logPrefix + " " + currThreadName;
LOG.info("Updating thread name to {}", newThreadName);
Thread.currentThread().setName(newThreadName);
}
}
public void resetThreadName() {
final String sessionId = getSessionId();
final String logPrefix = getConf().getLogIdVar(sessionId);
final String currThreadName = Thread.currentThread().getName();
if (currThreadName.contains(logPrefix)) {
final String[] names = currThreadName.split(logPrefix);
LOG.info("Resetting thread name to {}", names[names.length - 1]);
Thread.currentThread().setName(names[names.length - 1].trim());
}
}
/**
* Initialize the transaction manager. This is done lazily to avoid hard wiring one
* transaction manager at the beginning of the session.
* @param conf Hive configuration to initialize transaction manager
* @return transaction manager
* @throws LockException
*/
public synchronized HiveTxnManager initTxnMgr(HiveConf conf) throws LockException {
// Only change txnMgr if the setting has changed
if (txnMgr != null &&
!txnMgr.getTxnManagerName().equals(conf.getVar(HiveConf.ConfVars.HIVE_TXN_MANAGER))) {
txnMgr.closeTxnManager();
txnMgr = null;
}
if (txnMgr == null) {
txnMgr = TxnManagerFactory.getTxnManagerFactory().getTxnManager(conf);
}
return txnMgr;
}
public HiveTxnManager getTxnMgr() {
return txnMgr;
}
public HadoopShims.HdfsEncryptionShim getHdfsEncryptionShim() throws HiveException {
try {
return getHdfsEncryptionShim(FileSystem.get(sessionConf));
}
catch(HiveException hiveException) {
throw hiveException;
}
catch(Exception exception) {
throw new HiveException(exception);
}
}
public HadoopShims.HdfsEncryptionShim getHdfsEncryptionShim(FileSystem fs) throws HiveException {
if (!hdfsEncryptionShims.containsKey(fs.getUri())) {
try {
if ("hdfs".equals(fs.getUri().getScheme())) {
hdfsEncryptionShims.put(fs.getUri(), ShimLoader.getHadoopShims().createHdfsEncryptionShim(fs, sessionConf));
} else {
LOG.info("Could not get hdfsEncryptionShim, it is only applicable to hdfs filesystem.");
}
} catch (Exception e) {
throw new HiveException(e);
}
}
return hdfsEncryptionShims.get(fs.getUri());
}
// SessionState is not available in runtime and Hive.get().getConf() is not safe to call
private static class SessionStates {
private SessionState state;
private HiveConf conf;
private void attach(SessionState state) {
this.state = state;
attach(state.getConf());
}
private void attach(HiveConf conf) {
this.conf = conf;
ClassLoader classLoader = conf.getClassLoader();
if (classLoader != null) {
Thread.currentThread().setContextClassLoader(classLoader);
}
}
}
/**
* Singleton Session object per thread.
*
**/
private static ThreadLocal tss = new ThreadLocal() {
@Override
protected SessionStates initialValue() {
return new SessionStates();
}
};
/**
* start a new session and set it to current session.
*/
public static SessionState start(HiveConf conf) {
SessionState ss = new SessionState(conf);
return start(ss);
}
/**
* Sets the given session state in the thread local var for sessions.
*/
public static void setCurrentSessionState(SessionState startSs) {
tss.get().attach(startSs);
}
public static void detachSession() {
tss.remove();
}
/**
* set current session to existing session object if a thread is running
* multiple sessions - it must call this method with the new session object
* when switching from one session to another.
*/
public static SessionState start(SessionState startSs) {
start(startSs, false, null);
return startSs;
}
public static void beginStart(SessionState startSs, LogHelper console) {
start(startSs, true, console);
}
public static void endStart(SessionState startSs)
throws CancellationException, InterruptedException {
if (startSs.tezSessionState == null) return;
startSs.tezSessionState.endOpen();
}
synchronized private static void start(SessionState startSs, boolean isAsync, LogHelper console) {
setCurrentSessionState(startSs);
if (startSs.isStarted) {
return;
}
startSs.isStarted = true;
if (startSs.hiveHist == null){
if (startSs.getConf().getBoolVar(HiveConf.ConfVars.HIVE_SESSION_HISTORY_ENABLED)) {
startSs.hiveHist = new HiveHistoryImpl(startSs);
} else {
// Hive history is disabled, create a no-op proxy
startSs.hiveHist = HiveHistoryProxyHandler.getNoOpHiveHistoryProxy();
}
}
// Get the following out of the way when you start the session these take a
// while and should be done when we start up.
try {
UserGroupInformation sessionUGI = Utils.getUGI();
FileSystem.get(startSs.sessionConf);
// Create scratch dirs for this session
startSs.createSessionDirs(sessionUGI.getShortUserName());
// Set temp file containing results to be sent to HiveClient
if (startSs.getTmpOutputFile() == null) {
try {
startSs.setTmpOutputFile(createTempFile(startSs.getConf()));
} catch (IOException e) {
throw new RuntimeException(e);
}
}
// Set temp file containing error output to be sent to client
if (startSs.getTmpErrOutputFile() == null) {
try {
startSs.setTmpErrOutputFile(createTempFile(startSs.getConf()));
} catch (IOException e) {
throw new RuntimeException(e);
}
}
} catch (RuntimeException e) {
throw e;
} catch (Exception e) {
// Catch-all due to some exec time dependencies on session state
// that would cause ClassNoFoundException otherwise
throw new RuntimeException(e);
}
String engine = HiveConf.getVar(startSs.getConf(), HiveConf.ConfVars.HIVE_EXECUTION_ENGINE);
if (!engine.equals("tez") || startSs.isHiveServerQuery) return;
try {
if (startSs.tezSessionState == null) {
startSs.setTezSession(new TezSessionState(startSs.getSessionId()));
}
if (startSs.tezSessionState.isOpen()) {
return;
}
if (startSs.tezSessionState.isOpening()) {
if (!isAsync) {
startSs.tezSessionState.endOpen();
}
return;
}
// Neither open nor opening.
if (!isAsync) {
startSs.tezSessionState.open(startSs.sessionConf); // should use conf on session start-up
} else {
startSs.tezSessionState.beginOpen(startSs.sessionConf, null, console);
}
} catch (Exception e) {
throw new RuntimeException(e);
}
}
/**
* Create dirs & session paths for this session:
* 1. HDFS scratch dir
* 2. Local scratch dir
* 3. Local downloaded resource dir
* 4. HDFS session path
* 5. hold a lock file in HDFS session dir to indicate the it is in use
* 6. Local session path
* 7. HDFS temp table space
* @param userName
* @throws IOException
*/
private void createSessionDirs(String userName) throws IOException {
HiveConf conf = getConf();
Path rootHDFSDirPath = createRootHDFSDir(conf);
// Now create session specific dirs
String scratchDirPermission = HiveConf.getVar(conf, HiveConf.ConfVars.SCRATCHDIRPERMISSION);
Path path;
// 1. HDFS scratch dir
path = new Path(rootHDFSDirPath, userName);
hdfsScratchDirURIString = path.toUri().toString();
createPath(conf, path, scratchDirPermission, false, false);
// 2. Local scratch dir
path = new Path(HiveConf.getVar(conf, HiveConf.ConfVars.LOCALSCRATCHDIR));
createPath(conf, path, scratchDirPermission, true, false);
// 3. Download resources dir
path = new Path(HiveConf.getVar(conf, HiveConf.ConfVars.DOWNLOADED_RESOURCES_DIR));
createPath(conf, path, scratchDirPermission, true, false);
// Finally, create session paths for this session
// Local & non-local tmp location is configurable. however it is the same across
// all external file systems
String sessionId = getSessionId();
// 4. HDFS session path
hdfsSessionPath = new Path(hdfsScratchDirURIString, sessionId);
createPath(conf, hdfsSessionPath, scratchDirPermission, false, true);
conf.set(HDFS_SESSION_PATH_KEY, hdfsSessionPath.toUri().toString());
// 5. hold a lock file in HDFS session dir to indicate the it is in use
if (conf.getBoolVar(HiveConf.ConfVars.HIVE_SCRATCH_DIR_LOCK)) {
FileSystem fs = hdfsSessionPath.getFileSystem(conf);
FSDataOutputStream hdfsSessionPathInfoFile = fs.create(new Path(hdfsSessionPath, INFO_FILE_NAME),
true);
hdfsSessionPathInfoFile.writeUTF("process: " + ManagementFactory.getRuntimeMXBean().getName()
+"\n");
hdfsSessionPathInfoFile.close();
hdfsSessionPathLockFile = fs.create(new Path(hdfsSessionPath, LOCK_FILE_NAME), true);
}
// 6. Local session path
localSessionPath = new Path(HiveConf.getVar(conf, HiveConf.ConfVars.LOCALSCRATCHDIR), sessionId);
createPath(conf, localSessionPath, scratchDirPermission, true, true);
conf.set(LOCAL_SESSION_PATH_KEY, localSessionPath.toUri().toString());
// 7. HDFS temp table space
hdfsTmpTableSpace = new Path(hdfsSessionPath, TMP_PREFIX);
// This is a sub-dir under the hdfsSessionPath. Will be removed along with that dir.
// Don't register with deleteOnExit
createPath(conf, hdfsTmpTableSpace, scratchDirPermission, false, false);
conf.set(TMP_TABLE_SPACE_KEY, hdfsTmpTableSpace.toUri().toString());
}
/**
* Create the root scratch dir on hdfs (if it doesn't already exist) and make it writable
* @param conf
* @return
* @throws IOException
*/
private Path createRootHDFSDir(HiveConf conf) throws IOException {
Path rootHDFSDirPath = new Path(HiveConf.getVar(conf, HiveConf.ConfVars.SCRATCHDIR));
FsPermission writableHDFSDirPermission = new FsPermission((short)00733);
FileSystem fs = rootHDFSDirPath.getFileSystem(conf);
if (!fs.exists(rootHDFSDirPath)) {
Utilities.createDirsWithPermission(conf, rootHDFSDirPath, writableHDFSDirPermission, true);
}
FsPermission currentHDFSDirPermission = fs.getFileStatus(rootHDFSDirPath).getPermission();
if (rootHDFSDirPath != null && rootHDFSDirPath.toUri() != null) {
String schema = rootHDFSDirPath.toUri().getScheme();
LOG.debug(
"HDFS root scratch dir: " + rootHDFSDirPath + " with schema " + schema + ", permission: " +
currentHDFSDirPermission);
} else {
LOG.debug(
"HDFS root scratch dir: " + rootHDFSDirPath + ", permission: " + currentHDFSDirPermission);
}
// If the root HDFS scratch dir already exists, make sure it is writeable.
if (!((currentHDFSDirPermission.toShort() & writableHDFSDirPermission
.toShort()) == writableHDFSDirPermission.toShort())) {
throw new RuntimeException("The root scratch dir: " + rootHDFSDirPath
+ " on HDFS should be writable. Current permissions are: " + currentHDFSDirPermission);
}
return rootHDFSDirPath;
}
/**
* Create a given path if it doesn't exist.
*
* @param conf
* @param path
* @param permission
* @param isLocal
* @param isCleanUp
* @return
* @throws IOException
*/
private static void createPath(HiveConf conf, Path path, String permission, boolean isLocal,
boolean isCleanUp) throws IOException {
FsPermission fsPermission = new FsPermission(permission);
FileSystem fs;
if (isLocal) {
fs = FileSystem.getLocal(conf);
} else {
fs = path.getFileSystem(conf);
}
if (!fs.exists(path)) {
fs.mkdirs(path, fsPermission);
String dirType = isLocal ? "local" : "HDFS";
LOG.info("Created " + dirType + " directory: " + path.toString());
}
if (isCleanUp) {
fs.deleteOnExit(path);
}
}
public String getHdfsScratchDirURIString() {
return hdfsScratchDirURIString;
}
public static Path getLocalSessionPath(Configuration conf) {
SessionState ss = SessionState.get();
if (ss == null) {
String localPathString = conf.get(LOCAL_SESSION_PATH_KEY);
Preconditions.checkNotNull(localPathString,
"Conf local session path expected to be non-null");
return new Path(localPathString);
}
Preconditions.checkNotNull(ss.localSessionPath,
"Local session path expected to be non-null");
return ss.localSessionPath;
}
public static Path getHDFSSessionPath(Configuration conf) {
SessionState ss = SessionState.get();
if (ss == null) {
String sessionPathString = conf.get(HDFS_SESSION_PATH_KEY);
Preconditions.checkNotNull(sessionPathString,
"Conf non-local session path expected to be non-null");
return new Path(sessionPathString);
}
Preconditions.checkNotNull(ss.hdfsSessionPath,
"Non-local session path expected to be non-null");
return ss.hdfsSessionPath;
}
public static Path getTempTableSpace(Configuration conf) {
SessionState ss = SessionState.get();
if (ss == null) {
String tempTablePathString = conf.get(TMP_TABLE_SPACE_KEY);
Preconditions.checkNotNull(tempTablePathString,
"Conf temp table path expected to be non-null");
return new Path(tempTablePathString);
}
return ss.getTempTableSpace();
}
public Path getTempTableSpace() {
Preconditions.checkNotNull(this.hdfsTmpTableSpace,
"Temp table path expected to be non-null");
return this.hdfsTmpTableSpace;
}
@VisibleForTesting
void releaseSessionLockFile() throws IOException {
if (hdfsSessionPath != null && hdfsSessionPathLockFile != null) {
hdfsSessionPathLockFile.close();
}
}
private void dropSessionPaths(Configuration conf) throws IOException {
if (hdfsSessionPath != null) {
if (hdfsSessionPathLockFile != null) {
try {
hdfsSessionPathLockFile.close();
} catch (IOException e) {
LOG.error("Failed while closing remoteFsSessionLockFile", e);
}
}
dropPathAndUnregisterDeleteOnExit(hdfsSessionPath, conf, false);
}
if (localSessionPath != null) {
dropPathAndUnregisterDeleteOnExit(localSessionPath, conf, true);
}
deleteTmpOutputFile();
deleteTmpErrOutputFile();
}
private void dropPathAndUnregisterDeleteOnExit(Path path, Configuration conf, boolean localFs) {
FileSystem fs = null;
try {
if (localFs) {
fs = FileSystem.getLocal(conf);
} else {
fs = path.getFileSystem(conf);
}
fs.cancelDeleteOnExit(path);
fs.delete(path, true);
LOG.info("Deleted directory: {} on fs with scheme {}", path, fs.getScheme());
} catch (IOException e) {
LOG.error("Failed to delete path at {} on fs with scheme {}", path,
(fs == null ? "Unknown-null" : fs.getScheme()), e);
}
}
/**
* Setup authentication and authorization plugins for this session.
*/
private void setupAuth() {
if (authenticator != null) {
// auth has been initialized
return;
}
try {
authenticator = HiveUtils.getAuthenticator(sessionConf,
HiveConf.ConfVars.HIVE_AUTHENTICATOR_MANAGER);
authenticator.setSessionState(this);
String clsStr = HiveConf.getVar(sessionConf, HiveConf.ConfVars.HIVE_AUTHORIZATION_MANAGER);
authorizer = HiveUtils.getAuthorizeProviderManager(sessionConf,
clsStr, authenticator, true);
if (authorizer == null) {
// if it was null, the new (V2) authorization plugin must be specified in
// config
HiveAuthorizerFactory authorizerFactory = HiveUtils.getAuthorizerFactory(sessionConf,
HiveConf.ConfVars.HIVE_AUTHORIZATION_MANAGER);
HiveAuthzSessionContext.Builder authzContextBuilder = new HiveAuthzSessionContext.Builder();
authzContextBuilder.setClientType(isHiveServerQuery() ? CLIENT_TYPE.HIVESERVER2
: CLIENT_TYPE.HIVECLI);
authzContextBuilder.setSessionString(getSessionId());
authorizerV2 = authorizerFactory.createHiveAuthorizer(new HiveMetastoreClientFactoryImpl(),
sessionConf, authenticator, authzContextBuilder.build());
setAuthorizerV2Config();
}
// create the create table grants with new config
createTableGrants = CreateTableAutomaticGrant.create(sessionConf);
} catch (HiveException e) {
LOG.error("Error setting up authorization: " + e.getMessage(), e);
throw new RuntimeException(e);
}
if(LOG.isDebugEnabled()){
Object authorizationClass = getActiveAuthorizer();
LOG.debug("Session is using authorization class " + authorizationClass.getClass());
}
return;
}
private void setAuthorizerV2Config() throws HiveException {
// avoid processing the same config multiple times, check marker
if (sessionConf.get(CONFIG_AUTHZ_SETTINGS_APPLIED_MARKER, "").equals(Boolean.TRUE.toString())) {
return;
}
String metastoreHook = sessionConf.get(ConfVars.METASTORE_FILTER_HOOK.name());
if (!ConfVars.METASTORE_FILTER_HOOK.getDefaultValue().equals(metastoreHook) &&
!AuthorizationMetaStoreFilterHook.class.getName().equals(metastoreHook)) {
LOG.warn(ConfVars.METASTORE_FILTER_HOOK.name() +
" will be ignored, since hive.security.authorization.manager" +
" is set to instance of HiveAuthorizerFactory.");
}
sessionConf.setVar(ConfVars.METASTORE_FILTER_HOOK,
AuthorizationMetaStoreFilterHook.class.getName());
authorizerV2.applyAuthorizationConfigPolicy(sessionConf);
// update config in Hive thread local as well and init the metastore client
try {
Hive.get(sessionConf).getMSC();
} catch (Exception e) {
// catch-all due to some exec time dependencies on session state
// that would cause ClassNoFoundException otherwise
throw new HiveException(e.getMessage(), e);
}
// set a marker that this conf has been processed.
sessionConf.set(CONFIG_AUTHZ_SETTINGS_APPLIED_MARKER, Boolean.TRUE.toString());
}
public Object getActiveAuthorizer() {
return getAuthorizationMode() == AuthorizationMode.V1 ?
getAuthorizer() : getAuthorizerV2();
}
public Class> getAuthorizerInterface() {
return getAuthorizationMode() == AuthorizationMode.V1 ?
HiveAuthorizationProvider.class : HiveAuthorizer.class;
}
public void setActiveAuthorizer(Object authorizer) {
if (authorizer instanceof HiveAuthorizationProvider) {
this.authorizer = (HiveAuthorizationProvider)authorizer;
} else if (authorizer instanceof HiveAuthorizer) {
this.authorizerV2 = (HiveAuthorizer) authorizer;
} else if (authorizer != null) {
throw new IllegalArgumentException("Invalid authorizer " + authorizer);
}
}
/**
* @param conf
* @return per-session temp file
* @throws IOException
*/
private static File createTempFile(HiveConf conf) throws IOException {
String lScratchDir = HiveConf.getVar(conf, HiveConf.ConfVars.LOCALSCRATCHDIR);
String sessionID = conf.getVar(HiveConf.ConfVars.HIVESESSIONID);
return FileUtils.createTempFile(lScratchDir, sessionID, ".pipeout");
}
/**
* get the current session.
*/
public static SessionState get() {
return tss.get().state;
}
public static HiveConf getSessionConf() {
SessionStates state = tss.get();
if (state.conf == null) {
state.attach(new HiveConf());
}
return state.conf;
}
public static Registry getRegistry() {
SessionState session = get();
return session != null ? session.registry : null;
}
public static Registry getRegistryForWrite() {
Registry registry = getRegistry();
if (registry == null) {
throw new RuntimeException("Function registery for session is not initialized");
}
return registry;
}
/**
* get hiveHistory object which does structured logging.
*
* @return The hive history object
*/
public HiveHistory getHiveHistory() {
return hiveHist;
}
/**
* Update the history if set hive.session.history.enabled
*
* @param historyEnabled
* @param ss
*/
public void updateHistory(boolean historyEnabled, SessionState ss) {
if (historyEnabled) {
// Uses a no-op proxy
if (ss.hiveHist.getHistFileName() == null) {
ss.hiveHist = new HiveHistoryImpl(ss);
}
} else {
if (ss.hiveHist.getHistFileName() != null) {
ss.hiveHist = HiveHistoryProxyHandler.getNoOpHiveHistoryProxy();
}
}
}
/**
* Create a session ID. Looks like:
* $user_$pid@$host_$date
* @return the unique string
*/
private static String makeSessionId() {
return UUID.randomUUID().toString();
}
public String getLastCommand() {
return lastCommand;
}
public void setLastCommand(String lastCommand) {
this.lastCommand = lastCommand;
}
/**
* This class provides helper routines to emit informational and error
* messages to the user and log4j files while obeying the current session's
* verbosity levels.
*
* NEVER write directly to the SessionStates standard output other than to
* emit result data DO use printInfo and printError provided by LogHelper to
* emit non result data strings.
*
* It is perfectly acceptable to have global static LogHelper objects (for
* example - once per module) LogHelper always emits info/error to current
* session as required.
*/
public static class LogHelper {
protected Logger LOG;
protected boolean isSilent;
public LogHelper(Logger LOG) {
this(LOG, false);
}
public LogHelper(Logger LOG, boolean isSilent) {
this.LOG = LOG;
this.isSilent = isSilent;
}
public PrintStream getOutStream() {
SessionState ss = SessionState.get();
return ((ss != null) && (ss.out != null)) ? ss.out : System.out;
}
public static PrintStream getInfoStream() {
SessionState ss = SessionState.get();
return ((ss != null) && (ss.info != null)) ? ss.info : getErrStream();
}
public static PrintStream getErrStream() {
SessionState ss = SessionState.get();
return ((ss != null) && (ss.err != null)) ? ss.err : System.err;
}
public PrintStream getChildOutStream() {
SessionState ss = SessionState.get();
return ((ss != null) && (ss.childOut != null)) ? ss.childOut : System.out;
}
public PrintStream getChildErrStream() {
SessionState ss = SessionState.get();
return ((ss != null) && (ss.childErr != null)) ? ss.childErr : System.err;
}
public boolean getIsSilent() {
SessionState ss = SessionState.get();
// use the session or the one supplied in constructor
return (ss != null) ? ss.getIsSilent() : isSilent;
}
public void logInfo(String info) {
logInfo(info, null);
}
public void logInfo(String info, String detail) {
LOG.info(info + StringUtils.defaultString(detail));
}
public void printInfo(String info) {
printInfo(info, null);
}
public void printInfo(String info, boolean isSilent) {
printInfo(info, null, isSilent);
}
public void printInfo(String info, String detail) {
printInfo(info, detail, getIsSilent());
}
public void printInfo(String info, String detail, boolean isSilent) {
if (!isSilent) {
getInfoStream().println(info);
}
LOG.info(info + StringUtils.defaultString(detail));
}
public void printInfoNoLog(String info) {
if (!getIsSilent()) {
getInfoStream().println(info);
}
}
public void printError(String error) {
printError(error, null);
}
public void printError(String error, String detail) {
getErrStream().println(error);
LOG.error(error + StringUtils.defaultString(detail));
}
}
private static LogHelper _console;
/**
* initialize or retrieve console object for SessionState.
*/
public static LogHelper getConsole() {
if (_console == null) {
Logger LOG = LoggerFactory.getLogger("SessionState");
_console = new LogHelper(LOG);
}
return _console;
}
/**
*
* @return username from current SessionState authenticator. username will be
* null if there is no current SessionState object or authenticator is
* null.
*/
public static String getUserFromAuthenticator() {
if (SessionState.get() != null && SessionState.get().getAuthenticator() != null) {
return SessionState.get().getAuthenticator().getUserName();
}
return null;
}
static void validateFiles(List newFiles) throws IllegalArgumentException {
SessionState ss = SessionState.get();
Configuration conf = (ss == null) ? new Configuration() : ss.getConf();
for (String newFile : newFiles) {
try {
if (Utilities.realFile(newFile, conf) == null) {
String message = newFile + " does not exist";
throw new IllegalArgumentException(message);
}
} catch (IOException e) {
String message = "Unable to validate " + newFile;
throw new IllegalArgumentException(message, e);
}
}
}
/**
* Load the jars under the path specified in hive.aux.jars.path property. Add
* the jars to the classpath so the local task can refer to them.
* @throws IOException
*/
public void loadAuxJars() throws IOException {
String[] jarPaths = StringUtils.split(sessionConf.getAuxJars(), ',');
if (ArrayUtils.isEmpty(jarPaths)) return;
URLClassLoader currentCLoader =
(URLClassLoader) SessionState.get().getConf().getClassLoader();
currentCLoader =
(URLClassLoader) Utilities.addToClassPath(currentCLoader, jarPaths);
sessionConf.setClassLoader(currentCLoader);
Thread.currentThread().setContextClassLoader(currentCLoader);
}
/**
* Reload the jars under the path specified in hive.reloadable.aux.jars.path property.
* @throws IOException
*/
public void loadReloadableAuxJars() throws IOException {
final Set reloadedAuxJars = new HashSet();
final String renewableJarPath = sessionConf.getVar(ConfVars.HIVERELOADABLEJARS);
// do nothing if this property is not specified or empty
if (renewableJarPath == null || renewableJarPath.isEmpty()) {
return;
}
Set jarPaths = FileUtils.getJarFilesByPath(renewableJarPath, sessionConf);
// load jars under the hive.reloadable.aux.jars.path
if(!jarPaths.isEmpty()){
reloadedAuxJars.addAll(jarPaths);
}
// remove the previous renewable jars
if (preReloadableAuxJars != null && !preReloadableAuxJars.isEmpty()) {
Utilities.removeFromClassPath(preReloadableAuxJars.toArray(new String[0]));
}
if (reloadedAuxJars != null && !reloadedAuxJars.isEmpty()) {
URLClassLoader currentCLoader =
(URLClassLoader) SessionState.get().getConf().getClassLoader();
currentCLoader =
(URLClassLoader) Utilities.addToClassPath(currentCLoader,
reloadedAuxJars.toArray(new String[0]));
sessionConf.setClassLoader(currentCLoader);
Thread.currentThread().setContextClassLoader(currentCLoader);
}
preReloadableAuxJars.clear();
preReloadableAuxJars.addAll(reloadedAuxJars);
}
static void registerJars(List newJars) throws IllegalArgumentException {
LogHelper console = getConsole();
try {
ClassLoader loader = Thread.currentThread().getContextClassLoader();
ClassLoader newLoader = Utilities.addToClassPath(loader, newJars.toArray(new String[0]));
Thread.currentThread().setContextClassLoader(newLoader);
SessionState.get().getConf().setClassLoader(newLoader);
console.printInfo("Added " + newJars + " to class path");
} catch (Exception e) {
String message = "Unable to register " + newJars;
throw new IllegalArgumentException(message, e);
}
}
static boolean unregisterJar(List jarsToUnregister) {
LogHelper console = getConsole();
try {
Utilities.removeFromClassPath(jarsToUnregister.toArray(new String[0]));
console.printInfo("Deleted " + jarsToUnregister + " from class path");
return true;
} catch (IOException e) {
console.printError("Unable to unregister " + jarsToUnregister
+ "\nException: " + e.getMessage(), "\n"
+ org.apache.hadoop.util.StringUtils.stringifyException(e));
return false;
}
}
public String getATSDomainId() {
return atsDomainId;
}
public void setATSDomainId(String domainId) {
this.atsDomainId = domainId;
}
/**
* ResourceType.
*
*/
public static enum ResourceType {
FILE,
JAR {
@Override
public void preHook(Set cur, List s) throws IllegalArgumentException {
super.preHook(cur, s);
registerJars(s);
}
@Override
public void postHook(Set cur, List s) {
unregisterJar(s);
}
},
ARCHIVE;
public void preHook(Set cur, List s) throws IllegalArgumentException {
validateFiles(s);
}
public void postHook(Set cur, List s) {
}
};
public static ResourceType find_resource_type(String s) {
s = s.trim().toUpperCase();
try {
return ResourceType.valueOf(s);
} catch (IllegalArgumentException e) {
}
// try singular
if (s.endsWith("S")) {
s = s.substring(0, s.length() - 1);
} else {
return null;
}
try {
return ResourceType.valueOf(s);
} catch (IllegalArgumentException e) {
}
return null;
}
public String add_resource(ResourceType t, String value) throws RuntimeException {
return add_resource(t, value, false);
}
public String add_resource(ResourceType t, String value, boolean convertToUnix)
throws RuntimeException {
List added = add_resources(t, Arrays.asList(value), convertToUnix);
if (added == null || added.isEmpty()) {
return null;
}
return added.get(0);
}
public List add_resources(ResourceType t, Collection values)
throws RuntimeException {
// By default don't convert to unix
return add_resources(t, values, false);
}
public List add_resources(ResourceType t, Collection values, boolean convertToUnix)
throws RuntimeException {
Set resourceSet = resourceMaps.getResourceSet(t);
Map> resourcePathMap = resourceMaps.getResourcePathMap(t);
Map> reverseResourcePathMap = resourceMaps.getReverseResourcePathMap(t);
List localized = new ArrayList();
try {
for (String value : values) {
String key;
//get the local path of downloaded jars.
List downloadedURLs = resolveAndDownload(value, convertToUnix);
if (ResourceDownloader.isIvyUri(value)) {
// get the key to store in map
key = ResourceDownloader.createURI(value).getAuthority();
} else {
// for local file and hdfs, key and value are same.
key = downloadedURLs.get(0).toString();
}
Set downloadedValues = new HashSet();
for (URI uri : downloadedURLs) {
String resourceValue = uri.toString();
downloadedValues.add(resourceValue);
localized.add(resourceValue);
if (reverseResourcePathMap.containsKey(resourceValue)) {
if (!reverseResourcePathMap.get(resourceValue).contains(key)) {
reverseResourcePathMap.get(resourceValue).add(key);
}
} else {
Set addSet = new HashSet();
addSet.add(key);
reverseResourcePathMap.put(resourceValue, addSet);
}
}
resourcePathMap.put(key, downloadedValues);
}
t.preHook(resourceSet, localized);
} catch (RuntimeException e) {
getConsole().printError(e.getMessage(), "\n" + org.apache.hadoop.util.StringUtils.stringifyException(e));
throw e;
} catch (URISyntaxException e) {
getConsole().printError(e.getMessage());
throw new RuntimeException(e);
} catch (IOException e) {
getConsole().printError(e.getMessage());
throw new RuntimeException(e);
}
getConsole().printInfo("Added resources: " + values);
resourceSet.addAll(localized);
return localized;
}
@VisibleForTesting
protected List resolveAndDownload(String value, boolean convertToUnix)
throws URISyntaxException, IOException {
return resourceDownloader.resolveAndDownload(value, convertToUnix);
}
public void delete_resources(ResourceType t, List values) {
Set resources = resourceMaps.getResourceSet(t);
if (resources == null || resources.isEmpty()) {
return;
}
Map> resourcePathMap = resourceMaps.getResourcePathMap(t);
Map> reverseResourcePathMap = resourceMaps.getReverseResourcePathMap(t);
List deleteList = new LinkedList();
for (String value : values) {
String key = value;
try {
if (ResourceDownloader.isIvyUri(value)) {
key = ResourceDownloader.createURI(value).getAuthority();
}
} catch (URISyntaxException e) {
throw new RuntimeException("Invalid uri string " + value + ", " + e.getMessage());
}
// get all the dependencies to delete
Set resourcePaths = resourcePathMap.get(key);
if (resourcePaths == null) {
return;
}
for (String resourceValue : resourcePaths) {
reverseResourcePathMap.get(resourceValue).remove(key);
// delete a dependency only if no other resource depends on it.
if (reverseResourcePathMap.get(resourceValue).isEmpty()) {
deleteList.add(resourceValue);
reverseResourcePathMap.remove(resourceValue);
}
}
resourcePathMap.remove(key);
}
t.postHook(resources, deleteList);
resources.removeAll(deleteList);
}
public Set list_resource(ResourceType t, List filter) {
Set orig = resourceMaps.getResourceSet(t);
if (orig == null) {
return null;
}
if (filter == null) {
return orig;
} else {
Set fnl = new HashSet();
for (String one : orig) {
if (filter.contains(one)) {
fnl.add(one);
}
}
return fnl;
}
}
public void delete_resources(ResourceType t) {
Set resources = resourceMaps.getResourceSet(t);
if (resources != null && !resources.isEmpty()) {
delete_resources(t, new ArrayList(resources));
resourceMaps.getResourceMap().remove(t);
}
}
public HiveAuthorizationProvider getAuthorizer() {
setupAuth();
return authorizer;
}
public void setAuthorizer(HiveAuthorizationProvider authorizer) {
this.authorizer = authorizer;
}
public HiveAuthorizer getAuthorizerV2() {
setupAuth();
return authorizerV2;
}
public HiveAuthenticationProvider getAuthenticator() {
setupAuth();
return authenticator;
}
public void setAuthenticator(HiveAuthenticationProvider authenticator) {
this.authenticator = authenticator;
}
public CreateTableAutomaticGrant getCreateTableGrants() {
setupAuth();
return createTableGrants;
}
public void setCreateTableGrants(CreateTableAutomaticGrant createTableGrants) {
this.createTableGrants = createTableGrants;
}
public Map getMapRedStats() {
return mapRedStats;
}
public void setMapRedStats(Map mapRedStats) {
this.mapRedStats = mapRedStats;
}
public void setStackTraces(Map>> stackTraces) {
this.stackTraces = stackTraces;
}
public Map>> getStackTraces() {
return stackTraces;
}
public Map getOverriddenConfigurations() {
if (overriddenConfigurations == null) {
// Must be deterministic order map for consistent q-test output across Java versions
overriddenConfigurations = new LinkedHashMap();
}
return overriddenConfigurations;
}
public void setOverriddenConfigurations(Map overriddenConfigurations) {
this.overriddenConfigurations = overriddenConfigurations;
}
public Map> getLocalMapRedErrors() {
return localMapRedErrors;
}
public void addLocalMapRedErrors(String id, List localMapRedErrors) {
if (!this.localMapRedErrors.containsKey(id)) {
this.localMapRedErrors.put(id, new ArrayList());
}
this.localMapRedErrors.get(id).addAll(localMapRedErrors);
}
public void setLocalMapRedErrors(Map> localMapRedErrors) {
this.localMapRedErrors = localMapRedErrors;
}
public String getCurrentDatabase() {
if (currentDatabase == null) {
currentDatabase = DEFAULT_DATABASE_NAME;
}
return currentDatabase;
}
public void setCurrentDatabase(String currentDatabase) {
this.currentDatabase = currentDatabase;
}
public void close() throws IOException {
registry.clear();
if (txnMgr != null) txnMgr.closeTxnManager();
JavaUtils.closeClassLoadersTo(sessionConf.getClassLoader(), parentLoader);
File resourceDir =
new File(getConf().getVar(HiveConf.ConfVars.DOWNLOADED_RESOURCES_DIR));
LOG.debug("Removing resource dir " + resourceDir);
try {
if (resourceDir.exists()) {
FileUtils.deleteDirectory(resourceDir);
}
} catch (IOException e) {
LOG.info("Error removing session resource dir " + resourceDir, e);
} finally {
detachSession();
}
try {
if (tezSessionState != null) {
TezSessionPoolManager.closeIfNotDefault(tezSessionState, false);
}
} catch (Exception e) {
LOG.info("Error closing tez session", e);
} finally {
setTezSession(null);
}
try {
closeSparkSession();
registry.closeCUDFLoaders();
dropSessionPaths(sessionConf);
unCacheDataNucleusClassLoaders();
} finally {
// removes the threadlocal variables, closes underlying HMS connection
Hive.closeCurrent();
}
progressMonitor = null;
}
private void unCacheDataNucleusClassLoaders() {
try {
Hive threadLocalHive = Hive.get(sessionConf);
if ((threadLocalHive != null) && (threadLocalHive.getMSC() != null)
&& (threadLocalHive.getMSC().isLocalMetaStore())) {
if (sessionConf.getVar(ConfVars.METASTORE_RAW_STORE_IMPL).equals(ObjectStore.class.getName())) {
ObjectStore.unCacheDataNucleusClassLoaders();
}
}
} catch (Exception e) {
LOG.info("Failed to remove classloaders from DataNucleus ", e);
}
}
public void closeSparkSession() {
if (sparkSession != null) {
try {
SparkSessionManagerImpl.getInstance().closeSession(sparkSession);
} catch (Exception ex) {
LOG.error("Error closing spark session.", ex);
} finally {
sparkSession = null;
}
}
}
public AuthorizationMode getAuthorizationMode(){
setupAuth();
if(authorizer != null){
return AuthorizationMode.V1;
}else if(authorizerV2 != null){
return AuthorizationMode.V2;
}
//should not happen - this should not get called before this.start() is called
throw new AssertionError("Authorization plugins not initialized!");
}
public boolean isAuthorizationModeV2(){
return getAuthorizationMode() == AuthorizationMode.V2;
}
/**
* @return Tries to return an instance of the class whose name is configured in
* hive.exec.perf.logger, but if it can't it just returns an instance of
* the base PerfLogger class
*
*/
public static PerfLogger getPerfLogger() {
return getPerfLogger(false);
}
/**
* @param resetPerfLogger
* @return Tries to return an instance of the class whose name is configured in
* hive.exec.perf.logger, but if it can't it just returns an instance of
* the base PerfLogger class
*
*/
public static PerfLogger getPerfLogger(boolean resetPerfLogger) {
SessionState ss = get();
if (ss == null) {
return PerfLogger.getPerfLogger(null, resetPerfLogger);
} else {
return PerfLogger.getPerfLogger(ss.getConf(), resetPerfLogger);
}
}
public TezSessionState getTezSession() {
return tezSessionState;
}
/** Called from TezTask to attach a TezSession to use to the threadlocal. Ugly pattern... */
public void setTezSession(TezSessionState session) {
if (tezSessionState == session) return; // The same object.
if (tezSessionState != null) {
tezSessionState.markFree();
tezSessionState = null;
}
if (session != null) {
session.markInUse();
}
tezSessionState = session;
}
public String getUserName() {
return userName;
}
/**
* If authorization mode is v2, then pass it through authorizer so that it can apply
* any security configuration changes.
*/
public void applyAuthorizationPolicy() throws HiveException {
setupAuth();
}
public Map> getTempTables() {
return tempTables;
}
public Map> getTempTableColStats() {
return tempTableColStats;
}
/**
* @return ip address for user running the query
*/
public String getUserIpAddress() {
return userIpAddress;
}
/**
* set the ip address for user running the query
* @param userIpAddress
*/
public void setUserIpAddress(String userIpAddress) {
this.userIpAddress = userIpAddress;
}
public SparkSession getSparkSession() {
return sparkSession;
}
public void setSparkSession(SparkSession sparkSession) {
this.sparkSession = sparkSession;
}
/**
* Get the next suffix to use in naming a temporary table created by insert...values
* @return suffix
*/
public String getNextValuesTempTableSuffix() {
return Integer.toString(nextValueTempTableSuffix++);
}
/**
* Initialize current timestamp, other necessary query initialization.
*/
public void setupQueryCurrentTimestamp() {
queryCurrentTimestamp = new Timestamp(System.currentTimeMillis());
// Provide a facility to set current timestamp during tests
if (sessionConf.getBoolVar(ConfVars.HIVE_IN_TEST)) {
String overrideTimestampString =
HiveConf.getVar(sessionConf, HiveConf.ConfVars.HIVETESTCURRENTTIMESTAMP, (String)null);
if (overrideTimestampString != null && overrideTimestampString.length() > 0) {
queryCurrentTimestamp = Timestamp.valueOf(overrideTimestampString);
}
}
}
/**
* Get query current timestamp
* @return
*/
public Timestamp getQueryCurrentTimestamp() {
return queryCurrentTimestamp;
}
public ResourceDownloader getResourceDownloader() {
return resourceDownloader;
}
public void setForwardedAddresses(List forwardedAddresses) {
this.forwardedAddresses = forwardedAddresses;
}
public List getForwardedAddresses() {
return forwardedAddresses;
}
/**
* Gets the comma-separated reloadable aux jars
* @return the list of reloadable aux jars
*/
public String getReloadableAuxJars() {
return StringUtils.join(preReloadableAuxJars, ',');
}
public void updateProgressedPercentage(final double percentage) {
this.progressMonitor = new ProgressMonitor() {
@Override
public List headers() {
return null;
}
@Override
public List> rows() {
return null;
}
@Override
public String footerSummary() {
return null;
}
@Override
public long startTime() {
return 0;
}
@Override
public String executionStatus() {
return null;
}
@Override
public double progressedPercentage() {
return percentage;
}
};
}
public void updateProgressMonitor(ProgressMonitor progressMonitor) {
this.progressMonitor = progressMonitor;
}
public ProgressMonitor getProgressMonitor() {
return progressMonitor;
}
}
class ResourceMaps {
private final Map> resource_map;
//Given jar to add is stored as key and all its transitive dependencies as value. Used for deleting transitive dependencies.
private final Map>> resource_path_map;
// stores all the downloaded resources as key and the jars which depend on these resources as values in form of a list. Used for deleting transitive dependencies.
private final Map>> reverse_resource_path_map;
public ResourceMaps() {
resource_map = new HashMap>();
resource_path_map = new HashMap>>();
reverse_resource_path_map = new HashMap>>();
}
public Map> getResourceMap() {
return resource_map;
}
public Set getResourceSet(SessionState.ResourceType t) {
Set result = resource_map.get(t);
if (result == null) {
result = new HashSet();
resource_map.put(t, result);
}
return result;
}
public Map> getResourcePathMap(SessionState.ResourceType t) {
Map> result = resource_path_map.get(t);
if (result == null) {
result = new HashMap>();
resource_path_map.put(t, result);
}
return result;
}
public Map> getReverseResourcePathMap(SessionState.ResourceType t) {
Map> result = reverse_resource_path_map.get(t);
if (result == null) {
result = new HashMap>();
reverse_resource_path_map.put(t, result);
}
return result;
}
}