org.apache.hive.service.cli.operation.ExecuteStatementOperation Maven / Gradle / Ivy
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hive.service.cli.operation;
import static org.apache.hadoop.hive.conf.Constants.MODE;
import static org.apache.hive.service.cli.operation.hplsql.HplSqlQueryExecutor.HPLSQL;
import static org.apache.hive.service.cli.operation.hplsql.HplSqlQueryExecutor.QUERY_EXECUTOR;
import java.sql.SQLException;
import java.util.Map;
import org.apache.hive.service.cli.operation.hplsql.BeelineConsole;
import org.apache.hadoop.hive.ql.exec.FunctionRegistry;
import org.apache.hadoop.hive.ql.parse.SemanticException;
import org.apache.hadoop.hive.ql.processors.CommandProcessor;
import org.apache.hadoop.hive.ql.processors.CommandProcessorFactory;
import org.apache.hadoop.hive.ql.session.SessionState;
import org.apache.hive.common.util.HiveStringUtils;
import org.apache.hive.hplsql.Conf;
import org.apache.hive.hplsql.Exec;
import org.apache.hive.hplsql.HplSqlSessionState;
import org.apache.hive.hplsql.ResultListener;
import org.apache.hive.hplsql.udf.Udf;
import org.apache.hive.service.cli.HiveSQLException;
import org.apache.hive.service.cli.OperationType;
import org.apache.hive.service.cli.operation.hplsql.BeelineConsole;
import org.apache.hive.service.cli.operation.hplsql.HplSqlOperation;
import org.apache.hive.service.cli.operation.hplsql.HplSqlQueryExecutor;
import org.apache.hive.service.cli.session.HiveSession;
public abstract class ExecuteStatementOperation extends Operation {
protected String statement = null;
public ExecuteStatementOperation(HiveSession parentSession, String statement,
Map confOverlay) {
super(parentSession, confOverlay, OperationType.EXECUTE_STATEMENT);
this.statement = statement;
}
public ExecuteStatementOperation(HiveSession parentSession, String statement, Map confOverlay, boolean runInBackground, boolean generateNewQueryId) {
super(parentSession, confOverlay, OperationType.EXECUTE_STATEMENT, generateNewQueryId);
this.statement = statement;
}
public String getStatement() {
return statement;
}
public static ExecuteStatementOperation newExecuteStatementOperation(HiveSession parentSession,
String statement, Map confOverlay, boolean runAsync, long queryTimeout)
throws HiveSQLException {
String cleanStatement = HiveStringUtils.removeComments(statement);
if (proceduralMode(confOverlay) && hplSqlMode()) {
if (SessionState.get().getDynamicVar(Exec.class) == null) {
Exec interpreter = new Exec(
new Conf(),
new BeelineConsole(),
new HplSqlQueryExecutor(parentSession),
parentSession.getMetaStoreClient(),
new HiveHplSqlSessionState(SessionState.get())
);
interpreter.init();
registerUdf();
SessionState.get().addDynamicVar(interpreter);
}
return new HplSqlOperation(parentSession, statement, confOverlay, runAsync, SessionState.get().getDynamicVar(Exec.class));
}
String[] tokens = cleanStatement.trim().split("\\s+");
CommandProcessor processor = null;
try {
processor = CommandProcessorFactory.getForHiveCommand(tokens, parentSession.getHiveConf());
} catch (SQLException e) {
throw new HiveSQLException(e.getMessage(), e.getSQLState(), e);
}
if (processor == null) {
// runAsync, queryTimeout makes sense only for a SQLOperation
// Pass the original statement to SQLOperation as sql parser can remove comments by itself
return new SQLOperation(parentSession, statement, confOverlay, runAsync, queryTimeout, hplSqlMode());
}
return new HiveCommandOperation(parentSession, cleanStatement, processor, confOverlay);
}
private static void registerUdf() throws HiveSQLException {
try {
if (FunctionRegistry.getTemporaryFunctionInfo(Udf.NAME) == null) {
FunctionRegistry.registerTemporaryUDF(Udf.NAME, org.apache.hive.hplsql.udf.Udf.class);
}
} catch (SemanticException e) {
throw new HiveSQLException(e);
}
}
private static boolean proceduralMode(Map confOverlay) {
return confOverlay != null && !HPLSQL.equals(confOverlay.get(QUERY_EXECUTOR));
}
public static boolean hplSqlMode() {
return HPLSQL.equalsIgnoreCase(SessionState.get().getHiveVariables().getOrDefault(MODE, ""));
}
private static class HiveHplSqlSessionState implements HplSqlSessionState {
private final SessionState sessionState;
public HiveHplSqlSessionState(SessionState sessionState) {
this.sessionState = sessionState;
}
@Override
public String currentUser() {
return sessionState.getUserName();
}
@Override
public String currentDatabase() {
return sessionState.getCurrentDatabase();
}
@Override
public String currentCatalog() {
return sessionState.getCurrentCatalog();
}
}
}
© 2015 - 2024 Weber Informatics LLC | Privacy Policy