org.apache.hadoop.hive.ql.exec.SecureCmdDoAs Maven / Gradle / Ivy
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hive.ql.exec;
import java.io.File;
import java.io.IOException;
import java.net.URI;
import java.net.URISyntaxException;
import java.util.Map;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.ql.metadata.HiveException;
import org.apache.hadoop.hive.shims.ShimLoader;
import org.apache.hadoop.security.Credentials;
import org.apache.hadoop.security.UserGroupInformation;
import com.facebook.presto.hive.$internal.org.slf4j.Logger;
import com.facebook.presto.hive.$internal.org.slf4j.LoggerFactory;
/**
* SecureCmdDoAs - Helper class for setting parameters and env necessary for
* being able to run child jvm as intended user.
* Used only when kerberos security is used
*
*/
public class SecureCmdDoAs {
private static final Logger LOG = LoggerFactory.getLogger(SecureCmdDoAs.class);
private final Path tokenPath;
private final File tokenFile;
public SecureCmdDoAs(HiveConf conf) throws HiveException, IOException{
// Get delegation token for user from filesystem and write the token along with
// metastore tokens into a file
String uname = UserGroupInformation.getLoginUser().getShortUserName();
FileSystem fs = FileSystem.get(conf);
Credentials cred = new Credentials();
ShimLoader.getHadoopShims().addDelegationTokens(fs, cred, uname); // ask default fs first
for (String uri : conf.getStringCollection("mapreduce.job.hdfs-servers")) {
try {
ShimLoader.getHadoopShims().addDelegationTokens(
FileSystem.get(new URI(uri), conf),
cred, uname);
} catch (URISyntaxException e) {
LOG.warn("Invalid URI in mapreduce.job.hdfs-servers:["+uri+"], ignoring.", e);
}
}
tokenFile = File.createTempFile("hive_hadoop_delegation_token", null);
tokenPath = new Path(tokenFile.toURI());
//write credential with token to file
cred.writeTokenStorageFile(tokenPath, conf);
}
public void addEnv(Map env){
env.put(UserGroupInformation.HADOOP_TOKEN_FILE_LOCATION,
tokenPath.toUri().getPath());
}
public void close() {
tokenFile.delete();
}
}