All Downloads are FREE. Search and download functionalities are using the official Maven repository.
Please wait. This can take some minutes ...
Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance.
Project price only 1 $
You can buy this project and download/modify it how often you want.
org.apache.hadoop.hive.ql.parse.ParseContext Maven / Gradle / Ivy
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hive.ql.parse;
import java.io.Serializable;
import java.util.HashMap;
import java.util.HashSet;
import java.util.List;
import java.util.Map;
import java.util.Set;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.ql.Context;
import org.apache.hadoop.hive.ql.QueryProperties;
import org.apache.hadoop.hive.ql.exec.AbstractMapJoinOperator;
import org.apache.hadoop.hive.ql.exec.FetchTask;
import org.apache.hadoop.hive.ql.exec.JoinOperator;
import org.apache.hadoop.hive.ql.exec.ListSinkOperator;
import org.apache.hadoop.hive.ql.exec.MapJoinOperator;
import org.apache.hadoop.hive.ql.exec.Operator;
import org.apache.hadoop.hive.ql.exec.ReduceSinkOperator;
import org.apache.hadoop.hive.ql.exec.SMBMapJoinOperator;
import org.apache.hadoop.hive.ql.exec.TableScanOperator;
import org.apache.hadoop.hive.ql.exec.Task;
import org.apache.hadoop.hive.ql.hooks.LineageInfo;
import org.apache.hadoop.hive.ql.hooks.ReadEntity;
import org.apache.hadoop.hive.ql.optimizer.ppr.PartitionPruner;
import org.apache.hadoop.hive.ql.optimizer.unionproc.UnionProcContext;
import org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.AnalyzeRewriteContext;
import org.apache.hadoop.hive.ql.plan.CreateTableDesc;
import org.apache.hadoop.hive.ql.plan.ExprNodeDesc;
import org.apache.hadoop.hive.ql.plan.FilterDesc.SampleDesc;
import org.apache.hadoop.hive.ql.plan.LoadFileDesc;
import org.apache.hadoop.hive.ql.plan.LoadTableDesc;
import org.apache.hadoop.hive.ql.plan.MapJoinDesc;
import org.apache.hadoop.hive.ql.plan.OperatorDesc;
import org.apache.hadoop.hive.ql.plan.TableDesc;
/**
* Parse Context: The current parse context. This is passed to the optimizer
* which then transforms the operator tree using the parse context. All the
* optimizations are performed sequentially and then the new parse context
* populated. Note that since the parse context contains the operator tree, it
* can be easily retrieved by the next optimization step or finally for task
* generation after the plan has been completely optimized.
*
**/
public class ParseContext {
private HashMap opToPartPruner;
private HashMap opToPartList;
private HashMap opToSamplePruner;
private Map> opToPartToSkewedPruner;
private HashMap> topOps;
private Set joinOps;
private Set mapJoinOps;
private Set smbMapJoinOps;
private List reduceSinkOperatorsAddedByEnforceBucketingSorting;
private HashMap nameToSplitSample;
private List loadTableWork;
private List loadFileWork;
private Context ctx;
private HiveConf conf;
private HashMap idToTableNameMap;
private int destTableId;
private UnionProcContext uCtx;
private List> listMapJoinOpsNoReducer; // list of map join
// operators with no
// reducer
private Map prunedPartitions;
private Map viewAliasToInput;
/**
* The lineage information.
*/
private LineageInfo lInfo;
private GlobalLimitCtx globalLimitCtx;
private HashSet semanticInputs;
private List> rootTasks;
private FetchTask fetchTask;
private QueryProperties queryProperties;
private TableDesc fetchTableDesc;
private Operator> fetchSource;
private ListSinkOperator fetchSink;
private AnalyzeRewriteContext analyzeRewrite;
private CreateTableDesc createTableDesc;
public ParseContext() {
}
/**
* @param conf
* @param qb
* current QB
* @param ast
* current parse tree
* @param opToPartPruner
* map from table scan operator to partition pruner
* @param opToPartList
* @param topOps
* list of operators for the top query
* @param opParseCtx
* operator parse context - contains a mapping from operator to
* operator parse state (row resolver etc.)
* @param joinOps
* context needed join processing (map join specifically)
* @param loadTableWork
* list of destination tables being loaded
* @param loadFileWork
* list of destination files being loaded
* @param ctx
* parse context
* @param idToTableNameMap
* @param uCtx
* @param destTableId
* @param listMapJoinOpsNoReducer
* list of map join operators with no reducer
* @param groupOpToInputTables
* @param prunedPartitions
* @param opToSamplePruner
* operator to sample pruner map
* @param globalLimitCtx
* @param nameToSplitSample
* @param rootTasks
*/
public ParseContext(
HiveConf conf,
HashMap opToPartPruner,
HashMap opToPartList,
HashMap> topOps,
Set joinOps,
Set smbMapJoinOps,
List loadTableWork, List loadFileWork,
Context ctx, HashMap idToTableNameMap, int destTableId,
UnionProcContext uCtx, List> listMapJoinOpsNoReducer,
Map prunedPartitions,
HashMap opToSamplePruner,
GlobalLimitCtx globalLimitCtx,
HashMap nameToSplitSample,
HashSet semanticInputs, List> rootTasks,
Map> opToPartToSkewedPruner,
Map viewAliasToInput,
List reduceSinkOperatorsAddedByEnforceBucketingSorting,
AnalyzeRewriteContext analyzeRewrite, CreateTableDesc createTableDesc,
QueryProperties queryProperties) {
this.conf = conf;
this.opToPartPruner = opToPartPruner;
this.opToPartList = opToPartList;
this.joinOps = joinOps;
this.smbMapJoinOps = smbMapJoinOps;
this.loadFileWork = loadFileWork;
this.loadTableWork = loadTableWork;
this.topOps = topOps;
this.ctx = ctx;
this.idToTableNameMap = idToTableNameMap;
this.destTableId = destTableId;
this.uCtx = uCtx;
this.listMapJoinOpsNoReducer = listMapJoinOpsNoReducer;
this.prunedPartitions = prunedPartitions;
this.opToSamplePruner = opToSamplePruner;
this.nameToSplitSample = nameToSplitSample;
this.globalLimitCtx = globalLimitCtx;
this.semanticInputs = semanticInputs;
this.rootTasks = rootTasks;
this.opToPartToSkewedPruner = opToPartToSkewedPruner;
this.viewAliasToInput = viewAliasToInput;
this.reduceSinkOperatorsAddedByEnforceBucketingSorting =
reduceSinkOperatorsAddedByEnforceBucketingSorting;
this.analyzeRewrite = analyzeRewrite;
this.createTableDesc = createTableDesc;
this.queryProperties = queryProperties;
}
/**
* @return the context
*/
public Context getContext() {
return ctx;
}
/**
* @param ctx
* the context to set
*/
public void setContext(Context ctx) {
this.ctx = ctx;
}
/**
* @return the hive conf
*/
public HiveConf getConf() {
return conf;
}
/**
* @param conf
* the conf to set
*/
public void setConf(HiveConf conf) {
this.conf = conf;
}
/**
* @return the opToPartPruner
*/
public HashMap getOpToPartPruner() {
return opToPartPruner;
}
/**
* @param opToPartPruner
* the opToPartPruner to set
*/
public void setOpToPartPruner(
HashMap opToPartPruner) {
this.opToPartPruner = opToPartPruner;
}
public HashMap getOpToPartList() {
return opToPartList;
}
public List getReduceSinkOperatorsAddedByEnforceBucketingSorting() {
return reduceSinkOperatorsAddedByEnforceBucketingSorting;
}
public void setReduceSinkOperatorsAddedByEnforceBucketingSorting(
List reduceSinkOperatorsAddedByEnforceBucketingSorting) {
this.reduceSinkOperatorsAddedByEnforceBucketingSorting =
reduceSinkOperatorsAddedByEnforceBucketingSorting;
}
/**
* @return the topOps
*/
public HashMap> getTopOps() {
return topOps;
}
/**
* @param topOps
* the topOps to set
*/
public void setTopOps(HashMap> topOps) {
this.topOps = topOps;
}
public HashMap getNameToSplitSample() {
return nameToSplitSample;
}
public void setNameToSplitSample(HashMap nameToSplitSample) {
this.nameToSplitSample = nameToSplitSample;
}
/**
* @return the loadTableWork
*/
public List getLoadTableWork() {
return loadTableWork;
}
/**
* @param loadTableWork
* the loadTableWork to set
*/
public void setLoadTableWork(List loadTableWork) {
this.loadTableWork = loadTableWork;
}
/**
* @return the loadFileWork
*/
public List getLoadFileWork() {
return loadFileWork;
}
/**
* @param loadFileWork
* the loadFileWork to set
*/
public void setLoadFileWork(List loadFileWork) {
this.loadFileWork = loadFileWork;
}
public HashMap getIdToTableNameMap() {
return idToTableNameMap;
}
public void setIdToTableNameMap(HashMap idToTableNameMap) {
this.idToTableNameMap = idToTableNameMap;
}
public int getDestTableId() {
return destTableId;
}
public void setDestTableId(int destTableId) {
this.destTableId = destTableId;
}
public UnionProcContext getUCtx() {
return uCtx;
}
public void setUCtx(UnionProcContext uCtx) {
this.uCtx = uCtx;
}
/**
* @return the joinOps
*/
public Set getJoinOps() {
return joinOps;
}
/**
* @param joinOps
* the joinOps to set
*/
public void setJoinOps(Set joinOps) {
this.joinOps = joinOps;
}
/**
* @return the listMapJoinOpsNoReducer
*/
public List> getListMapJoinOpsNoReducer() {
return listMapJoinOpsNoReducer;
}
/**
* @param listMapJoinOpsNoReducer
* the listMapJoinOpsNoReducer to set
*/
public void setListMapJoinOpsNoReducer(
List> listMapJoinOpsNoReducer) {
this.listMapJoinOpsNoReducer = listMapJoinOpsNoReducer;
}
/**
* @return the opToSamplePruner
*/
public HashMap getOpToSamplePruner() {
return opToSamplePruner;
}
/**
* @param opToSamplePruner
* the opToSamplePruner to set
*/
public void setOpToSamplePruner(
HashMap opToSamplePruner) {
this.opToSamplePruner = opToSamplePruner;
}
/**
* @return pruned partition map
*/
public Map getPrunedPartitions() {
return prunedPartitions;
}
/**
* @param prunedPartitions
*/
public void setPrunedPartitions(
Map prunedPartitions) {
this.prunedPartitions = prunedPartitions;
}
/**
* Sets the lineage information.
*
* @param lInfo The lineage information.
*/
public void setLineageInfo(LineageInfo lInfo) {
this.lInfo = lInfo;
}
/**
* Gets the associated lineage information.
*
* @return LineageInfo
*/
public LineageInfo getLineageInfo() {
return lInfo;
}
public Set getMapJoinOps() {
return mapJoinOps;
}
public void setMapJoinOps(Set mapJoinOps) {
this.mapJoinOps = mapJoinOps;
}
public Set getSmbMapJoinOps() {
return smbMapJoinOps;
}
public void setSmbMapJoinOps(Set smbMapJoinOps) {
this.smbMapJoinOps = smbMapJoinOps;
}
public GlobalLimitCtx getGlobalLimitCtx() {
return globalLimitCtx;
}
public void setGlobalLimitCtx(GlobalLimitCtx globalLimitCtx) {
this.globalLimitCtx = globalLimitCtx;
}
public HashSet getSemanticInputs() {
return semanticInputs;
}
public void replaceRootTask(Task extends Serializable> rootTask,
List extends Task extends Serializable>> tasks) {
this.rootTasks.remove(rootTask);
this.rootTasks.addAll(tasks);
}
public FetchTask getFetchTask() {
return fetchTask;
}
public void setFetchTask(FetchTask fetchTask) {
this.fetchTask = fetchTask;
}
public PrunedPartitionList getPrunedPartitions(String alias, TableScanOperator ts)
throws SemanticException {
PrunedPartitionList partsList = opToPartList.get(ts);
if (partsList == null) {
partsList = PartitionPruner.prune(ts, this, alias);
opToPartList.put(ts, partsList);
}
return partsList;
}
/**
* @return the opToPartToSkewedPruner
*/
public Map> getOpToPartToSkewedPruner() {
return opToPartToSkewedPruner;
}
/**
* @param opToPartToSkewedPruner
* the opToSkewedPruner to set
*/
public void setOpPartToSkewedPruner(
HashMap> opToPartToSkewedPruner) {
this.opToPartToSkewedPruner = opToPartToSkewedPruner;
}
public Map getViewAliasToInput() {
return viewAliasToInput;
}
public QueryProperties getQueryProperties() {
return queryProperties;
}
public void setQueryProperties(QueryProperties queryProperties) {
this.queryProperties = queryProperties;
}
public TableDesc getFetchTableDesc() {
return fetchTableDesc;
}
public void setFetchTabledesc(TableDesc fetchTableDesc) {
this.fetchTableDesc = fetchTableDesc;
}
public Operator> getFetchSource() {
return fetchSource;
}
public void setFetchSource(Operator> fetchSource) {
this.fetchSource = fetchSource;
}
public ListSinkOperator getFetchSink() {
return fetchSink;
}
public void setFetchSink(ListSinkOperator fetchSink) {
this.fetchSink = fetchSink;
}
public AnalyzeRewriteContext getAnalyzeRewrite() {
return this.analyzeRewrite;
}
public void setAnalyzeRewrite(AnalyzeRewriteContext analyzeRewrite) {
this.analyzeRewrite = analyzeRewrite;
}
public CreateTableDesc getCreateTable() {
return this.createTableDesc;
}
public void setCreateTable(CreateTableDesc createTableDesc) {
this.createTableDesc = createTableDesc;
}
}