org.apache.hadoop.hive.ql.exec.vector.VectorFileSinkOperator Maven / Gradle / Ivy
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hive.ql.exec.vector;
import java.util.Collection;
import java.util.concurrent.Future;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hive.ql.exec.FileSinkOperator;
import org.apache.hadoop.hive.ql.metadata.HiveException;
import org.apache.hadoop.hive.ql.plan.FileSinkDesc;
import org.apache.hadoop.hive.ql.plan.OperatorDesc;
import org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector;
/**
* File Sink operator implementation.
**/
public class VectorFileSinkOperator extends FileSinkOperator {
private static final long serialVersionUID = 1L;
private VectorizationContext vContext;
// The above members are initialized by the constructor and must not be
// transient.
//---------------------------------------------------------------------------
private transient boolean firstBatch;
private transient VectorExtractRowDynBatch vectorExtractRowDynBatch;
protected transient Object[] singleRow;
public VectorFileSinkOperator(VectorizationContext vContext,
OperatorDesc conf) {
super();
this.conf = (FileSinkDesc) conf;
this.vContext = vContext;
}
public VectorFileSinkOperator() {
}
@Override
protected Collection> initializeOp(Configuration hconf) throws HiveException {
// We need a input object inspector that is for the row we will extract out of the
// vectorized row batch, not for example, an original inspector for an ORC table, etc.
inputObjInspectors[0] =
VectorizedBatchUtil.convertToStandardStructObjectInspector((StructObjectInspector) inputObjInspectors[0]);
// Call FileSinkOperator with new input inspector.
Collection> result = super.initializeOp(hconf);
assert result.isEmpty();
firstBatch = true;
return result;
}
@Override
public void process(Object data, int tag) throws HiveException {
VectorizedRowBatch batch = (VectorizedRowBatch) data;
if (firstBatch) {
vectorExtractRowDynBatch = new VectorExtractRowDynBatch();
vectorExtractRowDynBatch.init((StructObjectInspector) inputObjInspectors[0], vContext.getProjectedColumns());
singleRow = new Object[vectorExtractRowDynBatch.getCount()];
firstBatch = false;
}
vectorExtractRowDynBatch.setBatchOnEntry(batch);
if (batch.selectedInUse) {
int selected[] = batch.selected;
for (int logical = 0 ; logical < batch.size; logical++) {
int batchIndex = selected[logical];
vectorExtractRowDynBatch.extractRow(batchIndex, singleRow);
super.process(singleRow, tag);
}
} else {
for (int batchIndex = 0 ; batchIndex < batch.size; batchIndex++) {
vectorExtractRowDynBatch.extractRow(batchIndex, singleRow);
super.process(singleRow, tag);
}
}
vectorExtractRowDynBatch.forgetBatchOnExit();
}
}