Please wait. This can take some minutes ...
Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance.
Project price only 1 $
You can buy this project and download/modify it how often you want.
org.apache.hadoop.hive.ql.exec.vector.expressions.aggregates.gen.VectorUDAFVarFinal Maven / Gradle / Ivy
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hive.ql.exec.vector.expressions.aggregates.gen;
import java.util.ArrayList;
import java.util.List;
import org.apache.hadoop.hive.ql.exec.Description;
import org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression;
import org.apache.hadoop.hive.ql.exec.vector.expressions.aggregates.VectorAggregateExpression;
import org.apache.hadoop.hive.ql.exec.vector.VectorAggregationBufferRow;
import org.apache.hadoop.hive.ql.exec.vector.VectorAggregationDesc;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
import org.apache.hadoop.hive.ql.exec.vector.ColumnVector;
import org.apache.hadoop.hive.ql.exec.vector.StructColumnVector;
import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
import org.apache.hadoop.hive.ql.exec.vector.DoubleColumnVector;
import org.apache.hadoop.hive.ql.metadata.HiveException;
import org.apache.hadoop.hive.ql.plan.AggregationDesc;
import org.apache.hadoop.hive.ql.plan.ExprNodeDesc;
import org.apache.hadoop.hive.ql.udf.generic.GenericUDAFEvaluator;
import org.apache.hadoop.hive.ql.udf.generic.GenericUDAFEvaluator.Mode;
import org.apache.hadoop.hive.ql.udf.generic.GenericUDAFVariance;
import org.apache.hadoop.hive.ql.udf.generic.GenericUDAFVariance.VarianceKind;
import org.apache.hadoop.hive.ql.util.JavaDataModel;
import com.google.common.base.Preconditions;
/**
* Generated from template VectorUDAFVarMerge.txt.
*/
@Description(name = "variance",
value = "_FUNC_(expr) - Returns the average value of expr (vectorized, type: )")
public class VectorUDAFVarFinal extends VectorAggregateExpression {
private static final long serialVersionUID = 1L;
/** class for storing the current aggregate value. */
static class Aggregation implements AggregationBuffer {
private static final long serialVersionUID = 1L;
transient private long mergeCount;
transient private double mergeSum;
transient private double mergeVariance;
/**
* Value is explicitly (re)initialized in reset()
*/
transient private boolean isNull = true;
public void merge(long partialCount, double partialSum, double partialVariance) {
if (isNull || mergeCount == 0) {
// Just copy the information since there is nothing so far.
mergeCount = partialCount;
mergeSum = partialSum;
mergeVariance = partialVariance;
isNull = false;
return;
}
if (partialCount > 0 && mergeCount > 0) {
// Merge the two partials.
mergeVariance +=
GenericUDAFVariance.calculateMerge(
partialCount, mergeCount, partialSum, mergeSum,
partialVariance, mergeVariance);
// Update these after calculation.
mergeCount += partialCount;
mergeSum += partialSum;
}
}
@Override
public int getVariableSize() {
throw new UnsupportedOperationException();
}
@Override
public void reset () {
isNull = true;
mergeCount = 0L;
mergeSum = 0;
mergeVariance = 0;
}
}
transient private VarianceKind varianceKind = VarianceKind.NONE;
// This constructor is used to momentarily create the object so match can be called.
public VectorUDAFVarFinal() {
super();
}
public VectorUDAFVarFinal(VectorAggregationDesc vecAggrDesc) {
super(vecAggrDesc);
Preconditions.checkState(this.mode == GenericUDAFEvaluator.Mode.FINAL);
init();
}
private void init() {
String aggregateName = vecAggrDesc.getAggrDesc().getGenericUDAFName();
varianceKind = VarianceKind.nameMap.get(aggregateName);
}
private Aggregation getCurrentAggregationBuffer(
VectorAggregationBufferRow[] aggregationBufferSets,
int bufferIndex,
int row) {
VectorAggregationBufferRow mySet = aggregationBufferSets[row];
Aggregation myagg = (Aggregation) mySet.getAggregationBuffer(bufferIndex);
return myagg;
}
@Override
public void aggregateInputSelection(
VectorAggregationBufferRow[] aggregationBufferSets,
int bufferIndex,
VectorizedRowBatch batch) throws HiveException {
int batchSize = batch.size;
if (batchSize == 0) {
return;
}
inputExpression.evaluate(batch);
StructColumnVector inputStructColVector =
(StructColumnVector) batch.cols[
this.inputExpression.getOutputColumnNum()];
ColumnVector[] fields = inputStructColVector.fields;
long[] countVector = ((LongColumnVector) fields[VARIANCE_COUNT_FIELD_INDEX]).vector;
double[] sumVector = ((DoubleColumnVector) fields[VARIANCE_SUM_FIELD_INDEX]).vector;
double[] varianceVector = ((DoubleColumnVector) fields[VARIANCE_VARIANCE_FIELD_INDEX]).vector;
if (inputStructColVector.noNulls) {
if (inputStructColVector.isRepeating) {
iterateNoNullsRepeatingWithAggregationSelection(
aggregationBufferSets, bufferIndex,
countVector[0], sumVector[0], varianceVector[0], batchSize);
} else {
if (batch.selectedInUse) {
iterateNoNullsSelectionWithAggregationSelection(
aggregationBufferSets, bufferIndex,
countVector, sumVector, varianceVector, batch.selected, batchSize);
} else {
iterateNoNullsWithAggregationSelection(
aggregationBufferSets, bufferIndex,
countVector, sumVector, varianceVector, batchSize);
}
}
} else {
if (inputStructColVector.isRepeating) {
if (batch.selectedInUse) {
iterateHasNullsRepeatingSelectionWithAggregationSelection(
aggregationBufferSets, bufferIndex,
countVector[0], sumVector[0], varianceVector[0], batchSize, batch.selected, inputStructColVector.isNull);
} else {
iterateHasNullsRepeatingWithAggregationSelection(
aggregationBufferSets, bufferIndex,
countVector[0], sumVector[0], varianceVector[0], batchSize, inputStructColVector.isNull);
}
} else {
if (batch.selectedInUse) {
iterateHasNullsSelectionWithAggregationSelection(
aggregationBufferSets, bufferIndex,
countVector, sumVector, varianceVector, batchSize, batch.selected, inputStructColVector.isNull);
} else {
iterateHasNullsWithAggregationSelection(
aggregationBufferSets, bufferIndex,
countVector, sumVector, varianceVector, batchSize, inputStructColVector.isNull);
}
}
}
}
private void iterateNoNullsRepeatingWithAggregationSelection(
VectorAggregationBufferRow[] aggregationBufferSets,
int bufferIndex,
long count,
double sum,
double variance,
int batchSize) {
for (int i=0; i < batchSize; ++i) {
Aggregation myagg = getCurrentAggregationBuffer(
aggregationBufferSets,
bufferIndex,
i);
myagg.merge(count, sum, variance);
}
}
private void iterateNoNullsSelectionWithAggregationSelection(
VectorAggregationBufferRow[] aggregationBufferSets,
int bufferIndex,
long[] countVector,
double[] sumVector,
double[] varianceVector,
int[] selection,
int batchSize) {
for (int i=0; i < batchSize; ++i) {
Aggregation myagg = getCurrentAggregationBuffer(
aggregationBufferSets,
bufferIndex,
i);
final int batchIndex = selection[i];
myagg.merge(countVector[batchIndex], sumVector[batchIndex], varianceVector[batchIndex]);
}
}
private void iterateNoNullsWithAggregationSelection(
VectorAggregationBufferRow[] aggregationBufferSets,
int bufferIndex,
long[] countVector,
double[] sumVector,
double[] varianceVector,
int batchSize) {
for (int i=0; i < batchSize; ++i) {
Aggregation myagg = getCurrentAggregationBuffer(
aggregationBufferSets,
bufferIndex,
i);
myagg.merge(countVector[i], sumVector[i], varianceVector[i]);
}
}
private void iterateHasNullsRepeatingSelectionWithAggregationSelection(
VectorAggregationBufferRow[] aggregationBufferSets,
int bufferIndex,
long count,
double sum,
double variance,
int batchSize,
int[] selection,
boolean[] isNull) {
if (isNull[0]) {
return;
}
for (int i=0; i < batchSize; ++i) {
Aggregation myagg = getCurrentAggregationBuffer(
aggregationBufferSets,
bufferIndex,
i);
myagg.merge(count, sum, variance);
}
}
private void iterateHasNullsRepeatingWithAggregationSelection(
VectorAggregationBufferRow[] aggregationBufferSets,
int bufferIndex,
long count,
double sum,
double variance,
int batchSize,
boolean[] isNull) {
if (isNull[0]) {
return;
}
for (int i = 0; i < batchSize; ++i) {
Aggregation myagg = getCurrentAggregationBuffer(
aggregationBufferSets,
bufferIndex,
i);
myagg.merge(count, sum, variance);
}
}
private void iterateHasNullsSelectionWithAggregationSelection(
VectorAggregationBufferRow[] aggregationBufferSets,
int bufferIndex,
long[] countVector,
double[] sumVector,
double[] varianceVector,
int batchSize,
int[] selection,
boolean[] isNull) {
for (int i = 0; i < batchSize; i++) {
final int batchIndex = selection[i];
if (!isNull[batchIndex]) {
Aggregation myagg = getCurrentAggregationBuffer(
aggregationBufferSets,
bufferIndex,
i);
myagg.merge(countVector[batchIndex], sumVector[batchIndex], varianceVector[batchIndex]);
}
}
}
private void iterateHasNullsWithAggregationSelection(
VectorAggregationBufferRow[] aggregationBufferSets,
int bufferIndex,
long[] countVector,
double[] sumVector,
double[] varianceVector,
int batchSize,
boolean[] isNull) {
for (int i=0; i < batchSize; ++i) {
if (!isNull[i]) {
Aggregation myagg = getCurrentAggregationBuffer(
aggregationBufferSets,
bufferIndex,
i);
myagg.merge(countVector[i], sumVector[i], varianceVector[i]);
}
}
}
@Override
public void aggregateInput(AggregationBuffer agg, VectorizedRowBatch batch)
throws HiveException {
inputExpression.evaluate(batch);
StructColumnVector inputStructColVector =
(StructColumnVector) batch.cols[
this.inputExpression.getOutputColumnNum()];
ColumnVector[] fields = inputStructColVector.fields;
long[] countVector = ((LongColumnVector) fields[VARIANCE_COUNT_FIELD_INDEX]).vector;
double[] sumVector = ((DoubleColumnVector) fields[VARIANCE_SUM_FIELD_INDEX]).vector;
double[] varianceVector = ((DoubleColumnVector) fields[VARIANCE_VARIANCE_FIELD_INDEX]).vector;
int batchSize = batch.size;
if (batchSize == 0) {
return;
}
Aggregation myagg = (Aggregation)agg;
if (inputStructColVector.isRepeating) {
if (inputStructColVector.noNulls || !inputStructColVector.isNull[0]) {
final long count = countVector[0];
final double sum = sumVector[0];
final double variance = varianceVector[0];
for (int i = 0; i < batchSize; i++) {
myagg.merge(count, sum, variance);
}
}
return;
}
if (!batch.selectedInUse && inputStructColVector.noNulls) {
iterateNoSelectionNoNulls(myagg, countVector, sumVector, varianceVector, batchSize);
} else if (!batch.selectedInUse) {
iterateNoSelectionHasNulls(myagg, countVector, sumVector, varianceVector, batchSize, inputStructColVector.isNull);
} else if (inputStructColVector.noNulls){
iterateSelectionNoNulls(myagg, countVector, sumVector, varianceVector, batchSize, batch.selected);
} else {
iterateSelectionHasNulls(myagg, countVector, sumVector, varianceVector, batchSize, inputStructColVector.isNull, batch.selected);
}
}
private void iterateSelectionHasNulls(
Aggregation myagg,
long[] countVector,
double[] sumVector,
double[] varianceVector,
int batchSize,
boolean[] isNull,
int[] selected) {
for (int i=0; i < batchSize; i++) {
int batchIndex = selected[i];
if (!isNull[batchIndex]) {
myagg.merge(countVector[batchIndex], sumVector[batchIndex], varianceVector[batchIndex]);
}
}
}
private void iterateSelectionNoNulls(
Aggregation myagg,
long[] countVector,
double[] sumVector,
double[] varianceVector,
int batchSize,
int[] selected) {
for (int i = 0; i< batchSize; ++i) {
final int batchIndex = selected[i];
myagg.merge(countVector[batchIndex], sumVector[batchIndex], varianceVector[batchIndex]);
}
}
private void iterateNoSelectionHasNulls(
Aggregation myagg,
long[] countVector,
double[] sumVector,
double[] varianceVector,
int batchSize,
boolean[] isNull) {
for(int i = 0; i < batchSize; i++) {
if (!isNull[i]) {
myagg.merge(countVector[i], sumVector[i], varianceVector[i]);
}
}
}
private void iterateNoSelectionNoNulls(
Aggregation myagg,
long[] countVector,
double[] sumVector,
double[] varianceVector,
int batchSize) {
for (int i=0;i 1) {
// Use the common variance family result calculation method.
result = GenericUDAFVariance.calculateVarianceFamilyResult(
myagg.mergeVariance, myagg.mergeCount, varianceKind);
} else {
// For one element the variance is always 0.
result = 0.0;
}
outputColVector.vector[batchIndex] = result;
}
}