io.deephaven.engine.table.impl.by.BigDecimalChunkedAvgOperator Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of deephaven-engine-table Show documentation
Show all versions of deephaven-engine-table Show documentation
Engine Table: Implementation and closely-coupled utilities
/**
* Copyright (c) 2016-2022 Deephaven Data Labs and Patent Pending
*/
package io.deephaven.engine.table.impl.by;
import io.deephaven.chunk.attributes.ChunkLengths;
import io.deephaven.chunk.attributes.ChunkPositions;
import io.deephaven.chunk.attributes.Values;
import io.deephaven.engine.table.ColumnSource;
import io.deephaven.engine.table.impl.sources.ObjectArraySource;
import io.deephaven.chunk.*;
import io.deephaven.engine.rowset.chunkattributes.RowKeys;
import org.apache.commons.lang3.mutable.MutableInt;
import java.math.BigDecimal;
import java.util.Collections;
import java.util.LinkedHashMap;
import java.util.Map;
import static io.deephaven.engine.table.impl.by.RollupConstants.*;
class BigDecimalChunkedAvgOperator implements IterativeChunkedAggregationOperator {
private final String name;
private final boolean exposeInternalColumns;
private final ObjectArraySource resultColumn = new ObjectArraySource<>(BigDecimal.class);
private final ObjectArraySource runningSum = new ObjectArraySource<>(BigDecimal.class);
private final NonNullCounter nonNullCount = new NonNullCounter();
BigDecimalChunkedAvgOperator(String name, boolean exposeInternalColumns) {
this.name = name;
this.exposeInternalColumns = exposeInternalColumns;
}
@Override
public void addChunk(BucketedContext context, Chunk extends Values> values,
LongChunk extends RowKeys> inputRowKeys, IntChunk destinations,
IntChunk startPositions, IntChunk length,
WritableBooleanChunk stateModified) {
final ObjectChunk asObjectChunk = values.asObjectChunk();
for (int ii = 0; ii < startPositions.size(); ++ii) {
final int startPosition = startPositions.get(ii);
final long destination = destinations.get(startPosition);
stateModified.set(ii, addChunk(asObjectChunk, destination, startPosition, length.get(ii)));
}
}
@Override
public void removeChunk(BucketedContext context, Chunk extends Values> values,
LongChunk extends RowKeys> inputRowKeys, IntChunk destinations,
IntChunk startPositions, IntChunk length,
WritableBooleanChunk stateModified) {
final ObjectChunk asObjectChunk = values.asObjectChunk();
for (int ii = 0; ii < startPositions.size(); ++ii) {
final int startPosition = startPositions.get(ii);
final long destination = destinations.get(startPosition);
stateModified.set(ii, removeChunk(asObjectChunk, destination, startPosition, length.get(ii)));
}
}
@Override
public boolean addChunk(SingletonContext context, int chunkSize, Chunk extends Values> values,
LongChunk extends RowKeys> inputRowKeys, long destination) {
return addChunk(values.asObjectChunk(), destination, 0, values.size());
}
@Override
public boolean removeChunk(SingletonContext context, int chunkSize, Chunk extends Values> values,
LongChunk extends RowKeys> inputRowKeys, long destination) {
return removeChunk(values.asObjectChunk(), destination, 0, values.size());
}
public boolean addChunk(ObjectChunk values, long destination, int chunkStart,
int chunkSize) {
final MutableInt chunkNonNullCount = new MutableInt();
final BigDecimal chunkSum =
SumBigDecimalChunk.sumBigDecimalChunk(values, chunkStart, chunkSize, chunkNonNullCount);
if (chunkNonNullCount.intValue() <= 0) {
return false;
}
final long newCount = nonNullCount.addNonNullUnsafe(destination, chunkNonNullCount.intValue());
final BigDecimal newSum;
final BigDecimal oldSum = runningSum.getUnsafe(destination);
if (oldSum == null) {
newSum = chunkSum;
} else {
newSum = oldSum.add(chunkSum);
}
runningSum.set(destination, newSum);
resultColumn.set(destination, newSum.divide(BigDecimal.valueOf(newCount), BigDecimal.ROUND_HALF_UP));
return true;
}
public boolean removeChunk(ObjectChunk values, long destination, int chunkStart,
int chunkSize) {
final MutableInt chunkNonNullCount = new MutableInt();
final BigDecimal chunkSum =
SumBigDecimalChunk.sumBigDecimalChunk(values, chunkStart, chunkSize, chunkNonNullCount);
if (chunkNonNullCount.intValue() <= 0) {
return false;
}
final long newCount = nonNullCount.addNonNullUnsafe(destination, -chunkNonNullCount.intValue());
if (newCount == 0) {
resultColumn.set(destination, null);
runningSum.set(destination, null);
} else {
final BigDecimal oldSum = runningSum.getUnsafe(destination);
final BigDecimal newSum = oldSum.subtract(chunkSum);
runningSum.set(destination, newSum);
resultColumn.set(destination, newSum.divide(BigDecimal.valueOf(newCount), BigDecimal.ROUND_HALF_UP));
}
return true;
}
@Override
public void ensureCapacity(long tableSize) {
resultColumn.ensureCapacity(tableSize);
nonNullCount.ensureCapacity(tableSize);
runningSum.ensureCapacity(tableSize);
}
@Override
public Map> getResultColumns() {
if (exposeInternalColumns) {
final Map> results = new LinkedHashMap<>();
results.put(name, resultColumn);
results.put(name + ROLLUP_RUNNING_SUM_COLUMN_ID + ROLLUP_COLUMN_SUFFIX, runningSum);
results.put(name + ROLLUP_NONNULL_COUNT_COLUMN_ID + ROLLUP_COLUMN_SUFFIX, nonNullCount.getColumnSource());
return results;
} else {
return Collections.singletonMap(name, resultColumn);
}
}
@Override
public void startTrackingPrevValues() {
resultColumn.startTrackingPrevValues();
if (exposeInternalColumns) {
runningSum.startTrackingPrevValues();
nonNullCount.startTrackingPrevValues();
}
}
}
© 2015 - 2024 Weber Informatics LLC | Privacy Policy