Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
/*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package io.trino.sql.planner;
import com.fasterxml.jackson.annotation.JsonCreator;
import com.fasterxml.jackson.annotation.JsonProperty;
import io.trino.operator.BucketPartitionFunction;
import io.trino.operator.PartitionFunction;
import io.trino.operator.PrecomputedHashGenerator;
import io.trino.spi.Page;
import io.trino.spi.connector.BucketFunction;
import io.trino.spi.connector.ConnectorPartitioningHandle;
import io.trino.spi.type.Type;
import io.trino.spi.type.TypeOperators;
import java.util.List;
import java.util.Objects;
import java.util.Optional;
import static com.google.common.base.MoreObjects.toStringHelper;
import static com.google.common.base.Preconditions.checkArgument;
import static io.trino.operator.InterpretedHashGenerator.createPagePrefixHashGenerator;
import static java.util.Objects.requireNonNull;
public final class SystemPartitioningHandle
implements ConnectorPartitioningHandle
{
enum SystemPartitioning
{
SINGLE,
FIXED,
SOURCE,
COORDINATOR_ONLY,
ARBITRARY
}
public static final PartitioningHandle SINGLE_DISTRIBUTION = createSystemPartitioning(SystemPartitioning.SINGLE, SystemPartitionFunction.SINGLE);
public static final PartitioningHandle COORDINATOR_DISTRIBUTION = createSystemPartitioning(SystemPartitioning.COORDINATOR_ONLY, SystemPartitionFunction.SINGLE);
public static final PartitioningHandle FIXED_HASH_DISTRIBUTION = createSystemPartitioning(SystemPartitioning.FIXED, SystemPartitionFunction.HASH);
public static final PartitioningHandle FIXED_ARBITRARY_DISTRIBUTION = createSystemPartitioning(SystemPartitioning.FIXED, SystemPartitionFunction.ROUND_ROBIN);
public static final PartitioningHandle FIXED_BROADCAST_DISTRIBUTION = createSystemPartitioning(SystemPartitioning.FIXED, SystemPartitionFunction.BROADCAST);
public static final PartitioningHandle SCALED_WRITER_ROUND_ROBIN_DISTRIBUTION = createScaledWriterSystemPartitioning(SystemPartitionFunction.ROUND_ROBIN);
public static final PartitioningHandle SCALED_WRITER_HASH_DISTRIBUTION = createScaledWriterSystemPartitioning(SystemPartitionFunction.HASH);
public static final PartitioningHandle SOURCE_DISTRIBUTION = createSystemPartitioning(SystemPartitioning.SOURCE, SystemPartitionFunction.UNKNOWN);
public static final PartitioningHandle ARBITRARY_DISTRIBUTION = createSystemPartitioning(SystemPartitioning.ARBITRARY, SystemPartitionFunction.UNKNOWN);
public static final PartitioningHandle FIXED_PASSTHROUGH_DISTRIBUTION = createSystemPartitioning(SystemPartitioning.FIXED, SystemPartitionFunction.UNKNOWN);
private static PartitioningHandle createSystemPartitioning(SystemPartitioning partitioning, SystemPartitionFunction function)
{
return new PartitioningHandle(Optional.empty(), Optional.empty(), new SystemPartitioningHandle(partitioning, function));
}
private static PartitioningHandle createScaledWriterSystemPartitioning(SystemPartitionFunction function)
{
return new PartitioningHandle(Optional.empty(), Optional.empty(), new SystemPartitioningHandle(SystemPartitioning.ARBITRARY, function), true);
}
private final SystemPartitioning partitioning;
private final SystemPartitionFunction function;
@JsonCreator
public SystemPartitioningHandle(
@JsonProperty("partitioning") SystemPartitioning partitioning,
@JsonProperty("function") SystemPartitionFunction function)
{
this.partitioning = requireNonNull(partitioning, "partitioning is null");
this.function = requireNonNull(function, "function is null");
}
@JsonProperty
public SystemPartitioning getPartitioning()
{
return partitioning;
}
@JsonProperty
public SystemPartitionFunction getFunction()
{
return function;
}
public String getPartitioningName()
{
return partitioning.name();
}
@Override
public boolean isSingleNode()
{
return partitioning == SystemPartitioning.COORDINATOR_ONLY || partitioning == SystemPartitioning.SINGLE;
}
@Override
public boolean isCoordinatorOnly()
{
return partitioning == SystemPartitioning.COORDINATOR_ONLY;
}
@Override
public boolean equals(Object o)
{
if (this == o) {
return true;
}
if (o == null || getClass() != o.getClass()) {
return false;
}
SystemPartitioningHandle that = (SystemPartitioningHandle) o;
return partitioning == that.partitioning &&
function == that.function;
}
@Override
public int hashCode()
{
return Objects.hash(partitioning, function);
}
@Override
public String toString()
{
if (partitioning == SystemPartitioning.FIXED || partitioning == SystemPartitioning.ARBITRARY) {
return function.toString();
}
return partitioning.toString();
}
public PartitionFunction getPartitionFunction(List partitionChannelTypes, boolean isHashPrecomputed, int[] bucketToPartition, TypeOperators typeOperators)
{
requireNonNull(partitionChannelTypes, "partitionChannelTypes is null");
requireNonNull(bucketToPartition, "bucketToPartition is null");
BucketFunction bucketFunction = function.createBucketFunction(partitionChannelTypes, isHashPrecomputed, bucketToPartition.length, typeOperators);
return new BucketPartitionFunction(bucketFunction, bucketToPartition);
}
public enum SystemPartitionFunction
{
SINGLE {
@Override
public BucketFunction createBucketFunction(List partitionChannelTypes, boolean isHashPrecomputed, int bucketCount, TypeOperators typeOperators)
{
checkArgument(bucketCount == 1, "Single partition can only have one bucket");
return new SingleBucketFunction();
}
},
HASH {
@Override
public BucketFunction createBucketFunction(List partitionChannelTypes, boolean isHashPrecomputed, int bucketCount, TypeOperators typeOperators)
{
if (isHashPrecomputed) {
return new HashBucketFunction(new PrecomputedHashGenerator(0), bucketCount);
}
return new HashBucketFunction(createPagePrefixHashGenerator(partitionChannelTypes, typeOperators), bucketCount);
}
},
ROUND_ROBIN {
@Override
public BucketFunction createBucketFunction(List partitionChannelTypes, boolean isHashPrecomputed, int bucketCount, TypeOperators typeOperators)
{
return new RoundRobinBucketFunction(bucketCount);
}
},
BROADCAST {
@Override
public BucketFunction createBucketFunction(List partitionChannelTypes, boolean isHashPrecomputed, int bucketCount, TypeOperators typeOperators)
{
throw new UnsupportedOperationException();
}
},
UNKNOWN {
@Override
public BucketFunction createBucketFunction(List partitionChannelTypes, boolean isHashPrecomputed, int bucketCount, TypeOperators typeOperators)
{
throw new UnsupportedOperationException();
}
};
public abstract BucketFunction createBucketFunction(List partitionChannelTypes,
boolean isHashPrecomputed,
int bucketCount,
TypeOperators typeOperators);
private static class SingleBucketFunction
implements BucketFunction
{
@Override
public int getBucket(Page page, int position)
{
return 0;
}
}
public static class RoundRobinBucketFunction
implements BucketFunction
{
private final int bucketCount;
private int counter;
public RoundRobinBucketFunction(int bucketCount)
{
checkArgument(bucketCount > 0, "bucketCount must be at least 1");
this.bucketCount = bucketCount;
}
@Override
public int getBucket(Page page, int position)
{
int bucket = counter % bucketCount;
counter = (counter + 1) & 0x7fff_ffff;
return bucket;
}
@Override
public String toString()
{
return toStringHelper(this)
.add("bucketCount", bucketCount)
.toString();
}
}
}
}