
org.apache.flink.runtime.executiongraph.IntermediateResultPartition Maven / Gradle / Ivy
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.flink.runtime.executiongraph;
import org.apache.flink.runtime.io.network.partition.ResultPartitionType;
import org.apache.flink.runtime.jobgraph.DistributionPattern;
import org.apache.flink.runtime.jobgraph.IntermediateResultPartitionID;
import org.apache.flink.runtime.scheduler.strategy.ConsumedPartitionGroup;
import org.apache.flink.runtime.scheduler.strategy.ConsumerVertexGroup;
import java.util.List;
import java.util.Optional;
import static org.apache.flink.util.Preconditions.checkState;
public class IntermediateResultPartition {
private static final int UNKNOWN = -1;
private final IntermediateResult totalResult;
private final ExecutionVertex producer;
private final IntermediateResultPartitionID partitionId;
private final EdgeManager edgeManager;
/** Number of subpartitions. Initialized lazily and will not change once set. */
private int numberOfSubpartitions = UNKNOWN;
/** Whether this partition has produced some data. */
private boolean hasDataProduced = false;
public IntermediateResultPartition(
IntermediateResult totalResult,
ExecutionVertex producer,
int partitionNumber,
EdgeManager edgeManager) {
this.totalResult = totalResult;
this.producer = producer;
this.partitionId = new IntermediateResultPartitionID(totalResult.getId(), partitionNumber);
this.edgeManager = edgeManager;
}
public ExecutionVertex getProducer() {
return producer;
}
public int getPartitionNumber() {
return partitionId.getPartitionNumber();
}
public IntermediateResult getIntermediateResult() {
return totalResult;
}
public IntermediateResultPartitionID getPartitionId() {
return partitionId;
}
public ResultPartitionType getResultType() {
return totalResult.getResultType();
}
public ConsumerVertexGroup getConsumerVertexGroup() {
Optional consumerVertexGroup = getConsumerVertexGroupOptional();
checkState(consumerVertexGroup.isPresent());
return consumerVertexGroup.get();
}
public Optional getConsumerVertexGroupOptional() {
return Optional.ofNullable(
getEdgeManager().getConsumerVertexGroupForPartition(partitionId));
}
public List getConsumedPartitionGroups() {
return getEdgeManager().getConsumedPartitionGroupsById(partitionId);
}
public int getNumberOfSubpartitions() {
if (numberOfSubpartitions == UNKNOWN) {
numberOfSubpartitions = computeNumberOfSubpartitions();
checkState(
numberOfSubpartitions > 0,
"Number of subpartitions is an unexpected value: " + numberOfSubpartitions);
}
return numberOfSubpartitions;
}
private int computeNumberOfSubpartitions() {
if (!getProducer().getExecutionGraphAccessor().isDynamic()) {
ConsumerVertexGroup consumerVertexGroup = getConsumerVertexGroup();
checkState(consumerVertexGroup.size() > 0);
// The produced data is partitioned among a number of subpartitions, one for each
// consuming sub task.
return consumerVertexGroup.size();
} else {
if (totalResult.isBroadcast()) {
// for dynamic graph and broadcast result, we only produced one subpartition,
// and all the downstream vertices should consume this subpartition.
return 1;
} else {
return computeNumberOfMaxPossiblePartitionConsumers();
}
}
}
private int computeNumberOfMaxPossiblePartitionConsumers() {
final ExecutionJobVertex consumerJobVertex =
getIntermediateResult().getConsumerExecutionJobVertex();
final DistributionPattern distributionPattern =
getIntermediateResult().getConsumingDistributionPattern();
// decide the max possible consumer job vertex parallelism
int maxConsumerJobVertexParallelism = consumerJobVertex.getParallelism();
if (maxConsumerJobVertexParallelism <= 0) {
checkState(
consumerJobVertex.getMaxParallelism() > 0,
"Neither the parallelism nor the max parallelism of a job vertex is set");
maxConsumerJobVertexParallelism = consumerJobVertex.getMaxParallelism();
}
// compute number of subpartitions according to the distribution pattern
if (distributionPattern == DistributionPattern.ALL_TO_ALL) {
return maxConsumerJobVertexParallelism;
} else {
int numberOfPartitions = getIntermediateResult().getNumParallelProducers();
return (int) Math.ceil(((double) maxConsumerJobVertexParallelism) / numberOfPartitions);
}
}
public void markDataProduced() {
hasDataProduced = true;
}
public boolean isConsumable() {
return hasDataProduced;
}
void resetForNewExecution() {
if (getResultType().isBlocking() && hasDataProduced) {
// A BLOCKING result partition with data produced means it is finished
// Need to add the running producer count of the result on resetting it
for (ConsumedPartitionGroup consumedPartitionGroup : getConsumedPartitionGroups()) {
consumedPartitionGroup.partitionUnfinished();
}
}
hasDataProduced = false;
for (ConsumedPartitionGroup consumedPartitionGroup : getConsumedPartitionGroups()) {
totalResult.clearCachedInformationForPartitionGroup(consumedPartitionGroup);
}
}
public void addConsumers(ConsumerVertexGroup consumers) {
getEdgeManager().connectPartitionWithConsumerVertexGroup(partitionId, consumers);
}
private EdgeManager getEdgeManager() {
return edgeManager;
}
void markFinished() {
// Sanity check that this is only called on blocking partitions.
if (!getResultType().isBlocking()) {
throw new IllegalStateException(
"Tried to mark a non-blocking result partition as finished");
}
// Sanity check to make sure a result partition cannot be marked as finished twice.
if (hasDataProduced) {
throw new IllegalStateException(
"Tried to mark a finished result partition as finished.");
}
hasDataProduced = true;
for (ConsumedPartitionGroup consumedPartitionGroup : getConsumedPartitionGroups()) {
consumedPartitionGroup.partitionFinished();
}
}
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy