org.apache.spark.sql.execution.OrderAwareCoalesceExec.scala Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of iceberg-spark-3.5_2.13 Show documentation
Show all versions of iceberg-spark-3.5_2.13 Show documentation
A table format for huge analytic datasets
The newest version!
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.apache.spark.sql.execution
import org.apache.spark.rdd.PartitionCoalescer
import org.apache.spark.rdd.RDD
import org.apache.spark.sql.catalyst.InternalRow
import org.apache.spark.sql.catalyst.expressions.Attribute
import org.apache.spark.sql.catalyst.expressions.SortOrder
import org.apache.spark.sql.catalyst.plans.physical.Partitioning
import org.apache.spark.sql.catalyst.plans.physical.SinglePartition
import org.apache.spark.sql.catalyst.plans.physical.UnknownPartitioning
case class OrderAwareCoalesceExec(
numPartitions: Int,
coalescer: PartitionCoalescer,
child: SparkPlan) extends UnaryExecNode {
override def output: Seq[Attribute] = child.output
override def outputOrdering: Seq[SortOrder] = child.outputOrdering
override def outputPartitioning: Partitioning = {
if (numPartitions == 1) SinglePartition else UnknownPartitioning(numPartitions)
}
protected override def doExecute(): RDD[InternalRow] = {
val result = child.execute()
if (numPartitions == 1 && result.getNumPartitions < 1) {
// make sure we don't output an RDD with 0 partitions,
// when claiming that we have a `SinglePartition`
// see CoalesceExec in Spark
new CoalesceExec.EmptyRDDWithPartitions(sparkContext, numPartitions)
} else {
result.coalesce(numPartitions, shuffle = false, Some(coalescer))
}
}
override protected def withNewChildInternal(newChild: SparkPlan): SparkPlan = {
copy(child = newChild)
}
}