org.apache.spark.sql.catalyst.analysis.PaimonAnalysis.scala Maven / Gradle / Ivy
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.spark.sql.catalyst.analysis
import org.apache.paimon.spark.SparkTable
import org.apache.paimon.spark.commands.{PaimonDynamicPartitionOverwriteCommand, PaimonTruncateTableCommand}
import org.apache.paimon.table.FileStoreTable
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.catalyst.analysis.PaimonRelation.isPaimonTable
import org.apache.spark.sql.catalyst.plans.logical.{LogicalPlan, MergeIntoTable, OverwritePartitionsDynamic, PaimonTableValuedFunctions, PaimonTableValueFunction, TruncatePartition, TruncateTable}
import org.apache.spark.sql.catalyst.rules.Rule
import org.apache.spark.sql.execution.datasources.v2.DataSourceV2Relation
class PaimonAnalysis(session: SparkSession) extends Rule[LogicalPlan] {
override def apply(plan: LogicalPlan): LogicalPlan = plan.resolveOperatorsDown {
case func: PaimonTableValueFunction if func.args.forall(_.resolved) =>
PaimonTableValuedFunctions.resolvePaimonTableValuedFunction(session, func)
case o @ PaimonDynamicPartitionOverwrite(r, d) if o.resolved =>
PaimonDynamicPartitionOverwriteCommand(r, d, o.query, o.writeOptions, o.isByName)
case merge: MergeIntoTable if isPaimonTable(merge.targetTable) && merge.childrenResolved =>
PaimonMergeIntoResolver(merge, session)
}
}
case class PaimonPostHocResolutionRules(session: SparkSession) extends Rule[LogicalPlan] {
override def apply(plan: LogicalPlan): LogicalPlan = {
plan match {
case t @ TruncateTable(PaimonRelation(table)) if t.resolved =>
PaimonTruncateTableCommand(table, Map.empty)
case t @ TruncatePartition(PaimonRelation(table), ResolvedPartitionSpec(names, ident, _))
if t.resolved =>
assert(names.length == ident.numFields, "Names and values of partition don't match")
val resolver = session.sessionState.conf.resolver
val schema = table.schema()
val partitionSpec = names.zipWithIndex.map {
case (name, index) =>
val field = schema.find(f => resolver(f.name, name)).getOrElse {
throw new RuntimeException(s"$name is not a valid partition column in $schema.")
}
(name -> ident.get(index, field.dataType).toString)
}.toMap
PaimonTruncateTableCommand(table, partitionSpec)
case _ => plan
}
}
}
object PaimonDynamicPartitionOverwrite {
def unapply(o: OverwritePartitionsDynamic): Option[(DataSourceV2Relation, FileStoreTable)] = {
if (o.query.resolved) {
o.table match {
case r: DataSourceV2Relation if r.table.isInstanceOf[SparkTable] =>
Some((r, r.table.asInstanceOf[SparkTable].getTable.asInstanceOf[FileStoreTable]))
case _ => None
}
} else {
None
}
}
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy