org.apache.spark.sql.avro.confluent.CatalystDataToConfluentAvro.scala Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of spark-extensions_2.11 Show documentation
Show all versions of spark-extensions_2.11 Show documentation
Spark extensions for SmartDataLakeBuilder
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.spark.sql.avro.confluent
import java.io.ByteArrayOutputStream
import java.nio.ByteBuffer
import org.apache.avro.generic.GenericDatumWriter
import org.apache.avro.io.{BinaryEncoder, EncoderFactory}
import org.apache.spark.internal.Logging
import org.apache.spark.sql.catalyst.expressions.codegen.{CodegenContext, ExprCode}
import org.apache.spark.sql.catalyst.expressions.{Expression, UnaryExpression}
import org.apache.spark.sql.types.{BinaryType, DataType}
// copied from org.apache.spark.sql.avro.*
case class CatalystDataToConfluentAvro(child: Expression, subject: String, confluentHelper: ConfluentClient, updateAllowed: Boolean, mutualReadCheck: Boolean = false) extends UnaryExpression with Logging {
if (!updateAllowed && mutualReadCheck) logWarning("mutualReadCheck is ignored if updateAllowed=false")
override def dataType: DataType = BinaryType
// prepare serializer and writer for avro schema of subject
case class SerializerTools(schemaId: Int, serializer: MyAvroSerializer, writer: GenericDatumWriter[Any])
@transient private lazy val tgt = {
// Avro schema is not serializable. We must be careful to not store it in an attribute of the class.
val newSchema = MySchemaConverters.toAvroType(child.dataType, child.nullable)
val (schemaId, schema) = if (updateAllowed) confluentHelper.setOrUpdateSchema(subject, newSchema, mutualReadCheck)
else confluentHelper.setOrGetSchema(subject, newSchema)
val serializer = new MyAvroSerializer(child.dataType, schema, child.nullable)
val writer = new GenericDatumWriter[Any](schema)
SerializerTools(schemaId, serializer, writer)
}
@transient private var encoder: BinaryEncoder = _
@transient private lazy val out = new ByteArrayOutputStream
/**
* Instantiate serializer and writer for schema compatibility check
*/
def test(): Unit = {
tgt // initialize lazy value
}
override def nullSafeEval(input: Any): Any = {
out.reset()
appendSchemaId(tgt.schemaId, out)
encoder = EncoderFactory.get().directBinaryEncoder(out, encoder)
val avroData = tgt.serializer.serialize(input)
tgt.writer.write(avroData, encoder)
encoder.flush()
out.toByteArray
}
override def prettyName: String = "to_confluent_avro"
override protected def doGenCode(ctx: CodegenContext, ev: ExprCode): ExprCode = {
val expr = ctx.addReferenceObj("this", this)
defineCodeGen(ctx, ev, input => s"(byte[]) $expr.nullSafeEval($input)")
}
private def appendSchemaId(id: Int, os:ByteArrayOutputStream): Unit = {
os.write(confluentHelper.CONFLUENT_MAGIC_BYTE)
os.write(ByteBuffer.allocate(Integer.BYTES).putInt(id).array())
}
}
© 2015 - 2024 Weber Informatics LLC | Privacy Policy