Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.apache.iceberg.flink.data;
import java.io.IOException;
import java.util.List;
import java.util.Map;
import java.util.function.Supplier;
import org.apache.avro.LogicalType;
import org.apache.avro.LogicalTypes;
import org.apache.avro.Schema;
import org.apache.avro.io.DatumReader;
import org.apache.avro.io.Decoder;
import org.apache.flink.table.data.RowData;
import org.apache.iceberg.avro.AvroSchemaWithTypeVisitor;
import org.apache.iceberg.avro.SupportsRowPosition;
import org.apache.iceberg.avro.ValueReader;
import org.apache.iceberg.avro.ValueReaders;
import org.apache.iceberg.data.avro.DecoderResolver;
import org.apache.iceberg.relocated.com.google.common.collect.ImmutableMap;
import org.apache.iceberg.types.Type;
import org.apache.iceberg.types.Types;
public class FlinkAvroReader implements DatumReader, SupportsRowPosition {
private final Schema readSchema;
private final ValueReader reader;
private Schema fileSchema = null;
public FlinkAvroReader(org.apache.iceberg.Schema expectedSchema, Schema readSchema) {
this(expectedSchema, readSchema, ImmutableMap.of());
}
@SuppressWarnings("unchecked")
public FlinkAvroReader(
org.apache.iceberg.Schema expectedSchema, Schema readSchema, Map constants) {
this.readSchema = readSchema;
this.reader =
(ValueReader)
AvroSchemaWithTypeVisitor.visit(expectedSchema, readSchema, new ReadBuilder(constants));
}
@Override
public void setSchema(Schema newFileSchema) {
this.fileSchema = Schema.applyAliases(newFileSchema, readSchema);
}
@Override
public RowData read(RowData reuse, Decoder decoder) throws IOException {
return DecoderResolver.resolveAndRead(decoder, readSchema, fileSchema, reader, reuse);
}
@Override
public void setRowPositionSupplier(Supplier posSupplier) {
if (reader instanceof SupportsRowPosition) {
((SupportsRowPosition) reader).setRowPositionSupplier(posSupplier);
}
}
private static class ReadBuilder extends AvroSchemaWithTypeVisitor> {
private final Map idToConstant;
private ReadBuilder(Map idToConstant) {
this.idToConstant = idToConstant;
}
@Override
public ValueReader> record(
Types.StructType expected, Schema record, List names, List> fields) {
return FlinkValueReaders.struct(fields, expected.asStructType(), idToConstant);
}
@Override
public ValueReader> union(Type expected, Schema union, List> options) {
return ValueReaders.union(options);
}
@Override
public ValueReader> array(
Types.ListType expected, Schema array, ValueReader> elementReader) {
return FlinkValueReaders.array(elementReader);
}
@Override
public ValueReader> map(
Types.MapType expected, Schema map, ValueReader> keyReader, ValueReader> valueReader) {
return FlinkValueReaders.arrayMap(keyReader, valueReader);
}
@Override
public ValueReader> map(Types.MapType expected, Schema map, ValueReader> valueReader) {
return FlinkValueReaders.map(FlinkValueReaders.strings(), valueReader);
}
@Override
public ValueReader> primitive(Type.PrimitiveType expected, Schema primitive) {
LogicalType logicalType = primitive.getLogicalType();
if (logicalType != null) {
switch (logicalType.getName()) {
case "date":
return ValueReaders.ints();
case "time-micros":
return FlinkValueReaders.timeMicros();
case "timestamp-millis":
return FlinkValueReaders.timestampMills();
case "timestamp-micros":
return FlinkValueReaders.timestampMicros();
case "decimal":
LogicalTypes.Decimal decimal = (LogicalTypes.Decimal) logicalType;
return FlinkValueReaders.decimal(
ValueReaders.decimalBytesReader(primitive),
decimal.getPrecision(),
decimal.getScale());
case "uuid":
return FlinkValueReaders.uuids();
default:
throw new IllegalArgumentException("Unknown logical type: " + logicalType);
}
}
switch (primitive.getType()) {
case NULL:
return ValueReaders.nulls();
case BOOLEAN:
return ValueReaders.booleans();
case INT:
return ValueReaders.ints();
case LONG:
return ValueReaders.longs();
case FLOAT:
return ValueReaders.floats();
case DOUBLE:
return ValueReaders.doubles();
case STRING:
return FlinkValueReaders.strings();
case FIXED:
return ValueReaders.fixed(primitive.getFixedSize());
case BYTES:
return ValueReaders.bytes();
case ENUM:
return FlinkValueReaders.enums(primitive.getEnumSymbols());
default:
throw new IllegalArgumentException("Unsupported type: " + primitive);
}
}
}
}