org.apache.flink.formats.atomic.AtomicRowDataDeserializationSchema Maven / Gradle / Ivy
/*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.flink.formats.atomic;
import org.apache.flink.api.common.serialization.DeserializationSchema;
import org.apache.flink.api.common.typeinfo.TypeInformation;
import org.apache.flink.streaming.connectors.pulsar.internal.IncompatibleSchemaException;
import org.apache.flink.streaming.connectors.pulsar.internal.SimpleSchemaTranslator;
import org.apache.flink.streaming.connectors.pulsar.util.RowDataUtil;
import org.apache.flink.table.api.DataTypes;
import org.apache.flink.table.data.GenericRowData;
import org.apache.flink.table.data.RowData;
import org.apache.flink.table.runtime.typeutils.InternalTypeInfo;
import org.apache.flink.table.types.DataType;
import org.apache.flink.table.types.FieldsDataType;
import org.apache.flink.table.types.logical.RowType;
import org.apache.flink.table.types.utils.TypeConversions;
import org.apache.flink.types.RowKind;
import org.apache.pulsar.client.api.Schema;
import java.io.IOException;
import java.util.ArrayList;
import java.util.List;
/**
* rowDeserializationSchema for atomic type.
*/
public class AtomicRowDataDeserializationSchema implements DeserializationSchema {
private static final long serialVersionUID = -228294330688809195L;
private final String className;
private final boolean useExtendFields;
private final Class clazz;
public AtomicRowDataDeserializationSchema(String className, boolean useExtendFields) {
this.className = className;
this.useExtendFields = useExtendFields;
try {
this.clazz = Class.forName(className);
} catch (ClassNotFoundException e) {
throw new RuntimeException(e);
}
}
public boolean isUseExtendFields() {
return useExtendFields;
}
@Override
public RowData deserialize(byte[] message) throws IOException {
DataType dataType = TypeConversions.fromClassToDataType(clazz).
orElseThrow(() -> new IllegalStateException(clazz.getCanonicalName() + "cant cast to flink dataType"));
try {
Schema schema = SimpleSchemaTranslator.sqlType2PulsarSchema(dataType);
Object data = schema.decode(message);
final GenericRowData rowData = new GenericRowData(RowKind.INSERT, 1);
RowDataUtil.setField(rowData, 0, data);
return rowData;
} catch (IncompatibleSchemaException e) {
throw new RuntimeException(e);
}
}
@Override
public boolean isEndOfStream(RowData nextElement) {
return false;
}
@Override
public TypeInformation getProducedType() {
/* DataType dataType = TypeConversions.fromClassToDataType(clazz).
orElseThrow(() -> new IllegalStateException(clazz.getCanonicalName() + "cant cast to flink dataType"));
RowType.RowField rowField = new RowType.RowField("value", dataType.getLogicalType());
List fields = Collections.singletonList(rowField);
return (TypeInformation) TypeConversions.fromDataTypeToLegacyInfo(TypeConversions.fromLogicalToDataType(new RowType(fields)));*/
//return (TypeInformation) TypeConversions.fromDataTypeToLegacyInfo(dataType);
List mainSchema = new ArrayList<>();
DataType dataType = TypeConversions.fromClassToDataType(clazz).
orElseThrow(() -> new IllegalStateException(clazz.getCanonicalName() + "cant cast to flink dataType"));
if (dataType instanceof FieldsDataType) {
FieldsDataType fieldsDataType = (FieldsDataType) dataType;
RowType rowType = (RowType) fieldsDataType.getLogicalType();
List fieldNames = rowType.getFieldNames();
for (int i = 0; i < fieldNames.size(); i++) {
org.apache.flink.table.types.logical.LogicalType logicalType = rowType.getTypeAt(i);
DataTypes.Field field =
DataTypes.FIELD(fieldNames.get(i), TypeConversions.fromLogicalToDataType(logicalType));
mainSchema.add(field);
}
} else {
mainSchema.add(DataTypes.FIELD("value", dataType));
}
if (useExtendFields) {
mainSchema.addAll(SimpleSchemaTranslator.METADATA_FIELDS);
}
FieldsDataType fieldsDataType = (FieldsDataType) DataTypes.ROW(mainSchema.toArray(new DataTypes.Field[0]));
return InternalTypeInfo.of(fieldsDataType.getLogicalType());
}
/**
* Builder for {@link AtomicRowDataDeserializationSchema}.
*/
public static class Builder {
private final String className;
private boolean useExtendFields;
public Builder(String className) {
this.className = className;
}
public AtomicRowDataDeserializationSchema.Builder useExtendFields(boolean useExtendFields) {
this.useExtendFields = useExtendFields;
return this;
}
public AtomicRowDataDeserializationSchema build() {
return new AtomicRowDataDeserializationSchema(className, useExtendFields);
}
}
@Override
public boolean equals(Object o) {
if (this == o) {
return true;
}
if (o == null || getClass() != o.getClass()) {
return false;
}
AtomicRowDataDeserializationSchema that = (AtomicRowDataDeserializationSchema) o;
if (useExtendFields != that.useExtendFields) {
return false;
}
return className.equals(that.className);
}
@Override
public int hashCode() {
int result = className.hashCode();
result = 31 * result + (useExtendFields ? 1 : 0);
return result;
}
}