Please wait. This can take some minutes ...
Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance.
Project price only 1 $
You can buy this project and download/modify it how often you want.
io.prestosql.tests.kafka.TestKafkaAvroSmokeTest Maven / Gradle / Ivy
/*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package io.prestosql.tests.kafka;
import com.google.common.collect.ImmutableList;
import com.google.common.collect.ImmutableMap;
import io.prestosql.tempto.ProductTest;
import io.prestosql.tempto.Requirement;
import io.prestosql.tempto.RequirementsProvider;
import io.prestosql.tempto.Requires;
import io.prestosql.tempto.configuration.Configuration;
import io.prestosql.tempto.fulfillment.table.kafka.KafkaMessage;
import io.prestosql.tempto.fulfillment.table.kafka.KafkaTableDefinition;
import io.prestosql.tempto.fulfillment.table.kafka.ListKafkaDataSource;
import io.prestosql.tempto.query.QueryResult;
import org.apache.avro.Schema;
import org.apache.avro.file.DataFileWriter;
import org.apache.avro.generic.GenericData;
import org.apache.avro.generic.GenericDatumWriter;
import org.apache.avro.generic.GenericRecord;
import org.testng.annotations.Test;
import java.io.ByteArrayOutputStream;
import java.io.File;
import java.io.IOException;
import java.io.UncheckedIOException;
import java.util.Map;
import static io.prestosql.tempto.assertions.QueryAssert.Row.row;
import static io.prestosql.tempto.assertions.QueryAssert.assertThat;
import static io.prestosql.tempto.fulfillment.table.TableRequirements.immutableTable;
import static io.prestosql.tempto.fulfillment.table.kafka.KafkaMessageContentsBuilder.contentsBuilder;
import static io.prestosql.tempto.query.QueryExecutor.query;
import static io.prestosql.tests.TestGroups.KAFKA;
import static io.prestosql.tests.TestGroups.PROFILE_SPECIFIC_TESTS;
import static java.lang.String.format;
public class TestKafkaAvroSmokeTest
extends ProductTest
{
private static final String KAFKA_CATALOG = "kafka";
private static final String ALL_DATATYPES_AVRO_TABLE_NAME = "product_tests.all_datatypes_avro";
private static final String ALL_DATATYPES_AVRO_TOPIC_NAME = "all_datatypes_avro";
private static final String ALL_DATATYPE_SCHEMA_PATH = "/docker/presto-product-tests/conf/presto/etc/catalog/kafka/all_datatypes_avro_schema.avsc";
private static final String ALL_NULL_AVRO_TABLE_NAME = "product_tests.all_null_avro";
private static final String ALL_NULL_AVRO_TOPIC_NAME = "all_null_avro";
private static final String STRUCTURAL_AVRO_TABLE_NAME = "product_tests.structural_datatype_avro";
private static final String STRUCTURAL_AVRO_TOPIC_NAME = "structural_datatype_avro";
private static final String STRUCTURAL_SCHEMA_PATH = "/docker/presto-product-tests/conf/presto/etc/catalog/kafka/structural_datatype_avro_schema.avsc";
// kafka-connectors requires tables to be predefined in presto configuration
// the requirements here will be used to verify that table actually exists and to
// create topics and propagate them with data
private static class AllDataTypesAvroTable
implements RequirementsProvider
{
@Override
public Requirement getRequirements(Configuration configuration)
{
ImmutableMap record = ImmutableMap.of(
"a_varchar", "foobar",
"a_bigint", 127L,
"a_double", 234.567,
"a_boolean", true);
return createAvroTable(ALL_DATATYPE_SCHEMA_PATH, ALL_DATATYPES_AVRO_TABLE_NAME, ALL_DATATYPES_AVRO_TOPIC_NAME, record);
}
}
private static Requirement createAvroTable(String schemaPath, String tableName, String topicName, ImmutableMap record)
{
try {
Schema schema = new Schema.Parser().parse(new File(schemaPath));
byte[] avroData = convertRecordToAvro(schema, record);
return immutableTable(new KafkaTableDefinition(
tableName,
topicName,
new ListKafkaDataSource(ImmutableList.of(
new KafkaMessage(
contentsBuilder()
.appendBytes(avroData)
.build()))),
1,
1));
}
catch (IOException e) {
throw new UncheckedIOException(e);
}
}
private static byte[] convertRecordToAvro(Schema schema, Map values)
{
ByteArrayOutputStream outputStream = new ByteArrayOutputStream();
GenericData.Record record = new GenericData.Record(schema);
values.forEach(record::put);
try (DataFileWriter dataFileWriter = new DataFileWriter<>(new GenericDatumWriter<>(schema))) {
dataFileWriter.create(schema, outputStream);
dataFileWriter.append(record);
}
catch (IOException e) {
throw new UncheckedIOException("Failed to convert to Avro.", e);
}
return outputStream.toByteArray();
}
@Test(groups = {KAFKA, PROFILE_SPECIFIC_TESTS})
@Requires(AllDataTypesAvroTable.class)
public void testSelectPrimitiveDataType()
{
QueryResult queryResult = query(format("select * from %s.%s", KAFKA_CATALOG, ALL_DATATYPES_AVRO_TABLE_NAME));
assertThat(queryResult).containsOnly(row(
"foobar",
127,
234.567,
true));
}
private static class NullDataAvroTable
implements RequirementsProvider
{
@Override
public Requirement getRequirements(Configuration configuration)
{
return createAvroTable(ALL_DATATYPE_SCHEMA_PATH, ALL_NULL_AVRO_TABLE_NAME, ALL_NULL_AVRO_TOPIC_NAME, ImmutableMap.of());
}
}
@Test(groups = {KAFKA, PROFILE_SPECIFIC_TESTS})
@Requires(NullDataAvroTable.class)
public void testNullType()
{
QueryResult queryResult = query(format("select * from %s.%s", KAFKA_CATALOG, ALL_NULL_AVRO_TABLE_NAME));
assertThat(queryResult).containsOnly(row(
null,
null,
null,
null));
}
private static class StructuralDataTypeTable
implements RequirementsProvider
{
@Override
public Requirement getRequirements(Configuration configuration)
{
ImmutableMap record = ImmutableMap.of(
"a_array", ImmutableList.of(100L, 102L),
"a_map", ImmutableMap.of("key1", "value1"));
return createAvroTable(STRUCTURAL_SCHEMA_PATH, STRUCTURAL_AVRO_TABLE_NAME, STRUCTURAL_AVRO_TOPIC_NAME, record);
}
}
@Test(groups = {KAFKA, PROFILE_SPECIFIC_TESTS})
@Requires(StructuralDataTypeTable.class)
public void testSelectStructuralDataType()
{
QueryResult queryResult = query(format("SELECT a[1], a[2], m['key1'] FROM (SELECT c_array as a, c_map as m FROM %s.%s) t", KAFKA_CATALOG, STRUCTURAL_AVRO_TABLE_NAME));
assertThat(queryResult).containsOnly(row(100, 102, "value1"));
}
}