All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.bakdata.kafka.ErrorUtil Maven / Gradle / Ivy

There is a newer version: 1.5.0
Show newest version
/*
 * MIT License
 *
 * Copyright (c) 2022 bakdata
 *
 * Permission is hereby granted, free of charge, to any person obtaining a copy
 * of this software and associated documentation files (the "Software"), to deal
 * in the Software without restriction, including without limitation the rights
 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
 * copies of the Software, and to permit persons to whom the Software is
 * furnished to do so, subject to the following conditions:
 *
 * The above copyright notice and this permission notice shall be included in all
 * copies or substantial portions of the Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
 * SOFTWARE.
 */

package com.bakdata.kafka;

import java.io.ByteArrayOutputStream;
import java.io.IOException;
import java.io.OutputStream;
import java.net.SocketTimeoutException;
import java.nio.charset.StandardCharsets;
import java.util.Objects;
import lombok.experimental.UtilityClass;
import lombok.extern.slf4j.Slf4j;
import org.apache.avro.Schema;
import org.apache.avro.generic.GenericContainer;
import org.apache.avro.generic.GenericDatumWriter;
import org.apache.avro.generic.GenericRecord;
import org.apache.avro.io.DatumWriter;
import org.apache.avro.io.EncoderFactory;
import org.apache.avro.io.JsonEncoder;
import org.apache.avro.specific.SpecificDatumWriter;
import org.apache.avro.specific.SpecificRecord;
import org.apache.kafka.common.errors.RecordTooLargeException;
import org.apache.kafka.common.errors.SerializationException;

/**
 * This class provides utility methods for dealing with errors in Kafka streams, such as serializing values to string
 * and classifying errors as recoverable.
 */
@Slf4j
@UtilityClass
public class ErrorUtil {

    private static final String ORG_APACHE_KAFKA_COMMON_ERRORS = "org.apache.kafka.common.errors";

    /**
     * Check if an exception is recoverable and thus should be thrown so that the process is restarted by the execution
     * environment.
     * 

Recoverable errors are: *

    *
  • {@link #isRecoverableKafkaError(Exception)} *
* * @param e exception * @return whether exception is recoverable or not */ public static boolean isRecoverable(final Exception e) { return isRecoverableKafkaError(e); } /** * Check if an exception is thrown by Kafka, i.e., located in package {@code org.apache.kafka.common.errors}, and is * recoverable. *

Non-recoverable Kafka errors are: *

    *
  • {@link RecordTooLargeException} *
  • {@link SerializationException} which is not caused by timeout *
* * @param e exception * @return whether exception is thrown by Kafka and recoverable or not */ public static boolean isRecoverableKafkaError(final Exception e) { if (ORG_APACHE_KAFKA_COMMON_ERRORS.equals(e.getClass().getPackageName())) { if (e instanceof RecordTooLargeException) { return false; } if (e instanceof SerializationException) { // socket timeouts usually indicate that the schema registry is temporarily down return e.getCause() instanceof SocketTimeoutException; } return true; } return false; } /** * Convert an object to {@code String}. {@code SpecificRecord} will be serialized using * {@link #toString(SpecificRecord)}. {@code GenericRecord} will be serialized using * {@link #toString(GenericRecord)}. * * @param o object to be serialized * @return {@code String} representation of record */ public static String toString(final Object o) { final Object o1; if (o instanceof SpecificRecord) { o1 = toString((SpecificRecord) o); } else if (o instanceof GenericRecord) { o1 = toString((GenericRecord) o); } else { o1 = o; } return Objects.toString(o1); } /** * Convert a {@code SpecificRecord} to {@code String} using JSON serialization. * * @param specificRecord record to be serialized * @return JSON representation of record or record if an error occurred */ private static Object toString(final SpecificRecord specificRecord) { try { return writeAsJson(specificRecord); } catch (final IOException ex) { log.warn("Failed to write to json", ex); return specificRecord; } } private static String writeAsJson(final SpecificRecord itemRecord) throws IOException { final Schema targetSchema = itemRecord.getSchema(); final DatumWriter writer = new SpecificDatumWriter<>(targetSchema); return writeJson(itemRecord, writer); } /** * Convert a {@code GenericRecord} to {@code String} using JSON serialization. * * @param genericRecord record to be serialized * @return JSON representation of record or record if an error occurred */ private static Object toString(final GenericRecord genericRecord) { try { return writeAsJson(genericRecord); } catch (final IOException ex) { log.warn("Failed to write to json", ex); return genericRecord; } } private static String writeAsJson(final GenericRecord itemRecord) throws IOException { final Schema targetSchema = itemRecord.getSchema(); final DatumWriter writer = new GenericDatumWriter<>(targetSchema); return writeJson(itemRecord, writer); } private static String writeJson(final T itemRecord, final DatumWriter writer) throws IOException { try (final ByteArrayOutputStream out = new ByteArrayOutputStream()) { writeJson(itemRecord, writer, out); return out.toString(StandardCharsets.ISO_8859_1); } } private static void writeJson(final T itemRecord, final DatumWriter writer, final OutputStream out) throws IOException { final JsonEncoder jsonEncoder = EncoderFactory.get().jsonEncoder(itemRecord.getSchema(), out); writer.write(itemRecord, jsonEncoder); jsonEncoder.flush(); } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy