All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.kafka.common.message.KafkaLZ4BlockInputStream Maven / Gradle / Ivy

The newest version!
/**
 * Licensed to the Apache Software Foundation (ASF) under one or more
 * contributor license agreements.  See the NOTICE file distributed with
 * this work for additional information regarding copyright ownership.
 * The ASF licenses this file to You under the Apache License, Version 2.0
 * (the "License"); you may not use this file except in compliance with
 * the License.  You may obtain a copy of the License at
 * 
 *    http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package org.apache.kafka.common.message;

import static org.apache.kafka.common.message.KafkaLZ4BlockOutputStream.LZ4_FRAME_INCOMPRESSIBLE_MASK;
import static org.apache.kafka.common.message.KafkaLZ4BlockOutputStream.LZ4_MAX_HEADER_LENGTH;
import static org.apache.kafka.common.message.KafkaLZ4BlockOutputStream.MAGIC;

import java.io.FilterInputStream;
import java.io.IOException;
import java.io.InputStream;

import org.apache.kafka.common.message.KafkaLZ4BlockOutputStream.BD;
import org.apache.kafka.common.message.KafkaLZ4BlockOutputStream.FLG;
import org.apache.kafka.common.utils.Utils;

import net.jpountz.lz4.LZ4Exception;
import net.jpountz.lz4.LZ4Factory;
import net.jpountz.lz4.LZ4SafeDecompressor;
import net.jpountz.xxhash.XXHash32;
import net.jpountz.xxhash.XXHashFactory;

/**
 * A partial implementation of the v1.4.1 LZ4 Frame format.
 * 
 * @see LZ4 Framing Format Spec
 */
public final class KafkaLZ4BlockInputStream extends FilterInputStream {

  public static final String PREMATURE_EOS = "Stream ended prematurely";
  public static final String NOT_SUPPORTED = "Stream unsupported";
  public static final String BLOCK_HASH_MISMATCH = "Block checksum mismatch";
  public static final String DESCRIPTOR_HASH_MISMATCH = "Stream frame descriptor corrupted";
  
  private final LZ4SafeDecompressor decompressor;
  private final XXHash32 checksum;
  private final byte[] buffer;
  private final byte[] compressedBuffer;
  private final int maxBlockSize;
  private FLG flg;
  private BD bd;
  private int bufferOffset;
  private int bufferSize;
  private boolean finished;
  
  /**
   * Create a new {@link InputStream} that will decompress data using the LZ4 algorithm.
   * 
   * @param in The stream to decompress
   * @throws IOException
   */
  public KafkaLZ4BlockInputStream(InputStream in) throws IOException {
    super(in);
    decompressor = LZ4Factory.fastestInstance().safeDecompressor();
    checksum = XXHashFactory.fastestInstance().hash32();
    readHeader();
    maxBlockSize = bd.getBlockMaximumSize();
    buffer = new byte[maxBlockSize];
    compressedBuffer = new byte[maxBlockSize];
    bufferOffset = 0;
    bufferSize = 0;
    finished = false;
  }
  
  /**
   * Reads the magic number and frame descriptor from the underlying {@link InputStream}.
   *  
   * @throws IOException
   */
  private void readHeader() throws IOException {
    byte[] header = new byte[LZ4_MAX_HEADER_LENGTH];
    
    // read first 6 bytes into buffer to check magic and FLG/BD descriptor flags
    bufferOffset = 6;
    if (in.read(header, 0, bufferOffset) != bufferOffset) {
      throw new IOException(PREMATURE_EOS);
    }

    if (MAGIC != Utils.readUnsignedIntLE(header, bufferOffset-6)) {
      throw new IOException(NOT_SUPPORTED);
    }
    flg = FLG.fromByte(header[bufferOffset-2]);
    bd = BD.fromByte(header[bufferOffset-1]);
    // TODO read uncompressed content size, update flg.validate()
    // TODO read dictionary id, update flg.validate()
    
    // check stream descriptor hash
    byte hash = (byte) ((checksum.hash(header, 0, bufferOffset, 0) >> 8) & 0xFF);
    header[bufferOffset++] = (byte) in.read();
    if (hash != header[bufferOffset-1]) {
      throw new IOException(DESCRIPTOR_HASH_MISMATCH);
    }
  }

  /**
   * Decompresses (if necessary) buffered data, optionally computes and validates a XXHash32 checksum, 
   * and writes the result to a buffer.
   * 
   * @throws IOException
   */
  private void readBlock() throws IOException {
    int blockSize = Utils.readUnsignedIntLE(in);

    // Check for EndMark
    if (blockSize == 0) {
      finished = true;
      // TODO implement content checksum, update flg.validate()
      return;
    } else if (blockSize > maxBlockSize) {
      throw new IOException(String.format("Block size %s exceeded max: %s", blockSize, maxBlockSize));
    }
    
    boolean compressed = (blockSize & LZ4_FRAME_INCOMPRESSIBLE_MASK) == 0;
    byte[] bufferToRead;
    if (compressed) {
      bufferToRead = compressedBuffer;
    } else {
      blockSize &= ~LZ4_FRAME_INCOMPRESSIBLE_MASK;
      bufferToRead = buffer;
      bufferSize = blockSize;
    }
    
    if (in.read(bufferToRead, 0, blockSize) != blockSize) {
      throw new IOException(PREMATURE_EOS);
    }

    // verify checksum
    if (flg.isBlockChecksumSet() && Utils.readUnsignedIntLE(in) != checksum.hash(bufferToRead, 0, blockSize, 0)) {
      throw new IOException(BLOCK_HASH_MISMATCH);
    }

    if (compressed) {
      try {
        bufferSize = decompressor.decompress(compressedBuffer, 0, blockSize, buffer, 0, maxBlockSize);
      } catch (LZ4Exception e) {
        throw new IOException(e);
      }
    }

    bufferOffset = 0;
  }
 
  @Override
  public int read() throws IOException {
    if (finished) {
      return -1;
    }
    if (available() == 0) {
      readBlock();
    }
    if (finished) {
      return -1;
    }
    int value = buffer[bufferOffset++] & 0xFF;
    
    return value;
  }

  @Override
  public int read(byte b[], int off, int len) throws IOException {
    net.jpountz.util.Utils.checkRange(b, off, len);
    if (finished) {
      return -1;
    }
    if (available() == 0) {
      readBlock();
    }
    if (finished) {
      return -1;
    }
    len = Math.min(len, available());
    System.arraycopy(buffer, bufferOffset, b, off, len);
    bufferOffset += len;
    return len;
  }

  @Override
  public long skip(long n) throws IOException {
    if (finished) {
      return 0;
    }
    if (available() == 0) {
      readBlock();
    }
    if (finished) {
      return 0;
    }
    n = Math.min(n, available());
    bufferOffset += n;
    return n;
  }

  @Override
  public int available() throws IOException {
    return bufferSize - bufferOffset;
  }

  @Override
  public void close() throws IOException {
      in.close();
  }

  @Override
  public synchronized void mark(int readlimit) {
      throw new RuntimeException("mark not supported");
  }

  @Override
  public synchronized void reset() throws IOException {
      throw new RuntimeException("reset not supported");
  }

  @Override
  public boolean markSupported() {
      return false;
  }

}




© 2015 - 2024 Weber Informatics LLC | Privacy Policy