All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.parquet.hadoop.codec.Lz4RawCodec Maven / Gradle / Ivy

There is a newer version: 1.14.4
Show newest version
/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *   http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing,
 * software distributed under the License is distributed on an
 * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 * KIND, either express or implied.  See the License for the
 * specific language governing permissions and limitations
 * under the License.
 */
package org.apache.parquet.hadoop.codec;

import org.apache.hadoop.conf.Configurable;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.io.compress.CompressionCodec;
import org.apache.hadoop.io.compress.CompressionInputStream;
import org.apache.hadoop.io.compress.CompressionOutputStream;
import org.apache.hadoop.io.compress.Compressor;
import org.apache.hadoop.io.compress.Decompressor;

import java.io.IOException;
import java.io.InputStream;
import java.io.OutputStream;

/**
 * Lz4 raw compression codec for Parquet. This codec type has been introduced
 * into the parquet format since version 2.9.0. It differs from the Lz4Codec
 * shipped with Apache Hadoop by removing the light frame header which includes
 * 4 byte for uncompressed length and 4 byte for compressed length. In the
 * Apache Arrow implementation, these two Lz4 codecs are recognized as LZ4_RAW
 * and LZ4_HADOOP to minimize the confusion at its best. Please check the link
 * below for reference.
 * https://github.com/apache/parquet-format/blob/master/Compression.md
 */
public class Lz4RawCodec implements Configurable, CompressionCodec {

  private Configuration conf;

  // Hadoop config for how big to make intermediate buffers.
  public static final String BUFFER_SIZE_CONFIG = "io.file.buffer.size";

  private static final int DEFAULT_BUFFER_SIZE_CONFIG = 4 * 1024;

  @Override
  public void setConf(Configuration conf) {
    this.conf = conf;
  }

  @Override
  public Configuration getConf() {
    return conf;
  }

  @Override
  public Compressor createCompressor() {
    return new Lz4RawCompressor();
  }

  @Override
  public Decompressor createDecompressor() {
    return new Lz4RawDecompressor();
  }

  @Override
  public CompressionInputStream createInputStream(InputStream stream)
    throws IOException {
    return createInputStream(stream, createDecompressor());
  }

  @Override
  public CompressionInputStream createInputStream(InputStream stream,
                                                  Decompressor decompressor) throws IOException {
    return new NonBlockedDecompressorStream(stream, decompressor,
      conf.getInt(BUFFER_SIZE_CONFIG, DEFAULT_BUFFER_SIZE_CONFIG));
  }

  @Override
  public CompressionOutputStream createOutputStream(OutputStream stream)
    throws IOException {
    return createOutputStream(stream, createCompressor());
  }

  @Override
  public CompressionOutputStream createOutputStream(OutputStream stream,
                                                    Compressor compressor) throws IOException {
    return new NonBlockedCompressorStream(stream, compressor,
      conf.getInt(BUFFER_SIZE_CONFIG, DEFAULT_BUFFER_SIZE_CONFIG));
  }

  @Override
  public Class getCompressorType() {
    return Lz4RawCompressor.class;
  }

  @Override
  public Class getDecompressorType() {
    return Lz4RawDecompressor.class;
  }

  @Override
  public String getDefaultExtension() {
    return ".lz4";
  }
}




© 2015 - 2024 Weber Informatics LLC | Privacy Policy