All Downloads are FREE. Search and download functionalities are using the official Maven repository.

io.netty.handler.codec.compression.LzfEncoder Maven / Gradle / Ivy

Go to download

This artifact provides a single jar that contains all classes required to use remote EJB and JMS, including all dependencies. It is intended for use by those not using maven, maven users should just import the EJB and JMS BOM's instead (shaded JAR's cause lots of problems with maven, as it is very easy to inadvertently end up with different versions on classes on the class path).

The newest version!
/*
 * Copyright 2014 The Netty Project
 *
 * The Netty Project licenses this file to you under the Apache License,
 * version 2.0 (the "License"); you may not use this file except in compliance
 * with the License. You may obtain a copy of the License at:
 *
 *   https://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
 * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
 * License for the specific language governing permissions and limitations
 * under the License.
 */
package io.netty.handler.codec.compression;

import com.ning.compress.BufferRecycler;
import com.ning.compress.lzf.ChunkEncoder;
import com.ning.compress.lzf.LZFChunk;
import com.ning.compress.lzf.LZFEncoder;
import com.ning.compress.lzf.util.ChunkEncoderFactory;
import io.netty.buffer.ByteBuf;
import io.netty.channel.ChannelHandlerContext;
import io.netty.handler.codec.MessageToByteEncoder;
import io.netty.util.internal.PlatformDependent;

import static com.ning.compress.lzf.LZFChunk.MAX_CHUNK_LEN;

/**
 * Compresses a {@link ByteBuf} using the LZF format.
 * 

* See original LZF package * and LZF format for full description. */ public class LzfEncoder extends MessageToByteEncoder { /** * Minimum block size ready for compression. Blocks with length * less than {@link #MIN_BLOCK_TO_COMPRESS} will write as uncompressed. */ private static final int MIN_BLOCK_TO_COMPRESS = 16; private static final boolean DEFAULT_SAFE = !PlatformDependent.hasUnsafe(); /** * Compress threshold for LZF format. When the amount of input data is less than compressThreshold, * we will construct an uncompressed output according to the LZF format. *

* When the value is less than {@see ChunkEncoder#MIN_BLOCK_TO_COMPRESS}, since LZF will not compress data * that is less than {@see ChunkEncoder#MIN_BLOCK_TO_COMPRESS}, compressThreshold will not work. */ private final int compressThreshold; /** * Underlying decoder in use. */ private final ChunkEncoder encoder; /** * Object that handles details of buffer recycling. */ private final BufferRecycler recycler; /** * Creates a new LZF encoder with the most optimal available methods for underlying data access. * It will "unsafe" instance if one can be used on current JVM. * It should be safe to call this constructor as implementations are dynamically loaded; however, on some * non-standard platforms it may be necessary to use {@link #LzfEncoder(boolean)} with {@code true} param. */ public LzfEncoder() { this(DEFAULT_SAFE); } /** * Creates a new LZF encoder with specified encoding instance. * * @param safeInstance If {@code true} encoder will use {@link ChunkEncoder} that only uses * standard JDK access methods, and should work on all Java platforms and JVMs. * Otherwise encoder will try to use highly optimized {@link ChunkEncoder} * implementation that uses Sun JDK's {@link sun.misc.Unsafe} * class (which may be included by other JDK's as well). * @deprecated Use the constructor without the {@code safeInstance} parameter. */ @Deprecated public LzfEncoder(boolean safeInstance) { this(safeInstance, MAX_CHUNK_LEN); } /** * Creates a new LZF encoder with specified encoding instance and compressThreshold. * * @param safeInstance If {@code true} encoder will use {@link ChunkEncoder} that only uses standard * JDK access methods, and should work on all Java platforms and JVMs. * Otherwise encoder will try to use highly optimized {@link ChunkEncoder} * implementation that uses Sun JDK's {@link sun.misc.Unsafe} * class (which may be included by other JDK's as well). * @param totalLength Expected total length of content to compress; only matters for outgoing messages * that is smaller than maximum chunk size (64k), to optimize encoding hash tables. * @deprecated Use the constructor without the {@code safeInstance} parameter. */ @Deprecated public LzfEncoder(boolean safeInstance, int totalLength) { this(safeInstance, totalLength, MIN_BLOCK_TO_COMPRESS); } /** * Creates a new LZF encoder with specified total length of encoded chunk. You can configure it to encode * your data flow more efficient if you know the average size of messages that you send. * * @param totalLength Expected total length of content to compress; * only matters for outgoing messages that is smaller than maximum chunk size (64k), * to optimize encoding hash tables. */ public LzfEncoder(int totalLength) { this(DEFAULT_SAFE, totalLength); } /** * Creates a new LZF encoder with specified settings. * * @param totalLength Expected total length of content to compress; only matters for outgoing messages * that is smaller than maximum chunk size (64k), to optimize encoding hash tables. * @param compressThreshold Compress threshold for LZF format. When the amount of input data is less than * compressThreshold, we will construct an uncompressed output according * to the LZF format. */ public LzfEncoder(int totalLength, int compressThreshold) { this(DEFAULT_SAFE, totalLength, compressThreshold); } /** * Creates a new LZF encoder with specified settings. * * @param safeInstance If {@code true} encoder will use {@link ChunkEncoder} that only uses standard JDK * access methods, and should work on all Java platforms and JVMs. * Otherwise encoder will try to use highly optimized {@link ChunkEncoder} * implementation that uses Sun JDK's {@link sun.misc.Unsafe} * class (which may be included by other JDK's as well). * @param totalLength Expected total length of content to compress; only matters for outgoing messages * that is smaller than maximum chunk size (64k), to optimize encoding hash tables. * @param compressThreshold Compress threshold for LZF format. When the amount of input data is less than * compressThreshold, we will construct an uncompressed output according * to the LZF format. * @deprecated Use the constructor without the {@code safeInstance} parameter. */ @Deprecated public LzfEncoder(boolean safeInstance, int totalLength, int compressThreshold) { super(false); if (totalLength < MIN_BLOCK_TO_COMPRESS || totalLength > MAX_CHUNK_LEN) { throw new IllegalArgumentException("totalLength: " + totalLength + " (expected: " + MIN_BLOCK_TO_COMPRESS + '-' + MAX_CHUNK_LEN + ')'); } if (compressThreshold < MIN_BLOCK_TO_COMPRESS) { // not a suitable value. throw new IllegalArgumentException("compressThreshold:" + compressThreshold + " expected >=" + MIN_BLOCK_TO_COMPRESS); } this.compressThreshold = compressThreshold; this.encoder = safeInstance ? ChunkEncoderFactory.safeNonAllocatingInstance(totalLength) : ChunkEncoderFactory.optimalNonAllocatingInstance(totalLength); this.recycler = BufferRecycler.instance(); } @Override protected void encode(ChannelHandlerContext ctx, ByteBuf in, ByteBuf out) throws Exception { final int length = in.readableBytes(); final int idx = in.readerIndex(); final byte[] input; final int inputPtr; if (in.hasArray()) { input = in.array(); inputPtr = in.arrayOffset() + idx; } else { input = recycler.allocInputBuffer(length); in.getBytes(idx, input, 0, length); inputPtr = 0; } // Estimate may apparently under-count by one in some cases. final int maxOutputLength = LZFEncoder.estimateMaxWorkspaceSize(length) + 1; out.ensureWritable(maxOutputLength); final byte[] output; final int outputPtr; if (out.hasArray()) { output = out.array(); outputPtr = out.arrayOffset() + out.writerIndex(); } else { output = new byte[maxOutputLength]; outputPtr = 0; } final int outputLength; if (length >= compressThreshold) { // compress. outputLength = encodeCompress(input, inputPtr, length, output, outputPtr); } else { // not compress. outputLength = encodeNonCompress(input, inputPtr, length, output, outputPtr); } if (out.hasArray()) { out.writerIndex(out.writerIndex() + outputLength); } else { out.writeBytes(output, 0, outputLength); } in.skipBytes(length); if (!in.hasArray()) { recycler.releaseInputBuffer(input); } } private int encodeCompress(byte[] input, int inputPtr, int length, byte[] output, int outputPtr) { return LZFEncoder.appendEncoded(encoder, input, inputPtr, length, output, outputPtr) - outputPtr; } private static int lzfEncodeNonCompress(byte[] input, int inputPtr, int length, byte[] output, int outputPtr) { int left = length; int chunkLen = Math.min(LZFChunk.MAX_CHUNK_LEN, left); outputPtr = LZFChunk.appendNonCompressed(input, inputPtr, chunkLen, output, outputPtr); left -= chunkLen; if (left < 1) { return outputPtr; } inputPtr += chunkLen; do { chunkLen = Math.min(left, LZFChunk.MAX_CHUNK_LEN); outputPtr = LZFChunk.appendNonCompressed(input, inputPtr, chunkLen, output, outputPtr); inputPtr += chunkLen; left -= chunkLen; } while (left > 0); return outputPtr; } /** * Use lzf uncompressed format to encode a piece of input. */ private static int encodeNonCompress(byte[] input, int inputPtr, int length, byte[] output, int outputPtr) { return lzfEncodeNonCompress(input, inputPtr, length, output, outputPtr) - outputPtr; } @Override public void handlerRemoved(ChannelHandlerContext ctx) throws Exception { encoder.close(); super.handlerRemoved(ctx); } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy