All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.lucene.codecs.compressing.CompressingStoredFieldsFormat Maven / Gradle / Ivy

There is a newer version: 9.10.0
Show newest version
package org.apache.lucene.codecs.compressing;

/*
 * Licensed to the Apache Software Foundation (ASF) under one or more
 * contributor license agreements.  See the NOTICE file distributed with
 * this work for additional information regarding copyright ownership.
 * The ASF licenses this file to You under the Apache License, Version 2.0
 * (the "License"); you may not use this file except in compliance with
 * the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

import java.io.IOException;

import org.apache.lucene.codecs.CodecUtil;
import org.apache.lucene.codecs.StoredFieldsFormat;
import org.apache.lucene.codecs.StoredFieldsReader;
import org.apache.lucene.codecs.StoredFieldsWriter;
import org.apache.lucene.index.FieldInfos;
import org.apache.lucene.index.MergePolicy;
import org.apache.lucene.index.SegmentInfo;
import org.apache.lucene.store.Directory;
import org.apache.lucene.store.IOContext;


/**
 * A {@link StoredFieldsFormat} that compresses documents in chunks in
 * order to improve the compression ratio.
 * 

* For a chunk size of chunkSize bytes, this {@link StoredFieldsFormat} * does not support documents larger than (231 - chunkSize) * bytes. *

* For optimal performance, you should use a {@link MergePolicy} that returns * segments that have the biggest byte size first. * @lucene.experimental */ public class CompressingStoredFieldsFormat extends StoredFieldsFormat { private final String formatName; private final String segmentSuffix; private final CompressionMode compressionMode; private final int chunkSize; private final int maxDocsPerChunk; private final int blockSize; /** * Create a new {@link CompressingStoredFieldsFormat} with an empty segment * suffix. * * @see CompressingStoredFieldsFormat#CompressingStoredFieldsFormat(String, String, CompressionMode, int, int, int) */ public CompressingStoredFieldsFormat(String formatName, CompressionMode compressionMode, int chunkSize, int maxDocsPerChunk, int blockSize) { this(formatName, "", compressionMode, chunkSize, maxDocsPerChunk, blockSize); } /** * Create a new {@link CompressingStoredFieldsFormat}. *

* formatName is the name of the format. This name will be used * in the file formats to perform * {@link CodecUtil#checkIndexHeader codec header checks}. *

* segmentSuffix is the segment suffix. This suffix is added to * the result file name only if it's not the empty string. *

* The compressionMode parameter allows you to choose between * compression algorithms that have various compression and decompression * speeds so that you can pick the one that best fits your indexing and * searching throughput. You should never instantiate two * {@link CompressingStoredFieldsFormat}s that have the same name but * different {@link CompressionMode}s. *

* chunkSize is the minimum byte size of a chunk of documents. * A value of 1 can make sense if there is redundancy across * fields. * maxDocsPerChunk is an upperbound on how many docs may be stored * in a single chunk. This is to bound the cpu costs for highly compressible data. *

* Higher values of chunkSize should improve the compression * ratio but will require more memory at indexing time and might make document * loading a little slower (depending on the size of your OS cache compared * to the size of your index). * * @param formatName the name of the {@link StoredFieldsFormat} * @param compressionMode the {@link CompressionMode} to use * @param chunkSize the minimum number of bytes of a single chunk of stored documents * @param maxDocsPerChunk the maximum number of documents in a single chunk * @param blockSize the number of chunks to store in an index block * @see CompressionMode */ public CompressingStoredFieldsFormat(String formatName, String segmentSuffix, CompressionMode compressionMode, int chunkSize, int maxDocsPerChunk, int blockSize) { this.formatName = formatName; this.segmentSuffix = segmentSuffix; this.compressionMode = compressionMode; if (chunkSize < 1) { throw new IllegalArgumentException("chunkSize must be >= 1"); } this.chunkSize = chunkSize; if (maxDocsPerChunk < 1) { throw new IllegalArgumentException("maxDocsPerChunk must be >= 1"); } this.maxDocsPerChunk = maxDocsPerChunk; if (blockSize < 1) { throw new IllegalArgumentException("blockSize must be >= 1"); } this.blockSize = blockSize; } @Override public StoredFieldsReader fieldsReader(Directory directory, SegmentInfo si, FieldInfos fn, IOContext context) throws IOException { return new CompressingStoredFieldsReader(directory, si, segmentSuffix, fn, context, formatName, compressionMode); } @Override public StoredFieldsWriter fieldsWriter(Directory directory, SegmentInfo si, IOContext context) throws IOException { return new CompressingStoredFieldsWriter(directory, si, segmentSuffix, context, formatName, compressionMode, chunkSize, maxDocsPerChunk, blockSize); } @Override public String toString() { return getClass().getSimpleName() + "(compressionMode=" + compressionMode + ", chunkSize=" + chunkSize + ", maxDocsPerChunk=" + maxDocsPerChunk + ", blockSize=" + blockSize + ")"; } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy