All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.lucene.store.instantiated.InstantiatedIndex Maven / Gradle / Ivy

There is a newer version: 3.6.2
Show newest version
package org.apache.lucene.store.instantiated;

/**
 * Copyright 2006 The Apache Software Foundation
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

import java.io.IOException;
import java.io.Serializable;
import java.util.ArrayList;
import java.util.Collection;
import java.util.HashMap;
import java.util.HashSet;
import java.util.List;
import java.util.Map;
import java.util.Set;

import org.apache.lucene.analysis.Analyzer;
import org.apache.lucene.document.Document;
import org.apache.lucene.document.Field;
import org.apache.lucene.index.IndexReader;
import org.apache.lucene.index.Term;
import org.apache.lucene.index.TermEnum;
import org.apache.lucene.index.TermPositionVector;
import org.apache.lucene.index.TermPositions;
import org.apache.lucene.util.BitVector;

/**
 * Represented as a coupled graph of class instances, this
 * all-in-memory index store implementation delivers search
 * results up to a 100 times faster than the file-centric RAMDirectory
 * at the cost of greater RAM consumption.
 * 

* WARNING: This contrib is experimental and the APIs may change without warning. *

* There are no read and write locks in this store. * {@link InstantiatedIndexReader} {@link InstantiatedIndexReader#isCurrent()} all the time * and {@link org.apache.lucene.store.instantiated.InstantiatedIndexWriter} * will attempt to update instances of the object graph in memory * at the same time as a searcher is reading from it. * * Consider using InstantiatedIndex as if it was immutable. */ public class InstantiatedIndex implements Serializable { private static final long serialVersionUID = 1l; private long version = System.currentTimeMillis(); private InstantiatedDocument[] documentsByNumber; private BitVector deletedDocuments; private Map> termsByFieldAndText; private InstantiatedTerm[] orderedTerms; private Map normsByFieldNameAndDocumentNumber; private FieldSettings fieldSettings; /** * Creates an empty instantiated index for you to fill with data using an {@link org.apache.lucene.store.instantiated.InstantiatedIndexWriter}. */ public InstantiatedIndex() { initialize(); } void initialize() { // todo: clear index without loosing memory (uncouple stuff) termsByFieldAndText = new HashMap>(); fieldSettings = new FieldSettings(); orderedTerms = new InstantiatedTerm[0]; documentsByNumber = new InstantiatedDocument[0]; normsByFieldNameAndDocumentNumber = new HashMap(); } /** * Creates a new instantiated index that looks just like the index in a specific state as represented by a reader. * * @param sourceIndexReader the source index this new instantiated index will be copied from. * @throws IOException if the source index is not optimized, or when accessing the source. */ public InstantiatedIndex(IndexReader sourceIndexReader) throws IOException { this(sourceIndexReader, null); } /** * Creates a new instantiated index that looks just like the index in a specific state as represented by a reader. * * @param sourceIndexReader the source index this new instantiated index will be copied from. * @param fields fields to be added, or null for all * @throws IOException if the source index is not optimized, or when accessing the source. */ public InstantiatedIndex(IndexReader sourceIndexReader, Set fields) throws IOException { if (!sourceIndexReader.isOptimized()) { System.out.println(("Source index is not optimized.")); //throw new IOException("Source index is not optimized."); } initialize(); Collection allFieldNames = sourceIndexReader.getFieldNames(IndexReader.FieldOption.ALL); // load field options Collection indexedNames = sourceIndexReader.getFieldNames(IndexReader.FieldOption.INDEXED); for (String name : indexedNames) { FieldSetting setting = fieldSettings.get(name, true); setting.indexed = true; } Collection indexedNoVecNames = sourceIndexReader.getFieldNames(IndexReader.FieldOption.INDEXED_NO_TERMVECTOR); for (String name : indexedNoVecNames) { FieldSetting setting = fieldSettings.get(name, true); setting.storeTermVector = false; setting.indexed = true; } Collection indexedVecNames = sourceIndexReader.getFieldNames(IndexReader.FieldOption.INDEXED_WITH_TERMVECTOR); for (String name : indexedVecNames) { FieldSetting setting = fieldSettings.get(name, true); setting.storeTermVector = true; setting.indexed = true; } Collection payloadNames = sourceIndexReader.getFieldNames(IndexReader.FieldOption.STORES_PAYLOADS); for (String name : payloadNames) { FieldSetting setting = fieldSettings.get(name, true); setting.storePayloads = true; } Collection termVecNames = sourceIndexReader.getFieldNames(IndexReader.FieldOption.TERMVECTOR); for (String name : termVecNames) { FieldSetting setting = fieldSettings.get(name, true); setting.storeTermVector = true; } Collection termVecOffsetNames = sourceIndexReader.getFieldNames(IndexReader.FieldOption.TERMVECTOR_WITH_OFFSET); for (String name : termVecOffsetNames) { FieldSetting setting = fieldSettings.get(name, true); setting.storeOffsetWithTermVector = true; } Collection termVecPosNames = sourceIndexReader.getFieldNames(IndexReader.FieldOption.TERMVECTOR_WITH_POSITION); for (String name : termVecPosNames) { FieldSetting setting = fieldSettings.get(name, true); setting.storePositionWithTermVector = true; } Collection termVecPosOffNames = sourceIndexReader.getFieldNames(IndexReader.FieldOption.TERMVECTOR_WITH_POSITION_OFFSET); for (String name : termVecPosOffNames) { FieldSetting setting = fieldSettings.get(name, true); setting.storeOffsetWithTermVector = true; setting.storePositionWithTermVector = true; } Collection unindexedNames = sourceIndexReader.getFieldNames(IndexReader.FieldOption.UNINDEXED); for (String name : unindexedNames) { FieldSetting setting = fieldSettings.get(name, true); setting.indexed = false; } documentsByNumber = new InstantiatedDocument[sourceIndexReader.maxDoc()]; if (sourceIndexReader.hasDeletions()) { deletedDocuments = new BitVector(sourceIndexReader.maxDoc()); } // create documents for (int i = 0; i < sourceIndexReader.maxDoc(); i++) { if (sourceIndexReader.hasDeletions() && sourceIndexReader.isDeleted(i)) { deletedDocuments.set(i); } else { InstantiatedDocument document = new InstantiatedDocument(); // copy stored fields from source reader Document sourceDocument = sourceIndexReader.document(i); for (Field field : (List) sourceDocument.getFields()) { if (fields == null || fields.contains(field.name())) { document.getDocument().add(field); } } document.setDocumentNumber(i); documentsByNumber[i] = document; for (Field field : (List) document.getDocument().getFields()) { if (fields == null || fields.contains(field.name())) { if (field.isTermVectorStored()) { if (document.getVectorSpace() == null) { document.setVectorSpace(new HashMap>()); } document.getVectorSpace().put(field.name(), new ArrayList()); } } } } } // create norms for (String fieldName : allFieldNames) { if (fields == null || fields.contains(fieldName)) { getNormsByFieldNameAndDocumentNumber().put(fieldName, sourceIndexReader.norms(fieldName)); } } // create terms for (String fieldName : allFieldNames) { if (fields == null || fields.contains(fieldName)) { getTermsByFieldAndText().put(fieldName, new HashMap(5000)); } } List terms = new ArrayList(5000 * getTermsByFieldAndText().size()); TermEnum termEnum = sourceIndexReader.terms(); while (termEnum.next()) { if (fields == null || fields.contains(termEnum.term().field())) { // todo skipto if not using field InstantiatedTerm instantiatedTerm = new InstantiatedTerm(termEnum.term().field(), termEnum.term().text()); getTermsByFieldAndText().get(termEnum.term().field()).put(termEnum.term().text(), instantiatedTerm); instantiatedTerm.setTermIndex(terms.size()); terms.add(instantiatedTerm); instantiatedTerm.setAssociatedDocuments(new InstantiatedTermDocumentInformation[termEnum.docFreq()]); } } termEnum.close(); orderedTerms = terms.toArray(new InstantiatedTerm[terms.size()]); // create term-document informations for (InstantiatedTerm term : orderedTerms) { TermPositions termPositions = sourceIndexReader.termPositions(term.getTerm()); int position = 0; while (termPositions.next()) { InstantiatedDocument document = documentsByNumber[termPositions.doc()]; byte[][] payloads = new byte[termPositions.freq()][]; int[] positions = new int[termPositions.freq()]; for (int i = 0; i < termPositions.freq(); i++) { positions[i] = termPositions.nextPosition(); if (termPositions.isPayloadAvailable()) { payloads[i] = new byte[termPositions.getPayloadLength()]; termPositions.getPayload(payloads[i], 0); } } InstantiatedTermDocumentInformation termDocumentInformation = new InstantiatedTermDocumentInformation(term, document, positions, payloads); term.getAssociatedDocuments()[position++] = termDocumentInformation; if (document.getVectorSpace() != null && document.getVectorSpace().containsKey(term.field())) { document.getVectorSpace().get(term.field()).add(termDocumentInformation); } // termDocumentInformation.setIndexFromTerm(indexFromTerm++); } } // load offsets to term-document informations for (InstantiatedDocument document : getDocumentsByNumber()) { if (document == null) { continue; // deleted } for (Field field : (List) document.getDocument().getFields()) { if (field.isTermVectorStored() && field.isStoreOffsetWithTermVector()) { TermPositionVector termPositionVector = (TermPositionVector) sourceIndexReader.getTermFreqVector(document.getDocumentNumber(), field.name()); if (termPositionVector != null) { for (int i = 0; i < termPositionVector.getTerms().length; i++) { String token = termPositionVector.getTerms()[i]; InstantiatedTerm term = findTerm(field.name(), token); InstantiatedTermDocumentInformation termDocumentInformation = term.getAssociatedDocument(document.getDocumentNumber()); termDocumentInformation.setTermOffsets(termPositionVector.getOffsets(i)); } } } } } } public InstantiatedIndexWriter indexWriterFactory(Analyzer analyzer, boolean create) throws IOException { return new InstantiatedIndexWriter(this, analyzer, create); } public InstantiatedIndexReader indexReaderFactory() throws IOException { return new InstantiatedIndexReader(this); } public void close() throws IOException { // todo: decouple everything } InstantiatedTerm findTerm(Term term) { return findTerm(term.field(), term.text()); } InstantiatedTerm findTerm(String field, String text) { Map termsByField = termsByFieldAndText.get(field); if (termsByField == null) { return null; } else { return termsByField.get(text); } } public Map> getTermsByFieldAndText() { return termsByFieldAndText; } public InstantiatedTerm[] getOrderedTerms() { return orderedTerms; } public InstantiatedDocument[] getDocumentsByNumber() { return documentsByNumber; } public Map getNormsByFieldNameAndDocumentNumber() { return normsByFieldNameAndDocumentNumber; } void setNormsByFieldNameAndDocumentNumber(Map normsByFieldNameAndDocumentNumber) { this.normsByFieldNameAndDocumentNumber = normsByFieldNameAndDocumentNumber; } public BitVector getDeletedDocuments() { return deletedDocuments; } void setDeletedDocuments(BitVector deletedDocuments) { this.deletedDocuments = deletedDocuments; } void setOrderedTerms(InstantiatedTerm[] orderedTerms) { this.orderedTerms = orderedTerms; } void setDocumentsByNumber(InstantiatedDocument[] documentsByNumber) { this.documentsByNumber = documentsByNumber; } public long getVersion() { return version; } void setVersion(long version) { this.version = version; } FieldSettings getFieldSettings() { return fieldSettings; } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy