org.apache.lucene.analysis.CachingTokenStream Maven / Gradle / Ivy
Show all versions of elasticsearch-analysis-combo Show documentation
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.lucene.analysis;
import org.apache.lucene.util.Attribute;
import org.apache.lucene.util.AttributeSource;
import java.io.IOException;
import java.util.Iterator;
import java.util.LinkedList;
import java.util.List;
/**
* This class can be used if the token attributes of a TokenStream
* are intended to be consumed more than once. It caches
* all token attribute states locally in a List.
*
* CachingTokenStream implements the optional method
* {@link TokenStream#reset()}, which repositions the
* stream to the first Token.
*
*
It differs from Lucene's {@link CachingTokenFilter} in that it
* creates a new AttributeSource, unrelated to the provided input.
* This permits free reuse of the TokenStream, and after
* {@link #fillCache()} has been called, the underlying input
* TokenStream is available for garbage collection.
* {@link CachingTokenFilter} has no such feature because it keeps
* the same {@link AttributeSource} and if the source analyzer
* reuses the {@link TokenStream} in a call of
* {@link Analyzer#tokenStream(String, java.io.Reader)}, then a
* call on {@link org.apache.lucene.analysis.TokenStream#incrementToken()}
* will affect the current attribute state of the "cached" instance.
*/
public final class CachingTokenStream extends TokenStream {
protected List cache = null;
protected Iterator iterator = null;
protected AttributeSource.State finalState;
protected TokenStream input;
/**
* Create a new CachingTokenStream around input
,
* caching its token attributes, which can be replayed again
* after a call to {@link #reset()}.
*/
public CachingTokenStream(TokenStream input) {
this.input = input;
Iterator> attrIter = this.input.getAttributeClassesIterator();
while (attrIter.hasNext())
addAttribute(attrIter.next());
}
public void fillCache() throws IOException {
if (cache != null)
return;
cache = new LinkedList();
while(input.incrementToken()) {
cache.add(input.captureState());
}
// capture final state
input.end();
finalState = input.captureState();
input = null; // don't hold on the input
iterator = cache.iterator();
}
@Override
public final boolean incrementToken() throws IOException {
// fill cache lazily
fillCache();
if (!iterator.hasNext()) {
// the cache is exhausted, return false
return false;
}
// Since the TokenFilter can be reset, the tokens need to be preserved as immutable.
restoreState(iterator.next());
return true;
}
@Override
public final void end() {
if (finalState != null) {
restoreState(finalState);
}
}
/**
* Rewinds the iterator to the beginning of the cached list.
*
* Note that this does not call reset() on the wrapped tokenstream ever, even
* the first time. You should reset() the inner tokenstream before wrapping
* it with CachingTokenStream.
*/
@Override
public void reset() {
if(cache != null) {
iterator = cache.iterator();
}
}
}