All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.lucene.search.Weight Maven / Gradle / Ivy

/*
 * Licensed to the Apache Software Foundation (ASF) under one or more
 * contributor license agreements.  See the NOTICE file distributed with
 * this work for additional information regarding copyright ownership.
 * The ASF licenses this file to You under the Apache License, Version 2.0
 * (the "License"); you may not use this file except in compliance with
 * the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.lucene.search;


import java.io.IOException;
import java.util.Set;
import java.util.Arrays;

import org.apache.lucene.index.IndexReaderContext;
import org.apache.lucene.index.LeafReader;
import org.apache.lucene.index.LeafReaderContext;
import org.apache.lucene.index.Term;
import org.apache.lucene.util.Bits;

/**
 * Expert: Calculate query weights and build query scorers.
 * 

* The purpose of {@link Weight} is to ensure searching does not modify a * {@link Query}, so that a {@link Query} instance can be reused.
* {@link IndexSearcher} dependent state of the query should reside in the * {@link Weight}.
* {@link org.apache.lucene.index.LeafReader} dependent state should reside in the {@link Scorer}. *

* Since {@link Weight} creates {@link Scorer} instances for a given * {@link org.apache.lucene.index.LeafReaderContext} ({@link #scorer(org.apache.lucene.index.LeafReaderContext)}) * callers must maintain the relationship between the searcher's top-level * {@link IndexReaderContext} and the context used to create a {@link Scorer}. *

* A Weight is used in the following way: *

    *
  1. A Weight is constructed by a top-level query, given a * IndexSearcher ({@link Query#createWeight(IndexSearcher, ScoreMode, float)}). *
  2. A Scorer is constructed by * {@link #scorer(org.apache.lucene.index.LeafReaderContext)}. *
* * @since 2.9 */ public abstract class Weight implements SegmentCacheable { protected final Query parentQuery; /** Sole constructor, typically invoked by sub-classes. * @param query the parent query */ protected Weight(Query query) { this.parentQuery = query; } /** * Expert: adds all terms occurring in this query to the terms set. If the * {@link Weight} was created with {@code needsScores == true} then this * method will only extract terms which are used for scoring, otherwise it * will extract all terms which are used for matching. * * @deprecated Use {@link Query#visit(QueryVisitor)} with {@link QueryVisitor#termCollector(Set)} */ @Deprecated public abstract void extractTerms(Set terms); /** * Returns {@link Matches} for a specific document, or {@code null} if the document * does not match the parent query * * A query match that contains no position information (for example, a Point or * DocValues query) will return {@link MatchesUtils#MATCH_WITH_NO_TERMS} * * @param context the reader's context to create the {@link Matches} for * @param doc the document's id relative to the given context's reader * @lucene.experimental */ public Matches matches(LeafReaderContext context, int doc) throws IOException { ScorerSupplier scorerSupplier = scorerSupplier(context); if (scorerSupplier == null) { return null; } Scorer scorer = scorerSupplier.get(1); final TwoPhaseIterator twoPhase = scorer.twoPhaseIterator(); if (twoPhase == null) { if (scorer.iterator().advance(doc) != doc) { return null; } } else { if (twoPhase.approximation().advance(doc) != doc || twoPhase.matches() == false) { return null; } } return MatchesUtils.MATCH_WITH_NO_TERMS; } /** * An explanation of the score computation for the named document. * * @param context the readers context to create the {@link Explanation} for. * @param doc the document's id relative to the given context's reader * @return an Explanation for the score * @throws IOException if an {@link IOException} occurs */ public abstract Explanation explain(LeafReaderContext context, int doc) throws IOException; /** The query that this concerns. */ public final Query getQuery() { return parentQuery; } /** * Returns a {@link Scorer} which can iterate in order over all matching * documents and assign them a score. *

* NOTE: null can be returned if no documents will be scored by this * query. *

* NOTE: The returned {@link Scorer} does not have * {@link LeafReader#getLiveDocs()} applied, they need to be checked on top. * * @param context * the {@link org.apache.lucene.index.LeafReaderContext} for which to return the {@link Scorer}. * * @return a {@link Scorer} which scores documents in/out-of order. * @throws IOException if there is a low-level I/O error */ public abstract Scorer scorer(LeafReaderContext context) throws IOException; /** * Optional method. * Get a {@link ScorerSupplier}, which allows to know the cost of the {@link Scorer} * before building it. The default implementation calls {@link #scorer} and * builds a {@link ScorerSupplier} wrapper around it. * @see #scorer */ public ScorerSupplier scorerSupplier(LeafReaderContext context) throws IOException { final Scorer scorer = scorer(context); if (scorer == null) { return null; } return new ScorerSupplier() { @Override public Scorer get(long leadCost) { return scorer; } @Override public long cost() { return scorer.iterator().cost(); } }; } /** * Optional method, to return a {@link BulkScorer} to * score the query and send hits to a {@link Collector}. * Only queries that have a different top-level approach * need to override this; the default implementation * pulls a normal {@link Scorer} and iterates and * collects the resulting hits which are not marked as deleted. * * @param context * the {@link org.apache.lucene.index.LeafReaderContext} for which to return the {@link Scorer}. * * @return a {@link BulkScorer} which scores documents and * passes them to a collector. * @throws IOException if there is a low-level I/O error */ public BulkScorer bulkScorer(LeafReaderContext context) throws IOException { Scorer scorer = scorer(context); if (scorer == null) { // No docs match return null; } // This impl always scores docs in order, so we can // ignore scoreDocsInOrder: return new DefaultBulkScorer(scorer); } /** Just wraps a Scorer and performs top scoring using it. * @lucene.internal */ protected static class DefaultBulkScorer extends BulkScorer { private final Scorer scorer; private final DocIdSetIterator iterator; private final TwoPhaseIterator twoPhase; /** Sole constructor. */ public DefaultBulkScorer(Scorer scorer) { if (scorer == null) { throw new NullPointerException(); } this.scorer = scorer; this.iterator = scorer.iterator(); this.twoPhase = scorer.twoPhaseIterator(); } @Override public long cost() { return iterator.cost(); } @Override public int score(LeafCollector collector, Bits acceptDocs, int min, int max) throws IOException { collector.setScorer(scorer); DocIdSetIterator scorerIterator = twoPhase == null ? iterator : twoPhase.approximation(); DocIdSetIterator competitiveIterator = collector.competitiveIterator(); DocIdSetIterator filteredIterator; if (competitiveIterator == null) { filteredIterator = scorerIterator; } else { // Wrap CompetitiveIterator and ScorerIterator start with (i.e., calling nextDoc()) the last // visited docID because ConjunctionDISI might have advanced to it in the previous // scoreRange, but we didn't process due to the range limit of scoreRange. if (scorerIterator.docID() != -1) { scorerIterator = new StartDISIWrapper(scorerIterator); } if (competitiveIterator.docID() != -1) { competitiveIterator = new StartDISIWrapper(competitiveIterator); } // filter scorerIterator to keep only competitive docs as defined by collector filteredIterator = ConjunctionDISI.intersectIterators(Arrays.asList(scorerIterator, competitiveIterator)); } if (filteredIterator.docID() == -1 && min == 0 && max == DocIdSetIterator.NO_MORE_DOCS) { scoreAll(collector, filteredIterator, twoPhase, acceptDocs); return DocIdSetIterator.NO_MORE_DOCS; } else { int doc = filteredIterator.docID(); if (doc < min) { doc = filteredIterator.advance(min); } return scoreRange(collector, filteredIterator, twoPhase, acceptDocs, doc, max); } } /** Specialized method to bulk-score a range of hits; we * separate this from {@link #scoreAll} to help out * hotspot. * See LUCENE-5487 */ static int scoreRange(LeafCollector collector, DocIdSetIterator iterator, TwoPhaseIterator twoPhase, Bits acceptDocs, int currentDoc, int end) throws IOException { if (twoPhase == null) { while (currentDoc < end) { if (acceptDocs == null || acceptDocs.get(currentDoc)) { collector.collect(currentDoc); } currentDoc = iterator.nextDoc(); } return currentDoc; } else { while (currentDoc < end) { if ((acceptDocs == null || acceptDocs.get(currentDoc)) && twoPhase.matches()) { collector.collect(currentDoc); } currentDoc = iterator.nextDoc(); } return currentDoc; } } /** Specialized method to bulk-score all hits; we * separate this from {@link #scoreRange} to help out * hotspot. * See LUCENE-5487 */ static void scoreAll(LeafCollector collector, DocIdSetIterator iterator, TwoPhaseIterator twoPhase, Bits acceptDocs) throws IOException { if (twoPhase == null) { for (int doc = iterator.nextDoc(); doc != DocIdSetIterator.NO_MORE_DOCS; doc = iterator.nextDoc()) { if (acceptDocs == null || acceptDocs.get(doc)) { collector.collect(doc); } } } else { // The scorer has an approximation, so run the approximation first, then check acceptDocs, then confirm for (int doc = iterator.nextDoc(); doc != DocIdSetIterator.NO_MORE_DOCS; doc = iterator.nextDoc()) { if ((acceptDocs == null || acceptDocs.get(doc)) && twoPhase.matches()) { collector.collect(doc); } } } } } /** Wraps an internal docIdSetIterator for it to start with the last visited docID */ private static class StartDISIWrapper extends DocIdSetIterator { private final DocIdSetIterator in; private final int startDocID; private int docID = -1; StartDISIWrapper(DocIdSetIterator in) { this.in = in; this.startDocID = in.docID(); } @Override public int docID() { return docID; } @Override public int nextDoc() throws IOException { return advance(docID + 1); } @Override public int advance(int target) throws IOException { if (target <= startDocID) { return docID = startDocID; } return docID = in.advance(target); } @Override public long cost() { return in.cost(); } } }





© 2015 - 2025 Weber Informatics LLC | Privacy Policy