Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
/*
* Licensed to Elasticsearch under one or more contributor
* license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright
* ownership. Elasticsearch licenses this file to you under
* the Apache License, Version 2.0 (the "License"); you may
* not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.elasticsearch.search.facet.terms.strings;
import com.google.common.collect.ImmutableSet;
import org.apache.lucene.index.AtomicReaderContext;
import org.apache.lucene.search.Scorer;
import org.apache.lucene.util.BytesRef;
import org.elasticsearch.index.fielddata.BytesValues;
import org.elasticsearch.index.fielddata.IndexFieldData;
import org.elasticsearch.index.fielddata.ordinals.Ordinals;
import org.elasticsearch.script.SearchScript;
import org.elasticsearch.search.facet.FacetExecutor;
import org.elasticsearch.search.facet.InternalFacet;
import org.elasticsearch.search.facet.terms.TermsFacet;
import org.elasticsearch.search.internal.SearchContext;
import java.io.IOException;
import java.util.regex.Pattern;
/**
*
*/
public class TermsStringFacetExecutor extends FacetExecutor {
private final IndexFieldData indexFieldData;
private final TermsFacet.ComparatorType comparatorType;
private final SearchScript script;
private final int shardSize;
private final int size;
// the aggregation map
long missing;
long total;
private final boolean allTerms;
private final HashedAggregator aggregator;
public TermsStringFacetExecutor(IndexFieldData indexFieldData, int size, int shardSize, TermsFacet.ComparatorType comparatorType, boolean allTerms, SearchContext context,
ImmutableSet excluded, Pattern pattern, SearchScript script) {
this.indexFieldData = indexFieldData;
this.size = size;
this.shardSize = shardSize;
this.comparatorType = comparatorType;
this.script = script;
this.allTerms = allTerms;
if (excluded.isEmpty() && pattern == null && script == null) {
aggregator = new HashedAggregator();
} else {
aggregator = new HashedScriptAggregator(excluded, pattern, script);
}
if (allTerms) {
loadAllTerms(context, indexFieldData, aggregator);
}
}
@Override
public Collector collector() {
return new Collector(aggregator, allTerms);
}
@Override
public InternalFacet buildFacet(String facetName) {
try {
return HashedAggregator.buildFacet(facetName, size, shardSize, missing, total, comparatorType, aggregator);
} finally {
aggregator.release();
}
}
final class Collector extends FacetExecutor.Collector {
private final HashedAggregator aggregator;
private final boolean allTerms;
private BytesValues values;
Collector(HashedAggregator aggregator, boolean allTerms) {
this.aggregator = aggregator;
this.allTerms = allTerms;
}
@Override
public void setScorer(Scorer scorer) throws IOException {
if (script != null) {
script.setScorer(scorer);
}
}
@Override
public void setNextReader(AtomicReaderContext context) throws IOException {
values = indexFieldData.load(context).getBytesValues(true);
if (script != null) {
script.setNextReader(context);
}
}
@Override
public void collect(int doc) throws IOException {
aggregator.onDoc(doc, values);
}
@Override
public void postCollection() {
TermsStringFacetExecutor.this.missing = aggregator.missing();
TermsStringFacetExecutor.this.total = aggregator.total();
}
}
static void loadAllTerms(SearchContext context, IndexFieldData indexFieldData, HashedAggregator aggregator) {
for (AtomicReaderContext readerContext : context.searcher().getTopReaderContext().leaves()) {
int maxDoc = readerContext.reader().maxDoc();
if (indexFieldData instanceof IndexFieldData.WithOrdinals) {
BytesValues.WithOrdinals values = ((IndexFieldData.WithOrdinals) indexFieldData).load(readerContext).getBytesValues(false);
Ordinals.Docs ordinals = values.ordinals();
// 0 = docs with no value for field, so start from 1 instead
for (long ord = Ordinals.MIN_ORDINAL; ord < ordinals.getMaxOrd(); ord++) {
BytesRef value = values.getValueByOrd(ord);
aggregator.addValue(value, value.hashCode(), values);
}
} else {
BytesValues values = indexFieldData.load(readerContext).getBytesValues(true);
for (int docId = 0; docId < maxDoc; docId++) {
final int size = values.setDocument(docId);
for (int i = 0; i < size; i++) {
final BytesRef value = values.nextValue();
aggregator.addValue(value, values.currentValueHash(), values);
}
}
}
}
}
}