Please wait. This can take some minutes ...
Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance.
Project price only 1 $
You can buy this project and download/modify it how often you want.
org.janusgraph.diskstorage.lucene.LuceneIndex Maven / Gradle / Ivy
// Copyright 2017 JanusGraph Authors
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package org.janusgraph.diskstorage.lucene;
import com.google.common.base.Preconditions;
import com.google.common.collect.Iterators;
import com.google.common.collect.Sets;
import org.apache.commons.io.FileUtils;
import org.apache.commons.lang3.StringUtils;
import org.apache.lucene.analysis.Analyzer;
import org.apache.lucene.analysis.CachingTokenFilter;
import org.apache.lucene.analysis.tokenattributes.OffsetAttribute;
import org.apache.lucene.analysis.tokenattributes.TermToBytesRefAttribute;
import org.apache.lucene.document.Document;
import org.apache.lucene.document.DoubleDocValuesField;
import org.apache.lucene.document.DoublePoint;
import org.apache.lucene.document.Field;
import org.apache.lucene.document.IntPoint;
import org.apache.lucene.document.LongPoint;
import org.apache.lucene.document.NumericDocValuesField;
import org.apache.lucene.document.SortedDocValuesField;
import org.apache.lucene.document.StoredField;
import org.apache.lucene.document.StringField;
import org.apache.lucene.document.TextField;
import org.apache.lucene.index.DirectoryReader;
import org.apache.lucene.index.IndexNotFoundException;
import org.apache.lucene.index.IndexReader;
import org.apache.lucene.index.IndexWriter;
import org.apache.lucene.index.IndexWriterConfig;
import org.apache.lucene.index.IndexableField;
import org.apache.lucene.index.Term;
import org.apache.lucene.queryparser.classic.ParseException;
import org.apache.lucene.queryparser.classic.QueryParser;
import org.apache.lucene.search.BooleanClause;
import org.apache.lucene.search.BooleanQuery;
import org.apache.lucene.search.BooleanQuery.Builder;
import org.apache.lucene.search.DocValuesFieldExistsQuery;
import org.apache.lucene.search.FuzzyQuery;
import org.apache.lucene.search.IndexSearcher;
import org.apache.lucene.search.MatchAllDocsQuery;
import org.apache.lucene.search.MatchNoDocsQuery;
import org.apache.lucene.search.NormsFieldExistsQuery;
import org.apache.lucene.search.PrefixQuery;
import org.apache.lucene.search.Query;
import org.apache.lucene.search.RegexpQuery;
import org.apache.lucene.search.Sort;
import org.apache.lucene.search.SortField;
import org.apache.lucene.search.TermQuery;
import org.apache.lucene.search.TermRangeQuery;
import org.apache.lucene.search.TopDocs;
import org.apache.lucene.search.TopFieldDocs;
import org.apache.lucene.spatial.SpatialStrategy;
import org.apache.lucene.spatial.prefix.PrefixTreeStrategy;
import org.apache.lucene.spatial.prefix.RecursivePrefixTreeStrategy;
import org.apache.lucene.spatial.prefix.tree.QuadPrefixTree;
import org.apache.lucene.spatial.prefix.tree.SpatialPrefixTree;
import org.apache.lucene.spatial.query.SpatialArgs;
import org.apache.lucene.spatial.query.SpatialOperation;
import org.apache.lucene.spatial.vector.PointVectorStrategy;
import org.apache.lucene.store.Directory;
import org.apache.lucene.store.FSDirectory;
import org.apache.lucene.util.BytesRef;
import org.janusgraph.core.Cardinality;
import org.janusgraph.core.attribute.Cmp;
import org.janusgraph.core.attribute.Geo;
import org.janusgraph.core.attribute.Geoshape;
import org.janusgraph.core.attribute.Text;
import org.janusgraph.core.schema.Mapping;
import org.janusgraph.diskstorage.BackendException;
import org.janusgraph.diskstorage.BaseTransaction;
import org.janusgraph.diskstorage.BaseTransactionConfig;
import org.janusgraph.diskstorage.BaseTransactionConfigurable;
import org.janusgraph.diskstorage.PermanentBackendException;
import org.janusgraph.diskstorage.TemporaryBackendException;
import org.janusgraph.diskstorage.configuration.Configuration;
import org.janusgraph.diskstorage.indexing.IndexEntry;
import org.janusgraph.diskstorage.indexing.IndexFeatures;
import org.janusgraph.diskstorage.indexing.IndexMutation;
import org.janusgraph.diskstorage.indexing.IndexProvider;
import org.janusgraph.diskstorage.indexing.IndexQuery;
import org.janusgraph.diskstorage.indexing.KeyInformation;
import org.janusgraph.diskstorage.indexing.RawQuery;
import org.janusgraph.graphdb.configuration.GraphDatabaseConfiguration;
import org.janusgraph.graphdb.database.serialize.AttributeUtils;
import org.janusgraph.graphdb.internal.Order;
import org.janusgraph.graphdb.query.JanusGraphPredicate;
import org.janusgraph.graphdb.query.QueryUtil;
import org.janusgraph.graphdb.query.condition.And;
import org.janusgraph.graphdb.query.condition.Condition;
import org.janusgraph.graphdb.query.condition.Not;
import org.janusgraph.graphdb.query.condition.Or;
import org.janusgraph.graphdb.query.condition.PredicateCondition;
import org.janusgraph.graphdb.tinkerpop.optimize.step.Aggregation;
import org.janusgraph.graphdb.types.ParameterType;
import org.janusgraph.util.system.IOUtils;
import org.locationtech.spatial4j.context.SpatialContext;
import org.locationtech.spatial4j.shape.Shape;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.io.File;
import java.io.IOException;
import java.nio.file.DirectoryStream;
import java.nio.file.Files;
import java.nio.file.Path;
import java.nio.file.Paths;
import java.time.Instant;
import java.util.AbstractMap.SimpleEntry;
import java.util.ArrayList;
import java.util.Collections;
import java.util.Date;
import java.util.HashMap;
import java.util.Iterator;
import java.util.List;
import java.util.Map;
import java.util.Optional;
import java.util.Set;
import java.util.UUID;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.locks.ReentrantLock;
import java.util.function.Function;
import java.util.stream.Collectors;
import java.util.stream.Stream;
/**
* @author Matthias Broecheler ([email protected] )
*/
public class LuceneIndex implements IndexProvider {
private static final Logger log = LoggerFactory.getLogger(LuceneIndex.class);
static final String DOCID = "_____elementid";
private static final String STRING_SUFFIX = "_____s";
private static final String GEOID = "_____geo";
private static final Set FIELDS_TO_LOAD = Sets.newHashSet(DOCID);
private static final IndexFeatures LUCENE_FEATURES = new IndexFeatures.Builder()
.setDefaultStringMapping(Mapping.TEXT)
.supportedStringMappings(Mapping.TEXT, Mapping.STRING, Mapping.TEXTSTRING)
.supportsCardinality(Cardinality.SINGLE)
.supportsCardinality(Cardinality.LIST)
.supportsCardinality(Cardinality.SET)
.supportsCustomAnalyzer()
.supportsNanoseconds()
.supportsGeoContains()
.supportNotQueryNormalForm()
.build();
/**
* Default tree levels used when creating the prefix tree.
*/
public static final int DEFAULT_GEO_MAX_LEVELS = 20;
/**
* Default measure of shape precision used when creating the prefix tree.
*/
public static final double DEFAULT_GEO_DIST_ERROR_PCT = 0.025;
private static final Map SPATIAL_PREDICATES = spatialPredicates();
private final Map writers = new HashMap<>(4);
private final ReentrantLock writerLock = new ReentrantLock();
private final Map spatial = new ConcurrentHashMap<>(12);
private final SpatialContext ctx = Geoshape.getSpatialContext();
private final String basePath;
/**
* lazy cache for the delegating analyzers used for writing or querrying for each store
*/
private final Map delegatingAnalyzers = new HashMap<>();
public LuceneIndex(Configuration config) {
final String dir = config.get(GraphDatabaseConfiguration.INDEX_DIRECTORY);
final File directory = new File(dir);
if ((!directory.exists() && !directory.mkdirs()) || !directory.isDirectory() || !directory.canWrite()) {
throw new IllegalArgumentException("Cannot access or write to directory: " + dir);
}
basePath = directory.getAbsolutePath();
log.debug("Configured Lucene to use base directory [{}]", basePath);
}
private Directory getStoreDirectory(String store) throws BackendException {
Preconditions.checkArgument(StringUtils.isAlphanumeric(store), "Invalid store name: %s", store);
final String dir = basePath + File.separator + store;
try {
final File path = new File(dir);
if ((!path.exists() && !path.mkdirs()) || !path.isDirectory() || !path.canWrite()) {
throw new PermanentBackendException("Cannot access or write to directory: " + dir);
}
log.debug("Opening store directory [{}]", path);
return FSDirectory.open(path.toPath());
} catch (final IOException e) {
throw new PermanentBackendException("Could not open directory: " + dir, e);
}
}
private IndexWriter getWriter(String store, KeyInformation.IndexRetriever informations) throws BackendException {
Preconditions.checkArgument(writerLock.isHeldByCurrentThread());
IndexWriter writer = writers.get(store);
if (writer == null) {
final LuceneCustomAnalyzer analyzer = delegatingAnalyzerFor(store, informations);
final IndexWriterConfig iwc = new IndexWriterConfig(analyzer);
iwc.setOpenMode(IndexWriterConfig.OpenMode.CREATE_OR_APPEND);
try {
writer = new IndexWriter(getStoreDirectory(store), iwc);
writers.put(store, writer);
} catch (final IOException e) {
throw new PermanentBackendException("Could not create writer", e);
}
}
return writer;
}
private SpatialStrategy getSpatialStrategy(String key, KeyInformation ki) {
SpatialStrategy strategy = spatial.get(key);
final Mapping mapping = Mapping.getMapping(ki);
final int maxLevels = ParameterType.INDEX_GEO_MAX_LEVELS.findParameter(ki.getParameters(),
DEFAULT_GEO_MAX_LEVELS);
final double distErrorPct = ParameterType.INDEX_GEO_DIST_ERROR_PCT.findParameter(ki.getParameters(),
DEFAULT_GEO_DIST_ERROR_PCT);
if (strategy == null) {
synchronized (spatial) {
if (!spatial.containsKey(key)) {
// SpatialPrefixTree grid = new GeohashPrefixTree(ctx, GEO_MAX_LEVELS);
// strategy = new RecursivePrefixTreeStrategy(grid, key);
if (mapping == Mapping.DEFAULT) {
strategy = PointVectorStrategy.newInstance(ctx, key);
} else {
final SpatialPrefixTree grid = new QuadPrefixTree(ctx, maxLevels);
strategy = new RecursivePrefixTreeStrategy(grid, key);
((PrefixTreeStrategy) strategy).setDistErrPct(distErrorPct);
}
spatial.put(key, strategy);
} else return spatial.get(key);
}
}
return strategy;
}
private static Map spatialPredicates() {
return Collections.unmodifiableMap(Stream.of(
new SimpleEntry<>(Geo.WITHIN, SpatialOperation.IsWithin),
new SimpleEntry<>(Geo.CONTAINS, SpatialOperation.Contains),
new SimpleEntry<>(Geo.INTERSECT, SpatialOperation.Intersects),
new SimpleEntry<>(Geo.DISJOINT, SpatialOperation.IsDisjointTo))
.collect(Collectors.toMap(SimpleEntry::getKey, SimpleEntry::getValue)));
}
@Override
public void register(String store, String key, KeyInformation information, BaseTransaction tx) throws BackendException {
final Class> dataType = information.getDataType();
final Mapping map = Mapping.getMapping(information);
Preconditions.checkArgument(map == Mapping.DEFAULT || AttributeUtils.isString(dataType) ||
(map == Mapping.PREFIX_TREE && AttributeUtils.isGeo(dataType)),
"Specified illegal mapping [%s] for data type [%s]", map, dataType);
}
@Override
public void mutate(Map> mutations, KeyInformation.IndexRetriever information, BaseTransaction tx) throws BackendException {
final Transaction ltx = (Transaction) tx;
writerLock.lock();
try {
for (final Map.Entry> stores : mutations.entrySet()) {
mutateStores(stores, information);
}
ltx.postCommit();
} catch (final IOException e) {
throw new TemporaryBackendException("Could not update Lucene index", e);
} finally {
writerLock.unlock();
}
}
private void mutateStores(Map.Entry> stores, KeyInformation.IndexRetriever information) throws IOException, BackendException {
IndexReader reader = null;
try {
final String storeName = stores.getKey();
final IndexWriter writer = getWriter(storeName, information);
reader = DirectoryReader.open(writer, true, true);
final IndexSearcher searcher = new IndexSearcher(reader);
final KeyInformation.StoreRetriever storeRetriever = information.get(storeName);
for (final Map.Entry entry : stores.getValue().entrySet()) {
final String documentId = entry.getKey();
final IndexMutation mutation = entry.getValue();
if (mutation.isDeleted()) {
if (log.isTraceEnabled())
log.trace("Deleted entire document [{}]", documentId);
writer.deleteDocuments(new Term(DOCID, documentId));
continue;
}
final Document doc = retrieveOrCreate(documentId, searcher);
Preconditions.checkNotNull(doc);
for (final IndexEntry del : mutation.getDeletions()) {
Preconditions.checkArgument(!del.hasMetaData(), "Lucene index does not support indexing meta data: %s", del);
String fieldName = del.field;
if (log.isTraceEnabled()) {
log.trace("Removing field [{}] on document [{}]", fieldName, documentId);
}
KeyInformation ki = storeRetriever.get(fieldName);
removeField(doc, del, ki);
}
addToDocument(doc, mutation.getAdditions(), storeRetriever, false);
//write the old document to the index with the modifications
writer.updateDocument(new Term(DOCID, documentId), doc);
}
writer.commit();
} finally {
IOUtils.closeQuietly(reader);
}
}
@Override
public void restore(Map>> documents, KeyInformation.IndexRetriever information, BaseTransaction tx) throws BackendException {
writerLock.lock();
try {
for (final Map.Entry>> stores : documents.entrySet()) {
IndexReader reader = null;
try {
final String store = stores.getKey();
final IndexWriter writer = getWriter(store, information);
final KeyInformation.StoreRetriever storeRetriever = information.get(store);
reader = DirectoryReader.open(writer, true, true);
final IndexSearcher searcher = new IndexSearcher(reader);
for (final Map.Entry> entry : stores.getValue().entrySet()) {
final String docID = entry.getKey();
final List content = entry.getValue();
if (content == null || content.isEmpty()) {
if (log.isTraceEnabled())
log.trace("Deleting document [{}]", docID);
writer.deleteDocuments(new Term(DOCID, docID));
continue;
}
final Document doc = retrieveOrCreate(docID, searcher);
Iterators.removeIf(doc.iterator(), field -> !field.name().equals(DOCID));
addToDocument(doc, content, storeRetriever, true);
//write the old document to the index with the modifications
writer.updateDocument(new Term(DOCID, docID), doc);
}
writer.commit();
} finally {
IOUtils.closeQuietly(reader);
}
}
tx.commit();
} catch (final IOException e) {
throw new TemporaryBackendException("Could not update Lucene index", e);
} finally {
writerLock.unlock();
}
}
private Document retrieveOrCreate(String docID, IndexSearcher searcher) throws IOException {
final Document doc;
final TopDocs hits = searcher.search(new TermQuery(new Term(DOCID, docID)), 10);
if (hits.scoreDocs.length > 1)
throw new IllegalArgumentException("More than one document found for document id: " + docID);
if (hits.scoreDocs.length == 0) {
if (log.isTraceEnabled())
log.trace("Creating new document for [{}]", docID);
doc = new Document();
doc.add(new StringField(DOCID, docID, Field.Store.YES));
} else {
if (log.isTraceEnabled())
log.trace("Updating existing document for [{}]", docID);
final int docId = hits.scoreDocs[0].doc;
doc = searcher.doc(docId);
}
return doc;
}
private void addToDocument(Document doc, List content, final KeyInformation.StoreRetriever information, boolean isNew) {
Preconditions.checkNotNull(doc);
for (final IndexEntry e : content) {
Preconditions.checkArgument(!e.hasMetaData(), "Lucene index does not support indexing meta data: %s", e);
if (log.isTraceEnabled()) {
log.trace("Adding field [{}] on document [{}]", e.field, doc.get(DOCID));
}
KeyInformation ki = information.get(e.field);
if (!(isNew || ki.getCardinality() == Cardinality.LIST)) {
removeField(doc, e, ki);
}
doc.add(buildStoreField(e.field, e.value, Mapping.getMapping(ki)));
getDualFieldName(e.field, ki)
.ifPresent(dualFieldName -> doc.add(buildStoreField(dualFieldName, e.value, getDualMapping(ki))));
}
buildIndexFields(doc, information).forEach(doc::add);
}
private void removeField(Document doc, IndexEntry e, KeyInformation ki) {
removeFieldIfNeeded(doc, e.field, e.value, ki);
getDualFieldName(e.field, ki)
.ifPresent(dualFieldName -> removeFieldIfNeeded(doc, dualFieldName, e.value, ki));
}
private void removeFieldIfNeeded(Document doc, String fieldName, Object fieldValue, KeyInformation ki) {
boolean isSingle = ki.getCardinality() == Cardinality.SINGLE;
Iterator it = doc.iterator();
while (it.hasNext()) {
IndexableField field = it.next();
if (!fieldName.equals(field.name())) {
continue;
}
if (isSingle || convertToStringValue(ki, fieldValue).equals(field.stringValue())) {
it.remove();
break;
}
}
}
private String convertToStringValue(final KeyInformation ki, Object value) {
String converted;
if (value instanceof Number) {
converted = value.toString();
} else if (AttributeUtils.isString(value)) {
Mapping mapping = Mapping.getMapping(ki);
if (mapping == Mapping.DEFAULT || mapping == Mapping.TEXT || mapping == Mapping.TEXTSTRING) {
converted = ((String) value).toLowerCase();
} else {
converted = (String) value;
}
} else if (value instanceof Date) {
converted = String.valueOf(((Date) value).getTime());
} else if (value instanceof Instant) {
converted = String.valueOf(((Instant) value).toEpochMilli());
} else if (value instanceof Boolean) {
converted = String.valueOf(((Boolean) value) ? 1 : 0);
} else if (value instanceof UUID) {
converted = value.toString();
} else {
throw new IllegalArgumentException("Unsupported type: " + value);
}
return converted;
}
// NOTE: new SET/LIST store fields must be sync with convertToStringValue
private Field buildStoreField(final String fieldName, final Object value, final Mapping mapping) {
final Field field;
if (value instanceof Number) {
if (AttributeUtils.isWholeNumber((Number) value)) {
field = new StoredField(fieldName, ((Number) value).longValue());
} else { //double or float
field = new StoredField(fieldName, ((Number) value).doubleValue());
}
} else if (AttributeUtils.isString(value)) {
final String str = (String) value;
switch (mapping) {
case DEFAULT:
case TEXTSTRING:
case TEXT:
// lowering the case for case insensitive text search
field = new TextField(fieldName, str.toLowerCase(), Field.Store.YES);
break;
case STRING:
// if this field uses a custom analyzer, it must be stored as a TextField
// (or the analyzer, even if it is a KeywordAnalyzer won't be used)
field = new TextField(fieldName, str, Field.Store.YES);
break;
default:
throw new IllegalArgumentException("Illegal mapping specified: " + mapping);
}
} else if (value instanceof Geoshape) {
field = new StoredField(fieldName, GEOID + value);
} else if (value instanceof Date) {
field = new StoredField(fieldName, (((Date) value).getTime()));
} else if (value instanceof Instant) {
field = new StoredField(fieldName, (((Instant) value).toEpochMilli()));
} else if (value instanceof Boolean) {
field = new StoredField(fieldName, ((Boolean) value) ? 1 : 0);
} else if (value instanceof UUID) {
field = new TextField(fieldName, value.toString(), Field.Store.YES);
} else {
throw new IllegalArgumentException("Unsupported type: " + value);
}
return field;
}
private List buildIndexFields(final Document doc, final KeyInformation.StoreRetriever information) {
List fields = new ArrayList<>();
for (IndexableField field : doc.getFields()) {
String fieldName = field.name();
if (fieldName.equals(DOCID)) {
continue;
}
KeyInformation ki = information.get(getOrigFieldName(fieldName));
boolean isPossibleSortIndex = ki.getCardinality() == Cardinality.SINGLE;
Class> dataType = ki.getDataType();
if (AttributeUtils.isWholeNumber(dataType)) {
long value = field.numericValue().longValue();
fields.add(new LongPoint(fieldName, value));
if (isPossibleSortIndex) {
fields.add(new NumericDocValuesField(fieldName, value));
}
} else if (AttributeUtils.isDecimal(dataType)) {
double value = field.numericValue().doubleValue();
fields.add(new DoublePoint(fieldName, value));
if (isPossibleSortIndex) {
fields.add(new DoubleDocValuesField(fieldName, value));
}
} else if (AttributeUtils.isString(dataType)) {
final Mapping mapping = Mapping.getMapping(ki);
if ((mapping == Mapping.STRING || mapping == Mapping.TEXTSTRING) && isPossibleSortIndex) {
fields.add(new SortedDocValuesField(fieldName, new BytesRef(field.stringValue())));
}
} else if (AttributeUtils.isGeo(dataType)) {
if (log.isTraceEnabled())
log.trace("Updating geo-indexes for key {}", fieldName);
Shape shape;
try {
shape = Geoshape.fromWkt(field.stringValue().substring(GEOID.length())).getShape();
} catch (java.text.ParseException e) {
throw new IllegalArgumentException("Geoshape was not parsable", e);
}
final SpatialStrategy spatialStrategy = getSpatialStrategy(fieldName, ki);
Collections.addAll(fields, spatialStrategy.createIndexableFields(shape));
} else if (dataType.equals(Date.class) || dataType.equals(Instant.class)) {
long value = field.numericValue().longValue();
fields.add(new LongPoint(fieldName, value));
if (isPossibleSortIndex) {
fields.add(new NumericDocValuesField(fieldName, value));
}
} else if (dataType.equals(Boolean.class)) {
fields.add(new IntPoint(fieldName, field.numericValue().intValue() == 1 ? 1 : 0));
if (isPossibleSortIndex) {
fields.add(new NumericDocValuesField(fieldName, field.numericValue().intValue()));
}
}
}
return fields;
}
private static Sort getSortOrder(List orders, KeyInformation.StoreRetriever information) {
final Sort sort = new Sort();
if (!orders.isEmpty()) {
final SortField[] fields = new SortField[orders.size()];
for (int i = 0; i < orders.size(); i++) {
final IndexQuery.OrderEntry order = orders.get(i);
SortField.Type sortType = null;
final Class dataType = order.getDatatype();
if (AttributeUtils.isString(dataType)) sortType = SortField.Type.STRING;
else if (AttributeUtils.isWholeNumber(dataType)) sortType = SortField.Type.LONG;
else if (AttributeUtils.isDecimal(dataType)) sortType = SortField.Type.DOUBLE;
else if (dataType.equals(Instant.class) || dataType.equals(Date.class)) sortType = SortField.Type.LONG;
else if (dataType.equals(Boolean.class)) sortType = SortField.Type.LONG;
else
Preconditions.checkArgument(false, "Unsupported order specified on field [%s] with datatype [%s]", order.getKey(), dataType);
KeyInformation ki = information.get(order.getKey());
String fieldKey;
if (Mapping.getMapping(ki) == Mapping.TEXTSTRING) {
fieldKey = getDualFieldName(order.getKey(), ki).orElse(order.getKey());
} else {
fieldKey = order.getKey();
}
fields[i] = new SortField(fieldKey, sortType, order.getOrder() == Order.DESC);
}
sort.setSort(fields);
}
return sort;
}
@Override
public Stream query(IndexQuery query, KeyInformation.IndexRetriever information, BaseTransaction tx) throws BackendException {
//Construct query
final String store = query.getStore();
final LuceneCustomAnalyzer delegatingAnalyzer = delegatingAnalyzerFor(store, information);
final SearchParams searchParams = convertQuery(query.getCondition(), information.get(store), delegatingAnalyzer);
try {
final IndexSearcher searcher = ((Transaction) tx).getSearcher(query.getStore());
if (searcher == null) {
return Collections.unmodifiableList(new ArrayList()).stream(); //Index does not yet exist
}
Query q = searchParams.getQuery();
if (null == q)
q = new MatchAllDocsQuery();
final long time = System.currentTimeMillis();
final TopDocs docs;
int limit = query.hasLimit() ? query.getLimit() : Integer.MAX_VALUE - 1;
if (query.getOrder().isEmpty()) {
docs = searcher.search(q, limit);
} else {
docs = searcher.search(q, limit, getSortOrder(query.getOrder(), information.get(store)));
}
log.debug("Executed query [{}] in {} ms", q, System.currentTimeMillis() - time);
final List result = new ArrayList<>(docs.scoreDocs.length);
for (int i = 0; i < docs.scoreDocs.length; i++) {
final IndexableField field = searcher.doc(docs.scoreDocs[i].doc, FIELDS_TO_LOAD).getField(DOCID);
result.add(field == null ? null : field.stringValue());
}
return result.stream();
} catch (final IOException e) {
throw new TemporaryBackendException("Could not execute Lucene query", e);
}
}
private static Query numericQuery(String key, Cmp relation, Number value) {
switch (relation) {
case EQUAL:
return AttributeUtils.isWholeNumber(value) ?
LongPoint.newRangeQuery(key, value.longValue(), value.longValue()) :
DoublePoint.newRangeQuery(key, value.doubleValue(), value.doubleValue());
case NOT_EQUAL:
final BooleanQuery.Builder q = new BooleanQuery.Builder();
if (AttributeUtils.isWholeNumber(value)) {
q.add(LongPoint.newRangeQuery(key, Long.MIN_VALUE, Math.addExact(value.longValue(), -1)), BooleanClause.Occur.SHOULD);
q.add(LongPoint.newRangeQuery(key, Math.addExact(value.longValue(), 1), Long.MAX_VALUE), BooleanClause.Occur.SHOULD);
} else {
q.add(DoublePoint.newRangeQuery(key, Double.MIN_VALUE, DoublePoint.nextDown(value.doubleValue())), BooleanClause.Occur.SHOULD);
q.add(DoublePoint.newRangeQuery(key, DoublePoint.nextUp(value.doubleValue()), Double.MAX_VALUE), BooleanClause.Occur.SHOULD);
}
return q.build();
case LESS_THAN:
return (AttributeUtils.isWholeNumber(value)) ?
LongPoint.newRangeQuery(key, Long.MIN_VALUE, Math.addExact(value.longValue(), -1)) :
DoublePoint.newRangeQuery(key, Double.MIN_VALUE, DoublePoint.nextDown(value.doubleValue()));
case LESS_THAN_EQUAL:
return (AttributeUtils.isWholeNumber(value)) ?
LongPoint.newRangeQuery(key, Long.MIN_VALUE, value.longValue()) :
DoublePoint.newRangeQuery(key, Double.MIN_VALUE, value.doubleValue());
case GREATER_THAN:
return (AttributeUtils.isWholeNumber(value)) ?
LongPoint.newRangeQuery(key, Math.addExact(value.longValue(), 1), Long.MAX_VALUE) :
DoublePoint.newRangeQuery(key, DoublePoint.nextUp(value.doubleValue()), Double.MAX_VALUE);
case GREATER_THAN_EQUAL:
return (AttributeUtils.isWholeNumber(value)) ?
LongPoint.newRangeQuery(key, value.longValue(), Long.MAX_VALUE) :
DoublePoint.newRangeQuery(key, value.doubleValue(), Double.MAX_VALUE);
default:
throw new IllegalArgumentException("Unexpected relation: " + relation);
}
}
// adapted from SolrIndex
private List> customTokenize(Analyzer analyzer, String fieldName, String value) {
Map> stemsByOffset = new HashMap<>();
try (CachingTokenFilter stream = new CachingTokenFilter(analyzer.tokenStream(fieldName, value))) {
final OffsetAttribute offsetAtt = stream.getAttribute(OffsetAttribute.class);
final TermToBytesRefAttribute termAtt = stream.getAttribute(TermToBytesRefAttribute.class);
stream.reset();
while (stream.incrementToken()) {
int offset = offsetAtt.startOffset();
String stem = termAtt.getBytesRef().utf8ToString();
List stemList = stemsByOffset.get(offset);
if(stemList == null){
stemList = new ArrayList<>();
stemsByOffset.put(offset, stemList);
}
stemList.add(stem);
}
return new ArrayList<>(stemsByOffset.values());
} catch (IOException e) {
throw new IllegalArgumentException(e.getMessage(), e);
}
}
private void tokenize(SearchParams params, final Mapping mapping, final LuceneCustomAnalyzer delegatingAnalyzer, String value, String key, JanusGraphPredicate janusgraphPredicate) {
final Analyzer analyzer = delegatingAnalyzer.getWrappedAnalyzer(key);
final List> terms = customTokenize(analyzer, key, value);
if (terms.isEmpty()) {
// This might happen with very short terms
if (janusgraphPredicate == Text.CONTAINS_PREFIX) {
final Term term;
if (mapping == Mapping.STRING) {
term = new Term(key, value);
} else {
term = new Term(key, value.toLowerCase());
}
params.addQuery(new PrefixQuery(term), BooleanClause.Occur.MUST);
}
} else if (terms.size() == 1) {
if (janusgraphPredicate == Cmp.EQUAL || janusgraphPredicate == Text.CONTAINS) {
params.addQuery(combineTerms(key, terms.get(0), TermQuery::new));
} else if (janusgraphPredicate == Cmp.NOT_EQUAL) {
final BooleanQuery.Builder q = new BooleanQuery.Builder();
q.add(new MatchAllDocsQuery(), BooleanClause.Occur.MUST);
q.add(combineTerms(key, terms.get(0), TermQuery::new), BooleanClause.Occur.MUST_NOT);
params.addQuery(q.build(), BooleanClause.Occur.MUST);
final SearchParams existParams = new SearchParams();
addExistsQuery(existParams, key);
params.addParams(existParams, BooleanClause.Occur.MUST);
} else if (janusgraphPredicate == Text.CONTAINS_PREFIX) {
List preparedTerms = new ArrayList<>(terms.get(0));
if (mapping != Mapping.STRING) {
preparedTerms = terms.get(0).stream().map(String::toLowerCase).collect(Collectors.toList());
}
params.addQuery(combineTerms(key, preparedTerms, PrefixQuery::new), BooleanClause.Occur.MUST);
} else throw new IllegalArgumentException("LuceneIndex does not support this predicate with 1 token : " + janusgraphPredicate);
} else {
// at the moment, this is only walked for EQUAL, NOT_EQUAL and Text.CONTAINS (String and Text mappings)
final BooleanQuery.Builder q = new BooleanQuery.Builder();
BooleanClause.Occur occur;
if (janusgraphPredicate == Cmp.NOT_EQUAL) {
q.add(new MatchAllDocsQuery(), BooleanClause.Occur.MUST);
occur = BooleanClause.Occur.MUST_NOT;
} else {
occur = BooleanClause.Occur.MUST;
}
for (final List stems : terms) {
q.add(combineTerms(key, stems, TermQuery::new), occur);
}
params.addQuery(q.build());
}
}
private void addExistsQuery(final SearchParams params, final String key) {
// some fields like Integer omit norms but have docValues
params.addQuery(new DocValuesFieldExistsQuery(key), BooleanClause.Occur.SHOULD);
// some fields like Text have no docValue but have norms
params.addQuery(new NormsFieldExistsQuery(key), BooleanClause.Occur.SHOULD);
}
private Query combineTerms(String key, List terms, Function queryCreator) {
if (terms.size() > 1) {
final Builder q = new Builder();
for (String term : terms) {
q.add(queryCreator.apply(new Term(key, term)), BooleanClause.Occur.SHOULD);
}
return q.build();
} else if (terms.size() == 1){
return queryCreator.apply(new Term(key, terms.get(0)));
} else {
return new MatchNoDocsQuery("No terms for key " + key);
}
}
private LuceneCustomAnalyzer delegatingAnalyzerFor(String store, KeyInformation.IndexRetriever information2) {
if (!delegatingAnalyzers.containsKey(store)) {
delegatingAnalyzers.put(store, new LuceneCustomAnalyzer(store, information2, Analyzer.PER_FIELD_REUSE_STRATEGY));
}
return delegatingAnalyzers.get(store);
}
private SearchParams convertQuery(Condition> condition, final KeyInformation.StoreRetriever information, final LuceneCustomAnalyzer delegatingAnalyzer) {
final SearchParams params = new SearchParams();
if (condition instanceof PredicateCondition) {
final PredicateCondition atom = (PredicateCondition) condition;
Object value = atom.getValue();
final String key = atom.getKey();
KeyInformation ki = information.get(key);
final JanusGraphPredicate janusgraphPredicate = atom.getPredicate();
if (value == null && janusgraphPredicate == Cmp.NOT_EQUAL) {
addExistsQuery(params, key);
} else if (value instanceof Number) {
Preconditions.checkArgument(janusgraphPredicate instanceof Cmp, "Relation not supported on numeric types: %s", janusgraphPredicate);
params.addQuery(numericQuery(key, (Cmp) janusgraphPredicate, (Number) value));
} else if (value instanceof String) {
if (janusgraphPredicate == Cmp.LESS_THAN) {
params.addQuery(TermRangeQuery.newStringRange(key, null, value.toString(), false, false));
} else if (janusgraphPredicate == Cmp.LESS_THAN_EQUAL) {
params.addQuery(TermRangeQuery.newStringRange(key, null, value.toString(), false, true));
} else if (janusgraphPredicate == Cmp.GREATER_THAN) {
params.addQuery(TermRangeQuery.newStringRange(key, value.toString(), null, false, false));
} else if (janusgraphPredicate == Cmp.GREATER_THAN_EQUAL) {
params.addQuery(TermRangeQuery.newStringRange(key, value.toString(), null, true, false));
} else {
final Mapping map = Mapping.getMapping(ki);
final String stringFieldKey;
if (Mapping.getMapping(ki) == Mapping.TEXTSTRING) {
stringFieldKey = getDualFieldName(key, ki).orElse(key);
} else {
stringFieldKey = key;
}
if ((map == Mapping.DEFAULT || map == Mapping.TEXT) && !Text.HAS_CONTAINS.contains(janusgraphPredicate))
throw new IllegalArgumentException("Text mapped string values only support CONTAINS queries and not: " + janusgraphPredicate);
if (map == Mapping.STRING && Text.HAS_CONTAINS.contains(janusgraphPredicate))
throw new IllegalArgumentException("String mapped string values do not support CONTAINS queries: " + janusgraphPredicate);
if (janusgraphPredicate == Text.CONTAINS) {
tokenize(params, map, delegatingAnalyzer, ((String) value).toLowerCase(), key, janusgraphPredicate);
} else if (janusgraphPredicate == Text.CONTAINS_PREFIX) {
tokenize(params, map, delegatingAnalyzer, (String) value, key, janusgraphPredicate);
} else if (janusgraphPredicate == Text.PREFIX) {
params.addQuery(new PrefixQuery(new Term(stringFieldKey, (String) value)));
} else if (janusgraphPredicate == Text.REGEX) {
final RegexpQuery rq = new RegexpQuery(new Term(stringFieldKey, (String) value));
params.addQuery(rq);
} else if (janusgraphPredicate == Text.CONTAINS_REGEX) {
// This is terrible -- there is probably a better way
// putting this to lowercase because Text search is supposed to be case insensitive
final RegexpQuery rq = new RegexpQuery(new Term(key, ".*" + (((String) value).toLowerCase()) + ".*"));
params.addQuery(rq);
} else if (janusgraphPredicate == Cmp.EQUAL || janusgraphPredicate == Cmp.NOT_EQUAL) {
tokenize(params, map, delegatingAnalyzer, (String) value, stringFieldKey, janusgraphPredicate);
} else if (janusgraphPredicate == Text.FUZZY) {
params.addQuery(new FuzzyQuery(new Term(stringFieldKey, (String) value), Text.getMaxEditDistance((String) value)));
} else if (janusgraphPredicate == Text.CONTAINS_FUZZY) {
value = ((String) value).toLowerCase();
final Builder b = new BooleanQuery.Builder();
for (final String term : Text.tokenize((String) value)) {
b.add(new FuzzyQuery(new Term(key, term), Text.getMaxEditDistance(term)), BooleanClause.Occur.MUST);
}
params.addQuery(b.build());
} else
throw new IllegalArgumentException("Relation is not supported for string value: " + janusgraphPredicate);
}
} else if (value instanceof Geoshape) {
Preconditions.checkArgument(janusgraphPredicate instanceof Geo, "Relation not supported on geo types: %s", janusgraphPredicate);
final Shape shape = ((Geoshape) value).getShape();
final SpatialOperation spatialOp = SPATIAL_PREDICATES.get(janusgraphPredicate);
final SpatialArgs args = new SpatialArgs(spatialOp, shape);
params.addQuery(getSpatialStrategy(key, information.get(key)).makeQuery(args));
} else if (value instanceof Date) {
Preconditions.checkArgument(janusgraphPredicate instanceof Cmp, "Relation not supported on date types: %s", janusgraphPredicate);
params.addQuery(numericQuery(key, (Cmp) janusgraphPredicate, ((Date) value).getTime()));
} else if (value instanceof Instant) {
Preconditions.checkArgument(janusgraphPredicate instanceof Cmp, "Relation not supported on instant types: %s", janusgraphPredicate);
params.addQuery(numericQuery(key, (Cmp) janusgraphPredicate, ((Instant) value).toEpochMilli()));
} else if (value instanceof Boolean) {
Preconditions.checkArgument(janusgraphPredicate instanceof Cmp, "Relation not supported on boolean types: %s", janusgraphPredicate);
final int intValue;
switch ((Cmp) janusgraphPredicate) {
case EQUAL:
intValue = ((Boolean) value) ? 1 : 0;
params.addQuery(IntPoint.newRangeQuery(key, intValue, intValue));
break;
case NOT_EQUAL:
intValue = ((Boolean) value) ? 0 : 1;
params.addQuery(IntPoint.newRangeQuery(key, intValue, intValue));
break;
default:
throw new IllegalArgumentException("Boolean types only support EQUAL or NOT_EQUAL");
}
} else if (value instanceof UUID) {
Preconditions.checkArgument(janusgraphPredicate instanceof Cmp, "Relation not supported on UUID types: %s", janusgraphPredicate);
if (janusgraphPredicate == Cmp.EQUAL) {
params.addQuery(new TermQuery(new Term(key, value.toString())));
} else if (janusgraphPredicate == Cmp.NOT_EQUAL) {
final SearchParams existParams = new SearchParams();
addExistsQuery(existParams, key);
params.addParams(existParams, BooleanClause.Occur.MUST);
final BooleanQuery.Builder q = new BooleanQuery.Builder();
q.add(new MatchAllDocsQuery(), BooleanClause.Occur.MUST);
q.add(new TermQuery(new Term(key, value.toString())), BooleanClause.Occur.MUST_NOT);
params.addQuery(q.build());
} else {
throw new IllegalArgumentException("Relation is not supported for UUID type: " + janusgraphPredicate);
}
} else {
throw new IllegalArgumentException("Unsupported type: " + value);
}
} else if (condition instanceof Not) {
final SearchParams childParams = convertQuery(((Not) condition).getChild(), information, delegatingAnalyzer);
params.addQuery(new MatchAllDocsQuery(), BooleanClause.Occur.MUST);
params.addParams(childParams, BooleanClause.Occur.MUST_NOT);
} else if (condition instanceof And) {
for (final Condition c : condition.getChildren()) {
final SearchParams childParams = convertQuery(c, information, delegatingAnalyzer);
params.addParams(childParams, BooleanClause.Occur.MUST);
}
} else if (condition instanceof Or) {
for (final Condition c : condition.getChildren()) {
final SearchParams childParams = convertQuery(c, information, delegatingAnalyzer);
params.addParams(childParams, BooleanClause.Occur.SHOULD);
}
} else throw new IllegalArgumentException("Invalid condition: " + condition);
return params;
}
private QueryParser getQueryParser(String store, KeyInformation.IndexRetriever information) {
final Analyzer analyzer = delegatingAnalyzerFor(store, information);
final NumericTranslationQueryParser parser = new NumericTranslationQueryParser(information.get(store), "_all", analyzer);
parser.setAllowLeadingWildcard(true);
return parser;
}
@Override
public Stream> query(RawQuery query, KeyInformation.IndexRetriever information, BaseTransaction tx) throws BackendException {
final Query q;
try {
q = getQueryParser(query.getStore(), information).parse(query.getQuery());
// Lucene query parser does not take additional parameters so any parameters on the RawQuery are ignored.
} catch (final ParseException e) {
throw new PermanentBackendException("Could not parse raw query: " + query.getQuery(), e);
}
try {
final IndexSearcher searcher = ((Transaction) tx).getSearcher(query.getStore());
if (searcher == null) {
return Collections.unmodifiableList(new ArrayList>()).stream(); //Index does not yet exist
}
final long time = System.currentTimeMillis();
//TODO: can we make offset more efficient in Lucene?
final int offset = query.getOffset();
int adjustedLimit = query.hasLimit() ? query.getLimit() : Integer.MAX_VALUE - 1;
if (adjustedLimit < Integer.MAX_VALUE - 1 - offset) adjustedLimit += offset;
else adjustedLimit = Integer.MAX_VALUE - 1;
final TopDocs docs;
if (query.getOrders().isEmpty()) {
docs = searcher.search(q, adjustedLimit);
} else {
docs = searcher.search(q, adjustedLimit, getSortOrder(query.getOrders(), information.get(query.getStore())));
}
log.debug("Executed query [{}] in {} ms", q, System.currentTimeMillis() - time);
final List> result = new ArrayList<>(docs.scoreDocs.length);
for (int i = offset; i < docs.scoreDocs.length; i++) {
final IndexableField field = searcher.doc(docs.scoreDocs[i].doc, FIELDS_TO_LOAD).getField(DOCID);
result.add(new RawQuery.Result<>(field == null ? null : field.stringValue(), docs.scoreDocs[i].score));
}
return result.stream();
} catch (final IOException e) {
throw new TemporaryBackendException("Could not execute Lucene query", e);
}
}
@Override
public Number queryAggregation(IndexQuery query, KeyInformation.IndexRetriever information, BaseTransaction tx, Aggregation aggregation) throws BackendException {
//Construct query
final String store = query.getStore();
final LuceneCustomAnalyzer delegatingAnalyzer = delegatingAnalyzerFor(store, information);
final SearchParams searchParams = convertQuery(query.getCondition(), information.get(store), delegatingAnalyzer);
try {
final IndexSearcher searcher = ((Transaction) tx).getSearcher(query.getStore());
if (searcher == null) {
return 0L; //Index does not yet exist
}
Query q = searchParams.getQuery();
switch (aggregation.getType()) {
case COUNT: return QueryUtil.applyQueryLimitAfterCount(executeCount(searcher, q), query);
case MIN: return executeMin(searcher, q, aggregation.getFieldName(), aggregation.getDataType());
case MAX: return executeMax(searcher, q, aggregation.getFieldName(), aggregation.getDataType());
case AVG: return executeAvg(searcher, q, aggregation.getFieldName());
case SUM: return executeSum(searcher, q, aggregation.getFieldName(), aggregation.getDataType());
default: throw new UnsupportedOperationException();
}
} catch (final IOException e) {
throw new TemporaryBackendException("Could not execute Lucene query", e);
}
}
private long executeCount(IndexSearcher searcher, Query query) throws IOException {
final long time = System.currentTimeMillis();
// We ignore offset and limit for totals
final TopDocs docs = searcher.search(query, 1);
log.debug("Executed query [{}] in {} ms", query, System.currentTimeMillis() - time);
return docs.totalHits.value;
}
private SortField.Type sortFieldType(Class fieldType) {
if (fieldType != null) {
if (Long.class.isAssignableFrom(fieldType)) return SortField.Type.LONG;
else if (Float.class.isAssignableFrom(fieldType)) return SortField.Type.FLOAT;
else if (Double.class.isAssignableFrom(fieldType)) return SortField.Type.DOUBLE;
}
return SortField.Type.INT;
}
private Number adaptNumberType(Number value, Class extends Number> expectedType) {
if (expectedType == null) return value;
else if (Byte.class.isAssignableFrom(expectedType)) return value.byteValue();
else if (Short.class.isAssignableFrom(expectedType)) return value.shortValue();
else if (Integer.class.isAssignableFrom(expectedType)) return value.intValue();
else if (Long.class.isAssignableFrom(expectedType)) return value.longValue();
else if (Float.class.isAssignableFrom(expectedType)) return value.floatValue();
else if (Double.class.isAssignableFrom(expectedType)) return value.doubleValue();
else return value.doubleValue();
}
private Number executeMin(IndexSearcher searcher, Query query, String fieldName, Class fieldType) throws IOException {
final TopFieldDocs docs = searcher.search(query, 1, new Sort(new SortField(fieldName, sortFieldType(fieldType))));
final IndexableField field = searcher.doc(docs.scoreDocs[0].doc, Sets.newHashSet(fieldName)).getField(fieldName);
return adaptNumberType(field.numericValue(), fieldType);
}
private Number executeMax(IndexSearcher searcher, Query query, String fieldName, Class fieldType) throws IOException {
final TopFieldDocs docs = searcher.search(query, 1, new Sort(new SortField(fieldName, sortFieldType(fieldType), true)));
final IndexableField field = searcher.doc(docs.scoreDocs[0].doc, Sets.newHashSet(fieldName)).getField(fieldName);
return adaptNumberType(field.numericValue(), fieldType);
}
private Number executeSum(IndexSearcher searcher, Query query, String fieldName, Class fieldType) throws IOException {
SumCollector collector = new SumCollector(fieldName, searcher);
searcher.search(query, collector);
if (Float.class.isAssignableFrom(fieldType) || Double.class.isAssignableFrom(fieldType))
return collector.getValue();
else
return (long)collector.getValue();
}
private double executeAvg(IndexSearcher searcher, Query query, String fieldName) throws IOException {
return ((double)executeSum(searcher, query, fieldName, Double.class)) / executeCount(searcher, query);
}
@Override
public Long totals(RawQuery query, KeyInformation.IndexRetriever information, BaseTransaction tx) throws BackendException {
final Query q;
try {
q = getQueryParser(query.getStore(), information).parse(query.getQuery());
} catch (final ParseException e) {
throw new PermanentBackendException("Could not parse raw query: " + query.getQuery(), e);
}
try {
final IndexSearcher searcher = ((Transaction) tx).getSearcher(query.getStore());
if (searcher == null) return 0L; //Index does not yet exist
final long time = System.currentTimeMillis();
// Lucene doesn't like limits of 0. Also, it doesn't efficiently build a total list.
final TopDocs docs = searcher.search(q, 1);
log.debug("Executed query [{}] in {} ms", q, System.currentTimeMillis() - time);
return QueryUtil.applyOffsetWithQueryLimitAfterCount(docs.totalHits.value, query.getOffset(), query);
} catch (final IOException e) {
throw new TemporaryBackendException("Could not execute Lucene query", e);
}
}
@Override
public BaseTransactionConfigurable beginTransaction(BaseTransactionConfig config) throws BackendException {
return new Transaction(config);
}
@Override
public boolean supports(KeyInformation information, JanusGraphPredicate janusgraphPredicate) {
final Class> dataType = information.getDataType();
final Mapping mapping = Mapping.getMapping(information);
if (mapping != Mapping.DEFAULT && !AttributeUtils.isString(dataType) &&
!(mapping == Mapping.PREFIX_TREE && AttributeUtils.isGeo(dataType))) return false;
if (Number.class.isAssignableFrom(dataType)) {
return janusgraphPredicate instanceof Cmp;
} else if (dataType == Geoshape.class) {
if (information.getCardinality() != Cardinality.SINGLE) return false;
return janusgraphPredicate == Geo.INTERSECT || janusgraphPredicate == Geo.WITHIN || janusgraphPredicate == Geo.CONTAINS;
} else if (AttributeUtils.isString(dataType)) {
switch (mapping) {
case DEFAULT:
case TEXT:
return janusgraphPredicate == Text.CONTAINS || janusgraphPredicate == Text.CONTAINS_PREFIX || janusgraphPredicate == Text.CONTAINS_FUZZY; // || janusgraphPredicate == Text.CONTAINS_REGEX;
case STRING:
return janusgraphPredicate instanceof Cmp || janusgraphPredicate == Text.PREFIX || janusgraphPredicate == Text.REGEX || janusgraphPredicate == Text.FUZZY;
case TEXTSTRING:
return janusgraphPredicate instanceof Text || janusgraphPredicate instanceof Cmp;
}
} else if (dataType == Date.class || dataType == Instant.class) {
return janusgraphPredicate instanceof Cmp;
} else if (dataType == Boolean.class) {
return janusgraphPredicate == Cmp.EQUAL || janusgraphPredicate == Cmp.NOT_EQUAL;
} else if (dataType == UUID.class) {
return janusgraphPredicate == Cmp.EQUAL || janusgraphPredicate == Cmp.NOT_EQUAL;
}
return false;
}
@Override
public boolean supports(KeyInformation information) {
final Class> dataType = information.getDataType();
final Mapping mapping = Mapping.getMapping(information);
if (Number.class.isAssignableFrom(dataType) || dataType == Date.class || dataType == Instant.class || dataType == Boolean.class || dataType == UUID.class) {
return mapping == Mapping.DEFAULT;
} else if (AttributeUtils.isString(dataType)) {
return mapping == Mapping.DEFAULT || mapping == Mapping.STRING || mapping == Mapping.TEXT || mapping == Mapping.TEXTSTRING;
} else if (AttributeUtils.isGeo(dataType)) {
return information.getCardinality() == Cardinality.SINGLE && (mapping == Mapping.DEFAULT || mapping == Mapping.PREFIX_TREE);
}
return false;
}
@Override
public String mapKey2Field(String key, KeyInformation information) {
IndexProvider.checkKeyValidity(key);
return key.replace(' ', REPLACEMENT_CHAR);
}
@Override
public IndexFeatures getFeatures() {
return LUCENE_FEATURES;
}
@Override
public void close() throws BackendException {
try {
for (final IndexWriter w : writers.values()) w.close();
} catch (final IOException e) {
throw new PermanentBackendException("Could not close writers", e);
}
}
@Override
public void clearStorage() throws BackendException {
try {
FileUtils.deleteDirectory(new File(basePath));
} catch (final IOException e) {
throw new PermanentBackendException("Could not delete lucene directory: " + basePath, e);
}
}
@Override
public void clearStore(String storeName) throws BackendException {
throw new PermanentBackendException("Lucene index does not yet support deleting single stores.");
}
@Override
public boolean exists() throws BackendException {
if (Files.exists(Paths.get(basePath))) {
try (final DirectoryStream dirStream = Files.newDirectoryStream(Paths.get(basePath))) {
return dirStream.iterator().hasNext();
} catch (final IOException e) {
throw new PermanentBackendException("Could not read lucene directory: " + basePath, e);
}
} else {
return false;
}
}
static String getOrigFieldName(String fieldName) {
if (isDualFieldName(fieldName)) {
return fieldName.replaceAll(STRING_SUFFIX, "");
} else {
return fieldName;
}
}
static boolean isDualFieldName(String fieldName) {
return fieldName.endsWith(STRING_SUFFIX);
}
static Optional getDualFieldName(String fieldKey, KeyInformation ki) {
if (AttributeUtils.isString(ki.getDataType()) && Mapping.getMapping(ki) == Mapping.TEXTSTRING) {
return Optional.of(fieldKey + STRING_SUFFIX);
}
return Optional.empty();
}
static Mapping getDualMapping(KeyInformation ki) {
if (AttributeUtils.isString(ki.getDataType()) && Mapping.getMapping(ki) == Mapping.TEXTSTRING) {
return Mapping.STRING;
}
return Mapping.DEFAULT;
}
private class Transaction implements BaseTransactionConfigurable {
private final BaseTransactionConfig config;
private final Set updatedStores = Sets.newHashSet();
private final Map searchers = new HashMap<>(4);
private Transaction(BaseTransactionConfig config) {
this.config = config;
}
private synchronized IndexSearcher getSearcher(String store) throws BackendException {
IndexSearcher searcher = searchers.get(store);
if (searcher == null) {
final IndexReader reader;
try {
reader = DirectoryReader.open(getStoreDirectory(store));
searcher = new IndexSearcher(reader);
} catch (final IndexNotFoundException e) {
searcher = null;
} catch (final IOException e) {
throw new PermanentBackendException("Could not open index reader on store: " + store, e);
}
searchers.put(store, searcher);
}
return searcher;
}
public void postCommit() throws BackendException {
close();
searchers.clear();
}
@Override
public void commit() throws BackendException {
close();
}
@Override
public void rollback() throws BackendException {
close();
}
private void close() throws BackendException {
try {
for (final IndexSearcher searcher : searchers.values()) {
if (searcher != null) searcher.getIndexReader().close();
}
} catch (final IOException e) {
throw new PermanentBackendException("Could not close searcher", e);
}
}
@Override
public BaseTransactionConfig getConfiguration() {
return config;
}
}
/**
* Encapsulates a Lucene Query that express a JanusGraph {@link org.janusgraph.graphdb.query.Query} using Lucene's
* abstractions. This object's state is mutable.
*/
private static class SearchParams {
private final BooleanQuery.Builder qb = new BooleanQuery.Builder();
private void addQuery(Query newQuery) {
addQuery(newQuery, BooleanClause.Occur.MUST);
}
private void addQuery(Query newQuery, BooleanClause.Occur occur) {
qb.add(newQuery, occur);
}
private void addParams(SearchParams other, BooleanClause.Occur occur) {
final Query otherQuery = other.getQuery();
if (null != otherQuery)
addQuery(otherQuery, occur);
}
private Query getQuery() {
final BooleanQuery q = qb.build();
if (0 == q.clauses().size()) {
return null;
}
return q;
}
}
}