![JAR search and dependency download from the Maven repository](/logo.png)
com.arangodb.model.AqlQueryOptions Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of core Show documentation
Show all versions of core Show documentation
Core module for ArangoDB Java Driver
/*
* DISCLAIMER
*
* Copyright 2016 ArangoDB GmbH, Cologne, Germany
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*
* Copyright holder is ArangoDB GmbH, Cologne, Germany
*/
package com.arangodb.model;
import com.arangodb.ArangoCursor;
import com.arangodb.internal.serde.UserDataInside;
import com.fasterxml.jackson.annotation.JsonAnyGetter;
import com.fasterxml.jackson.annotation.JsonIgnore;
import com.fasterxml.jackson.annotation.JsonInclude;
import java.util.*;
/**
* @author Mark Vollmary
* @author Michele Rastelli
*/
public final class AqlQueryOptions extends TransactionalOptions implements Cloneable {
private Boolean allowDirtyRead;
private Integer batchSize;
private Map bindVars;
private Boolean cache;
private Boolean count;
private Long memoryLimit;
private Options options;
private String query;
private Integer ttl;
@Override
AqlQueryOptions getThis() {
return this;
}
public Boolean getAllowDirtyRead() {
return allowDirtyRead;
}
/**
* Sets the header {@code x-arango-allow-dirty-read} to {@code true} to allow the Coordinator to ask any shard
* replica for the data, not only the shard leader. This may result in “dirty reads”.
*
* The header is ignored if this operation is part of a Stream Transaction
* ({@link AqlQueryOptions#streamTransactionId(String)}). The header set when creating the transaction decides
* about dirty reads for the entire transaction, not the individual read operations.
*
* @param allowDirtyRead Set to {@code true} allows reading from followers in an active-failover setup.
* @return this
* @see API
* Documentation
*/
public AqlQueryOptions allowDirtyRead(final Boolean allowDirtyRead) {
this.allowDirtyRead = allowDirtyRead;
return this;
}
public Integer getBatchSize() {
return batchSize;
}
/**
* @param batchSize maximum number of result documents to be transferred from the server to the client in one
* roundtrip. If this attribute is not set, a server-controlled default value will be used.
* A batchSize value of 0 is disallowed.
* @return this
*/
public AqlQueryOptions batchSize(final Integer batchSize) {
this.batchSize = batchSize;
return this;
}
@UserDataInside
public Map getBindVars() {
return bindVars;
}
/**
* @param bindVars A map with key/value pairs representing the bind parameters. For a bind variable {@code @var} in
* the query, specify the value using an attribute with the name {@code var}. For a collection bind
* variable {@code @@coll}, use {@code @coll} as the attribute name.
* @return this
*/
AqlQueryOptions bindVars(final Map bindVars) {
this.bindVars = bindVars;
return this;
}
public Boolean getCache() {
return cache;
}
/**
* @param cache flag to determine whether the AQL query results cache shall be used. If set to false, then any
* query cache lookup will be skipped for the query. If set to true, it will lead to the query cache
* being checked for the query if the query cache mode is either on or demand.
* @return this
*/
public AqlQueryOptions cache(final Boolean cache) {
this.cache = cache;
return this;
}
public Boolean getCount() {
return count;
}
/**
* @param count indicates whether the number of documents in the result set should be returned and made accessible
* via {@link ArangoCursor#getCount()}. Calculating the {@code count} attribute might have a
* performance impact for some queries in the future so this option is turned off by default, and
* {@code count} is only returned when requested.
* @return this
*/
public AqlQueryOptions count(final Boolean count) {
this.count = count;
return this;
}
public Long getMemoryLimit() {
return memoryLimit;
}
/**
* @param memoryLimit the maximum number of memory (measured in bytes) that the query is allowed to use. If set,
* then the query will fail with error {@code resource limit exceeded} in case it allocates too
* much memory. A value of {@code 0} indicates that there is no memory limit.
* @return this
* @since ArangoDB 3.1.0
*/
public AqlQueryOptions memoryLimit(final Long memoryLimit) {
this.memoryLimit = memoryLimit;
return this;
}
public Options getOptions() {
if (options == null) {
options = new Options();
}
return options;
}
/**
* @param options extra options for the query
* @return this
*/
public AqlQueryOptions options(final Options options) {
this.options = options;
return this;
}
public String getQuery() {
return query;
}
/**
* @param query the query to be executed
* @return this
*/
public AqlQueryOptions query(final String query) {
this.query = query;
return this;
}
public Integer getTtl() {
return ttl;
}
/**
* @param ttl The time-to-live for the cursor (in seconds). If the result set is small enough (less than or equal
* to batchSize) then results are returned right away. Otherwise, they are stored in memory and will be
* accessible via the cursor with respect to the ttl. The cursor will be removed on the server
* automatically after the specified amount of time. This is useful to ensure garbage collection of
* cursors that are not fully fetched by clients.
*
* If not set, a server-defined value will be used (default: 30 seconds).
*
* The time-to-live is renewed upon every access to the cursor.
* @return this
*/
public AqlQueryOptions ttl(final Integer ttl) {
this.ttl = ttl;
return this;
}
@Override
public AqlQueryOptions clone() {
try {
AqlQueryOptions clone = (AqlQueryOptions) super.clone();
clone.bindVars = bindVars != null ? new HashMap<>(bindVars) : null;
clone.options = options != null ? options.clone() : null;
return clone;
} catch (CloneNotSupportedException e) {
throw new AssertionError();
}
}
public static final class Options implements Cloneable {
private Map customOptions;
private Boolean allPlans;
private Boolean allowDirtyReads;
private Boolean allowRetry;
private Boolean failOnWarning;
private Boolean fillBlockCache;
private String forceOneShardAttributeValue;
private Boolean fullCount;
private Long intermediateCommitCount;
private Long intermediateCommitSize;
private Integer maxDNFConditionMembers;
private Integer maxNodesPerCallstack;
private Integer maxNumberOfPlans;
private Double maxRuntime;
private Long maxTransactionSize;
private Long maxWarningCount;
private Optimizer optimizer;
private Boolean profile;
private Double satelliteSyncWait;
private Collection shardIds;
private Boolean skipInaccessibleCollections;
private Long spillOverThresholdMemoryUsage;
private Long spillOverThresholdNumRows;
private Boolean stream;
@JsonInclude
@JsonAnyGetter
public Map getCustomOptions() {
if (customOptions == null) {
customOptions = new HashMap<>();
}
return customOptions;
}
public void setCustomOption(String key, Object value) {
getCustomOptions().put(key, value);
}
public Boolean getAllPlans() {
return allPlans;
}
public Boolean getAllowDirtyReads() {
return allowDirtyReads;
}
public Boolean getAllowRetry() {
return allowRetry;
}
public Boolean getFailOnWarning() {
return failOnWarning;
}
public Boolean getFillBlockCache() {
return fillBlockCache;
}
public String getForceOneShardAttributeValue() {
return forceOneShardAttributeValue;
}
public Boolean getFullCount() {
return fullCount;
}
public Long getIntermediateCommitCount() {
return intermediateCommitCount;
}
public Long getIntermediateCommitSize() {
return intermediateCommitSize;
}
public Integer getMaxDNFConditionMembers() {
return maxDNFConditionMembers;
}
public Integer getMaxNodesPerCallstack() {
return maxNodesPerCallstack;
}
public Integer getMaxNumberOfPlans() {
return maxNumberOfPlans;
}
/**
* @deprecated for removal, use {@link Options#getMaxNumberOfPlans()} instead
*/
@Deprecated
@JsonIgnore
public Integer getMaxPlans() {
return getMaxNumberOfPlans();
}
public Double getMaxRuntime() {
return maxRuntime;
}
public Long getMaxTransactionSize() {
return maxTransactionSize;
}
public Long getMaxWarningCount() {
return maxWarningCount;
}
public Optimizer getOptimizer() {
if (optimizer == null) {
optimizer = new Optimizer();
}
return optimizer;
}
public Boolean getProfile() {
return profile;
}
public Double getSatelliteSyncWait() {
return satelliteSyncWait;
}
public Collection getShardIds() {
return shardIds;
}
public Boolean getSkipInaccessibleCollections() {
return skipInaccessibleCollections;
}
public Long getSpillOverThresholdMemoryUsage() {
return spillOverThresholdMemoryUsage;
}
public Long getSpillOverThresholdNumRows() {
return spillOverThresholdNumRows;
}
public Boolean getStream() {
return stream;
}
public void setAllPlans(Boolean allPlans) {
this.allPlans = allPlans;
}
public void setAllowDirtyReads(Boolean allowDirtyReads) {
this.allowDirtyReads = allowDirtyReads;
}
public void setAllowRetry(Boolean allowRetry) {
this.allowRetry = allowRetry;
}
public void setFailOnWarning(Boolean failOnWarning) {
this.failOnWarning = failOnWarning;
}
public void setFillBlockCache(Boolean fillBlockCache) {
this.fillBlockCache = fillBlockCache;
}
public void setForceOneShardAttributeValue(String forceOneShardAttributeValue) {
this.forceOneShardAttributeValue = forceOneShardAttributeValue;
}
public void setFullCount(Boolean fullCount) {
this.fullCount = fullCount;
}
public void setIntermediateCommitCount(Long intermediateCommitCount) {
this.intermediateCommitCount = intermediateCommitCount;
}
public void setIntermediateCommitSize(Long intermediateCommitSize) {
this.intermediateCommitSize = intermediateCommitSize;
}
public void setMaxDNFConditionMembers(Integer maxDNFConditionMembers) {
this.maxDNFConditionMembers = maxDNFConditionMembers;
}
public void setMaxNodesPerCallstack(Integer maxNodesPerCallstack) {
this.maxNodesPerCallstack = maxNodesPerCallstack;
}
public void setMaxNumberOfPlans(Integer maxNumberOfPlans) {
this.maxNumberOfPlans = maxNumberOfPlans;
}
public void setMaxRuntime(Double maxRuntime) {
this.maxRuntime = maxRuntime;
}
public void setMaxTransactionSize(Long maxTransactionSize) {
this.maxTransactionSize = maxTransactionSize;
}
public void setMaxWarningCount(Long maxWarningCount) {
this.maxWarningCount = maxWarningCount;
}
public void setOptimizer(Optimizer optimizer) {
this.optimizer = optimizer;
}
public void setProfile(Boolean profile) {
this.profile = profile;
}
public void setSatelliteSyncWait(Double satelliteSyncWait) {
this.satelliteSyncWait = satelliteSyncWait;
}
public void setShardIds(Collection shardIds) {
this.shardIds = shardIds;
}
public void setSkipInaccessibleCollections(Boolean skipInaccessibleCollections) {
this.skipInaccessibleCollections = skipInaccessibleCollections;
}
public void setSpillOverThresholdMemoryUsage(Long spillOverThresholdMemoryUsage) {
this.spillOverThresholdMemoryUsage = spillOverThresholdMemoryUsage;
}
public void setSpillOverThresholdNumRows(Long spillOverThresholdNumRows) {
this.spillOverThresholdNumRows = spillOverThresholdNumRows;
}
public void setStream(Boolean stream) {
this.stream = stream;
}
@Override
public Options clone() {
try {
Options clone = (Options) super.clone();
clone.customOptions = customOptions != null ? new HashMap<>(customOptions) : null;
clone.optimizer = optimizer != null ? optimizer.clone() : null;
clone.shardIds = shardIds != null ? new ArrayList<>(shardIds) : null;
return clone;
} catch (CloneNotSupportedException e) {
throw new AssertionError();
}
}
}
public static final class Optimizer implements Cloneable {
private Collection rules;
public Collection getRules() {
return rules;
}
public void setRules(Collection rules) {
this.rules = rules;
}
@Override
public Optimizer clone() {
try {
Optimizer clone = (Optimizer) super.clone();
clone.rules = rules != null ? new ArrayList<>(rules) : null;
return clone;
} catch (CloneNotSupportedException e) {
throw new AssertionError();
}
}
}
// ------------------------------------
// --- accessors for nested options ---
// ------------------------------------
@JsonIgnore
public Map getCustomOptions() {
return getOptions().getCustomOptions();
}
/**
* Set an additional custom option in the form of key-value pair.
*
* @param key option name
* @param value option value
* @return this
*/
public AqlQueryOptions customOption(String key, Object value) {
getOptions().setCustomOption(key, value);
return this;
}
@JsonIgnore
public Boolean getAllowDirtyReads() {
return getOptions().getAllowDirtyReads();
}
/**
* @param allowDirtyReads If you set this option to true and execute the query against a cluster deployment, then
* the Coordinator is allowed to read from any shard replica and not only from the leader.
* You may observe data inconsistencies (dirty reads) when reading from followers, namely
* obsolete revisions of documents because changes have not yet been replicated to the
* follower, as well as changes to documents before they are officially committed on the
* leader. This feature is only available in the Enterprise Edition.
* @return this
*/
public AqlQueryOptions allowDirtyReads(final Boolean allowDirtyReads) {
getOptions().setAllowDirtyReads(allowDirtyReads);
return this;
}
@JsonIgnore
public Boolean getAllowRetry() {
return getOptions().getAllowRetry();
}
/**
* @param allowRetry Set this option to true to make it possible to retry fetching the latest batch from a cursor.
*
* This makes possible to safely retry invoking {@link com.arangodb.ArangoCursor#next()} in
* case of I/O exceptions (which are actually thrown as {@link com.arangodb.ArangoDBException}
* with cause {@link java.io.IOException})
*
* If set to false (default), then it is not safe to retry invoking
* {@link com.arangodb.ArangoCursor#next()} in case of I/O exceptions, since the request to
* fetch the next batch is not idempotent (i.e. the cursor may advance multiple times on the
* server).
*
* Note: once you successfully received the last batch, you should call
* {@link com.arangodb.ArangoCursor#close()} so that the server does not unnecessary keep the
* batch until the cursor times out ({@link AqlQueryOptions#ttl(Integer)}).
* @return this
* @since ArangoDB 3.11
*/
public AqlQueryOptions allowRetry(final Boolean allowRetry) {
getOptions().setAllowRetry(allowRetry);
return this;
}
@JsonIgnore
public Boolean getFailOnWarning() {
return getOptions().getFailOnWarning();
}
/**
* @param failOnWarning When set to true, the query will throw an exception and abort instead of producing a
* warning. This option should be used during development to catch potential issues early.
* When the attribute is set to false, warnings will not be propagated to exceptions and will
* be returned with the query result. There is also a server configuration option
* --query.fail-on-warning for setting the default value for failOnWarning so it does not
* need to be set on a per-query level.
* @return this
*/
public AqlQueryOptions failOnWarning(final Boolean failOnWarning) {
getOptions().setFailOnWarning(failOnWarning);
return this;
}
@JsonIgnore
public Boolean getFillBlockCache() {
return getOptions().getFillBlockCache();
}
/**
* @param fillBlockCache if set to true
or not specified, this will make the query store
* the data it reads via the RocksDB storage engine in the RocksDB block cache. This is
* usually the desired behavior. The option can be set to false
for queries that
* are known to either read a lot of data that would thrash the block cache, or for queries
* that read data known to be outside of the hot set. By setting the option
* to false
, data read by the query will not make it into the RocksDB block
* cache if it is not already in there, thus leaving more room for the actual hot set.
* @return this
* @since ArangoDB 3.8.1
*/
public AqlQueryOptions fillBlockCache(final Boolean fillBlockCache) {
getOptions().setFillBlockCache(fillBlockCache);
return this;
}
@JsonIgnore
public String getForceOneShardAttributeValue() {
return getOptions().getForceOneShardAttributeValue();
}
/**
* @param forceOneShardAttributeValue This query option can be used in complex queries in case the query optimizer
* cannot automatically detect that the query can be limited to only a single
* server (e.g. in a disjoint smart graph case).
*
* If the option is set incorrectly, i.e. to a wrong shard key value, then the
* query may be shipped to a wrong DB server and may not return results (i.e.
* empty result set).
*
* Use at your own risk.
* @return this
*/
public AqlQueryOptions forceOneShardAttributeValue(final String forceOneShardAttributeValue) {
getOptions().setForceOneShardAttributeValue(forceOneShardAttributeValue);
return this;
}
@JsonIgnore
public Boolean getFullCount() {
return getOptions().getFullCount();
}
/**
* @param fullCount if set to true and the query contains a LIMIT clause, then the result will have an extra
* attribute
* with the sub-attributes stats and fullCount, { ... , "extra": { "stats": { "fullCount": 123 }
* } }. The
* fullCount attribute will contain the number of documents in the result before the last LIMIT
* in the
* query was applied. It can be used to count the number of documents that match certain filter
* criteria,
* but only return a subset of them, in one go. It is thus similar to MySQL's
* SQL_CALC_FOUND_ROWS hint.
* Note that setting the option will disable a few LIMIT optimizations and may lead to more
* documents
* being processed, and thus make queries run longer. Note that the fullCount attribute will
* only be
* present in the result if the query has a LIMIT clause and the LIMIT clause is actually used
* in the
* query.
* @return this
*/
public AqlQueryOptions fullCount(final Boolean fullCount) {
getOptions().setFullCount(fullCount);
return this;
}
@JsonIgnore
public Long getIntermediateCommitCount() {
return getOptions().getIntermediateCommitCount();
}
/**
* @param intermediateCommitCount Maximum number of operations after which an intermediate commit is performed
* automatically. Honored by
* the RocksDB storage engine only.
* @return this
* @since ArangoDB 3.2.0
*/
public AqlQueryOptions intermediateCommitCount(final Long intermediateCommitCount) {
getOptions().setIntermediateCommitCount(intermediateCommitCount);
return this;
}
@JsonIgnore
public Long getIntermediateCommitSize() {
return getOptions().getIntermediateCommitSize();
}
/**
* @param intermediateCommitSize Maximum total size of operations after which an intermediate commit is performed
* automatically.
* Honored by the RocksDB storage engine only.
* @return this
* @since ArangoDB 3.2.0
*/
public AqlQueryOptions intermediateCommitSize(final Long intermediateCommitSize) {
getOptions().setIntermediateCommitSize(intermediateCommitSize);
return this;
}
@JsonIgnore
public Integer getMaxDNFConditionMembers() {
return getOptions().getMaxDNFConditionMembers();
}
/**
* @param maxDNFConditionMembers A threshold for the maximum number of OR sub-nodes in the internal representation
* of an AQL FILTER condition.
*
* Yon can use this option to limit the computation time and memory usage when
* converting complex AQL FILTER conditions into the internal DNF (disjunctive normal
* form) format. FILTER conditions with a lot of logical branches (AND, OR, NOT) can
* take a large amount of processing time and memory. This query option limits the
* computation time and memory usage for such conditions.
*
* Once the threshold value is reached during the DNF conversion of a FILTER
* condition, the conversion is aborted, and the query continues with a simplified
* internal representation of the condition, which cannot be used for index lookups.
*
* You can set the threshold globally instead of per query with the
* --query.max-dnf-condition-members startup option.
* @return this
*/
public AqlQueryOptions maxDNFConditionMembers(final Integer maxDNFConditionMembers) {
getOptions().setMaxDNFConditionMembers(maxDNFConditionMembers);
return this;
}
@JsonIgnore
public Integer getMaxNodesPerCallstack() {
return getOptions().getMaxNodesPerCallstack();
}
/**
* @param maxNodesPerCallstack The number of execution nodes in the query plan after that stack splitting is
* performed to avoid a potential stack overflow. Defaults to the configured value of
* the startup option --query.max-nodes-per-callstack.
*
* This option is only useful for testing and debugging and normally does not need any
* adjustment.
* @return this
*/
public AqlQueryOptions maxNodesPerCallstack(final Integer maxNodesPerCallstack) {
getOptions().setMaxNodesPerCallstack(maxNodesPerCallstack);
return this;
}
@JsonIgnore
public Integer getMaxNumberOfPlans() {
return getOptions().getMaxNumberOfPlans();
}
/**
* @param maxNumberOfPlans Limits the maximum number of plans that are created by the AQL query optimizer.
* @return this
*/
public AqlQueryOptions maxNumberOfPlans(final Integer maxNumberOfPlans) {
getOptions().setMaxNumberOfPlans(maxNumberOfPlans);
return this;
}
/**
* @deprecated for removal, use {@link AqlQueryOptions#getMaxNumberOfPlans()} instead
*/
@Deprecated
@JsonIgnore
public Integer getMaxPlans() {
return getMaxNumberOfPlans();
}
/**
* @param maxPlans Limits the maximum number of plans that are created by the AQL query optimizer.
* @return this
* @deprecated for removal, use {@link AqlQueryOptions#maxNumberOfPlans(Integer)} instead
*/
@Deprecated
public AqlQueryOptions maxPlans(final Integer maxPlans) {
return maxNumberOfPlans(maxPlans);
}
@JsonIgnore
public Double getMaxRuntime() {
return getOptions().getMaxRuntime();
}
/**
* @param maxRuntime The query has to be executed within the given runtime or it will be killed. The value is specified
* in seconds. The default value is 0.0 (no timeout).
* @return this
*/
public AqlQueryOptions maxRuntime(final Double maxRuntime) {
getOptions().setMaxRuntime(maxRuntime);
return this;
}
@JsonIgnore
public Long getMaxTransactionSize() {
return getOptions().getMaxTransactionSize();
}
/**
* @param maxTransactionSize Transaction size limit in bytes. Honored by the RocksDB storage engine only.
* @return this
* @since ArangoDB 3.2.0
*/
public AqlQueryOptions maxTransactionSize(final Long maxTransactionSize) {
getOptions().setMaxTransactionSize(maxTransactionSize);
return this;
}
@JsonIgnore
public Long getMaxWarningCount() {
return getOptions().getMaxWarningCount();
}
/**
* @param maxWarningCount Limits the maximum number of warnings a query will return. The number of warnings a
* query will return
* is limited to 10 by default, but that number can be increased or decreased by setting
* this attribute.
* @return this
* @since ArangoDB 3.2.0
*/
public AqlQueryOptions maxWarningCount(final Long maxWarningCount) {
getOptions().setMaxWarningCount(maxWarningCount);
return this;
}
@JsonIgnore
public Optimizer getOptimizer() {
return getOptions().getOptimizer();
}
/**
* @param optimizer Options related to the query optimizer.
* @return this
*/
public AqlQueryOptions optimizer(final Optimizer optimizer) {
getOptions().setOptimizer(optimizer);
return this;
}
@JsonIgnore
public Boolean getProfile() {
return getOptions().getProfile();
}
/**
* @param profile If set to true, then the additional query profiling information will be returned in the
* sub-attribute
* profile of the extra return attribute if the query result is not served from the query cache.
* @return this
*/
public AqlQueryOptions profile(final Boolean profile) {
getOptions().setProfile(profile);
return this;
}
@JsonIgnore
public Double getSatelliteSyncWait() {
return getOptions().getSatelliteSyncWait();
}
/**
* @param satelliteSyncWait This enterprise parameter allows to configure how long a DBServer will have time to
* bring the
* satellite collections involved in the query into sync. The default value is 60.0
* (seconds). When the
* max time has been reached the query will be stopped.
* @return this
* @since ArangoDB 3.2.0
*/
public AqlQueryOptions satelliteSyncWait(final Double satelliteSyncWait) {
getOptions().setSatelliteSyncWait(satelliteSyncWait);
return this;
}
@JsonIgnore
public Collection getShardIds() {
return getOptions().getShardIds();
}
/**
* Restrict query to shards by given ids. This is an internal option. Use at your own risk.
*
* @param shardIds
* @return this
*/
public AqlQueryOptions shardIds(final String... shardIds) {
getOptions().setShardIds(Arrays.asList(shardIds));
return this;
}
@JsonIgnore
public Boolean getSkipInaccessibleCollections() {
return getOptions().getSkipInaccessibleCollections();
}
/**
* @param skipInaccessibleCollections AQL queries (especially graph traversals) will treat collection to which a
* user has no access rights
* as if these collections were empty. Instead of returning a forbidden access
* error, your queries will
* execute normally. This is intended to help with certain use-cases: A graph
* contains several
* collections and different users execute AQL queries on that graph. You can
* now naturally limit the
* accessible results by changing the access rights of users on collections.
* This feature is only
* available in the Enterprise Edition.
* @return this
* @since ArangoDB 3.2.0
*/
public AqlQueryOptions skipInaccessibleCollections(final Boolean skipInaccessibleCollections) {
getOptions().setSkipInaccessibleCollections(skipInaccessibleCollections);
return this;
}
@JsonIgnore
public Long getSpillOverThresholdMemoryUsage() {
return getOptions().getSpillOverThresholdMemoryUsage();
}
/**
* @param spillOverThresholdMemoryUsage This option allows queries to store intermediate and final results
* temporarily on disk if the amount of memory used (in bytes) exceeds the
* specified value. This is used for decreasing the memory usage during the
* query execution.
*
* This option only has an effect on queries that use the SORT operation but
* without a LIMIT, and if you enable the spillover feature by setting a path
* for the directory to store the temporary data in with the
* --temp.intermediate-results-path startup option.
*
* Default value: 128MB.
*
* Spilling data from RAM onto disk is an experimental feature and is turned
* off by default. The query results are still built up entirely in RAM on
* Coordinators and single servers for non-streaming queries. To avoid the
* buildup of the entire query result in RAM, use a streaming query (see the
* stream option).
* @return this
*/
public AqlQueryOptions spillOverThresholdMemoryUsage(final Long spillOverThresholdMemoryUsage) {
getOptions().setSpillOverThresholdMemoryUsage(spillOverThresholdMemoryUsage);
return this;
}
@JsonIgnore
public Long getSpillOverThresholdNumRows() {
return getOptions().getSpillOverThresholdNumRows();
}
/**
* @param spillOverThresholdNumRows This option allows queries to store intermediate and final results temporarily
* on disk if the number of rows produced by the query exceeds the specified value.
* This is used for decreasing the memory usage during the query execution. In a
* query that iterates over a collection that contains documents, each row is a
* document, and in a query that iterates over temporary values
* (i.e. FOR i IN 1..100), each row is one of such temporary values.
*
* This option only has an effect on queries that use the SORT operation but
* without a LIMIT, and if you enable the spillover feature by setting a path for
* the directory to store the temporary data in with the
* --temp.intermediate-results-path startup option.
*
* Default value: 5000000 rows.
*
* Spilling data from RAM onto disk is an experimental feature and is turned off
* by default. The query results are still built up entirely in RAM on Coordinators
* and single servers for non-streaming queries. To avoid the buildup of the entire
* query result in RAM, use a streaming query (see the stream option).
* @return this
*/
public AqlQueryOptions spillOverThresholdNumRows(final Long spillOverThresholdNumRows) {
getOptions().setSpillOverThresholdNumRows(spillOverThresholdNumRows);
return this;
}
@JsonIgnore
public Boolean getStream() {
return getOptions().getStream();
}
/**
* @param stream Specify true and the query will be executed in a streaming fashion. The query result is not
* stored on
* the server, but calculated on the fly. Beware: long-running queries will need to hold the
* collection
* locks for as long as the query cursor exists. When set to false a query will be executed right
* away in
* its entirety. In that case query results are either returned right away (if the resultset is small
* enough), or stored on the arangod instance and accessible via the cursor API (with respect to the
* ttl). It is advisable to only use this option on short-running queries or without exclusive locks
* (write-locks on MMFiles). Please note that the query options cache, count and fullCount will not
* work
* on streaming queries. Additionally query statistics, warnings and profiling data will only be
* available after the query is finished. The default value is false
* @return this
* @since ArangoDB 3.4.0
*/
public AqlQueryOptions stream(final Boolean stream) {
getOptions().setStream(stream);
return this;
}
@JsonIgnore
public Collection getRules() {
return getOptions().getOptimizer().getRules();
}
/**
* @param rules A list of to-be-included or to-be-excluded optimizer rules can be put into this attribute,
* telling the
* optimizer to include or exclude specific rules. To disable a rule, prefix its name with a -, to
* enable
* a rule, prefix it with a +. There is also a pseudo-rule all, which will match all optimizer rules
* @return this
*/
public AqlQueryOptions rules(final Collection rules) {
getOptions().getOptimizer().setRules(rules);
return this;
}
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy