com.mongodb.internal.operation.AggregateOperationImpl Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of mongodb-driver-core Show documentation
Show all versions of mongodb-driver-core Show documentation
The Java operations layer for the MongoDB Java Driver. Third parties can ' +
'wrap this layer to provide custom higher-level APIs
/*
* Copyright 2008-present MongoDB, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.mongodb.internal.operation;
import com.mongodb.MongoNamespace;
import com.mongodb.client.model.Collation;
import com.mongodb.connection.ConnectionDescription;
import com.mongodb.internal.async.AsyncBatchCursor;
import com.mongodb.internal.async.SingleResultCallback;
import com.mongodb.internal.binding.AsyncReadBinding;
import com.mongodb.internal.binding.ReadBinding;
import com.mongodb.internal.client.model.AggregationLevel;
import com.mongodb.internal.connection.QueryResult;
import com.mongodb.internal.session.SessionContext;
import com.mongodb.lang.Nullable;
import org.bson.BsonArray;
import org.bson.BsonBoolean;
import org.bson.BsonDocument;
import org.bson.BsonInt32;
import org.bson.BsonInt64;
import org.bson.BsonString;
import org.bson.BsonValue;
import org.bson.codecs.Decoder;
import java.util.Arrays;
import java.util.List;
import java.util.concurrent.TimeUnit;
import static com.mongodb.assertions.Assertions.assertNotNull;
import static com.mongodb.assertions.Assertions.isTrueArgument;
import static com.mongodb.assertions.Assertions.notNull;
import static com.mongodb.internal.async.ErrorHandlingResultCallback.errorHandlingCallback;
import static com.mongodb.internal.operation.AsyncOperationHelper.CommandReadTransformerAsync;
import static com.mongodb.internal.operation.AsyncOperationHelper.executeRetryableReadAsync;
import static com.mongodb.internal.operation.CommandOperationHelper.CommandCreator;
import static com.mongodb.internal.operation.OperationHelper.LOGGER;
import static com.mongodb.internal.operation.OperationHelper.cursorDocumentToQueryResult;
import static com.mongodb.internal.operation.OperationReadConcernHelper.appendReadConcernToCommand;
import static com.mongodb.internal.operation.SyncOperationHelper.CommandReadTransformer;
import static com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead;
class AggregateOperationImpl implements AsyncReadOperation>, ReadOperation> {
private static final String RESULT = "result";
private static final String CURSOR = "cursor";
private static final String FIRST_BATCH = "firstBatch";
private static final List FIELD_NAMES_WITH_RESULT = Arrays.asList(RESULT, FIRST_BATCH);
private final MongoNamespace namespace;
private final List pipeline;
private final Decoder decoder;
private final AggregateTarget aggregateTarget;
private final PipelineCreator pipelineCreator;
private boolean retryReads;
private Boolean allowDiskUse;
private Integer batchSize;
private Collation collation;
private BsonValue comment;
private BsonValue hint;
private long maxAwaitTimeMS;
private long maxTimeMS;
private BsonDocument variables;
AggregateOperationImpl(final MongoNamespace namespace, final List pipeline, final Decoder decoder,
final AggregationLevel aggregationLevel) {
this(namespace, pipeline, decoder, defaultAggregateTarget(notNull("aggregationLevel", aggregationLevel),
notNull("namespace", namespace).getCollectionName()), defaultPipelineCreator(pipeline));
}
AggregateOperationImpl(final MongoNamespace namespace, final List pipeline, final Decoder decoder,
final AggregateTarget aggregateTarget, final PipelineCreator pipelineCreator) {
this.namespace = notNull("namespace", namespace);
this.pipeline = notNull("pipeline", pipeline);
this.decoder = notNull("decoder", decoder);
this.aggregateTarget = notNull("aggregateTarget", aggregateTarget);
this.pipelineCreator = notNull("pipelineCreator", pipelineCreator);
}
MongoNamespace getNamespace() {
return namespace;
}
List getPipeline() {
return pipeline;
}
Decoder getDecoder() {
return decoder;
}
Boolean getAllowDiskUse() {
return allowDiskUse;
}
AggregateOperationImpl allowDiskUse(@Nullable final Boolean allowDiskUse) {
this.allowDiskUse = allowDiskUse;
return this;
}
Integer getBatchSize() {
return batchSize;
}
AggregateOperationImpl batchSize(@Nullable final Integer batchSize) {
this.batchSize = batchSize;
return this;
}
long getMaxAwaitTime(final TimeUnit timeUnit) {
notNull("timeUnit", timeUnit);
return timeUnit.convert(maxAwaitTimeMS, TimeUnit.MILLISECONDS);
}
AggregateOperationImpl maxAwaitTime(final long maxAwaitTime, final TimeUnit timeUnit) {
notNull("timeUnit", timeUnit);
isTrueArgument("maxAwaitTime >= 0", maxAwaitTime >= 0);
this.maxAwaitTimeMS = TimeUnit.MILLISECONDS.convert(maxAwaitTime, timeUnit);
return this;
}
long getMaxTime(final TimeUnit timeUnit) {
notNull("timeUnit", timeUnit);
return timeUnit.convert(maxTimeMS, TimeUnit.MILLISECONDS);
}
AggregateOperationImpl maxTime(final long maxTime, final TimeUnit timeUnit) {
notNull("timeUnit", timeUnit);
isTrueArgument("maxTime >= 0", maxTime >= 0);
this.maxTimeMS = TimeUnit.MILLISECONDS.convert(maxTime, timeUnit);
return this;
}
Collation getCollation() {
return collation;
}
AggregateOperationImpl collation(@Nullable final Collation collation) {
this.collation = collation;
return this;
}
@Nullable
BsonValue getComment() {
return comment;
}
AggregateOperationImpl comment(@Nullable final BsonValue comment) {
this.comment = comment;
return this;
}
AggregateOperationImpl let(@Nullable final BsonDocument variables) {
this.variables = variables;
return this;
}
AggregateOperationImpl retryReads(final boolean retryReads) {
this.retryReads = retryReads;
return this;
}
boolean getRetryReads() {
return retryReads;
}
@Nullable
BsonValue getHint() {
return hint;
}
AggregateOperationImpl hint(@Nullable final BsonValue hint) {
isTrueArgument("BsonString or BsonDocument", hint == null || hint.isDocument() || hint.isString());
this.hint = hint;
return this;
}
@Override
public BatchCursor execute(final ReadBinding binding) {
return executeRetryableRead(binding, namespace.getDatabaseName(), getCommandCreator(binding.getSessionContext()),
CommandResultDocumentCodec.create(decoder, FIELD_NAMES_WITH_RESULT), transformer(), retryReads);
}
@Override
public void executeAsync(final AsyncReadBinding binding, final SingleResultCallback> callback) {
SingleResultCallback> errHandlingCallback = errorHandlingCallback(callback, LOGGER);
executeRetryableReadAsync(binding, namespace.getDatabaseName(), getCommandCreator(binding.getSessionContext()),
CommandResultDocumentCodec.create(this.decoder, FIELD_NAMES_WITH_RESULT), asyncTransformer(), retryReads,
errHandlingCallback);
}
private CommandCreator getCommandCreator(final SessionContext sessionContext) {
return (serverDescription, connectionDescription) -> getCommand(sessionContext, connectionDescription.getMaxWireVersion());
}
BsonDocument getCommand(final SessionContext sessionContext, final int maxWireVersion) {
BsonDocument commandDocument = new BsonDocument("aggregate", aggregateTarget.create());
appendReadConcernToCommand(sessionContext, maxWireVersion, commandDocument);
commandDocument.put("pipeline", pipelineCreator.create());
if (maxTimeMS > 0) {
commandDocument.put("maxTimeMS", maxTimeMS > Integer.MAX_VALUE
? new BsonInt64(maxTimeMS) : new BsonInt32((int) maxTimeMS));
}
BsonDocument cursor = new BsonDocument();
if (batchSize != null) {
cursor.put("batchSize", new BsonInt32(batchSize));
}
commandDocument.put(CURSOR, cursor);
if (allowDiskUse != null) {
commandDocument.put("allowDiskUse", BsonBoolean.valueOf(allowDiskUse));
}
if (collation != null) {
commandDocument.put("collation", collation.asDocument());
}
if (comment != null) {
commandDocument.put("comment", comment);
}
if (hint != null) {
commandDocument.put("hint", hint);
}
if (variables != null) {
commandDocument.put("let", variables);
}
return commandDocument;
}
private QueryResult createQueryResult(final BsonDocument result, final ConnectionDescription description) {
assertNotNull(result);
return cursorDocumentToQueryResult(result.getDocument(CURSOR), description.getServerAddress());
}
private CommandReadTransformer> transformer() {
return (result, source, connection) -> {
QueryResult queryResult = createQueryResult(result, connection.getDescription());
return new QueryBatchCursor<>(queryResult, 0, batchSize != null ? batchSize : 0, maxAwaitTimeMS, decoder, comment,
source, connection, result);
};
}
private CommandReadTransformerAsync> asyncTransformer() {
return (result, source, connection) -> {
QueryResult queryResult = createQueryResult(result, connection.getDescription());
return new AsyncQueryBatchCursor<>(queryResult, 0, batchSize != null ? batchSize : 0, maxAwaitTimeMS, decoder,
comment, source, connection, result);
};
}
interface AggregateTarget {
BsonValue create();
}
interface PipelineCreator {
BsonArray create();
}
private static AggregateTarget defaultAggregateTarget(final AggregationLevel aggregationLevel, final String collectionName) {
return () -> {
if (aggregationLevel == AggregationLevel.DATABASE) {
return new BsonInt32(1);
} else {
return new BsonString(collectionName);
}
};
}
private static PipelineCreator defaultPipelineCreator(final List pipeline) {
return () -> new BsonArray(pipeline);
}
}