All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.amazonaws.services.sqs.buffered.QueueBufferConfig Maven / Gradle / Ivy

Go to download

The AWS Java SDK for Amazon SQS module holds the client classes that are used for communicating with Amazon Simple Queue Service

There is a newer version: 1.12.778
Show newest version
/*
 * Copyright 2012-2024 Amazon.com, Inc. or its affiliates. All Rights Reserved.
 *
 * Licensed under the Apache License, Version 2.0 (the "License").
 * You may not use this file except in compliance with the License.
 * A copy of the License is located at
 *
 *  http://aws.amazon.com/apache2.0
 *
 * or in the "license" file accompanying this file. This file is distributed
 * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
 * express or implied. See the License for the specific language governing
 * permissions and limitations under the License.
 */

package com.amazonaws.services.sqs.buffered;

import java.util.ArrayList;
import java.util.Collections;
import java.util.List;

import com.amazonaws.AmazonClientException;
import com.amazonaws.handlers.AsyncHandler;
import com.amazonaws.services.sqs.model.DeleteMessageRequest;
import com.amazonaws.services.sqs.model.DeleteMessageResult;

public class QueueBufferConfig {

    public static final int MAX_BATCH_SIZE_DEFAULT = 10;

    /** the maximum number of entries in a batch command */
    private int maxBatchSize;

    /** Updated as the service now supports messages of size max 256 KiB. */
    public static final long SERVICE_MAX_BATCH_SIZE_BYTES = 256 * 1024;

    /**
     * The maximum time (milliseconds) a send batch is held open for additional outbound requests.
     * The longer this timeout, the longer messages wait for other messages to be added to the
     * batch. Increasing this timeout reduces the number of calls made and increases throughput, but
     * also increases average message latency.
     */
    private long maxBatchOpenMs;

    /** 200 milliseconds */
    public static final long MAX_BATCH_OPEN_MS_DEFAULT = 200;

    /**
     * If true, even synchronous calls to delete messages will be made using background
     * asynchronous batches. The client will return results indicating that the messages were deleted successfully
     * even if the background calls eventually fail; the actual errors will be logged instead.
     * This can be beneficial for decreasing message acknowledgement latency at the cost of potential
     * duplicate messages (which can be produced by SQS itself anyway).
     */
    private boolean deleteInBackground = false;

    /**
     * Should we use long polling or not?
     */
    private boolean longPoll;

    /** true */
    private static final boolean LONG_POLL_DEFAULT = true;

    /**
     * The maximum number of concurrent batches for each type of outbound request. The greater the
     * number, the greater the throughput that can be achieved (at the expense of consuming more
     * threads).
     */
    private int maxInflightOutboundBatches;

    /** 5 batches */
    public static final int MAX_INFLIGHT_OUTBOUND_BATCHES_DEFAULT = 5;

    /**
     * The maximum number of concurrent receive message batches. The greater this number, the faster
     * the queue will be pulling messages from the SQS servers (at the expense of consuming more
     * threads).
     */
    private int maxInflightReceiveBatches;

    /** 10 batches */
    public static final int MAX_INFLIGHT_RECEIVE_BATCHES_DEFAULT = 10;

    /**
     * If more than that number of completed receive batches are waiting in the buffer, the querying
     * for new messages will stop. The larger this number, the more messages the buffer queue will
     * pre-fetch and keep in the buffer on the client side, and the faster receive requests will be
     * satisfied. The visibility timeout of a pre-fetched message starts at the point of pre-fetch,
     * which means that while the message is in the local buffer it is unavailable for other clients
     * to process, and when this client retrieves it, part of the visibility timeout may have
     * already expired. The number of messages prefetched will not exceed maxBatchSize *
     * maxDoneReceiveBatches.
     */
    private int maxDoneReceiveBatches;

    /** 10 batches */
    public static final int MAX_DONE_RECEIVE_BATCHES_DEFAULT = 10;

    /**
     * Maximum permitted size of a SendMessage or SendMessageBatch message, in bytes
     */
    private long maxBatchSizeBytes;

    /** 256 kilobytes */
    public static final long MAX_BATCH_SIZE_BYTES_DEFAULT = SERVICE_MAX_BATCH_SIZE_BYTES;

    /**
     * Custom visibility timeout to use when retrieving messages from SQS. If set to a value greater
     * than zero, this timeout will override the default visibility timeout set on the SQS queue.
     * Set it to -1 to use the default visiblity timeout of the queue. Visibility timeout of 0
     * seconds is not supported.
     */
    private int visibilityTimeoutSeconds;

    /** -1, which means use the visibility timeout of the queue */
    public static final int VISIBILITY_TIMEOUT_SECONDS_DEFAULT = -1;

    /**
     * Specifies the amount of time, in seconds, the receive call will block on the server waiting
     * for messages to arrive if the queue is empty when the receive call is first made. This
     * setting has no effect if long polling is disabled.
     */
    private int longPollWaitTimeoutSeconds;

    public static final int LONGPOLL_WAIT_TIMEOUT_SECONDS_DEFAULT = 20;

    /**
     * Configures the minimum wait time for incoming receive message requests. Without a non-zero
     * minimum wait time, threads can easily waste CPU time busy-waiting against empty local buffers.
     * Avoid setting this to 0 unless you are confident threads will do useful work in-between
     * each call to receive messages!
     * 

* This will be applied to both requests that explicitly set WaitTimeSeconds and * those that inherit the ReceiveMessageWaitTimeSeconds queue attribute. */ private int minReceiveWaitTimeMs = MIN_RECEIVE_WAIT_TIME_MS_DEFAULT; /** 50 ms, which is in the ballpark for typical latency contacting a remote service like SQS */ public static final int MIN_RECEIVE_WAIT_TIME_MS_DEFAULT = 50; /** * Specifies the message attributes receive calls will request. Only receive message requests that * request the same set of attributes will be satisfied from the receive buffers. *

* The default value is an empty list, so any receive requests that require message attributes * will not be fulfilled from buffers. */ private List receiveMessageAttributeNames = RECEIVE_MESSAGE_ATTRIBUTE_NAMES_DEFAULT; public static final List RECEIVE_MESSAGE_ATTRIBUTE_NAMES_DEFAULT = Collections.emptyList(); /** * Specifies the attributes receive calls will request. Only receive message requests that * request the same set of attributes will be satisfied from the receive buffers. *

* The default value is an empty list, so any receive requests that require attributes * will not be fulfilled from buffers. */ private List receiveAttributeNames = RECEIVE_ATTRIBUTE_NAMES_DEFAULT; public static final List RECEIVE_ATTRIBUTE_NAMES_DEFAULT = Collections.emptyList(); /** * If set, prefetching will be scaled with the number of in-flight incoming receive requests * made to the client. The advantage of this is reducing the number of outgoing requests * made to SQS when incoming requests are reduced: in particular, if all incoming requests * stop no future requests to SQS will be made. The disadvantage is increased latency when * incoming requests first start occurring. */ private boolean adaptivePrefetching = ADAPTIVE_PREFETCHING_DEFAULT; public static final boolean ADAPTIVE_PREFETCHING_DEFAULT = false; /** * Option to configure flushOnShutdown. Enabling this option will flush the pending requests in the * {@link SendQueueBuffer} during shutdown. *

* The default value is false which indicates flushOnShutdown is disabled. *

*/ private boolean flushOnShutdown = false; public QueueBufferConfig(long maxBatchOpenMs, int maxInflightOutboundBatches, int maxInflightReceiveBatches, int maxDoneReceiveBatches, boolean paramLongPoll, long maxBatchSizeBytes, int visibilityTimeout, int longPollTimeout, int maxBatch) { super(); this.maxBatchOpenMs = maxBatchOpenMs; this.maxInflightOutboundBatches = maxInflightOutboundBatches; this.maxInflightReceiveBatches = maxInflightReceiveBatches; this.maxDoneReceiveBatches = maxDoneReceiveBatches; this.longPoll = paramLongPoll; this.maxBatchSizeBytes = maxBatchSizeBytes; this.visibilityTimeoutSeconds = visibilityTimeout; this.longPollWaitTimeoutSeconds = longPollTimeout; this.maxBatchSize = maxBatch; } public QueueBufferConfig() { this(MAX_BATCH_OPEN_MS_DEFAULT, MAX_INFLIGHT_OUTBOUND_BATCHES_DEFAULT, MAX_INFLIGHT_RECEIVE_BATCHES_DEFAULT, MAX_DONE_RECEIVE_BATCHES_DEFAULT, LONG_POLL_DEFAULT, MAX_BATCH_SIZE_BYTES_DEFAULT, VISIBILITY_TIMEOUT_SECONDS_DEFAULT, LONGPOLL_WAIT_TIMEOUT_SECONDS_DEFAULT, MAX_BATCH_SIZE_DEFAULT); } /** copy constructor */ public QueueBufferConfig(QueueBufferConfig other) { longPoll = other.longPoll; longPollWaitTimeoutSeconds = other.longPollWaitTimeoutSeconds; minReceiveWaitTimeMs = other.minReceiveWaitTimeMs; maxBatchOpenMs = other.maxBatchOpenMs; maxBatchSize = other.maxBatchSize; maxBatchSizeBytes = other.maxBatchSizeBytes; maxDoneReceiveBatches = other.maxDoneReceiveBatches; maxInflightOutboundBatches = other.maxInflightOutboundBatches; maxInflightReceiveBatches = other.maxInflightReceiveBatches; visibilityTimeoutSeconds = other.visibilityTimeoutSeconds; flushOnShutdown = other.flushOnShutdown; receiveAttributeNames = other.receiveAttributeNames; receiveMessageAttributeNames = other.receiveMessageAttributeNames; adaptivePrefetching = other.adaptivePrefetching; deleteInBackground = other.deleteInBackground; } @Override public String toString() { return "QueueBufferConfig [maxBatchSize=" + maxBatchSize + ", maxBatchOpenMs=" + maxBatchOpenMs + ", longPoll=" + longPoll + ", maxInflightOutboundBatches=" + maxInflightOutboundBatches + ", maxInflightReceiveBatches=" + maxInflightReceiveBatches + ", maxDoneReceiveBatches=" + maxDoneReceiveBatches + ", maxBatchSizeBytes=" + maxBatchSizeBytes + ", visibilityTimeoutSeconds=" + visibilityTimeoutSeconds + ", longPollWaitTimeoutSeconds=" + longPollWaitTimeoutSeconds + "]"; } /** * The maximum time (milliseconds) a send batch is held open for additional outbound requests. * The longer this timeout, the longer messages wait for other messages to be added to the * batch. Increasing this timeout reduces the number of calls made and increases throughput, but * also increases average message latency. */ public long getMaxBatchOpenMs() { return maxBatchOpenMs; } /** * The maximum time (milliseconds) a send batch is held open for additional outbound requests. * The longer this timeout, the longer messages wait for other messages to be added to the * batch. Increasing this timeout reduces the number of calls made and increases throughput, but * also increases average message latency. */ public void setMaxBatchOpenMs(long maxBatchOpenMs) { this.maxBatchOpenMs = maxBatchOpenMs; } /** * The maximum time (milliseconds) a send batch is held open for additional outbound requests. * The longer this timeout, the longer messages wait for other messages to be added to the * batch. Increasing this timeout reduces the number of calls made and increases throughput, but * also increases average message latency. */ public QueueBufferConfig withMaxBatchOpenMs(long maxBatchOpenMs) { setMaxBatchOpenMs(maxBatchOpenMs); return this; } /** * If set, even synchronous calls to delete messages will be made using background * asynchronous batches. The client will return results indicating that the messages were deleted successfully * even if the background calls eventually fail; the actual result of the deletions will be reported * through the given handler instead (often just logging errors). This can be beneficial for decreasing message * acknowledgement latency at the cost of potential duplicate messages (which can be produced by SQS itself anyway). */ public boolean isDeleteInBackground() { return deleteInBackground; } /** * If set, even synchronous calls to delete messages will be made using background * asynchronous batches. The client will return results indicating that the messages were deleted successfully * even if the background calls eventually fail; any errors result of the deletions will be reported * through the given handler instead (often just logging errors). This can be beneficial for decreasing message * acknowledgement latency at the cost of potential duplicate messages (which can be produced by SQS itself anyway). */ public void setDeleteInBackground(boolean deleteInBackground) { this.deleteInBackground = deleteInBackground; } /** * If set, even synchronous calls to delete messages will be made using background * asynchronous batches. The client will return results indicating that the messages were deleted successfully * even if the background calls eventually fail; the actual result of the deletions will be reported * through the given handler instead (often just logging errors). This can be beneficial for decreasing message * acknowledgement latency at the cost of potential duplicate messages (which can be produced by SQS itself anyway). */ public QueueBufferConfig withDeleteInBackground(boolean deleteInBackground) { setDeleteInBackground(deleteInBackground); return this; } /** * @return true if the queue buffer will use long polling while retrieving messages from the * SQS server, false otherwise. */ public boolean isLongPoll() { return longPoll; } /** * Specify "true" for receive requests to use long polling. */ public void setLongPoll(boolean longPoll) { this.longPoll = longPoll; } /** * Specify "true" for receive requests to use long polling. */ public QueueBufferConfig withLongPoll(boolean longPoll) { setLongPoll(longPoll); return this; } /** * The maximum number of concurrent batches for each type of outbound request. The greater the * number, the greater the throughput that can be achieved (at the expense of consuming more * threads). */ public int getMaxInflightOutboundBatches() { return maxInflightOutboundBatches; } /** * The maximum number of concurrent batches for each type of outbound request. The greater the * number, the greater the throughput that can be achieved (at the expense of consuming more * threads). */ public void setMaxInflightOutboundBatches(int maxInflightOutboundBatches) { this.maxInflightOutboundBatches = maxInflightOutboundBatches; } /** * The maximum number of concurrent batches for each type of outbound request. The greater the * number, the greater the throughput that can be achieved (at the expense of consuming more * threads). */ public QueueBufferConfig withMaxInflightOutboundBatches(int maxInflightOutboundBatches) { setMaxInflightOutboundBatches(maxInflightOutboundBatches); return this; } /** * The maximum number of concurrent receive message batches. The greater this number, the faster * the queue will be pulling messages from the SQS servers (at the expense of consuming more * threads). */ public int getMaxInflightReceiveBatches() { return maxInflightReceiveBatches; } /** * The maximum number of concurrent receive message batches. The greater this number, the faster * the queue will be pulling messages from the SQS servers (at the expense of consuming more * threads). */ public void setMaxInflightReceiveBatches(int maxInflightReceiveBatches) { this.maxInflightReceiveBatches = maxInflightReceiveBatches; } /** * The maximum number of concurrent receive message batches. The greater this number, the faster * the queue will be pulling messages from the SQS servers (at the expense of consuming more * threads). */ public QueueBufferConfig withMaxInflightReceiveBatches(int maxInflightReceiveBatches) { setMaxInflightReceiveBatches(maxInflightReceiveBatches); return this; } /** * If more than that number of completed receive batches are waiting in the buffer, the querying * for new messages will stop.
* The larger this number, the more messages the queue buffer will pre-fetch and keep in the * buffer on the client side, and the faster receive requests will be satisfied.
* The visibility timeout of a pre-fetched message starts at the point of pre-fetch, which means * that while the message is in the local buffer it is unavailable for other clients to process, * and when this client retrieves it, part of the visibility timeout may have already expired.
* The number of messages prefetched will not exceed 10 * maxDoneReceiveBatches, as there can be * a maximum of 10 messages per batch.
*/ public int getMaxDoneReceiveBatches() { return maxDoneReceiveBatches; } /** * If more than that number of completed receive batches are waiting in the buffer, the querying * for new messages will stop. The larger this number, the more messages the buffer queue will * pre-fetch and keep in the buffer on the client side, and the faster receive requests will be * satisfied. The visibility timeout of a pre-fetched message starts at the point of pre-fetch, * which means that while the message is in the local buffer it is unavailable for other clients * to process, and when this client retrieves it, part of the visibility timeout may have * already expired. The number of messages prefetched will not exceed maxBatchSize * * maxDoneReceiveBatches. */ public void setMaxDoneReceiveBatches(int maxDoneReceiveBatches) { this.maxDoneReceiveBatches = maxDoneReceiveBatches; } /** * If more than that number of completed receive batches are waiting in the buffer, the querying * for new messages will stop. The larger this number, the more messages the buffer queue will * pre-fetch and keep in the buffer on the client side, and the faster receive requests will be * satisfied. The visibility timeout of a pre-fetched message starts at the point of pre-fetch, * which means that while the message is in the local buffer it is unavailable for other clients * to process, and when this client retrieves it, part of the visibility timeout may have * already expired. The number of messages prefetched will not exceed maxBatchSize * * maxDoneReceiveBatches. */ public QueueBufferConfig withMaxDoneReceiveBatches(int maxDoneReceiveBatches) { setMaxDoneReceiveBatches(maxDoneReceiveBatches); return this; } /** * Maximum permitted size of a SendMessage or SendMessageBatch message, in bytes. This setting * is also enforced on the server, and if this client submits a request of a size larger than * the server can support, the server will reject the request. */ public long getMaxBatchSizeBytes() { return maxBatchSizeBytes; } /** * Maximum permitted size of a SendMessage or SendMessageBatch message, in bytes. This setting * is also enforced on the server, and if this client submits a request of a size larger than * the server can support, the server will reject the request. * * @throws IllegalArgumentException * if the size being set is greater than the service allowed size for message body. */ public void setMaxBatchSizeBytes(long maxBatchSizeBytes) { if (maxBatchSizeBytes > SERVICE_MAX_BATCH_SIZE_BYTES) { throw new IllegalArgumentException( "Maximum Size of the message cannot be greater than the allowed limit of " + SERVICE_MAX_BATCH_SIZE_BYTES + " bytes"); } this.maxBatchSizeBytes = maxBatchSizeBytes; } /** * Maximum permitted size of a SendMessage or SendMessageBatch message, in bytes. This setting * is also enforced on the server, and if this client submits a request of a size larger than * the server can support, the server will reject the request. * * @throws IllegalArgumentException * if the size being set is greater than the service allowed size for message body. */ public QueueBufferConfig withMaxBatchSizeBytes(long maxBatchSizeBytes) { setMaxBatchSizeBytes(maxBatchSizeBytes); return this; } /** * Custom visibility timeout to use when retrieving messages from SQS. If set to a value greater * than zero, this timeout will override the default visibility timeout set on the SQS queue. * Set it to -1 to use the default visiblity timeout of the queue. Visibility timeout of 0 * seconds is not supported. */ public int getVisibilityTimeoutSeconds() { return visibilityTimeoutSeconds; } /** * Custom visibility timeout to use when retrieving messages from SQS. If set to a value greater * than zero, this timeout will override the default visibility timeout set on the SQS queue. * Set it to -1 to use the default visiblity timeout of the queue. Visibility timeout of 0 * seconds is not supported. */ public void setVisibilityTimeoutSeconds(int visibilityTimeoutSeconds) { this.visibilityTimeoutSeconds = visibilityTimeoutSeconds; } /** * Custom visibility timeout to use when retrieving messages from SQS. If set to a value greater * than zero, this timeout will override the default visibility timeout set on the SQS queue. * Set it to -1 to use the default visiblity timeout of the queue. Visibility timeout of 0 * seconds is not supported. */ public QueueBufferConfig withVisibilityTimeoutSeconds(int visibilityTimeoutSeconds) { setVisibilityTimeoutSeconds(visibilityTimeoutSeconds); return this; } /** * Specifies the amount of time, in seconds, the receive call will block on the server waiting * for messages to arrive if the queue is empty when the receive call is first made. This * setting has no effect if long polling is disabled. */ public void setLongPollWaitTimeoutSeconds(int longPollWaitTimeoutSeconds) { this.longPollWaitTimeoutSeconds = longPollWaitTimeoutSeconds; } /** * Specifies the amount of time, in seconds, the receive call will block on the server waiting * for messages to arrive if the queue is empty when the receive call is first made. This * setting has no effect if long polling is disabled. */ public int getLongPollWaitTimeoutSeconds() { return longPollWaitTimeoutSeconds; } /** * Specifies the amount of time, in seconds, the receive call will block on the server waiting * for messages to arrive if the queue is empty when the receive call is first made. This * setting has no effect if long polling is disabled. */ public QueueBufferConfig withLongPollWaitTimeoutSeconds(int longPollWaitTimeoutSeconds) { setLongPollWaitTimeoutSeconds(longPollWaitTimeoutSeconds); return this; } /** * Configures the minimum wait time for incoming receive message requests. Without a non-zero * minimum wait time, threads can easily waste CPU time busy-waiting against empty local buffers. * Avoid setting this to 0 unless you are confident threads will do useful work in-between * each call to receive messages! *

* This will be applied to both requests that explicitly set WaitTimeSeconds and * those that inherit the ReceiveMessageWaitTimeSeconds queue attribute. */ public int getMinReceiveWaitTimeMs() { return minReceiveWaitTimeMs; } /** * Configures the minimum wait time for incoming receive message requests. Without a non-zero * minimum wait time, threads can easily waste CPU time busy-waiting against empty local buffers. * Avoid setting this to 0 unless you are confident threads will do useful work in-between * each call to receive messages! *

* This will be applied to both requests that explicitly set WaitTimeSeconds and * those that inherit the ReceiveMessageWaitTimeSeconds queue attribute. */ public void setMinReceiveWaitTimeMs(int minReceiveWaitTimeMs) { this.minReceiveWaitTimeMs = minReceiveWaitTimeMs; } /** * Configures the minimum wait time for incoming receive message requests. Without a non-zero * minimum wait time, threads can easily waste CPU time busy-waiting against empty local buffers. * Avoid setting this to 0 unless you are confident threads will do useful work in-between * each call to receive messages! *

* This will be applied to both requests that explicitly set WaitTimeSeconds and * those that inherit the ReceiveMessageWaitTimeSeconds queue attribute. */ public QueueBufferConfig withMinReceiveWaitTimeMs(int minReceiveWaitTimeMs) { setMinReceiveWaitTimeMs(minReceiveWaitTimeMs); return this; } /** * Specifies the maximum number of entries the buffered client will put in a single batch * request. */ public int getMaxBatchSize() { return maxBatchSize; } /** * Specifies the maximum number of entries the buffered client will put in a single batch * request. */ public void setMaxBatchSize(int maxBatchSize) { this.maxBatchSize = maxBatchSize; } /** * Specifies the maximum number of entries the buffered client will put in a single batch * request. */ public QueueBufferConfig withMaxBatchSize(int maxBatchSize) { setMaxBatchSize(maxBatchSize); return this; } /** * Specifies the attributes receive calls will request. Only receive message requests that * request the same set of attributes will be satisfied from the receive buffers. *

* The default value is an empty list, so any receive requests that require attributes * will not be fulfilled from buffers. */ public List getReceiveAttributeNames() { return receiveAttributeNames; } /** * Specifies the attributes receive calls will request. Only receive message requests that * request the same set of attributes will be satisfied from the receive buffers. *

* The default value is an empty list, so any receive requests that require attributes * will not be fulfilled from buffers. */ public void setReceiveAttributeNames(List receiveAttributeNames) { if (receiveAttributeNames == null) { this.receiveAttributeNames = Collections.emptyList(); } else { this.receiveAttributeNames = Collections.unmodifiableList(new ArrayList(receiveAttributeNames)); } } /** * Specifies the attributes receive calls will request. Only receive message requests that * request the same set of attributes will be satisfied from the receive buffers. *

* The default value is an empty list, so any receive requests that require attributes * will not be fulfilled from buffers. */ public QueueBufferConfig withReceiveAttributeNames(List receiveAttributes) { setReceiveAttributeNames(receiveAttributes); return this; } /** * Specifies the message attributes receive calls will request. Only receive message requests that * request the same set of attributes will be satisfied from the receive buffers. *

* The default value is an empty list, so any receive requests that require message attributes * will not be fulfilled from buffers. */ public List getReceiveMessageAttributeNames() { return receiveMessageAttributeNames; } /** * Specifies the message attributes receive calls will request. Only receive message requests that * request the same set of attributes will be satisfied from the receive buffers. *

* The default value is an empty list, so any receive requests that require message attributes * will not be fulfilled from buffers. */ public void setReceiveMessageAttributeNames(List receiveMessageAttributeNames) { if (receiveMessageAttributeNames == null) { this.receiveMessageAttributeNames = Collections.emptyList(); } else { this.receiveMessageAttributeNames = Collections.unmodifiableList(new ArrayList(receiveMessageAttributeNames)); } } /** * Specifies the message attributes receive calls will request. Only receive message requests that * request the same set of attributes will be satisfied from the receive buffers. *

* The default value is an empty list, so any receive requests that require message attributes * will not be fulfilled from buffers. */ public QueueBufferConfig withReceiveMessageAttributeNames(List receiveMessageAttributes) { setReceiveMessageAttributeNames(receiveMessageAttributes); return this; } /** * If set, prefetching will be scaled with the number of in-flight incoming receive requests * made to the client. The advantage of this is reducing the number of outgoing requests * made to SQS when incoming requests are reduced: in particular, if all incoming requests * stop no future requests to SQS will be made. The disadvantage is increased latency when * incoming requests first start occurring. */ public void setAdaptivePrefetching(boolean adaptivePrefetching) { this.adaptivePrefetching = adaptivePrefetching; } /** * If set, prefetching will be scaled with the number of in-flight incoming receive requests * made to the client. The advantage of this is reducing the number of outgoing requests * made to SQS when incoming requests are reduced: in particular, if all incoming requests * stop no future requests to SQS will be made. The disadvantage is increased latency when * incoming requests first start occurring. */ public boolean isAdapativePrefetching() { return adaptivePrefetching; } /** * If set, prefetching will be scaled with the number of in-flight incoming receive requests * made to the client. The advantage of this is reducing the number of outgoing requests * made to SQS when incoming requests are reduced: in particular, if all incoming requests * stop no future requests to SQS will be made. The disadvantage is increased latency when * incoming requests first start occurring. */ public QueueBufferConfig withAdapativePrefetching(boolean adaptivePrefetching) { setAdaptivePrefetching(adaptivePrefetching); return this; } /** * Returns the flushOnShutdown value. The default value is false which indicates flushOnShutdown is disabled. * * Enabling this option will flush the pending requests in the {@link SendQueueBuffer} during shutdown. * * @return true if flushOnShutdown is enabled, otherwise false. */ public boolean isFlushOnShutdown() { return flushOnShutdown; } /** * Sets the flushOnShutdown option. The default value is false which indicates flushOnShutdown is disabled. * * Enabling this option will flush the pending requests in the {@link SendQueueBuffer} during shutdown. * * @param flushOnShutdown boolean value to configure flushOnShutdown. */ public void setFlushOnShutdown(boolean flushOnShutdown) { this.flushOnShutdown = flushOnShutdown; } /** * Sets the flushOnShutdown option. The default value is false which indicates flushOnShutdown is disabled. * * Enabling this option will flush the pending requests in the {@link SendQueueBuffer} during shutdown. * * @param flushOnShutdown boolean value to configure flushOnShutdown. * @return This object for method chaining. */ public QueueBufferConfig withFlushOnShutdown(boolean flushOnShutdown) { setFlushOnShutdown(flushOnShutdown); return this; } /** * this method checks the config for validity. If the config is deemed to be invalid, an * informative exception is thrown. * * @throws AmazonClientException * with a message explaining why the config was invalid */ void validate() { if (visibilityTimeoutSeconds == 0) { throw new AmazonClientException("Visibility timeout value may not be equal to zero "); } } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy