com.amazonaws.services.kinesis.AmazonKinesisAsync Maven / Gradle / Ivy
Show all versions of aws-java-sdk-kinesis Show documentation
/*
* Copyright 2014-2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License"). You may not use this file except in compliance with
* the License. A copy of the License is located at
*
* http://aws.amazon.com/apache2.0
*
* or in the "license" file accompanying this file. This file is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
* CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions
* and limitations under the License.
*/
package com.amazonaws.services.kinesis;
import javax.annotation.Generated;
import com.amazonaws.services.kinesis.model.*;
/**
* Interface for accessing Kinesis asynchronously. Each asynchronous method will return a Java Future object
* representing the asynchronous operation; overloads which accept an {@code AsyncHandler} can be used to receive
* notification when an asynchronous operation completes.
*
* Note: Do not directly implement this interface, new methods are added to it regularly. Extend from
* {@link com.amazonaws.services.kinesis.AbstractAmazonKinesisAsync} instead.
*
*
* Amazon Kinesis Data Streams Service API Reference
*
* Amazon Kinesis Data Streams is a managed service that scales elastically for real-time processing of streaming big
* data.
*
*/
@Generated("com.amazonaws:aws-java-sdk-code-generator")
public interface AmazonKinesisAsync extends AmazonKinesis {
/**
*
* Adds or updates tags for the specified Kinesis data stream. Each time you invoke this operation, you can specify
* up to 10 tags. If you want to add more than 10 tags to your stream, you can invoke this operation multiple times.
* In total, each stream can have up to 50 tags.
*
*
* If tags have already been assigned to the stream, AddTagsToStream
overwrites any existing tags that
* correspond to the specified tag keys.
*
*
* AddTagsToStream has a limit of five transactions per second per account.
*
*
* @param addTagsToStreamRequest
* Represents the input for AddTagsToStream
.
* @return A Java Future containing the result of the AddTagsToStream operation returned by the service.
* @sample AmazonKinesisAsync.AddTagsToStream
* @see AWS API
* Documentation
*/
java.util.concurrent.Future addTagsToStreamAsync(AddTagsToStreamRequest addTagsToStreamRequest);
/**
*
* Adds or updates tags for the specified Kinesis data stream. Each time you invoke this operation, you can specify
* up to 10 tags. If you want to add more than 10 tags to your stream, you can invoke this operation multiple times.
* In total, each stream can have up to 50 tags.
*
*
* If tags have already been assigned to the stream, AddTagsToStream
overwrites any existing tags that
* correspond to the specified tag keys.
*
*
* AddTagsToStream has a limit of five transactions per second per account.
*
*
* @param addTagsToStreamRequest
* Represents the input for AddTagsToStream
.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the AddTagsToStream operation returned by the service.
* @sample AmazonKinesisAsyncHandler.AddTagsToStream
* @see AWS API
* Documentation
*/
java.util.concurrent.Future addTagsToStreamAsync(AddTagsToStreamRequest addTagsToStreamRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Creates a Kinesis data stream. A stream captures and transports data records that are continuously emitted from
* different data sources or producers. Scale-out within a stream is explicitly supported by means of shards,
* which are uniquely identified groups of data records in a stream.
*
*
* You specify and control the number of shards that a stream is composed of. Each shard can support reads up to
* five transactions per second, up to a maximum data read total of 2 MB per second. Each shard can support writes
* up to 1,000 records per second, up to a maximum data write total of 1 MB per second. If the amount of data input
* increases or decreases, you can add or remove shards.
*
*
* The stream name identifies the stream. The name is scoped to the AWS account used by the application. It is also
* scoped by AWS Region. That is, two streams in two different accounts can have the same name, and two streams in
* the same account, but in two different Regions, can have the same name.
*
*
* CreateStream
is an asynchronous operation. Upon receiving a CreateStream
request,
* Kinesis Data Streams immediately returns and sets the stream status to CREATING
. After the stream is
* created, Kinesis Data Streams sets the stream status to ACTIVE
. You should perform read and write
* operations only on an ACTIVE
stream.
*
*
* You receive a LimitExceededException
when making a CreateStream
request when you try to
* do one of the following:
*
*
* -
*
* Have more than five streams in the CREATING
state at any point in time.
*
*
* -
*
* Create more shards than are authorized for your account.
*
*
*
*
* For the default shard limit for an AWS account, see Amazon Kinesis Data Streams
* Limits in the Amazon Kinesis Data Streams Developer Guide. To increase this limit, contact AWS Support.
*
*
* You can use DescribeStream
to check the stream status, which is returned in
* StreamStatus
.
*
*
* CreateStream has a limit of five transactions per second per account.
*
*
* @param createStreamRequest
* Represents the input for CreateStream
.
* @return A Java Future containing the result of the CreateStream operation returned by the service.
* @sample AmazonKinesisAsync.CreateStream
* @see AWS API
* Documentation
*/
java.util.concurrent.Future createStreamAsync(CreateStreamRequest createStreamRequest);
/**
*
* Creates a Kinesis data stream. A stream captures and transports data records that are continuously emitted from
* different data sources or producers. Scale-out within a stream is explicitly supported by means of shards,
* which are uniquely identified groups of data records in a stream.
*
*
* You specify and control the number of shards that a stream is composed of. Each shard can support reads up to
* five transactions per second, up to a maximum data read total of 2 MB per second. Each shard can support writes
* up to 1,000 records per second, up to a maximum data write total of 1 MB per second. If the amount of data input
* increases or decreases, you can add or remove shards.
*
*
* The stream name identifies the stream. The name is scoped to the AWS account used by the application. It is also
* scoped by AWS Region. That is, two streams in two different accounts can have the same name, and two streams in
* the same account, but in two different Regions, can have the same name.
*
*
* CreateStream
is an asynchronous operation. Upon receiving a CreateStream
request,
* Kinesis Data Streams immediately returns and sets the stream status to CREATING
. After the stream is
* created, Kinesis Data Streams sets the stream status to ACTIVE
. You should perform read and write
* operations only on an ACTIVE
stream.
*
*
* You receive a LimitExceededException
when making a CreateStream
request when you try to
* do one of the following:
*
*
* -
*
* Have more than five streams in the CREATING
state at any point in time.
*
*
* -
*
* Create more shards than are authorized for your account.
*
*
*
*
* For the default shard limit for an AWS account, see Amazon Kinesis Data Streams
* Limits in the Amazon Kinesis Data Streams Developer Guide. To increase this limit, contact AWS Support.
*
*
* You can use DescribeStream
to check the stream status, which is returned in
* StreamStatus
.
*
*
* CreateStream has a limit of five transactions per second per account.
*
*
* @param createStreamRequest
* Represents the input for CreateStream
.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the CreateStream operation returned by the service.
* @sample AmazonKinesisAsyncHandler.CreateStream
* @see AWS API
* Documentation
*/
java.util.concurrent.Future createStreamAsync(CreateStreamRequest createStreamRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the CreateStream operation.
*
* @see #createStreamAsync(CreateStreamRequest)
*/
java.util.concurrent.Future createStreamAsync(String streamName, Integer shardCount);
/**
* Simplified method form for invoking the CreateStream operation with an AsyncHandler.
*
* @see #createStreamAsync(CreateStreamRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future createStreamAsync(String streamName, Integer shardCount,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Decreases the Kinesis data stream's retention period, which is the length of time data records are accessible
* after they are added to the stream. The minimum value of a stream's retention period is 24 hours.
*
*
* This operation may result in lost data. For example, if the stream's retention period is 48 hours and is
* decreased to 24 hours, any data already in the stream that is older than 24 hours is inaccessible.
*
*
* @param decreaseStreamRetentionPeriodRequest
* Represents the input for DecreaseStreamRetentionPeriod.
* @return A Java Future containing the result of the DecreaseStreamRetentionPeriod operation returned by the
* service.
* @sample AmazonKinesisAsync.DecreaseStreamRetentionPeriod
* @see AWS API Documentation
*/
java.util.concurrent.Future decreaseStreamRetentionPeriodAsync(
DecreaseStreamRetentionPeriodRequest decreaseStreamRetentionPeriodRequest);
/**
*
* Decreases the Kinesis data stream's retention period, which is the length of time data records are accessible
* after they are added to the stream. The minimum value of a stream's retention period is 24 hours.
*
*
* This operation may result in lost data. For example, if the stream's retention period is 48 hours and is
* decreased to 24 hours, any data already in the stream that is older than 24 hours is inaccessible.
*
*
* @param decreaseStreamRetentionPeriodRequest
* Represents the input for DecreaseStreamRetentionPeriod.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the DecreaseStreamRetentionPeriod operation returned by the
* service.
* @sample AmazonKinesisAsyncHandler.DecreaseStreamRetentionPeriod
* @see AWS API Documentation
*/
java.util.concurrent.Future decreaseStreamRetentionPeriodAsync(
DecreaseStreamRetentionPeriodRequest decreaseStreamRetentionPeriodRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Deletes a Kinesis data stream and all its shards and data. You must shut down any applications that are operating
* on the stream before you delete the stream. If an application attempts to operate on a deleted stream, it
* receives the exception ResourceNotFoundException
.
*
*
* If the stream is in the ACTIVE
state, you can delete it. After a DeleteStream
request,
* the specified stream is in the DELETING
state until Kinesis Data Streams completes the deletion.
*
*
* Note: Kinesis Data Streams might continue to accept data read and write operations, such as
* PutRecord, PutRecords, and GetRecords, on a stream in the DELETING
state until
* the stream deletion is complete.
*
*
* When you delete a stream, any shards in that stream are also deleted, and any tags are dissociated from the
* stream.
*
*
* You can use the DescribeStream operation to check the state of the stream, which is returned in
* StreamStatus
.
*
*
* DeleteStream has a limit of five transactions per second per account.
*
*
* @param deleteStreamRequest
* Represents the input for DeleteStream.
* @return A Java Future containing the result of the DeleteStream operation returned by the service.
* @sample AmazonKinesisAsync.DeleteStream
* @see AWS API
* Documentation
*/
java.util.concurrent.Future deleteStreamAsync(DeleteStreamRequest deleteStreamRequest);
/**
*
* Deletes a Kinesis data stream and all its shards and data. You must shut down any applications that are operating
* on the stream before you delete the stream. If an application attempts to operate on a deleted stream, it
* receives the exception ResourceNotFoundException
.
*
*
* If the stream is in the ACTIVE
state, you can delete it. After a DeleteStream
request,
* the specified stream is in the DELETING
state until Kinesis Data Streams completes the deletion.
*
*
* Note: Kinesis Data Streams might continue to accept data read and write operations, such as
* PutRecord, PutRecords, and GetRecords, on a stream in the DELETING
state until
* the stream deletion is complete.
*
*
* When you delete a stream, any shards in that stream are also deleted, and any tags are dissociated from the
* stream.
*
*
* You can use the DescribeStream operation to check the state of the stream, which is returned in
* StreamStatus
.
*
*
* DeleteStream has a limit of five transactions per second per account.
*
*
* @param deleteStreamRequest
* Represents the input for DeleteStream.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the DeleteStream operation returned by the service.
* @sample AmazonKinesisAsyncHandler.DeleteStream
* @see AWS API
* Documentation
*/
java.util.concurrent.Future deleteStreamAsync(DeleteStreamRequest deleteStreamRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the DeleteStream operation.
*
* @see #deleteStreamAsync(DeleteStreamRequest)
*/
java.util.concurrent.Future deleteStreamAsync(String streamName);
/**
* Simplified method form for invoking the DeleteStream operation with an AsyncHandler.
*
* @see #deleteStreamAsync(DeleteStreamRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future deleteStreamAsync(String streamName,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* To deregister a consumer, provide its ARN. Alternatively, you can provide the ARN of the data stream and the name
* you gave the consumer when you registered it. You may also provide all three parameters, as long as they don't
* conflict with each other. If you don't know the name or ARN of the consumer that you want to deregister, you can
* use the ListStreamConsumers operation to get a list of the descriptions of all the consumers that are
* currently registered with a given data stream. The description of a consumer contains its name and ARN.
*
*
* This operation has a limit of five transactions per second per account.
*
*
* @param deregisterStreamConsumerRequest
* @return A Java Future containing the result of the DeregisterStreamConsumer operation returned by the service.
* @sample AmazonKinesisAsync.DeregisterStreamConsumer
* @see AWS API Documentation
*/
java.util.concurrent.Future deregisterStreamConsumerAsync(DeregisterStreamConsumerRequest deregisterStreamConsumerRequest);
/**
*
* To deregister a consumer, provide its ARN. Alternatively, you can provide the ARN of the data stream and the name
* you gave the consumer when you registered it. You may also provide all three parameters, as long as they don't
* conflict with each other. If you don't know the name or ARN of the consumer that you want to deregister, you can
* use the ListStreamConsumers operation to get a list of the descriptions of all the consumers that are
* currently registered with a given data stream. The description of a consumer contains its name and ARN.
*
*
* This operation has a limit of five transactions per second per account.
*
*
* @param deregisterStreamConsumerRequest
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the DeregisterStreamConsumer operation returned by the service.
* @sample AmazonKinesisAsyncHandler.DeregisterStreamConsumer
* @see AWS API Documentation
*/
java.util.concurrent.Future deregisterStreamConsumerAsync(DeregisterStreamConsumerRequest deregisterStreamConsumerRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Describes the shard limits and usage for the account.
*
*
* If you update your account limits, the old limits might be returned for a few minutes.
*
*
* This operation has a limit of one transaction per second per account.
*
*
* @param describeLimitsRequest
* @return A Java Future containing the result of the DescribeLimits operation returned by the service.
* @sample AmazonKinesisAsync.DescribeLimits
* @see AWS API
* Documentation
*/
java.util.concurrent.Future describeLimitsAsync(DescribeLimitsRequest describeLimitsRequest);
/**
*
* Describes the shard limits and usage for the account.
*
*
* If you update your account limits, the old limits might be returned for a few minutes.
*
*
* This operation has a limit of one transaction per second per account.
*
*
* @param describeLimitsRequest
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the DescribeLimits operation returned by the service.
* @sample AmazonKinesisAsyncHandler.DescribeLimits
* @see AWS API
* Documentation
*/
java.util.concurrent.Future describeLimitsAsync(DescribeLimitsRequest describeLimitsRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Describes the specified Kinesis data stream.
*
*
* The information returned includes the stream name, Amazon Resource Name (ARN), creation time, enhanced metric
* configuration, and shard map. The shard map is an array of shard objects. For each shard object, there is the
* hash key and sequence number ranges that the shard spans, and the IDs of any earlier shards that played in a role
* in creating the shard. Every record ingested in the stream is identified by a sequence number, which is assigned
* when the record is put into the stream.
*
*
* You can limit the number of shards returned by each call. For more information, see Retrieving
* Shards from a Stream in the Amazon Kinesis Data Streams Developer Guide.
*
*
* There are no guarantees about the chronological order shards returned. To process shards in chronological order,
* use the ID of the parent shard to track the lineage to the oldest shard.
*
*
* This operation has a limit of 10 transactions per second per account.
*
*
* @param describeStreamRequest
* Represents the input for DescribeStream
.
* @return A Java Future containing the result of the DescribeStream operation returned by the service.
* @sample AmazonKinesisAsync.DescribeStream
* @see AWS API
* Documentation
*/
java.util.concurrent.Future describeStreamAsync(DescribeStreamRequest describeStreamRequest);
/**
*
* Describes the specified Kinesis data stream.
*
*
* The information returned includes the stream name, Amazon Resource Name (ARN), creation time, enhanced metric
* configuration, and shard map. The shard map is an array of shard objects. For each shard object, there is the
* hash key and sequence number ranges that the shard spans, and the IDs of any earlier shards that played in a role
* in creating the shard. Every record ingested in the stream is identified by a sequence number, which is assigned
* when the record is put into the stream.
*
*
* You can limit the number of shards returned by each call. For more information, see Retrieving
* Shards from a Stream in the Amazon Kinesis Data Streams Developer Guide.
*
*
* There are no guarantees about the chronological order shards returned. To process shards in chronological order,
* use the ID of the parent shard to track the lineage to the oldest shard.
*
*
* This operation has a limit of 10 transactions per second per account.
*
*
* @param describeStreamRequest
* Represents the input for DescribeStream
.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the DescribeStream operation returned by the service.
* @sample AmazonKinesisAsyncHandler.DescribeStream
* @see AWS API
* Documentation
*/
java.util.concurrent.Future describeStreamAsync(DescribeStreamRequest describeStreamRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the DescribeStream operation.
*
* @see #describeStreamAsync(DescribeStreamRequest)
*/
java.util.concurrent.Future describeStreamAsync(String streamName);
/**
* Simplified method form for invoking the DescribeStream operation with an AsyncHandler.
*
* @see #describeStreamAsync(DescribeStreamRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future describeStreamAsync(String streamName,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the DescribeStream operation.
*
* @see #describeStreamAsync(DescribeStreamRequest)
*/
java.util.concurrent.Future describeStreamAsync(String streamName, String exclusiveStartShardId);
/**
* Simplified method form for invoking the DescribeStream operation with an AsyncHandler.
*
* @see #describeStreamAsync(DescribeStreamRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future describeStreamAsync(String streamName, String exclusiveStartShardId,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the DescribeStream operation.
*
* @see #describeStreamAsync(DescribeStreamRequest)
*/
java.util.concurrent.Future describeStreamAsync(String streamName, Integer limit, String exclusiveStartShardId);
/**
* Simplified method form for invoking the DescribeStream operation with an AsyncHandler.
*
* @see #describeStreamAsync(DescribeStreamRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future describeStreamAsync(String streamName, Integer limit, String exclusiveStartShardId,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* To get the description of a registered consumer, provide the ARN of the consumer. Alternatively, you can provide
* the ARN of the data stream and the name you gave the consumer when you registered it. You may also provide all
* three parameters, as long as they don't conflict with each other. If you don't know the name or ARN of the
* consumer that you want to describe, you can use the ListStreamConsumers operation to get a list of the
* descriptions of all the consumers that are currently registered with a given data stream.
*
*
* This operation has a limit of 20 transactions per second per account.
*
*
* @param describeStreamConsumerRequest
* @return A Java Future containing the result of the DescribeStreamConsumer operation returned by the service.
* @sample AmazonKinesisAsync.DescribeStreamConsumer
* @see AWS
* API Documentation
*/
java.util.concurrent.Future describeStreamConsumerAsync(DescribeStreamConsumerRequest describeStreamConsumerRequest);
/**
*
* To get the description of a registered consumer, provide the ARN of the consumer. Alternatively, you can provide
* the ARN of the data stream and the name you gave the consumer when you registered it. You may also provide all
* three parameters, as long as they don't conflict with each other. If you don't know the name or ARN of the
* consumer that you want to describe, you can use the ListStreamConsumers operation to get a list of the
* descriptions of all the consumers that are currently registered with a given data stream.
*
*
* This operation has a limit of 20 transactions per second per account.
*
*
* @param describeStreamConsumerRequest
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the DescribeStreamConsumer operation returned by the service.
* @sample AmazonKinesisAsyncHandler.DescribeStreamConsumer
* @see AWS
* API Documentation
*/
java.util.concurrent.Future describeStreamConsumerAsync(DescribeStreamConsumerRequest describeStreamConsumerRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Provides a summarized description of the specified Kinesis data stream without the shard list.
*
*
* The information returned includes the stream name, Amazon Resource Name (ARN), status, record retention period,
* approximate creation time, monitoring, encryption details, and open shard count.
*
*
* @param describeStreamSummaryRequest
* @return A Java Future containing the result of the DescribeStreamSummary operation returned by the service.
* @sample AmazonKinesisAsync.DescribeStreamSummary
* @see AWS
* API Documentation
*/
java.util.concurrent.Future describeStreamSummaryAsync(DescribeStreamSummaryRequest describeStreamSummaryRequest);
/**
*
* Provides a summarized description of the specified Kinesis data stream without the shard list.
*
*
* The information returned includes the stream name, Amazon Resource Name (ARN), status, record retention period,
* approximate creation time, monitoring, encryption details, and open shard count.
*
*
* @param describeStreamSummaryRequest
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the DescribeStreamSummary operation returned by the service.
* @sample AmazonKinesisAsyncHandler.DescribeStreamSummary
* @see AWS
* API Documentation
*/
java.util.concurrent.Future describeStreamSummaryAsync(DescribeStreamSummaryRequest describeStreamSummaryRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Disables enhanced monitoring.
*
*
* @param disableEnhancedMonitoringRequest
* Represents the input for DisableEnhancedMonitoring.
* @return A Java Future containing the result of the DisableEnhancedMonitoring operation returned by the service.
* @sample AmazonKinesisAsync.DisableEnhancedMonitoring
* @see AWS API Documentation
*/
java.util.concurrent.Future disableEnhancedMonitoringAsync(
DisableEnhancedMonitoringRequest disableEnhancedMonitoringRequest);
/**
*
* Disables enhanced monitoring.
*
*
* @param disableEnhancedMonitoringRequest
* Represents the input for DisableEnhancedMonitoring.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the DisableEnhancedMonitoring operation returned by the service.
* @sample AmazonKinesisAsyncHandler.DisableEnhancedMonitoring
* @see AWS API Documentation
*/
java.util.concurrent.Future disableEnhancedMonitoringAsync(
DisableEnhancedMonitoringRequest disableEnhancedMonitoringRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Enables enhanced Kinesis data stream monitoring for shard-level metrics.
*
*
* @param enableEnhancedMonitoringRequest
* Represents the input for EnableEnhancedMonitoring.
* @return A Java Future containing the result of the EnableEnhancedMonitoring operation returned by the service.
* @sample AmazonKinesisAsync.EnableEnhancedMonitoring
* @see AWS API Documentation
*/
java.util.concurrent.Future enableEnhancedMonitoringAsync(EnableEnhancedMonitoringRequest enableEnhancedMonitoringRequest);
/**
*
* Enables enhanced Kinesis data stream monitoring for shard-level metrics.
*
*
* @param enableEnhancedMonitoringRequest
* Represents the input for EnableEnhancedMonitoring.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the EnableEnhancedMonitoring operation returned by the service.
* @sample AmazonKinesisAsyncHandler.EnableEnhancedMonitoring
* @see AWS API Documentation
*/
java.util.concurrent.Future enableEnhancedMonitoringAsync(EnableEnhancedMonitoringRequest enableEnhancedMonitoringRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Gets data records from a Kinesis data stream's shard.
*
*
* Specify a shard iterator using the ShardIterator
parameter. The shard iterator specifies the
* position in the shard from which you want to start reading data records sequentially. If there are no records
* available in the portion of the shard that the iterator points to, GetRecords returns an empty list. It
* might take multiple calls to get to a portion of the shard that contains records.
*
*
* You can scale by provisioning multiple shards per stream while considering service limits (for more information,
* see Amazon Kinesis Data
* Streams Limits in the Amazon Kinesis Data Streams Developer Guide). Your application should have one
* thread per shard, each reading continuously from its stream. To read from a stream continually, call
* GetRecords in a loop. Use GetShardIterator to get the shard iterator to specify in the first
* GetRecords call. GetRecords returns a new shard iterator in NextShardIterator
. Specify
* the shard iterator returned in NextShardIterator
in subsequent calls to GetRecords. If the
* shard has been closed, the shard iterator can't return more data and GetRecords returns null
* in NextShardIterator
. You can terminate the loop when the shard is closed, or when the shard
* iterator reaches the record with the sequence number or other attribute that marks it as the last record to
* process.
*
*
* Each data record can be up to 1 MiB in size, and each shard can read up to 2 MiB per second. You can ensure that
* your calls don't exceed the maximum supported size or throughput by using the Limit
parameter to
* specify the maximum number of records that GetRecords can return. Consider your average record size when
* determining this limit. The maximum number of records that can be returned per call is 10,000.
*
*
* The size of the data returned by GetRecords varies depending on the utilization of the shard. The maximum
* size of data that GetRecords can return is 10 MiB. If a call returns this amount of data, subsequent calls
* made within the next 5 seconds throw ProvisionedThroughputExceededException
. If there is
* insufficient provisioned throughput on the stream, subsequent calls made within the next 1 second throw
* ProvisionedThroughputExceededException
. GetRecords doesn't return any data when it throws an
* exception. For this reason, we recommend that you wait 1 second between calls to GetRecords. However, it's
* possible that the application will get exceptions for longer than 1 second.
*
*
* To detect whether the application is falling behind in processing, you can use the
* MillisBehindLatest
response attribute. You can also monitor the stream using CloudWatch metrics and
* other mechanisms (see Monitoring in
* the Amazon Kinesis Data Streams Developer Guide).
*
*
* Each Amazon Kinesis record includes a value, ApproximateArrivalTimestamp
, that is set when a stream
* successfully receives and stores a record. This is commonly referred to as a server-side time stamp, whereas a
* client-side time stamp is set when a data producer creates or sends the record to a stream (a data producer is
* any data source putting data records into a stream, for example with PutRecords). The time stamp has
* millisecond precision. There are no guarantees about the time stamp accuracy, or that the time stamp is always
* increasing. For example, records in a shard or across a stream might have time stamps that are out of order.
*
*
* This operation has a limit of five transactions per second per account.
*
*
* @param getRecordsRequest
* Represents the input for GetRecords.
* @return A Java Future containing the result of the GetRecords operation returned by the service.
* @sample AmazonKinesisAsync.GetRecords
* @see AWS API
* Documentation
*/
java.util.concurrent.Future getRecordsAsync(GetRecordsRequest getRecordsRequest);
/**
*
* Gets data records from a Kinesis data stream's shard.
*
*
* Specify a shard iterator using the ShardIterator
parameter. The shard iterator specifies the
* position in the shard from which you want to start reading data records sequentially. If there are no records
* available in the portion of the shard that the iterator points to, GetRecords returns an empty list. It
* might take multiple calls to get to a portion of the shard that contains records.
*
*
* You can scale by provisioning multiple shards per stream while considering service limits (for more information,
* see Amazon Kinesis Data
* Streams Limits in the Amazon Kinesis Data Streams Developer Guide). Your application should have one
* thread per shard, each reading continuously from its stream. To read from a stream continually, call
* GetRecords in a loop. Use GetShardIterator to get the shard iterator to specify in the first
* GetRecords call. GetRecords returns a new shard iterator in NextShardIterator
. Specify
* the shard iterator returned in NextShardIterator
in subsequent calls to GetRecords. If the
* shard has been closed, the shard iterator can't return more data and GetRecords returns null
* in NextShardIterator
. You can terminate the loop when the shard is closed, or when the shard
* iterator reaches the record with the sequence number or other attribute that marks it as the last record to
* process.
*
*
* Each data record can be up to 1 MiB in size, and each shard can read up to 2 MiB per second. You can ensure that
* your calls don't exceed the maximum supported size or throughput by using the Limit
parameter to
* specify the maximum number of records that GetRecords can return. Consider your average record size when
* determining this limit. The maximum number of records that can be returned per call is 10,000.
*
*
* The size of the data returned by GetRecords varies depending on the utilization of the shard. The maximum
* size of data that GetRecords can return is 10 MiB. If a call returns this amount of data, subsequent calls
* made within the next 5 seconds throw ProvisionedThroughputExceededException
. If there is
* insufficient provisioned throughput on the stream, subsequent calls made within the next 1 second throw
* ProvisionedThroughputExceededException
. GetRecords doesn't return any data when it throws an
* exception. For this reason, we recommend that you wait 1 second between calls to GetRecords. However, it's
* possible that the application will get exceptions for longer than 1 second.
*
*
* To detect whether the application is falling behind in processing, you can use the
* MillisBehindLatest
response attribute. You can also monitor the stream using CloudWatch metrics and
* other mechanisms (see Monitoring in
* the Amazon Kinesis Data Streams Developer Guide).
*
*
* Each Amazon Kinesis record includes a value, ApproximateArrivalTimestamp
, that is set when a stream
* successfully receives and stores a record. This is commonly referred to as a server-side time stamp, whereas a
* client-side time stamp is set when a data producer creates or sends the record to a stream (a data producer is
* any data source putting data records into a stream, for example with PutRecords). The time stamp has
* millisecond precision. There are no guarantees about the time stamp accuracy, or that the time stamp is always
* increasing. For example, records in a shard or across a stream might have time stamps that are out of order.
*
*
* This operation has a limit of five transactions per second per account.
*
*
* @param getRecordsRequest
* Represents the input for GetRecords.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the GetRecords operation returned by the service.
* @sample AmazonKinesisAsyncHandler.GetRecords
* @see AWS API
* Documentation
*/
java.util.concurrent.Future getRecordsAsync(GetRecordsRequest getRecordsRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Gets an Amazon Kinesis shard iterator. A shard iterator expires 5 minutes after it is returned to the requester.
*
*
* A shard iterator specifies the shard position from which to start reading data records sequentially. The position
* is specified using the sequence number of a data record in a shard. A sequence number is the identifier
* associated with every record ingested in the stream, and is assigned when a record is put into the stream. Each
* stream has one or more shards.
*
*
* You must specify the shard iterator type. For example, you can set the ShardIteratorType
parameter
* to read exactly from the position denoted by a specific sequence number by using the
* AT_SEQUENCE_NUMBER
shard iterator type. Alternatively, the parameter can read right after the
* sequence number by using the AFTER_SEQUENCE_NUMBER
shard iterator type, using sequence numbers
* returned by earlier calls to PutRecord, PutRecords, GetRecords, or DescribeStream. In
* the request, you can specify the shard iterator type AT_TIMESTAMP
to read records from an arbitrary
* point in time, TRIM_HORIZON
to cause ShardIterator
to point to the last untrimmed
* record in the shard in the system (the oldest data record in the shard), or LATEST
so that you
* always read the most recent data in the shard.
*
*
* When you read repeatedly from a stream, use a GetShardIterator request to get the first shard iterator for
* use in your first GetRecords request and for subsequent reads use the shard iterator returned by the
* GetRecords request in NextShardIterator
. A new shard iterator is returned by every
* GetRecords request in NextShardIterator
, which you use in the ShardIterator
* parameter of the next GetRecords request.
*
*
* If a GetShardIterator request is made too often, you receive a
* ProvisionedThroughputExceededException
. For more information about throughput limits, see
* GetRecords, and Streams Limits in the
* Amazon Kinesis Data Streams Developer Guide.
*
*
* If the shard is closed, GetShardIterator returns a valid iterator for the last sequence number of the
* shard. A shard can be closed as a result of using SplitShard or MergeShards.
*
*
* GetShardIterator has a limit of five transactions per second per account per open shard.
*
*
* @param getShardIteratorRequest
* Represents the input for GetShardIterator
.
* @return A Java Future containing the result of the GetShardIterator operation returned by the service.
* @sample AmazonKinesisAsync.GetShardIterator
* @see AWS API
* Documentation
*/
java.util.concurrent.Future getShardIteratorAsync(GetShardIteratorRequest getShardIteratorRequest);
/**
*
* Gets an Amazon Kinesis shard iterator. A shard iterator expires 5 minutes after it is returned to the requester.
*
*
* A shard iterator specifies the shard position from which to start reading data records sequentially. The position
* is specified using the sequence number of a data record in a shard. A sequence number is the identifier
* associated with every record ingested in the stream, and is assigned when a record is put into the stream. Each
* stream has one or more shards.
*
*
* You must specify the shard iterator type. For example, you can set the ShardIteratorType
parameter
* to read exactly from the position denoted by a specific sequence number by using the
* AT_SEQUENCE_NUMBER
shard iterator type. Alternatively, the parameter can read right after the
* sequence number by using the AFTER_SEQUENCE_NUMBER
shard iterator type, using sequence numbers
* returned by earlier calls to PutRecord, PutRecords, GetRecords, or DescribeStream. In
* the request, you can specify the shard iterator type AT_TIMESTAMP
to read records from an arbitrary
* point in time, TRIM_HORIZON
to cause ShardIterator
to point to the last untrimmed
* record in the shard in the system (the oldest data record in the shard), or LATEST
so that you
* always read the most recent data in the shard.
*
*
* When you read repeatedly from a stream, use a GetShardIterator request to get the first shard iterator for
* use in your first GetRecords request and for subsequent reads use the shard iterator returned by the
* GetRecords request in NextShardIterator
. A new shard iterator is returned by every
* GetRecords request in NextShardIterator
, which you use in the ShardIterator
* parameter of the next GetRecords request.
*
*
* If a GetShardIterator request is made too often, you receive a
* ProvisionedThroughputExceededException
. For more information about throughput limits, see
* GetRecords, and Streams Limits in the
* Amazon Kinesis Data Streams Developer Guide.
*
*
* If the shard is closed, GetShardIterator returns a valid iterator for the last sequence number of the
* shard. A shard can be closed as a result of using SplitShard or MergeShards.
*
*
* GetShardIterator has a limit of five transactions per second per account per open shard.
*
*
* @param getShardIteratorRequest
* Represents the input for GetShardIterator
.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the GetShardIterator operation returned by the service.
* @sample AmazonKinesisAsyncHandler.GetShardIterator
* @see AWS API
* Documentation
*/
java.util.concurrent.Future getShardIteratorAsync(GetShardIteratorRequest getShardIteratorRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the GetShardIterator operation.
*
* @see #getShardIteratorAsync(GetShardIteratorRequest)
*/
java.util.concurrent.Future getShardIteratorAsync(String streamName, String shardId, String shardIteratorType);
/**
* Simplified method form for invoking the GetShardIterator operation with an AsyncHandler.
*
* @see #getShardIteratorAsync(GetShardIteratorRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future getShardIteratorAsync(String streamName, String shardId, String shardIteratorType,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the GetShardIterator operation.
*
* @see #getShardIteratorAsync(GetShardIteratorRequest)
*/
java.util.concurrent.Future getShardIteratorAsync(String streamName, String shardId, String shardIteratorType,
String startingSequenceNumber);
/**
* Simplified method form for invoking the GetShardIterator operation with an AsyncHandler.
*
* @see #getShardIteratorAsync(GetShardIteratorRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future getShardIteratorAsync(String streamName, String shardId, String shardIteratorType,
String startingSequenceNumber, com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Increases the Kinesis data stream's retention period, which is the length of time data records are accessible
* after they are added to the stream. The maximum value of a stream's retention period is 168 hours (7 days).
*
*
* If you choose a longer stream retention period, this operation increases the time period during which records
* that have not yet expired are accessible. However, it does not make previous, expired data (older than the
* stream's previous retention period) accessible after the operation has been called. For example, if a stream's
* retention period is set to 24 hours and is increased to 168 hours, any data that is older than 24 hours remains
* inaccessible to consumer applications.
*
*
* @param increaseStreamRetentionPeriodRequest
* Represents the input for IncreaseStreamRetentionPeriod.
* @return A Java Future containing the result of the IncreaseStreamRetentionPeriod operation returned by the
* service.
* @sample AmazonKinesisAsync.IncreaseStreamRetentionPeriod
* @see AWS API Documentation
*/
java.util.concurrent.Future increaseStreamRetentionPeriodAsync(
IncreaseStreamRetentionPeriodRequest increaseStreamRetentionPeriodRequest);
/**
*
* Increases the Kinesis data stream's retention period, which is the length of time data records are accessible
* after they are added to the stream. The maximum value of a stream's retention period is 168 hours (7 days).
*
*
* If you choose a longer stream retention period, this operation increases the time period during which records
* that have not yet expired are accessible. However, it does not make previous, expired data (older than the
* stream's previous retention period) accessible after the operation has been called. For example, if a stream's
* retention period is set to 24 hours and is increased to 168 hours, any data that is older than 24 hours remains
* inaccessible to consumer applications.
*
*
* @param increaseStreamRetentionPeriodRequest
* Represents the input for IncreaseStreamRetentionPeriod.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the IncreaseStreamRetentionPeriod operation returned by the
* service.
* @sample AmazonKinesisAsyncHandler.IncreaseStreamRetentionPeriod
* @see AWS API Documentation
*/
java.util.concurrent.Future increaseStreamRetentionPeriodAsync(
IncreaseStreamRetentionPeriodRequest increaseStreamRetentionPeriodRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Lists the shards in a stream and provides information about each shard. This operation has a limit of 100
* transactions per second per data stream.
*
*
*
* This API is a new operation that is used by the Amazon Kinesis Client Library (KCL). If you have a fine-grained
* IAM policy that only allows specific operations, you must update your policy to allow calls to this API. For more
* information, see Controlling
* Access to Amazon Kinesis Data Streams Resources Using IAM.
*
*
*
* @param listShardsRequest
* @return A Java Future containing the result of the ListShards operation returned by the service.
* @sample AmazonKinesisAsync.ListShards
* @see AWS API
* Documentation
*/
java.util.concurrent.Future listShardsAsync(ListShardsRequest listShardsRequest);
/**
*
* Lists the shards in a stream and provides information about each shard. This operation has a limit of 100
* transactions per second per data stream.
*
*
*
* This API is a new operation that is used by the Amazon Kinesis Client Library (KCL). If you have a fine-grained
* IAM policy that only allows specific operations, you must update your policy to allow calls to this API. For more
* information, see Controlling
* Access to Amazon Kinesis Data Streams Resources Using IAM.
*
*
*
* @param listShardsRequest
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the ListShards operation returned by the service.
* @sample AmazonKinesisAsyncHandler.ListShards
* @see AWS API
* Documentation
*/
java.util.concurrent.Future listShardsAsync(ListShardsRequest listShardsRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Lists the consumers registered to receive data from a stream using enhanced fan-out, and provides information
* about each consumer.
*
*
* This operation has a limit of 10 transactions per second per account.
*
*
* @param listStreamConsumersRequest
* @return A Java Future containing the result of the ListStreamConsumers operation returned by the service.
* @sample AmazonKinesisAsync.ListStreamConsumers
* @see AWS
* API Documentation
*/
java.util.concurrent.Future listStreamConsumersAsync(ListStreamConsumersRequest listStreamConsumersRequest);
/**
*
* Lists the consumers registered to receive data from a stream using enhanced fan-out, and provides information
* about each consumer.
*
*
* This operation has a limit of 10 transactions per second per account.
*
*
* @param listStreamConsumersRequest
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the ListStreamConsumers operation returned by the service.
* @sample AmazonKinesisAsyncHandler.ListStreamConsumers
* @see AWS
* API Documentation
*/
java.util.concurrent.Future listStreamConsumersAsync(ListStreamConsumersRequest listStreamConsumersRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Lists your Kinesis data streams.
*
*
* The number of streams may be too large to return from a single call to ListStreams
. You can limit
* the number of returned streams using the Limit
parameter. If you do not specify a value for the
* Limit
parameter, Kinesis Data Streams uses the default limit, which is currently 10.
*
*
* You can detect if there are more streams available to list by using the HasMoreStreams
flag from the
* returned output. If there are more streams available, you can request more streams by using the name of the last
* stream returned by the ListStreams
request in the ExclusiveStartStreamName
parameter in
* a subsequent request to ListStreams
. The group of stream names returned by the subsequent request is
* then added to the list. You can continue this process until all the stream names have been collected in the list.
*
*
* ListStreams has a limit of five transactions per second per account.
*
*
* @param listStreamsRequest
* Represents the input for ListStreams
.
* @return A Java Future containing the result of the ListStreams operation returned by the service.
* @sample AmazonKinesisAsync.ListStreams
* @see AWS API
* Documentation
*/
java.util.concurrent.Future listStreamsAsync(ListStreamsRequest listStreamsRequest);
/**
*
* Lists your Kinesis data streams.
*
*
* The number of streams may be too large to return from a single call to ListStreams
. You can limit
* the number of returned streams using the Limit
parameter. If you do not specify a value for the
* Limit
parameter, Kinesis Data Streams uses the default limit, which is currently 10.
*
*
* You can detect if there are more streams available to list by using the HasMoreStreams
flag from the
* returned output. If there are more streams available, you can request more streams by using the name of the last
* stream returned by the ListStreams
request in the ExclusiveStartStreamName
parameter in
* a subsequent request to ListStreams
. The group of stream names returned by the subsequent request is
* then added to the list. You can continue this process until all the stream names have been collected in the list.
*
*
* ListStreams has a limit of five transactions per second per account.
*
*
* @param listStreamsRequest
* Represents the input for ListStreams
.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the ListStreams operation returned by the service.
* @sample AmazonKinesisAsyncHandler.ListStreams
* @see AWS API
* Documentation
*/
java.util.concurrent.Future listStreamsAsync(ListStreamsRequest listStreamsRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the ListStreams operation.
*
* @see #listStreamsAsync(ListStreamsRequest)
*/
java.util.concurrent.Future listStreamsAsync();
/**
* Simplified method form for invoking the ListStreams operation with an AsyncHandler.
*
* @see #listStreamsAsync(ListStreamsRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future listStreamsAsync(com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the ListStreams operation.
*
* @see #listStreamsAsync(ListStreamsRequest)
*/
java.util.concurrent.Future listStreamsAsync(String exclusiveStartStreamName);
/**
* Simplified method form for invoking the ListStreams operation with an AsyncHandler.
*
* @see #listStreamsAsync(ListStreamsRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future listStreamsAsync(String exclusiveStartStreamName,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the ListStreams operation.
*
* @see #listStreamsAsync(ListStreamsRequest)
*/
java.util.concurrent.Future listStreamsAsync(Integer limit, String exclusiveStartStreamName);
/**
* Simplified method form for invoking the ListStreams operation with an AsyncHandler.
*
* @see #listStreamsAsync(ListStreamsRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future listStreamsAsync(Integer limit, String exclusiveStartStreamName,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Lists the tags for the specified Kinesis data stream. This operation has a limit of five transactions per second
* per account.
*
*
* @param listTagsForStreamRequest
* Represents the input for ListTagsForStream
.
* @return A Java Future containing the result of the ListTagsForStream operation returned by the service.
* @sample AmazonKinesisAsync.ListTagsForStream
* @see AWS API
* Documentation
*/
java.util.concurrent.Future listTagsForStreamAsync(ListTagsForStreamRequest listTagsForStreamRequest);
/**
*
* Lists the tags for the specified Kinesis data stream. This operation has a limit of five transactions per second
* per account.
*
*
* @param listTagsForStreamRequest
* Represents the input for ListTagsForStream
.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the ListTagsForStream operation returned by the service.
* @sample AmazonKinesisAsyncHandler.ListTagsForStream
* @see AWS API
* Documentation
*/
java.util.concurrent.Future listTagsForStreamAsync(ListTagsForStreamRequest listTagsForStreamRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Merges two adjacent shards in a Kinesis data stream and combines them into a single shard to reduce the stream's
* capacity to ingest and transport data. Two shards are considered adjacent if the union of the hash key ranges for
* the two shards form a contiguous set with no gaps. For example, if you have two shards, one with a hash key range
* of 276...381 and the other with a hash key range of 382...454, then you could merge these two shards into a
* single shard that would have a hash key range of 276...454. After the merge, the single child shard receives data
* for all hash key values covered by the two parent shards.
*
*
* MergeShards
is called when there is a need to reduce the overall capacity of a stream because of
* excess capacity that is not being used. You must specify the shard to be merged and the adjacent shard for a
* stream. For more information about merging shards, see Merge Two
* Shards in the Amazon Kinesis Data Streams Developer Guide.
*
*
* If the stream is in the ACTIVE
state, you can call MergeShards
. If a stream is in the
* CREATING
, UPDATING
, or DELETING
state, MergeShards
returns a
* ResourceInUseException
. If the specified stream does not exist, MergeShards
returns a
* ResourceNotFoundException
.
*
*
* You can use DescribeStream to check the state of the stream, which is returned in
* StreamStatus
.
*
*
* MergeShards
is an asynchronous operation. Upon receiving a MergeShards
request, Amazon
* Kinesis Data Streams immediately returns a response and sets the StreamStatus
to
* UPDATING
. After the operation is completed, Kinesis Data Streams sets the StreamStatus
* to ACTIVE
. Read and write operations continue to work while the stream is in the
* UPDATING
state.
*
*
* You use DescribeStream to determine the shard IDs that are specified in the MergeShards
* request.
*
*
* If you try to operate on too many streams in parallel using CreateStream, DeleteStream,
* MergeShards
, or SplitShard, you receive a LimitExceededException
.
*
*
* MergeShards
has a limit of five transactions per second per account.
*
*
* @param mergeShardsRequest
* Represents the input for MergeShards
.
* @return A Java Future containing the result of the MergeShards operation returned by the service.
* @sample AmazonKinesisAsync.MergeShards
* @see AWS API
* Documentation
*/
java.util.concurrent.Future mergeShardsAsync(MergeShardsRequest mergeShardsRequest);
/**
*
* Merges two adjacent shards in a Kinesis data stream and combines them into a single shard to reduce the stream's
* capacity to ingest and transport data. Two shards are considered adjacent if the union of the hash key ranges for
* the two shards form a contiguous set with no gaps. For example, if you have two shards, one with a hash key range
* of 276...381 and the other with a hash key range of 382...454, then you could merge these two shards into a
* single shard that would have a hash key range of 276...454. After the merge, the single child shard receives data
* for all hash key values covered by the two parent shards.
*
*
* MergeShards
is called when there is a need to reduce the overall capacity of a stream because of
* excess capacity that is not being used. You must specify the shard to be merged and the adjacent shard for a
* stream. For more information about merging shards, see Merge Two
* Shards in the Amazon Kinesis Data Streams Developer Guide.
*
*
* If the stream is in the ACTIVE
state, you can call MergeShards
. If a stream is in the
* CREATING
, UPDATING
, or DELETING
state, MergeShards
returns a
* ResourceInUseException
. If the specified stream does not exist, MergeShards
returns a
* ResourceNotFoundException
.
*
*
* You can use DescribeStream to check the state of the stream, which is returned in
* StreamStatus
.
*
*
* MergeShards
is an asynchronous operation. Upon receiving a MergeShards
request, Amazon
* Kinesis Data Streams immediately returns a response and sets the StreamStatus
to
* UPDATING
. After the operation is completed, Kinesis Data Streams sets the StreamStatus
* to ACTIVE
. Read and write operations continue to work while the stream is in the
* UPDATING
state.
*
*
* You use DescribeStream to determine the shard IDs that are specified in the MergeShards
* request.
*
*
* If you try to operate on too many streams in parallel using CreateStream, DeleteStream,
* MergeShards
, or SplitShard, you receive a LimitExceededException
.
*
*
* MergeShards
has a limit of five transactions per second per account.
*
*
* @param mergeShardsRequest
* Represents the input for MergeShards
.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the MergeShards operation returned by the service.
* @sample AmazonKinesisAsyncHandler.MergeShards
* @see AWS API
* Documentation
*/
java.util.concurrent.Future mergeShardsAsync(MergeShardsRequest mergeShardsRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the MergeShards operation.
*
* @see #mergeShardsAsync(MergeShardsRequest)
*/
java.util.concurrent.Future mergeShardsAsync(String streamName, String shardToMerge, String adjacentShardToMerge);
/**
* Simplified method form for invoking the MergeShards operation with an AsyncHandler.
*
* @see #mergeShardsAsync(MergeShardsRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future mergeShardsAsync(String streamName, String shardToMerge, String adjacentShardToMerge,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Writes a single data record into an Amazon Kinesis data stream. Call PutRecord
to send data into the
* stream for real-time ingestion and subsequent processing, one record at a time. Each shard can support writes up
* to 1,000 records per second, up to a maximum data write total of 1 MB per second.
*
*
* You must specify the name of the stream that captures, stores, and transports the data; a partition key; and the
* data blob itself.
*
*
* The data blob can be any type of data; for example, a segment from a log file, geographic/location data, website
* clickstream data, and so on.
*
*
* The partition key is used by Kinesis Data Streams to distribute data across shards. Kinesis Data Streams
* segregates the data records that belong to a stream into multiple shards, using the partition key associated with
* each data record to determine the shard to which a given data record belongs.
*
*
* Partition keys are Unicode strings, with a maximum length limit of 256 characters for each key. An MD5 hash
* function is used to map partition keys to 128-bit integer values and to map associated data records to shards
* using the hash key ranges of the shards. You can override hashing the partition key to determine the shard by
* explicitly specifying a hash value using the ExplicitHashKey
parameter. For more information, see Adding Data to a Stream in the Amazon Kinesis Data Streams Developer Guide.
*
*
* PutRecord
returns the shard ID of where the data record was placed and the sequence number that was
* assigned to the data record.
*
*
* Sequence numbers increase over time and are specific to a shard within a stream, not across all shards within a
* stream. To guarantee strictly increasing ordering, write serially to a shard and use the
* SequenceNumberForOrdering
parameter. For more information, see Adding Data to a Stream in the Amazon Kinesis Data Streams Developer Guide.
*
*
* If a PutRecord
request cannot be processed because of insufficient provisioned throughput on the
* shard involved in the request, PutRecord
throws ProvisionedThroughputExceededException
.
*
*
* By default, data records are accessible for 24 hours from the time that they are added to a stream. You can use
* IncreaseStreamRetentionPeriod or DecreaseStreamRetentionPeriod to modify this retention period.
*
*
* @param putRecordRequest
* Represents the input for PutRecord
.
* @return A Java Future containing the result of the PutRecord operation returned by the service.
* @sample AmazonKinesisAsync.PutRecord
* @see AWS API
* Documentation
*/
java.util.concurrent.Future putRecordAsync(PutRecordRequest putRecordRequest);
/**
*
* Writes a single data record into an Amazon Kinesis data stream. Call PutRecord
to send data into the
* stream for real-time ingestion and subsequent processing, one record at a time. Each shard can support writes up
* to 1,000 records per second, up to a maximum data write total of 1 MB per second.
*
*
* You must specify the name of the stream that captures, stores, and transports the data; a partition key; and the
* data blob itself.
*
*
* The data blob can be any type of data; for example, a segment from a log file, geographic/location data, website
* clickstream data, and so on.
*
*
* The partition key is used by Kinesis Data Streams to distribute data across shards. Kinesis Data Streams
* segregates the data records that belong to a stream into multiple shards, using the partition key associated with
* each data record to determine the shard to which a given data record belongs.
*
*
* Partition keys are Unicode strings, with a maximum length limit of 256 characters for each key. An MD5 hash
* function is used to map partition keys to 128-bit integer values and to map associated data records to shards
* using the hash key ranges of the shards. You can override hashing the partition key to determine the shard by
* explicitly specifying a hash value using the ExplicitHashKey
parameter. For more information, see Adding Data to a Stream in the Amazon Kinesis Data Streams Developer Guide.
*
*
* PutRecord
returns the shard ID of where the data record was placed and the sequence number that was
* assigned to the data record.
*
*
* Sequence numbers increase over time and are specific to a shard within a stream, not across all shards within a
* stream. To guarantee strictly increasing ordering, write serially to a shard and use the
* SequenceNumberForOrdering
parameter. For more information, see Adding Data to a Stream in the Amazon Kinesis Data Streams Developer Guide.
*
*
* If a PutRecord
request cannot be processed because of insufficient provisioned throughput on the
* shard involved in the request, PutRecord
throws ProvisionedThroughputExceededException
.
*
*
* By default, data records are accessible for 24 hours from the time that they are added to a stream. You can use
* IncreaseStreamRetentionPeriod or DecreaseStreamRetentionPeriod to modify this retention period.
*
*
* @param putRecordRequest
* Represents the input for PutRecord
.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the PutRecord operation returned by the service.
* @sample AmazonKinesisAsyncHandler.PutRecord
* @see AWS API
* Documentation
*/
java.util.concurrent.Future putRecordAsync(PutRecordRequest putRecordRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the PutRecord operation.
*
* @see #putRecordAsync(PutRecordRequest)
*/
java.util.concurrent.Future putRecordAsync(String streamName, java.nio.ByteBuffer data, String partitionKey);
/**
* Simplified method form for invoking the PutRecord operation with an AsyncHandler.
*
* @see #putRecordAsync(PutRecordRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future putRecordAsync(String streamName, java.nio.ByteBuffer data, String partitionKey,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the PutRecord operation.
*
* @see #putRecordAsync(PutRecordRequest)
*/
java.util.concurrent.Future putRecordAsync(String streamName, java.nio.ByteBuffer data, String partitionKey,
String sequenceNumberForOrdering);
/**
* Simplified method form for invoking the PutRecord operation with an AsyncHandler.
*
* @see #putRecordAsync(PutRecordRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future putRecordAsync(String streamName, java.nio.ByteBuffer data, String partitionKey,
String sequenceNumberForOrdering, com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Writes multiple data records into a Kinesis data stream in a single call (also referred to as a
* PutRecords
request). Use this operation to send data into the stream for data ingestion and
* processing.
*
*
* Each PutRecords
request can support up to 500 records. Each record in the request can be as large as
* 1 MB, up to a limit of 5 MB for the entire request, including partition keys. Each shard can support writes up to
* 1,000 records per second, up to a maximum data write total of 1 MB per second.
*
*
* You must specify the name of the stream that captures, stores, and transports the data; and an array of request
* Records
, with each record in the array requiring a partition key and data blob. The record size
* limit applies to the total size of the partition key and data blob.
*
*
* The data blob can be any type of data; for example, a segment from a log file, geographic/location data, website
* clickstream data, and so on.
*
*
* The partition key is used by Kinesis Data Streams as input to a hash function that maps the partition key and
* associated data to a specific shard. An MD5 hash function is used to map partition keys to 128-bit integer values
* and to map associated data records to shards. As a result of this hashing mechanism, all data records with the
* same partition key map to the same shard within the stream. For more information, see Adding Data to a Stream in the Amazon Kinesis Data Streams Developer Guide.
*
*
* Each record in the Records
array may include an optional parameter, ExplicitHashKey
,
* which overrides the partition key to shard mapping. This parameter allows a data producer to determine explicitly
* the shard where the record is stored. For more information, see Adding Multiple Records with PutRecords in the Amazon Kinesis Data Streams Developer Guide.
*
*
* The PutRecords
response includes an array of response Records
. Each record in the
* response array directly correlates with a record in the request array using natural ordering, from the top to the
* bottom of the request and response. The response Records
array always includes the same number of
* records as the request array.
*
*
* The response Records
array includes both successfully and unsuccessfully processed records. Kinesis
* Data Streams attempts to process all records in each PutRecords
request. A single record failure
* does not stop the processing of subsequent records.
*
*
* A successfully processed record includes ShardId
and SequenceNumber
values. The
* ShardId
parameter identifies the shard in the stream where the record is stored. The
* SequenceNumber
parameter is an identifier assigned to the put record, unique to all records in the
* stream.
*
*
* An unsuccessfully processed record includes ErrorCode
and ErrorMessage
values.
* ErrorCode
reflects the type of error and can be one of the following values:
* ProvisionedThroughputExceededException
or InternalFailure
. ErrorMessage
* provides more detailed information about the ProvisionedThroughputExceededException
exception
* including the account ID, stream name, and shard ID of the record that was throttled. For more information about
* partially successful responses, see Adding Multiple Records with PutRecords in the Amazon Kinesis Data Streams Developer Guide.
*
*
* By default, data records are accessible for 24 hours from the time that they are added to a stream. You can use
* IncreaseStreamRetentionPeriod or DecreaseStreamRetentionPeriod to modify this retention period.
*
*
* @param putRecordsRequest
* A PutRecords
request.
* @return A Java Future containing the result of the PutRecords operation returned by the service.
* @sample AmazonKinesisAsync.PutRecords
* @see AWS API
* Documentation
*/
java.util.concurrent.Future putRecordsAsync(PutRecordsRequest putRecordsRequest);
/**
*
* Writes multiple data records into a Kinesis data stream in a single call (also referred to as a
* PutRecords
request). Use this operation to send data into the stream for data ingestion and
* processing.
*
*
* Each PutRecords
request can support up to 500 records. Each record in the request can be as large as
* 1 MB, up to a limit of 5 MB for the entire request, including partition keys. Each shard can support writes up to
* 1,000 records per second, up to a maximum data write total of 1 MB per second.
*
*
* You must specify the name of the stream that captures, stores, and transports the data; and an array of request
* Records
, with each record in the array requiring a partition key and data blob. The record size
* limit applies to the total size of the partition key and data blob.
*
*
* The data blob can be any type of data; for example, a segment from a log file, geographic/location data, website
* clickstream data, and so on.
*
*
* The partition key is used by Kinesis Data Streams as input to a hash function that maps the partition key and
* associated data to a specific shard. An MD5 hash function is used to map partition keys to 128-bit integer values
* and to map associated data records to shards. As a result of this hashing mechanism, all data records with the
* same partition key map to the same shard within the stream. For more information, see Adding Data to a Stream in the Amazon Kinesis Data Streams Developer Guide.
*
*
* Each record in the Records
array may include an optional parameter, ExplicitHashKey
,
* which overrides the partition key to shard mapping. This parameter allows a data producer to determine explicitly
* the shard where the record is stored. For more information, see Adding Multiple Records with PutRecords in the Amazon Kinesis Data Streams Developer Guide.
*
*
* The PutRecords
response includes an array of response Records
. Each record in the
* response array directly correlates with a record in the request array using natural ordering, from the top to the
* bottom of the request and response. The response Records
array always includes the same number of
* records as the request array.
*
*
* The response Records
array includes both successfully and unsuccessfully processed records. Kinesis
* Data Streams attempts to process all records in each PutRecords
request. A single record failure
* does not stop the processing of subsequent records.
*
*
* A successfully processed record includes ShardId
and SequenceNumber
values. The
* ShardId
parameter identifies the shard in the stream where the record is stored. The
* SequenceNumber
parameter is an identifier assigned to the put record, unique to all records in the
* stream.
*
*
* An unsuccessfully processed record includes ErrorCode
and ErrorMessage
values.
* ErrorCode
reflects the type of error and can be one of the following values:
* ProvisionedThroughputExceededException
or InternalFailure
. ErrorMessage
* provides more detailed information about the ProvisionedThroughputExceededException
exception
* including the account ID, stream name, and shard ID of the record that was throttled. For more information about
* partially successful responses, see Adding Multiple Records with PutRecords in the Amazon Kinesis Data Streams Developer Guide.
*
*
* By default, data records are accessible for 24 hours from the time that they are added to a stream. You can use
* IncreaseStreamRetentionPeriod or DecreaseStreamRetentionPeriod to modify this retention period.
*
*
* @param putRecordsRequest
* A PutRecords
request.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the PutRecords operation returned by the service.
* @sample AmazonKinesisAsyncHandler.PutRecords
* @see AWS API
* Documentation
*/
java.util.concurrent.Future putRecordsAsync(PutRecordsRequest putRecordsRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Registers a consumer with a Kinesis data stream. When you use this operation, the consumer you register can read
* data from the stream at a rate of up to 2 MiB per second. This rate is unaffected by the total number of
* consumers that read from the same stream.
*
*
* You can register up to 5 consumers per stream. A given consumer can only be registered with one stream.
*
*
* This operation has a limit of five transactions per second per account.
*
*
* @param registerStreamConsumerRequest
* @return A Java Future containing the result of the RegisterStreamConsumer operation returned by the service.
* @sample AmazonKinesisAsync.RegisterStreamConsumer
* @see AWS
* API Documentation
*/
java.util.concurrent.Future registerStreamConsumerAsync(RegisterStreamConsumerRequest registerStreamConsumerRequest);
/**
*
* Registers a consumer with a Kinesis data stream. When you use this operation, the consumer you register can read
* data from the stream at a rate of up to 2 MiB per second. This rate is unaffected by the total number of
* consumers that read from the same stream.
*
*
* You can register up to 5 consumers per stream. A given consumer can only be registered with one stream.
*
*
* This operation has a limit of five transactions per second per account.
*
*
* @param registerStreamConsumerRequest
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the RegisterStreamConsumer operation returned by the service.
* @sample AmazonKinesisAsyncHandler.RegisterStreamConsumer
* @see AWS
* API Documentation
*/
java.util.concurrent.Future registerStreamConsumerAsync(RegisterStreamConsumerRequest registerStreamConsumerRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Removes tags from the specified Kinesis data stream. Removed tags are deleted and cannot be recovered after this
* operation successfully completes.
*
*
* If you specify a tag that does not exist, it is ignored.
*
*
* RemoveTagsFromStream has a limit of five transactions per second per account.
*
*
* @param removeTagsFromStreamRequest
* Represents the input for RemoveTagsFromStream
.
* @return A Java Future containing the result of the RemoveTagsFromStream operation returned by the service.
* @sample AmazonKinesisAsync.RemoveTagsFromStream
* @see AWS
* API Documentation
*/
java.util.concurrent.Future removeTagsFromStreamAsync(RemoveTagsFromStreamRequest removeTagsFromStreamRequest);
/**
*
* Removes tags from the specified Kinesis data stream. Removed tags are deleted and cannot be recovered after this
* operation successfully completes.
*
*
* If you specify a tag that does not exist, it is ignored.
*
*
* RemoveTagsFromStream has a limit of five transactions per second per account.
*
*
* @param removeTagsFromStreamRequest
* Represents the input for RemoveTagsFromStream
.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the RemoveTagsFromStream operation returned by the service.
* @sample AmazonKinesisAsyncHandler.RemoveTagsFromStream
* @see AWS
* API Documentation
*/
java.util.concurrent.Future removeTagsFromStreamAsync(RemoveTagsFromStreamRequest removeTagsFromStreamRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Splits a shard into two new shards in the Kinesis data stream, to increase the stream's capacity to ingest and
* transport data. SplitShard
is called when there is a need to increase the overall capacity of a
* stream because of an expected increase in the volume of data records being ingested.
*
*
* You can also use SplitShard
when a shard appears to be approaching its maximum utilization; for
* example, the producers sending data into the specific shard are suddenly sending more than previously
* anticipated. You can also call SplitShard
to increase stream capacity, so that more Kinesis Data
* Streams applications can simultaneously read data from the stream for real-time processing.
*
*
* You must specify the shard to be split and the new hash key, which is the position in the shard where the shard
* gets split in two. In many cases, the new hash key might be the average of the beginning and ending hash key, but
* it can be any hash key value in the range being mapped into the shard. For more information, see Split a
* Shard in the Amazon Kinesis Data Streams Developer Guide.
*
*
* You can use DescribeStream to determine the shard ID and hash key values for the ShardToSplit
* and NewStartingHashKey
parameters that are specified in the SplitShard
request.
*
*
* SplitShard
is an asynchronous operation. Upon receiving a SplitShard
request, Kinesis
* Data Streams immediately returns a response and sets the stream status to UPDATING
. After the
* operation is completed, Kinesis Data Streams sets the stream status to ACTIVE
. Read and write
* operations continue to work while the stream is in the UPDATING
state.
*
*
* You can use DescribeStream
to check the status of the stream, which is returned in
* StreamStatus
. If the stream is in the ACTIVE
state, you can call
* SplitShard
. If a stream is in CREATING
or UPDATING
or
* DELETING
states, DescribeStream
returns a ResourceInUseException
.
*
*
* If the specified stream does not exist, DescribeStream
returns a
* ResourceNotFoundException
. If you try to create more shards than are authorized for your account,
* you receive a LimitExceededException
.
*
*
* For the default shard limit for an AWS account, see Kinesis Data Streams
* Limits in the Amazon Kinesis Data Streams Developer Guide. To increase this limit, contact AWS Support.
*
*
* If you try to operate on too many streams simultaneously using CreateStream, DeleteStream,
* MergeShards, and/or SplitShard, you receive a LimitExceededException
.
*
*
* SplitShard
has a limit of five transactions per second per account.
*
*
* @param splitShardRequest
* Represents the input for SplitShard
.
* @return A Java Future containing the result of the SplitShard operation returned by the service.
* @sample AmazonKinesisAsync.SplitShard
* @see AWS API
* Documentation
*/
java.util.concurrent.Future splitShardAsync(SplitShardRequest splitShardRequest);
/**
*
* Splits a shard into two new shards in the Kinesis data stream, to increase the stream's capacity to ingest and
* transport data. SplitShard
is called when there is a need to increase the overall capacity of a
* stream because of an expected increase in the volume of data records being ingested.
*
*
* You can also use SplitShard
when a shard appears to be approaching its maximum utilization; for
* example, the producers sending data into the specific shard are suddenly sending more than previously
* anticipated. You can also call SplitShard
to increase stream capacity, so that more Kinesis Data
* Streams applications can simultaneously read data from the stream for real-time processing.
*
*
* You must specify the shard to be split and the new hash key, which is the position in the shard where the shard
* gets split in two. In many cases, the new hash key might be the average of the beginning and ending hash key, but
* it can be any hash key value in the range being mapped into the shard. For more information, see Split a
* Shard in the Amazon Kinesis Data Streams Developer Guide.
*
*
* You can use DescribeStream to determine the shard ID and hash key values for the ShardToSplit
* and NewStartingHashKey
parameters that are specified in the SplitShard
request.
*
*
* SplitShard
is an asynchronous operation. Upon receiving a SplitShard
request, Kinesis
* Data Streams immediately returns a response and sets the stream status to UPDATING
. After the
* operation is completed, Kinesis Data Streams sets the stream status to ACTIVE
. Read and write
* operations continue to work while the stream is in the UPDATING
state.
*
*
* You can use DescribeStream
to check the status of the stream, which is returned in
* StreamStatus
. If the stream is in the ACTIVE
state, you can call
* SplitShard
. If a stream is in CREATING
or UPDATING
or
* DELETING
states, DescribeStream
returns a ResourceInUseException
.
*
*
* If the specified stream does not exist, DescribeStream
returns a
* ResourceNotFoundException
. If you try to create more shards than are authorized for your account,
* you receive a LimitExceededException
.
*
*
* For the default shard limit for an AWS account, see Kinesis Data Streams
* Limits in the Amazon Kinesis Data Streams Developer Guide. To increase this limit, contact AWS Support.
*
*
* If you try to operate on too many streams simultaneously using CreateStream, DeleteStream,
* MergeShards, and/or SplitShard, you receive a LimitExceededException
.
*
*
* SplitShard
has a limit of five transactions per second per account.
*
*
* @param splitShardRequest
* Represents the input for SplitShard
.
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the SplitShard operation returned by the service.
* @sample AmazonKinesisAsyncHandler.SplitShard
* @see AWS API
* Documentation
*/
java.util.concurrent.Future splitShardAsync(SplitShardRequest splitShardRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
* Simplified method form for invoking the SplitShard operation.
*
* @see #splitShardAsync(SplitShardRequest)
*/
java.util.concurrent.Future splitShardAsync(String streamName, String shardToSplit, String newStartingHashKey);
/**
* Simplified method form for invoking the SplitShard operation with an AsyncHandler.
*
* @see #splitShardAsync(SplitShardRequest, com.amazonaws.handlers.AsyncHandler)
*/
java.util.concurrent.Future splitShardAsync(String streamName, String shardToSplit, String newStartingHashKey,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Enables or updates server-side encryption using an AWS KMS key for a specified stream.
*
*
* Starting encryption is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns
* immediately and sets the status of the stream to UPDATING
. After the update is complete, Kinesis
* Data Streams sets the status of the stream back to ACTIVE
. Updating or applying encryption normally
* takes a few seconds to complete, but it can take minutes. You can continue to read and write data to your stream
* while its status is UPDATING
. Once the status of the stream is ACTIVE
, encryption
* begins for records written to the stream.
*
*
* API Limits: You can successfully apply a new AWS KMS key for server-side encryption 25 times in a rolling 24-hour
* period.
*
*
* Note: It can take up to 5 seconds after the stream is in an ACTIVE
status before all records written
* to the stream are encrypted. After you enable encryption, you can verify that encryption is applied by inspecting
* the API response from PutRecord
or PutRecords
.
*
*
* @param startStreamEncryptionRequest
* @return A Java Future containing the result of the StartStreamEncryption operation returned by the service.
* @sample AmazonKinesisAsync.StartStreamEncryption
* @see AWS
* API Documentation
*/
java.util.concurrent.Future startStreamEncryptionAsync(StartStreamEncryptionRequest startStreamEncryptionRequest);
/**
*
* Enables or updates server-side encryption using an AWS KMS key for a specified stream.
*
*
* Starting encryption is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns
* immediately and sets the status of the stream to UPDATING
. After the update is complete, Kinesis
* Data Streams sets the status of the stream back to ACTIVE
. Updating or applying encryption normally
* takes a few seconds to complete, but it can take minutes. You can continue to read and write data to your stream
* while its status is UPDATING
. Once the status of the stream is ACTIVE
, encryption
* begins for records written to the stream.
*
*
* API Limits: You can successfully apply a new AWS KMS key for server-side encryption 25 times in a rolling 24-hour
* period.
*
*
* Note: It can take up to 5 seconds after the stream is in an ACTIVE
status before all records written
* to the stream are encrypted. After you enable encryption, you can verify that encryption is applied by inspecting
* the API response from PutRecord
or PutRecords
.
*
*
* @param startStreamEncryptionRequest
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the StartStreamEncryption operation returned by the service.
* @sample AmazonKinesisAsyncHandler.StartStreamEncryption
* @see AWS
* API Documentation
*/
java.util.concurrent.Future startStreamEncryptionAsync(StartStreamEncryptionRequest startStreamEncryptionRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Disables server-side encryption for a specified stream.
*
*
* Stopping encryption is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns
* immediately and sets the status of the stream to UPDATING
. After the update is complete, Kinesis
* Data Streams sets the status of the stream back to ACTIVE
. Stopping encryption normally takes a few
* seconds to complete, but it can take minutes. You can continue to read and write data to your stream while its
* status is UPDATING
. Once the status of the stream is ACTIVE
, records written to the
* stream are no longer encrypted by Kinesis Data Streams.
*
*
* API Limits: You can successfully disable server-side encryption 25 times in a rolling 24-hour period.
*
*
* Note: It can take up to 5 seconds after the stream is in an ACTIVE
status before all records written
* to the stream are no longer subject to encryption. After you disabled encryption, you can verify that encryption
* is not applied by inspecting the API response from PutRecord
or PutRecords
.
*
*
* @param stopStreamEncryptionRequest
* @return A Java Future containing the result of the StopStreamEncryption operation returned by the service.
* @sample AmazonKinesisAsync.StopStreamEncryption
* @see AWS
* API Documentation
*/
java.util.concurrent.Future stopStreamEncryptionAsync(StopStreamEncryptionRequest stopStreamEncryptionRequest);
/**
*
* Disables server-side encryption for a specified stream.
*
*
* Stopping encryption is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns
* immediately and sets the status of the stream to UPDATING
. After the update is complete, Kinesis
* Data Streams sets the status of the stream back to ACTIVE
. Stopping encryption normally takes a few
* seconds to complete, but it can take minutes. You can continue to read and write data to your stream while its
* status is UPDATING
. Once the status of the stream is ACTIVE
, records written to the
* stream are no longer encrypted by Kinesis Data Streams.
*
*
* API Limits: You can successfully disable server-side encryption 25 times in a rolling 24-hour period.
*
*
* Note: It can take up to 5 seconds after the stream is in an ACTIVE
status before all records written
* to the stream are no longer subject to encryption. After you disabled encryption, you can verify that encryption
* is not applied by inspecting the API response from PutRecord
or PutRecords
.
*
*
* @param stopStreamEncryptionRequest
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the StopStreamEncryption operation returned by the service.
* @sample AmazonKinesisAsyncHandler.StopStreamEncryption
* @see AWS
* API Documentation
*/
java.util.concurrent.Future stopStreamEncryptionAsync(StopStreamEncryptionRequest stopStreamEncryptionRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
/**
*
* Updates the shard count of the specified stream to the specified number of shards.
*
*
* Updating the shard count is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns
* immediately and sets the status of the stream to UPDATING
. After the update is complete, Kinesis
* Data Streams sets the status of the stream back to ACTIVE
. Depending on the size of the stream, the
* scaling action could take a few minutes to complete. You can continue to read and write data to your stream while
* its status is UPDATING
.
*
*
* To update the shard count, Kinesis Data Streams performs splits or merges on individual shards. This can cause
* short-lived shards to be created, in addition to the final shards. We recommend that you double or halve the
* shard count, as this results in the fewest number of splits or merges.
*
*
* This operation has the following default limits. By default, you cannot do the following:
*
*
* -
*
* Scale more than twice per rolling 24-hour period per stream
*
*
* -
*
* Scale up to more than double your current shard count for a stream
*
*
* -
*
* Scale down below half your current shard count for a stream
*
*
* -
*
* Scale up to more than 500 shards in a stream
*
*
* -
*
* Scale a stream with more than 500 shards down unless the result is less than 500 shards
*
*
* -
*
* Scale up to more than the shard limit for your account
*
*
*
*
* For the default limits for an AWS account, see Streams Limits in the
* Amazon Kinesis Data Streams Developer Guide. To request an increase in the call rate limit, the shard
* limit for this API, or your overall shard limit, use the limits form.
*
*
* @param updateShardCountRequest
* @return A Java Future containing the result of the UpdateShardCount operation returned by the service.
* @sample AmazonKinesisAsync.UpdateShardCount
* @see AWS API
* Documentation
*/
java.util.concurrent.Future updateShardCountAsync(UpdateShardCountRequest updateShardCountRequest);
/**
*
* Updates the shard count of the specified stream to the specified number of shards.
*
*
* Updating the shard count is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns
* immediately and sets the status of the stream to UPDATING
. After the update is complete, Kinesis
* Data Streams sets the status of the stream back to ACTIVE
. Depending on the size of the stream, the
* scaling action could take a few minutes to complete. You can continue to read and write data to your stream while
* its status is UPDATING
.
*
*
* To update the shard count, Kinesis Data Streams performs splits or merges on individual shards. This can cause
* short-lived shards to be created, in addition to the final shards. We recommend that you double or halve the
* shard count, as this results in the fewest number of splits or merges.
*
*
* This operation has the following default limits. By default, you cannot do the following:
*
*
* -
*
* Scale more than twice per rolling 24-hour period per stream
*
*
* -
*
* Scale up to more than double your current shard count for a stream
*
*
* -
*
* Scale down below half your current shard count for a stream
*
*
* -
*
* Scale up to more than 500 shards in a stream
*
*
* -
*
* Scale a stream with more than 500 shards down unless the result is less than 500 shards
*
*
* -
*
* Scale up to more than the shard limit for your account
*
*
*
*
* For the default limits for an AWS account, see Streams Limits in the
* Amazon Kinesis Data Streams Developer Guide. To request an increase in the call rate limit, the shard
* limit for this API, or your overall shard limit, use the limits form.
*
*
* @param updateShardCountRequest
* @param asyncHandler
* Asynchronous callback handler for events in the lifecycle of the request. Users can provide an
* implementation of the callback methods in this interface to receive notification of successful or
* unsuccessful completion of the operation.
* @return A Java Future containing the result of the UpdateShardCount operation returned by the service.
* @sample AmazonKinesisAsyncHandler.UpdateShardCount
* @see AWS API
* Documentation
*/
java.util.concurrent.Future updateShardCountAsync(UpdateShardCountRequest updateShardCountRequest,
com.amazonaws.handlers.AsyncHandler asyncHandler);
}