com.amazonaws.services.kinesisfirehose.AmazonKinesisFirehoseClient Maven / Gradle / Ivy
Show all versions of aws-java-sdk-kinesis Show documentation
/*
* Copyright 2019-2024 Amazon.com, Inc. or its affiliates. All Rights Reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License"). You may not use this file except in compliance with
* the License. A copy of the License is located at
*
* http://aws.amazon.com/apache2.0
*
* or in the "license" file accompanying this file. This file is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
* CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions
* and limitations under the License.
*/
package com.amazonaws.services.kinesisfirehose;
import org.w3c.dom.*;
import java.net.*;
import java.util.*;
import javax.annotation.Generated;
import org.apache.commons.logging.*;
import com.amazonaws.*;
import com.amazonaws.annotation.SdkInternalApi;
import com.amazonaws.auth.*;
import com.amazonaws.handlers.*;
import com.amazonaws.http.*;
import com.amazonaws.internal.*;
import com.amazonaws.internal.auth.*;
import com.amazonaws.metrics.*;
import com.amazonaws.regions.*;
import com.amazonaws.transform.*;
import com.amazonaws.util.*;
import com.amazonaws.protocol.json.*;
import com.amazonaws.util.AWSRequestMetrics.Field;
import com.amazonaws.annotation.ThreadSafe;
import com.amazonaws.client.AwsSyncClientParams;
import com.amazonaws.client.builder.AdvancedConfig;
import com.amazonaws.services.kinesisfirehose.AmazonKinesisFirehoseClientBuilder;
import com.amazonaws.AmazonServiceException;
import com.amazonaws.services.kinesisfirehose.model.*;
import com.amazonaws.services.kinesisfirehose.model.transform.*;
/**
* Client for accessing Firehose. All service calls made using this client are blocking, and will not return until the
* service call completes.
*
* Amazon Kinesis Data Firehose API Reference
*
* Amazon Kinesis Data Firehose is a fully managed service that delivers real-time streaming data to destinations such
* as Amazon Simple Storage Service (Amazon S3), Amazon OpenSearch Service, Amazon Redshift, Splunk, and various other
* supportd destinations.
*
*/
@ThreadSafe
@Generated("com.amazonaws:aws-java-sdk-code-generator")
public class AmazonKinesisFirehoseClient extends AmazonWebServiceClient implements AmazonKinesisFirehose {
/** Provider for AWS credentials. */
private final AWSCredentialsProvider awsCredentialsProvider;
private static final Log log = LogFactory.getLog(AmazonKinesisFirehose.class);
/** Default signing name for the service. */
private static final String DEFAULT_SIGNING_NAME = "firehose";
/** Client configuration factory providing ClientConfigurations tailored to this client */
protected static final ClientConfigurationFactory configFactory = new ClientConfigurationFactory();
private final AdvancedConfig advancedConfig;
private static final com.amazonaws.protocol.json.SdkJsonProtocolFactory protocolFactory = new com.amazonaws.protocol.json.SdkJsonProtocolFactory(
new JsonClientMetadata()
.withProtocolVersion("1.1")
.withSupportsCbor(false)
.withSupportsIon(false)
.addErrorMetadata(
new JsonErrorShapeMetadata().withErrorCode("ConcurrentModificationException").withExceptionUnmarshaller(
com.amazonaws.services.kinesisfirehose.model.transform.ConcurrentModificationExceptionUnmarshaller.getInstance()))
.addErrorMetadata(
new JsonErrorShapeMetadata().withErrorCode("ResourceInUseException").withExceptionUnmarshaller(
com.amazonaws.services.kinesisfirehose.model.transform.ResourceInUseExceptionUnmarshaller.getInstance()))
.addErrorMetadata(
new JsonErrorShapeMetadata().withErrorCode("InvalidSourceException").withExceptionUnmarshaller(
com.amazonaws.services.kinesisfirehose.model.transform.InvalidSourceExceptionUnmarshaller.getInstance()))
.addErrorMetadata(
new JsonErrorShapeMetadata().withErrorCode("LimitExceededException").withExceptionUnmarshaller(
com.amazonaws.services.kinesisfirehose.model.transform.LimitExceededExceptionUnmarshaller.getInstance()))
.addErrorMetadata(
new JsonErrorShapeMetadata().withErrorCode("InvalidArgumentException").withExceptionUnmarshaller(
com.amazonaws.services.kinesisfirehose.model.transform.InvalidArgumentExceptionUnmarshaller.getInstance()))
.addErrorMetadata(
new JsonErrorShapeMetadata().withErrorCode("ResourceNotFoundException").withExceptionUnmarshaller(
com.amazonaws.services.kinesisfirehose.model.transform.ResourceNotFoundExceptionUnmarshaller.getInstance()))
.addErrorMetadata(
new JsonErrorShapeMetadata().withErrorCode("InvalidKMSResourceException").withExceptionUnmarshaller(
com.amazonaws.services.kinesisfirehose.model.transform.InvalidKMSResourceExceptionUnmarshaller.getInstance()))
.addErrorMetadata(
new JsonErrorShapeMetadata().withErrorCode("ServiceUnavailableException").withExceptionUnmarshaller(
com.amazonaws.services.kinesisfirehose.model.transform.ServiceUnavailableExceptionUnmarshaller.getInstance()))
.withBaseServiceExceptionClass(com.amazonaws.services.kinesisfirehose.model.AmazonKinesisFirehoseException.class));
/**
* Constructs a new client to invoke service methods on Firehose. A credentials provider chain will be used that
* searches for credentials in this order:
*
* - Environment Variables - AWS_ACCESS_KEY_ID and AWS_SECRET_KEY
* - Java System Properties - aws.accessKeyId and aws.secretKey
* - Instance profile credentials delivered through the Amazon EC2 metadata service
*
*
*
* All service calls made using this new client object are blocking, and will not return until the service call
* completes.
*
* @see DefaultAWSCredentialsProviderChain
* @deprecated use {@link AmazonKinesisFirehoseClientBuilder#defaultClient()}
*/
@Deprecated
public AmazonKinesisFirehoseClient() {
this(DefaultAWSCredentialsProviderChain.getInstance(), configFactory.getConfig());
}
/**
* Constructs a new client to invoke service methods on Firehose. A credentials provider chain will be used that
* searches for credentials in this order:
*
* - Environment Variables - AWS_ACCESS_KEY_ID and AWS_SECRET_KEY
* - Java System Properties - aws.accessKeyId and aws.secretKey
* - Instance profile credentials delivered through the Amazon EC2 metadata service
*
*
*
* All service calls made using this new client object are blocking, and will not return until the service call
* completes.
*
* @param clientConfiguration
* The client configuration options controlling how this client connects to Firehose (ex: proxy settings,
* retry counts, etc.).
*
* @see DefaultAWSCredentialsProviderChain
* @deprecated use {@link AmazonKinesisFirehoseClientBuilder#withClientConfiguration(ClientConfiguration)}
*/
@Deprecated
public AmazonKinesisFirehoseClient(ClientConfiguration clientConfiguration) {
this(DefaultAWSCredentialsProviderChain.getInstance(), clientConfiguration);
}
/**
* Constructs a new client to invoke service methods on Firehose using the specified AWS account credentials.
*
*
* All service calls made using this new client object are blocking, and will not return until the service call
* completes.
*
* @param awsCredentials
* The AWS credentials (access key ID and secret key) to use when authenticating with AWS services.
* @deprecated use {@link AmazonKinesisFirehoseClientBuilder#withCredentials(AWSCredentialsProvider)} for example:
* {@code AmazonKinesisFirehoseClientBuilder.standard().withCredentials(new AWSStaticCredentialsProvider(awsCredentials)).build();}
*/
@Deprecated
public AmazonKinesisFirehoseClient(AWSCredentials awsCredentials) {
this(awsCredentials, configFactory.getConfig());
}
/**
* Constructs a new client to invoke service methods on Firehose using the specified AWS account credentials and
* client configuration options.
*
*
* All service calls made using this new client object are blocking, and will not return until the service call
* completes.
*
* @param awsCredentials
* The AWS credentials (access key ID and secret key) to use when authenticating with AWS services.
* @param clientConfiguration
* The client configuration options controlling how this client connects to Firehose (ex: proxy settings,
* retry counts, etc.).
* @deprecated use {@link AmazonKinesisFirehoseClientBuilder#withCredentials(AWSCredentialsProvider)} and
* {@link AmazonKinesisFirehoseClientBuilder#withClientConfiguration(ClientConfiguration)}
*/
@Deprecated
public AmazonKinesisFirehoseClient(AWSCredentials awsCredentials, ClientConfiguration clientConfiguration) {
super(clientConfiguration);
this.awsCredentialsProvider = new StaticCredentialsProvider(awsCredentials);
this.advancedConfig = AdvancedConfig.EMPTY;
init();
}
/**
* Constructs a new client to invoke service methods on Firehose using the specified AWS account credentials
* provider.
*
*
* All service calls made using this new client object are blocking, and will not return until the service call
* completes.
*
* @param awsCredentialsProvider
* The AWS credentials provider which will provide credentials to authenticate requests with AWS services.
* @deprecated use {@link AmazonKinesisFirehoseClientBuilder#withCredentials(AWSCredentialsProvider)}
*/
@Deprecated
public AmazonKinesisFirehoseClient(AWSCredentialsProvider awsCredentialsProvider) {
this(awsCredentialsProvider, configFactory.getConfig());
}
/**
* Constructs a new client to invoke service methods on Firehose using the specified AWS account credentials
* provider and client configuration options.
*
*
* All service calls made using this new client object are blocking, and will not return until the service call
* completes.
*
* @param awsCredentialsProvider
* The AWS credentials provider which will provide credentials to authenticate requests with AWS services.
* @param clientConfiguration
* The client configuration options controlling how this client connects to Firehose (ex: proxy settings,
* retry counts, etc.).
* @deprecated use {@link AmazonKinesisFirehoseClientBuilder#withCredentials(AWSCredentialsProvider)} and
* {@link AmazonKinesisFirehoseClientBuilder#withClientConfiguration(ClientConfiguration)}
*/
@Deprecated
public AmazonKinesisFirehoseClient(AWSCredentialsProvider awsCredentialsProvider, ClientConfiguration clientConfiguration) {
this(awsCredentialsProvider, clientConfiguration, null);
}
/**
* Constructs a new client to invoke service methods on Firehose using the specified AWS account credentials
* provider, client configuration options, and request metric collector.
*
*
* All service calls made using this new client object are blocking, and will not return until the service call
* completes.
*
* @param awsCredentialsProvider
* The AWS credentials provider which will provide credentials to authenticate requests with AWS services.
* @param clientConfiguration
* The client configuration options controlling how this client connects to Firehose (ex: proxy settings,
* retry counts, etc.).
* @param requestMetricCollector
* optional request metric collector
* @deprecated use {@link AmazonKinesisFirehoseClientBuilder#withCredentials(AWSCredentialsProvider)} and
* {@link AmazonKinesisFirehoseClientBuilder#withClientConfiguration(ClientConfiguration)} and
* {@link AmazonKinesisFirehoseClientBuilder#withMetricsCollector(RequestMetricCollector)}
*/
@Deprecated
public AmazonKinesisFirehoseClient(AWSCredentialsProvider awsCredentialsProvider, ClientConfiguration clientConfiguration,
RequestMetricCollector requestMetricCollector) {
super(clientConfiguration, requestMetricCollector);
this.awsCredentialsProvider = awsCredentialsProvider;
this.advancedConfig = AdvancedConfig.EMPTY;
init();
}
public static AmazonKinesisFirehoseClientBuilder builder() {
return AmazonKinesisFirehoseClientBuilder.standard();
}
/**
* Constructs a new client to invoke service methods on Firehose using the specified parameters.
*
*
* All service calls made using this new client object are blocking, and will not return until the service call
* completes.
*
* @param clientParams
* Object providing client parameters.
*/
AmazonKinesisFirehoseClient(AwsSyncClientParams clientParams) {
this(clientParams, false);
}
/**
* Constructs a new client to invoke service methods on Firehose using the specified parameters.
*
*
* All service calls made using this new client object are blocking, and will not return until the service call
* completes.
*
* @param clientParams
* Object providing client parameters.
*/
AmazonKinesisFirehoseClient(AwsSyncClientParams clientParams, boolean endpointDiscoveryEnabled) {
super(clientParams);
this.awsCredentialsProvider = clientParams.getCredentialsProvider();
this.advancedConfig = clientParams.getAdvancedConfig();
init();
}
private void init() {
setServiceNameIntern(DEFAULT_SIGNING_NAME);
setEndpointPrefix(ENDPOINT_PREFIX);
// calling this.setEndPoint(...) will also modify the signer accordingly
setEndpoint("https://firehose.us-east-1.amazonaws.com");
HandlerChainFactory chainFactory = new HandlerChainFactory();
requestHandler2s.addAll(chainFactory.newRequestHandlerChain("/com/amazonaws/services/kinesisfirehose/request.handlers"));
requestHandler2s.addAll(chainFactory.newRequestHandler2Chain("/com/amazonaws/services/kinesisfirehose/request.handler2s"));
requestHandler2s.addAll(chainFactory.getGlobalHandlers());
}
/**
*
* Creates a Kinesis Data Firehose delivery stream.
*
*
* By default, you can create up to 50 delivery streams per Amazon Web Services Region.
*
*
* This is an asynchronous operation that immediately returns. The initial status of the delivery stream is
* CREATING
. After the delivery stream is created, its status is ACTIVE
and it now accepts
* data. If the delivery stream creation fails, the status transitions to CREATING_FAILED
. Attempts to
* send data to a delivery stream that is not in the ACTIVE
state cause an exception. To check the
* state of a delivery stream, use DescribeDeliveryStream.
*
*
* If the status of a delivery stream is CREATING_FAILED
, this status doesn't change, and you can't
* invoke CreateDeliveryStream
again on it. However, you can invoke the DeleteDeliveryStream
* operation to delete it.
*
*
* A Kinesis Data Firehose delivery stream can be configured to receive records directly from providers using
* PutRecord or PutRecordBatch, or it can be configured to use an existing Kinesis stream as its
* source. To specify a Kinesis data stream as input, set the DeliveryStreamType
parameter to
* KinesisStreamAsSource
, and provide the Kinesis stream Amazon Resource Name (ARN) and role ARN in the
* KinesisStreamSourceConfiguration
parameter.
*
*
* To create a delivery stream with server-side encryption (SSE) enabled, include
* DeliveryStreamEncryptionConfigurationInput in your request. This is optional. You can also invoke
* StartDeliveryStreamEncryption to turn on SSE for an existing delivery stream that doesn't have SSE
* enabled.
*
*
* A delivery stream is configured with a single destination, such as Amazon Simple Storage Service (Amazon S3),
* Amazon Redshift, Amazon OpenSearch Service, Amazon OpenSearch Serverless, Splunk, and any custom HTTP endpoint or
* HTTP endpoints owned by or supported by third-party service providers, including Datadog, Dynatrace,
* LogicMonitor, MongoDB, New Relic, and Sumo Logic. You must specify only one of the following destination
* configuration parameters: ExtendedS3DestinationConfiguration
,
* S3DestinationConfiguration
, ElasticsearchDestinationConfiguration
,
* RedshiftDestinationConfiguration
, or SplunkDestinationConfiguration
.
*
*
* When you specify S3DestinationConfiguration
, you can also provide the following optional values:
* BufferingHints, EncryptionConfiguration
, and CompressionFormat
. By default, if no
* BufferingHints
value is provided, Kinesis Data Firehose buffers data up to 5 MB or for 5 minutes,
* whichever condition is satisfied first. BufferingHints
is a hint, so there are some cases where the
* service cannot adhere to these conditions strictly. For example, record boundaries might be such that the size is
* a little over or under the configured buffering size. By default, no encryption is performed. We strongly
* recommend that you enable encryption to ensure secure data storage in Amazon S3.
*
*
* A few notes about Amazon Redshift as a destination:
*
*
* -
*
* An Amazon Redshift destination requires an S3 bucket as intermediate location. Kinesis Data Firehose first
* delivers data to Amazon S3 and then uses COPY
syntax to load data into an Amazon Redshift table.
* This is specified in the RedshiftDestinationConfiguration.S3Configuration
parameter.
*
*
* -
*
* The compression formats SNAPPY
or ZIP
cannot be specified in
* RedshiftDestinationConfiguration.S3Configuration
because the Amazon Redshift COPY
* operation that reads from the S3 bucket doesn't support these compression formats.
*
*
* -
*
* We strongly recommend that you use the user name and password you provide exclusively with Kinesis Data Firehose,
* and that the permissions for the account are restricted for Amazon Redshift INSERT
permissions.
*
*
*
*
* Kinesis Data Firehose assumes the IAM role that is configured as part of the destination. The role should allow
* the Kinesis Data Firehose principal to assume the role, and the role should have permissions that allow the
* service to deliver the data. For more information, see Grant Kinesis Data
* Firehose Access to an Amazon S3 Destination in the Amazon Kinesis Data Firehose Developer Guide.
*
*
* @param createDeliveryStreamRequest
* @return Result of the CreateDeliveryStream operation returned by the service.
* @throws InvalidArgumentException
* The specified input parameter has a value that is not valid.
* @throws LimitExceededException
* You have already reached the limit for a requested resource.
* @throws ResourceInUseException
* The resource is already in use and not available for this operation.
* @throws InvalidKMSResourceException
* Kinesis Data Firehose throws this exception when an attempt to put records or to start or stop delivery
* stream encryption fails. This happens when the KMS service throws one of the following exception types:
* AccessDeniedException
, InvalidStateException
, DisabledException
,
* or NotFoundException
.
* @sample AmazonKinesisFirehose.CreateDeliveryStream
* @see AWS
* API Documentation
*/
@Override
public CreateDeliveryStreamResult createDeliveryStream(CreateDeliveryStreamRequest request) {
request = beforeClientExecution(request);
return executeCreateDeliveryStream(request);
}
@SdkInternalApi
final CreateDeliveryStreamResult executeCreateDeliveryStream(CreateDeliveryStreamRequest createDeliveryStreamRequest) {
ExecutionContext executionContext = createExecutionContext(createDeliveryStreamRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new CreateDeliveryStreamRequestProtocolMarshaller(protocolFactory).marshall(super.beforeMarshalling(createDeliveryStreamRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "CreateDeliveryStream");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(
new JsonOperationMetadata().withPayloadJson(true).withHasStreamingSuccessResponse(false), new CreateDeliveryStreamResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
*
* Deletes a delivery stream and its data.
*
*
* To check the state of a delivery stream, use DescribeDeliveryStream. You can delete a delivery stream only
* if it is in one of the following states: ACTIVE
, DELETING
, CREATING_FAILED
* , or DELETING_FAILED
. You can't delete a delivery stream that is in the CREATING
state.
* While the deletion request is in process, the delivery stream is in the DELETING
state.
*
*
* While the delivery stream is in the DELETING
state, the service might continue to accept records,
* but it doesn't make any guarantees with respect to delivering the data. Therefore, as a best practice, first stop
* any applications that are sending records before you delete a delivery stream.
*
*
* @param deleteDeliveryStreamRequest
* @return Result of the DeleteDeliveryStream operation returned by the service.
* @throws ResourceInUseException
* The resource is already in use and not available for this operation.
* @throws ResourceNotFoundException
* The specified resource could not be found.
* @sample AmazonKinesisFirehose.DeleteDeliveryStream
* @see AWS
* API Documentation
*/
@Override
public DeleteDeliveryStreamResult deleteDeliveryStream(DeleteDeliveryStreamRequest request) {
request = beforeClientExecution(request);
return executeDeleteDeliveryStream(request);
}
@SdkInternalApi
final DeleteDeliveryStreamResult executeDeleteDeliveryStream(DeleteDeliveryStreamRequest deleteDeliveryStreamRequest) {
ExecutionContext executionContext = createExecutionContext(deleteDeliveryStreamRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new DeleteDeliveryStreamRequestProtocolMarshaller(protocolFactory).marshall(super.beforeMarshalling(deleteDeliveryStreamRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "DeleteDeliveryStream");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(
new JsonOperationMetadata().withPayloadJson(true).withHasStreamingSuccessResponse(false), new DeleteDeliveryStreamResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
*
* Describes the specified delivery stream and its status. For example, after your delivery stream is created, call
* DescribeDeliveryStream
to see whether the delivery stream is ACTIVE
and therefore ready
* for data to be sent to it.
*
*
* If the status of a delivery stream is CREATING_FAILED
, this status doesn't change, and you can't
* invoke CreateDeliveryStream again on it. However, you can invoke the DeleteDeliveryStream operation
* to delete it. If the status is DELETING_FAILED
, you can force deletion by invoking
* DeleteDeliveryStream again but with DeleteDeliveryStreamInput$AllowForceDelete set to true.
*
*
* @param describeDeliveryStreamRequest
* @return Result of the DescribeDeliveryStream operation returned by the service.
* @throws ResourceNotFoundException
* The specified resource could not be found.
* @sample AmazonKinesisFirehose.DescribeDeliveryStream
* @see AWS API Documentation
*/
@Override
public DescribeDeliveryStreamResult describeDeliveryStream(DescribeDeliveryStreamRequest request) {
request = beforeClientExecution(request);
return executeDescribeDeliveryStream(request);
}
@SdkInternalApi
final DescribeDeliveryStreamResult executeDescribeDeliveryStream(DescribeDeliveryStreamRequest describeDeliveryStreamRequest) {
ExecutionContext executionContext = createExecutionContext(describeDeliveryStreamRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new DescribeDeliveryStreamRequestProtocolMarshaller(protocolFactory).marshall(super.beforeMarshalling(describeDeliveryStreamRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "DescribeDeliveryStream");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(
new JsonOperationMetadata().withPayloadJson(true).withHasStreamingSuccessResponse(false),
new DescribeDeliveryStreamResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
*
* Lists your delivery streams in alphabetical order of their names.
*
*
* The number of delivery streams might be too large to return using a single call to
* ListDeliveryStreams
. You can limit the number of delivery streams returned, using the
* Limit
parameter. To determine whether there are more delivery streams to list, check the value of
* HasMoreDeliveryStreams
in the output. If there are more delivery streams to list, you can request
* them by calling this operation again and setting the ExclusiveStartDeliveryStreamName
parameter to
* the name of the last delivery stream returned in the last call.
*
*
* @param listDeliveryStreamsRequest
* @return Result of the ListDeliveryStreams operation returned by the service.
* @sample AmazonKinesisFirehose.ListDeliveryStreams
* @see AWS
* API Documentation
*/
@Override
public ListDeliveryStreamsResult listDeliveryStreams(ListDeliveryStreamsRequest request) {
request = beforeClientExecution(request);
return executeListDeliveryStreams(request);
}
@SdkInternalApi
final ListDeliveryStreamsResult executeListDeliveryStreams(ListDeliveryStreamsRequest listDeliveryStreamsRequest) {
ExecutionContext executionContext = createExecutionContext(listDeliveryStreamsRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new ListDeliveryStreamsRequestProtocolMarshaller(protocolFactory).marshall(super.beforeMarshalling(listDeliveryStreamsRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "ListDeliveryStreams");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(
new JsonOperationMetadata().withPayloadJson(true).withHasStreamingSuccessResponse(false), new ListDeliveryStreamsResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
*
* Lists the tags for the specified delivery stream. This operation has a limit of five transactions per second per
* account.
*
*
* @param listTagsForDeliveryStreamRequest
* @return Result of the ListTagsForDeliveryStream operation returned by the service.
* @throws ResourceNotFoundException
* The specified resource could not be found.
* @throws InvalidArgumentException
* The specified input parameter has a value that is not valid.
* @throws LimitExceededException
* You have already reached the limit for a requested resource.
* @sample AmazonKinesisFirehose.ListTagsForDeliveryStream
* @see AWS API Documentation
*/
@Override
public ListTagsForDeliveryStreamResult listTagsForDeliveryStream(ListTagsForDeliveryStreamRequest request) {
request = beforeClientExecution(request);
return executeListTagsForDeliveryStream(request);
}
@SdkInternalApi
final ListTagsForDeliveryStreamResult executeListTagsForDeliveryStream(ListTagsForDeliveryStreamRequest listTagsForDeliveryStreamRequest) {
ExecutionContext executionContext = createExecutionContext(listTagsForDeliveryStreamRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new ListTagsForDeliveryStreamRequestProtocolMarshaller(protocolFactory).marshall(super
.beforeMarshalling(listTagsForDeliveryStreamRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "ListTagsForDeliveryStream");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(
new JsonOperationMetadata().withPayloadJson(true).withHasStreamingSuccessResponse(false),
new ListTagsForDeliveryStreamResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
*
* Writes a single data record into an Amazon Kinesis Data Firehose delivery stream. To write multiple data records
* into a delivery stream, use PutRecordBatch. Applications using these operations are referred to as
* producers.
*
*
* By default, each delivery stream can take in up to 2,000 transactions per second, 5,000 records per second, or 5
* MB per second. If you use PutRecord and PutRecordBatch, the limits are an aggregate across these
* two operations for each delivery stream. For more information about limits and how to request an increase, see Amazon Kinesis Data Firehose Limits.
*
*
* Kinesis Data Firehose accumulates and publishes a particular metric for a customer account in one minute
* intervals. It is possible that the bursts of incoming bytes/records ingested to a delivery stream last only for a
* few seconds. Due to this, the actual spikes in the traffic might not be fully visible in the customer's 1 minute
* CloudWatch metrics.
*
*
* You must specify the name of the delivery stream and the data record when using PutRecord. The data record
* consists of a data blob that can be up to 1,000 KiB in size, and any kind of data. For example, it can be a
* segment from a log file, geographic location data, website clickstream data, and so on.
*
*
* Kinesis Data Firehose buffers records before delivering them to the destination. To disambiguate the data blobs
* at the destination, a common solution is to use delimiters in the data, such as a newline (\n
) or
* some other character unique within the data. This allows the consumer application to parse individual data items
* when reading the data from the destination.
*
*
* The PutRecord
operation returns a RecordId
, which is a unique string assigned to each
* record. Producer applications can use this ID for purposes such as auditability and investigation.
*
*
* If the PutRecord
operation throws a ServiceUnavailableException
, the API is
* automatically reinvoked (retried) 3 times. If the exception persists, it is possible that the throughput limits
* have been exceeded for the delivery stream.
*
*
* Re-invoking the Put API operations (for example, PutRecord and PutRecordBatch) can result in data duplicates. For
* larger data assets, allow for a longer time out before retrying Put API operations.
*
*
* Data records sent to Kinesis Data Firehose are stored for 24 hours from the time they are added to a delivery
* stream as it tries to send the records to the destination. If the destination is unreachable for more than 24
* hours, the data is no longer available.
*
*
*
* Don't concatenate two or more base64 strings to form the data fields of your records. Instead, concatenate the
* raw data, then perform base64 encoding.
*
*
*
* @param putRecordRequest
* @return Result of the PutRecord operation returned by the service.
* @throws ResourceNotFoundException
* The specified resource could not be found.
* @throws InvalidArgumentException
* The specified input parameter has a value that is not valid.
* @throws InvalidKMSResourceException
* Kinesis Data Firehose throws this exception when an attempt to put records or to start or stop delivery
* stream encryption fails. This happens when the KMS service throws one of the following exception types:
* AccessDeniedException
, InvalidStateException
, DisabledException
,
* or NotFoundException
.
* @throws InvalidSourceException
* Only requests from CloudWatch Logs are supported when CloudWatch Logs decompression is enabled.
* @throws ServiceUnavailableException
* The service is unavailable. Back off and retry the operation. If you continue to see the exception,
* throughput limits for the delivery stream may have been exceeded. For more information about limits and
* how to request an increase, see Amazon Kinesis Data Firehose
* Limits.
* @sample AmazonKinesisFirehose.PutRecord
* @see AWS API
* Documentation
*/
@Override
public PutRecordResult putRecord(PutRecordRequest request) {
request = beforeClientExecution(request);
return executePutRecord(request);
}
@SdkInternalApi
final PutRecordResult executePutRecord(PutRecordRequest putRecordRequest) {
ExecutionContext executionContext = createExecutionContext(putRecordRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new PutRecordRequestProtocolMarshaller(protocolFactory).marshall(super.beforeMarshalling(putRecordRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "PutRecord");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(new JsonOperationMetadata()
.withPayloadJson(true).withHasStreamingSuccessResponse(false), new PutRecordResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
*
* Writes multiple data records into a delivery stream in a single call, which can achieve higher throughput per
* producer than when writing single records. To write single data records into a delivery stream, use
* PutRecord. Applications using these operations are referred to as producers.
*
*
* Kinesis Data Firehose accumulates and publishes a particular metric for a customer account in one minute
* intervals. It is possible that the bursts of incoming bytes/records ingested to a delivery stream last only for a
* few seconds. Due to this, the actual spikes in the traffic might not be fully visible in the customer's 1 minute
* CloudWatch metrics.
*
*
* For information about service quota, see Amazon Kinesis Data Firehose Quota.
*
*
* Each PutRecordBatch request supports up to 500 records. Each record in the request can be as large as
* 1,000 KB (before base64 encoding), up to a limit of 4 MB for the entire request. These limits cannot be changed.
*
*
* You must specify the name of the delivery stream and the data record when using PutRecord. The data record
* consists of a data blob that can be up to 1,000 KB in size, and any kind of data. For example, it could be a
* segment from a log file, geographic location data, website clickstream data, and so on.
*
*
* Kinesis Data Firehose buffers records before delivering them to the destination. To disambiguate the data blobs
* at the destination, a common solution is to use delimiters in the data, such as a newline (\n
) or
* some other character unique within the data. This allows the consumer application to parse individual data items
* when reading the data from the destination.
*
*
* The PutRecordBatch response includes a count of failed records, FailedPutCount
, and an array
* of responses, RequestResponses
. Even if the PutRecordBatch call succeeds, the value of
* FailedPutCount
may be greater than 0, indicating that there are records for which the operation
* didn't succeed. Each entry in the RequestResponses
array provides additional information about the
* processed record. It directly correlates with a record in the request array using the same ordering, from the top
* to the bottom. The response array always includes the same number of records as the request array.
* RequestResponses
includes both successfully and unsuccessfully processed records. Kinesis Data
* Firehose tries to process all records in each PutRecordBatch request. A single record failure does not
* stop the processing of subsequent records.
*
*
* A successfully processed record includes a RecordId
value, which is unique for the record. An
* unsuccessfully processed record includes ErrorCode
and ErrorMessage
values.
* ErrorCode
reflects the type of error, and is one of the following values:
* ServiceUnavailableException
or InternalFailure
. ErrorMessage
provides more
* detailed information about the error.
*
*
* If there is an internal server error or a timeout, the write might have completed or it might have failed. If
* FailedPutCount
is greater than 0, retry the request, resending only those records that might have
* failed processing. This minimizes the possible duplicate records and also reduces the total bytes sent (and
* corresponding charges). We recommend that you handle any duplicates at the destination.
*
*
* If PutRecordBatch throws ServiceUnavailableException
, the API is automatically reinvoked
* (retried) 3 times. If the exception persists, it is possible that the throughput limits have been exceeded for
* the delivery stream.
*
*
* Re-invoking the Put API operations (for example, PutRecord and PutRecordBatch) can result in data duplicates. For
* larger data assets, allow for a longer time out before retrying Put API operations.
*
*
* Data records sent to Kinesis Data Firehose are stored for 24 hours from the time they are added to a delivery
* stream as it attempts to send the records to the destination. If the destination is unreachable for more than 24
* hours, the data is no longer available.
*
*
*
* Don't concatenate two or more base64 strings to form the data fields of your records. Instead, concatenate the
* raw data, then perform base64 encoding.
*
*
*
* @param putRecordBatchRequest
* @return Result of the PutRecordBatch operation returned by the service.
* @throws ResourceNotFoundException
* The specified resource could not be found.
* @throws InvalidArgumentException
* The specified input parameter has a value that is not valid.
* @throws InvalidKMSResourceException
* Kinesis Data Firehose throws this exception when an attempt to put records or to start or stop delivery
* stream encryption fails. This happens when the KMS service throws one of the following exception types:
* AccessDeniedException
, InvalidStateException
, DisabledException
,
* or NotFoundException
.
* @throws InvalidSourceException
* Only requests from CloudWatch Logs are supported when CloudWatch Logs decompression is enabled.
* @throws ServiceUnavailableException
* The service is unavailable. Back off and retry the operation. If you continue to see the exception,
* throughput limits for the delivery stream may have been exceeded. For more information about limits and
* how to request an increase, see Amazon Kinesis Data Firehose
* Limits.
* @sample AmazonKinesisFirehose.PutRecordBatch
* @see AWS API
* Documentation
*/
@Override
public PutRecordBatchResult putRecordBatch(PutRecordBatchRequest request) {
request = beforeClientExecution(request);
return executePutRecordBatch(request);
}
@SdkInternalApi
final PutRecordBatchResult executePutRecordBatch(PutRecordBatchRequest putRecordBatchRequest) {
ExecutionContext executionContext = createExecutionContext(putRecordBatchRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new PutRecordBatchRequestProtocolMarshaller(protocolFactory).marshall(super.beforeMarshalling(putRecordBatchRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "PutRecordBatch");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(
new JsonOperationMetadata().withPayloadJson(true).withHasStreamingSuccessResponse(false), new PutRecordBatchResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
*
* Enables server-side encryption (SSE) for the delivery stream.
*
*
* This operation is asynchronous. It returns immediately. When you invoke it, Kinesis Data Firehose first sets the
* encryption status of the stream to ENABLING
, and then to ENABLED
. The encryption status
* of a delivery stream is the Status
property in DeliveryStreamEncryptionConfiguration. If the
* operation fails, the encryption status changes to ENABLING_FAILED
. You can continue to read and
* write data to your delivery stream while the encryption status is ENABLING
, but the data is not
* encrypted. It can take up to 5 seconds after the encryption status changes to ENABLED
before all
* records written to the delivery stream are encrypted. To find out whether a record or a batch of records was
* encrypted, check the response elements PutRecordOutput$Encrypted and
* PutRecordBatchOutput$Encrypted, respectively.
*
*
* To check the encryption status of a delivery stream, use DescribeDeliveryStream.
*
*
* Even if encryption is currently enabled for a delivery stream, you can still invoke this operation on it to
* change the ARN of the CMK or both its type and ARN. If you invoke this method to change the CMK, and the old CMK
* is of type CUSTOMER_MANAGED_CMK
, Kinesis Data Firehose schedules the grant it had on the old CMK for
* retirement. If the new CMK is of type CUSTOMER_MANAGED_CMK
, Kinesis Data Firehose creates a grant
* that enables it to use the new CMK to encrypt and decrypt data and to manage the grant.
*
*
* For the KMS grant creation to be successful, Kinesis Data Firehose APIs
* StartDeliveryStreamEncryption
and CreateDeliveryStream
should not be called with
* session credentials that are more than 6 hours old.
*
*
* If a delivery stream already has encryption enabled and then you invoke this operation to change the ARN of the
* CMK or both its type and ARN and you get ENABLING_FAILED
, this only means that the attempt to change
* the CMK failed. In this case, encryption remains enabled with the old CMK.
*
*
* If the encryption status of your delivery stream is ENABLING_FAILED
, you can invoke this operation
* again with a valid CMK. The CMK must be enabled and the key policy mustn't explicitly deny the permission for
* Kinesis Data Firehose to invoke KMS encrypt and decrypt operations.
*
*
* You can enable SSE for a delivery stream only if it's a delivery stream that uses DirectPut
as its
* source.
*
*
* The StartDeliveryStreamEncryption
and StopDeliveryStreamEncryption
operations have a
* combined limit of 25 calls per delivery stream per 24 hours. For example, you reach the limit if you call
* StartDeliveryStreamEncryption
13 times and StopDeliveryStreamEncryption
12 times for
* the same delivery stream in a 24-hour period.
*
*
* @param startDeliveryStreamEncryptionRequest
* @return Result of the StartDeliveryStreamEncryption operation returned by the service.
* @throws ResourceNotFoundException
* The specified resource could not be found.
* @throws ResourceInUseException
* The resource is already in use and not available for this operation.
* @throws InvalidArgumentException
* The specified input parameter has a value that is not valid.
* @throws LimitExceededException
* You have already reached the limit for a requested resource.
* @throws InvalidKMSResourceException
* Kinesis Data Firehose throws this exception when an attempt to put records or to start or stop delivery
* stream encryption fails. This happens when the KMS service throws one of the following exception types:
* AccessDeniedException
, InvalidStateException
, DisabledException
,
* or NotFoundException
.
* @sample AmazonKinesisFirehose.StartDeliveryStreamEncryption
* @see AWS API Documentation
*/
@Override
public StartDeliveryStreamEncryptionResult startDeliveryStreamEncryption(StartDeliveryStreamEncryptionRequest request) {
request = beforeClientExecution(request);
return executeStartDeliveryStreamEncryption(request);
}
@SdkInternalApi
final StartDeliveryStreamEncryptionResult executeStartDeliveryStreamEncryption(StartDeliveryStreamEncryptionRequest startDeliveryStreamEncryptionRequest) {
ExecutionContext executionContext = createExecutionContext(startDeliveryStreamEncryptionRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new StartDeliveryStreamEncryptionRequestProtocolMarshaller(protocolFactory).marshall(super
.beforeMarshalling(startDeliveryStreamEncryptionRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "StartDeliveryStreamEncryption");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(
new JsonOperationMetadata().withPayloadJson(true).withHasStreamingSuccessResponse(false),
new StartDeliveryStreamEncryptionResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
*
* Disables server-side encryption (SSE) for the delivery stream.
*
*
* This operation is asynchronous. It returns immediately. When you invoke it, Kinesis Data Firehose first sets the
* encryption status of the stream to DISABLING
, and then to DISABLED
. You can continue to
* read and write data to your stream while its status is DISABLING
. It can take up to 5 seconds after
* the encryption status changes to DISABLED
before all records written to the delivery stream are no
* longer subject to encryption. To find out whether a record or a batch of records was encrypted, check the
* response elements PutRecordOutput$Encrypted and PutRecordBatchOutput$Encrypted, respectively.
*
*
* To check the encryption state of a delivery stream, use DescribeDeliveryStream.
*
*
* If SSE is enabled using a customer managed CMK and then you invoke StopDeliveryStreamEncryption
,
* Kinesis Data Firehose schedules the related KMS grant for retirement and then retires it after it ensures that it
* is finished delivering records to the destination.
*
*
* The StartDeliveryStreamEncryption
and StopDeliveryStreamEncryption
operations have a
* combined limit of 25 calls per delivery stream per 24 hours. For example, you reach the limit if you call
* StartDeliveryStreamEncryption
13 times and StopDeliveryStreamEncryption
12 times for
* the same delivery stream in a 24-hour period.
*
*
* @param stopDeliveryStreamEncryptionRequest
* @return Result of the StopDeliveryStreamEncryption operation returned by the service.
* @throws ResourceNotFoundException
* The specified resource could not be found.
* @throws ResourceInUseException
* The resource is already in use and not available for this operation.
* @throws InvalidArgumentException
* The specified input parameter has a value that is not valid.
* @throws LimitExceededException
* You have already reached the limit for a requested resource.
* @sample AmazonKinesisFirehose.StopDeliveryStreamEncryption
* @see AWS API Documentation
*/
@Override
public StopDeliveryStreamEncryptionResult stopDeliveryStreamEncryption(StopDeliveryStreamEncryptionRequest request) {
request = beforeClientExecution(request);
return executeStopDeliveryStreamEncryption(request);
}
@SdkInternalApi
final StopDeliveryStreamEncryptionResult executeStopDeliveryStreamEncryption(StopDeliveryStreamEncryptionRequest stopDeliveryStreamEncryptionRequest) {
ExecutionContext executionContext = createExecutionContext(stopDeliveryStreamEncryptionRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new StopDeliveryStreamEncryptionRequestProtocolMarshaller(protocolFactory).marshall(super
.beforeMarshalling(stopDeliveryStreamEncryptionRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "StopDeliveryStreamEncryption");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(
new JsonOperationMetadata().withPayloadJson(true).withHasStreamingSuccessResponse(false),
new StopDeliveryStreamEncryptionResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
*
* Adds or updates tags for the specified delivery stream. A tag is a key-value pair that you can define and assign
* to Amazon Web Services resources. If you specify a tag that already exists, the tag value is replaced with the
* value that you specify in the request. Tags are metadata. For example, you can add friendly names and
* descriptions or other types of information that can help you distinguish the delivery stream. For more
* information about tags, see Using Cost Allocation
* Tags in the Amazon Web Services Billing and Cost Management User Guide.
*
*
* Each delivery stream can have up to 50 tags.
*
*
* This operation has a limit of five transactions per second per account.
*
*
* @param tagDeliveryStreamRequest
* @return Result of the TagDeliveryStream operation returned by the service.
* @throws ResourceNotFoundException
* The specified resource could not be found.
* @throws ResourceInUseException
* The resource is already in use and not available for this operation.
* @throws InvalidArgumentException
* The specified input parameter has a value that is not valid.
* @throws LimitExceededException
* You have already reached the limit for a requested resource.
* @sample AmazonKinesisFirehose.TagDeliveryStream
* @see AWS API
* Documentation
*/
@Override
public TagDeliveryStreamResult tagDeliveryStream(TagDeliveryStreamRequest request) {
request = beforeClientExecution(request);
return executeTagDeliveryStream(request);
}
@SdkInternalApi
final TagDeliveryStreamResult executeTagDeliveryStream(TagDeliveryStreamRequest tagDeliveryStreamRequest) {
ExecutionContext executionContext = createExecutionContext(tagDeliveryStreamRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new TagDeliveryStreamRequestProtocolMarshaller(protocolFactory).marshall(super.beforeMarshalling(tagDeliveryStreamRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "TagDeliveryStream");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(
new JsonOperationMetadata().withPayloadJson(true).withHasStreamingSuccessResponse(false), new TagDeliveryStreamResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
*
* Removes tags from the specified delivery stream. Removed tags are deleted, and you can't recover them after this
* operation successfully completes.
*
*
* If you specify a tag that doesn't exist, the operation ignores it.
*
*
* This operation has a limit of five transactions per second per account.
*
*
* @param untagDeliveryStreamRequest
* @return Result of the UntagDeliveryStream operation returned by the service.
* @throws ResourceNotFoundException
* The specified resource could not be found.
* @throws ResourceInUseException
* The resource is already in use and not available for this operation.
* @throws InvalidArgumentException
* The specified input parameter has a value that is not valid.
* @throws LimitExceededException
* You have already reached the limit for a requested resource.
* @sample AmazonKinesisFirehose.UntagDeliveryStream
* @see AWS
* API Documentation
*/
@Override
public UntagDeliveryStreamResult untagDeliveryStream(UntagDeliveryStreamRequest request) {
request = beforeClientExecution(request);
return executeUntagDeliveryStream(request);
}
@SdkInternalApi
final UntagDeliveryStreamResult executeUntagDeliveryStream(UntagDeliveryStreamRequest untagDeliveryStreamRequest) {
ExecutionContext executionContext = createExecutionContext(untagDeliveryStreamRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new UntagDeliveryStreamRequestProtocolMarshaller(protocolFactory).marshall(super.beforeMarshalling(untagDeliveryStreamRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "UntagDeliveryStream");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(
new JsonOperationMetadata().withPayloadJson(true).withHasStreamingSuccessResponse(false), new UntagDeliveryStreamResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
*
* Updates the specified destination of the specified delivery stream.
*
*
* Use this operation to change the destination type (for example, to replace the Amazon S3 destination with Amazon
* Redshift) or change the parameters associated with a destination (for example, to change the bucket name of the
* Amazon S3 destination). The update might not occur immediately. The target delivery stream remains active while
* the configurations are updated, so data writes to the delivery stream can continue during this process. The
* updated configurations are usually effective within a few minutes.
*
*
* Switching between Amazon OpenSearch Service and other services is not supported. For an Amazon OpenSearch Service
* destination, you can only update to another Amazon OpenSearch Service destination.
*
*
* If the destination type is the same, Kinesis Data Firehose merges the configuration parameters specified with the
* destination configuration that already exists on the delivery stream. If any of the parameters are not specified
* in the call, the existing values are retained. For example, in the Amazon S3 destination, if
* EncryptionConfiguration is not specified, then the existing EncryptionConfiguration
is
* maintained on the destination.
*
*
* If the destination type is not the same, for example, changing the destination from Amazon S3 to Amazon Redshift,
* Kinesis Data Firehose does not merge any parameters. In this case, all parameters must be specified.
*
*
* Kinesis Data Firehose uses CurrentDeliveryStreamVersionId
to avoid race conditions and conflicting
* merges. This is a required field, and the service updates the configuration only if the existing configuration
* has a version ID that matches. After the update is applied successfully, the version ID is updated, and can be
* retrieved using DescribeDeliveryStream. Use the new version ID to set
* CurrentDeliveryStreamVersionId
in the next call.
*
*
* @param updateDestinationRequest
* @return Result of the UpdateDestination operation returned by the service.
* @throws InvalidArgumentException
* The specified input parameter has a value that is not valid.
* @throws ResourceInUseException
* The resource is already in use and not available for this operation.
* @throws ResourceNotFoundException
* The specified resource could not be found.
* @throws ConcurrentModificationException
* Another modification has already happened. Fetch VersionId
again and use it to update the
* destination.
* @sample AmazonKinesisFirehose.UpdateDestination
* @see AWS API
* Documentation
*/
@Override
public UpdateDestinationResult updateDestination(UpdateDestinationRequest request) {
request = beforeClientExecution(request);
return executeUpdateDestination(request);
}
@SdkInternalApi
final UpdateDestinationResult executeUpdateDestination(UpdateDestinationRequest updateDestinationRequest) {
ExecutionContext executionContext = createExecutionContext(updateDestinationRequest);
AWSRequestMetrics awsRequestMetrics = executionContext.getAwsRequestMetrics();
awsRequestMetrics.startEvent(Field.ClientExecuteTime);
Request request = null;
Response response = null;
try {
awsRequestMetrics.startEvent(Field.RequestMarshallTime);
try {
request = new UpdateDestinationRequestProtocolMarshaller(protocolFactory).marshall(super.beforeMarshalling(updateDestinationRequest));
// Binds the request metrics to the current request.
request.setAWSRequestMetrics(awsRequestMetrics);
request.addHandlerContext(HandlerContextKey.CLIENT_ENDPOINT, endpoint);
request.addHandlerContext(HandlerContextKey.ENDPOINT_OVERRIDDEN, isEndpointOverridden());
request.addHandlerContext(HandlerContextKey.SIGNING_REGION, getSigningRegion());
request.addHandlerContext(HandlerContextKey.SERVICE_ID, "Firehose");
request.addHandlerContext(HandlerContextKey.OPERATION_NAME, "UpdateDestination");
request.addHandlerContext(HandlerContextKey.ADVANCED_CONFIG, advancedConfig);
} finally {
awsRequestMetrics.endEvent(Field.RequestMarshallTime);
}
HttpResponseHandler> responseHandler = protocolFactory.createResponseHandler(
new JsonOperationMetadata().withPayloadJson(true).withHasStreamingSuccessResponse(false), new UpdateDestinationResultJsonUnmarshaller());
response = invoke(request, responseHandler, executionContext);
return response.getAwsResponse();
} finally {
endClientExecution(awsRequestMetrics, request, response);
}
}
/**
* Returns additional metadata for a previously executed successful, request, typically used for debugging issues
* where a service isn't acting as expected. This data isn't considered part of the result data returned by an
* operation, so it's available through this separate, diagnostic interface.
*
* Response metadata is only cached for a limited period of time, so if you need to access this extra diagnostic
* information for an executed request, you should use this method to retrieve it as soon as possible after
* executing the request.
*
* @param request
* The originally executed request
*
* @return The response metadata for the specified request, or null if none is available.
*/
public ResponseMetadata getCachedResponseMetadata(AmazonWebServiceRequest request) {
return client.getResponseMetadataForRequest(request);
}
/**
* Normal invoke with authentication. Credentials are required and may be overriden at the request level.
**/
private Response invoke(Request request, HttpResponseHandler> responseHandler,
ExecutionContext executionContext) {
return invoke(request, responseHandler, executionContext, null, null);
}
/**
* Normal invoke with authentication. Credentials are required and may be overriden at the request level.
**/
private Response invoke(Request request, HttpResponseHandler> responseHandler,
ExecutionContext executionContext, URI cachedEndpoint, URI uriFromEndpointTrait) {
executionContext.setCredentialsProvider(CredentialUtils.getCredentialsProvider(request.getOriginalRequest(), awsCredentialsProvider));
return doInvoke(request, responseHandler, executionContext, cachedEndpoint, uriFromEndpointTrait);
}
/**
* Invoke with no authentication. Credentials are not required and any credentials set on the client or request will
* be ignored for this operation.
**/
private Response anonymousInvoke(Request request,
HttpResponseHandler> responseHandler, ExecutionContext executionContext) {
return doInvoke(request, responseHandler, executionContext, null, null);
}
/**
* Invoke the request using the http client. Assumes credentials (or lack thereof) have been configured in the
* ExecutionContext beforehand.
**/
private Response doInvoke(Request request, HttpResponseHandler> responseHandler,
ExecutionContext executionContext, URI discoveredEndpoint, URI uriFromEndpointTrait) {
if (discoveredEndpoint != null) {
request.setEndpoint(discoveredEndpoint);
request.getOriginalRequest().getRequestClientOptions().appendUserAgent("endpoint-discovery");
} else if (uriFromEndpointTrait != null) {
request.setEndpoint(uriFromEndpointTrait);
} else {
request.setEndpoint(endpoint);
}
request.setTimeOffset(timeOffset);
HttpResponseHandler errorResponseHandler = protocolFactory.createErrorResponseHandler(new JsonErrorResponseMetadata());
return client.execute(request, responseHandler, errorResponseHandler, executionContext);
}
@com.amazonaws.annotation.SdkInternalApi
static com.amazonaws.protocol.json.SdkJsonProtocolFactory getProtocolFactory() {
return protocolFactory;
}
@Override
public void shutdown() {
super.shutdown();
}
}