All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.amazonaws.services.kinesis.model.PutRecordRequest Maven / Gradle / Ivy

Go to download

The AWS Java SDK for Amazon Kinesis module holds the client classes that are used for communicating with Amazon Kinesis Service

There is a newer version: 1.12.772
Show newest version
/*
 * Copyright 2010-2015 Amazon.com, Inc. or its affiliates. All Rights Reserved.
 * 
 * Licensed under the Apache License, Version 2.0 (the "License").
 * You may not use this file except in compliance with the License.
 * A copy of the License is located at
 * 
 *  http://aws.amazon.com/apache2.0
 * 
 * or in the "license" file accompanying this file. This file is distributed
 * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
 * express or implied. See the License for the specific language governing
 * permissions and limitations under the License.
 */
package com.amazonaws.services.kinesis.model;

import java.io.Serializable;

import com.amazonaws.AmazonWebServiceRequest;

/**
 * Container for the parameters to the {@link com.amazonaws.services.kinesis.AmazonKinesis#putRecord(PutRecordRequest) PutRecord operation}.
 * 

* Writes a single data record from a producer into an Amazon Kinesis * stream. Call PutRecord to send data from the producer * into the Amazon Kinesis stream for real-time ingestion and subsequent * processing, one record at a time. Each shard can support writes up to * 1,000 records per second, up to a maximum data write total of 1 MB per * second. *

*

* You must specify the name of the stream that captures, stores, and * transports the data; a partition key; and the data blob itself. *

*

* The data blob can be any type of data; for example, a segment from a * log file, geographic/location data, website clickstream data, and so * on. *

*

* The partition key is used by Amazon Kinesis to distribute data across * shards. Amazon Kinesis segregates the data records that belong to a * data stream into multiple shards, using the partition key associated * with each data record to determine which shard a given data record * belongs to. *

*

* Partition keys are Unicode strings, with a maximum length limit of 256 * characters for each key. An MD5 hash function is used to map partition * keys to 128-bit integer values and to map associated data records to * shards using the hash key ranges of the shards. You can override * hashing the partition key to determine the shard by explicitly * specifying a hash value using the ExplicitHashKey * parameter. For more information, see * Adding Data to a Stream * in the Amazon Kinesis Developer Guide . *

*

* PutRecord returns the shard ID of where the data record * was placed and the sequence number that was assigned to the data * record. *

*

* Sequence numbers generally increase over time. To guarantee strictly * increasing ordering, use the SequenceNumberForOrdering * parameter. For more information, see * Adding Data to a Stream * in the Amazon Kinesis Developer Guide . *

*

* If a PutRecord request cannot be processed because of * insufficient provisioned throughput on the shard involved in the * request, PutRecord throws * ProvisionedThroughputExceededException . *

*

* Data records are accessible for only 24 hours from the time that they * are added to an Amazon Kinesis stream. *

* * @see com.amazonaws.services.kinesis.AmazonKinesis#putRecord(PutRecordRequest) */ public class PutRecordRequest extends AmazonWebServiceRequest implements Serializable, Cloneable { /** * The name of the stream to put the data record into. *

* Constraints:
* Length: 1 - 128
* Pattern: [a-zA-Z0-9_.-]+
*/ private String streamName; /** * The data blob to put into the record, which is base64-encoded when the * blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total size * must not exceed the maximum record size (1 MB). *

* Constraints:
* Length: 0 - 1048576
*/ private java.nio.ByteBuffer data; /** * Determines which shard in the stream the data record is assigned to. * Partition keys are Unicode strings with a maximum length limit of 256 * characters for each key. Amazon Kinesis uses the partition key as * input to a hash function that maps the partition key and associated * data to a specific shard. Specifically, an MD5 hash function is used * to map partition keys to 128-bit integer values and to map associated * data records to shards. As a result of this hashing mechanism, all * data records with the same partition key will map to the same shard * within the stream. *

* Constraints:
* Length: 1 - 256
*/ private String partitionKey; /** * The hash value used to explicitly determine the shard the data record * is assigned to by overriding the partition key hash. *

* Constraints:
* Pattern: 0|([1-9]\d{0,38})
*/ private String explicitHashKey; /** * Guarantees strictly increasing sequence numbers, for puts from the * same client and to the same partition key. Usage: set the * SequenceNumberForOrdering of record n to the * sequence number of record n-1 (as returned in the result when * putting record n-1). If this parameter is not set, records will * be coarsely ordered based on arrival time. *

* Constraints:
* Pattern: 0|([1-9]\d{0,128})
*/ private String sequenceNumberForOrdering; /** * The name of the stream to put the data record into. *

* Constraints:
* Length: 1 - 128
* Pattern: [a-zA-Z0-9_.-]+
* * @return The name of the stream to put the data record into. */ public String getStreamName() { return streamName; } /** * The name of the stream to put the data record into. *

* Constraints:
* Length: 1 - 128
* Pattern: [a-zA-Z0-9_.-]+
* * @param streamName The name of the stream to put the data record into. */ public void setStreamName(String streamName) { this.streamName = streamName; } /** * The name of the stream to put the data record into. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 1 - 128
* Pattern: [a-zA-Z0-9_.-]+
* * @param streamName The name of the stream to put the data record into. * * @return A reference to this updated object so that method calls can be chained * together. */ public PutRecordRequest withStreamName(String streamName) { this.streamName = streamName; return this; } /** * The data blob to put into the record, which is base64-encoded when the * blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total size * must not exceed the maximum record size (1 MB). *

* Constraints:
* Length: 0 - 1048576
* * @return The data blob to put into the record, which is base64-encoded when the * blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total size * must not exceed the maximum record size (1 MB). */ public java.nio.ByteBuffer getData() { return data; } /** * The data blob to put into the record, which is base64-encoded when the * blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total size * must not exceed the maximum record size (1 MB). *

* Constraints:
* Length: 0 - 1048576
* * @param data The data blob to put into the record, which is base64-encoded when the * blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total size * must not exceed the maximum record size (1 MB). */ public void setData(java.nio.ByteBuffer data) { this.data = data; } /** * The data blob to put into the record, which is base64-encoded when the * blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total size * must not exceed the maximum record size (1 MB). *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 0 - 1048576
* * @param data The data blob to put into the record, which is base64-encoded when the * blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total size * must not exceed the maximum record size (1 MB). * * @return A reference to this updated object so that method calls can be chained * together. */ public PutRecordRequest withData(java.nio.ByteBuffer data) { this.data = data; return this; } /** * Determines which shard in the stream the data record is assigned to. * Partition keys are Unicode strings with a maximum length limit of 256 * characters for each key. Amazon Kinesis uses the partition key as * input to a hash function that maps the partition key and associated * data to a specific shard. Specifically, an MD5 hash function is used * to map partition keys to 128-bit integer values and to map associated * data records to shards. As a result of this hashing mechanism, all * data records with the same partition key will map to the same shard * within the stream. *

* Constraints:
* Length: 1 - 256
* * @return Determines which shard in the stream the data record is assigned to. * Partition keys are Unicode strings with a maximum length limit of 256 * characters for each key. Amazon Kinesis uses the partition key as * input to a hash function that maps the partition key and associated * data to a specific shard. Specifically, an MD5 hash function is used * to map partition keys to 128-bit integer values and to map associated * data records to shards. As a result of this hashing mechanism, all * data records with the same partition key will map to the same shard * within the stream. */ public String getPartitionKey() { return partitionKey; } /** * Determines which shard in the stream the data record is assigned to. * Partition keys are Unicode strings with a maximum length limit of 256 * characters for each key. Amazon Kinesis uses the partition key as * input to a hash function that maps the partition key and associated * data to a specific shard. Specifically, an MD5 hash function is used * to map partition keys to 128-bit integer values and to map associated * data records to shards. As a result of this hashing mechanism, all * data records with the same partition key will map to the same shard * within the stream. *

* Constraints:
* Length: 1 - 256
* * @param partitionKey Determines which shard in the stream the data record is assigned to. * Partition keys are Unicode strings with a maximum length limit of 256 * characters for each key. Amazon Kinesis uses the partition key as * input to a hash function that maps the partition key and associated * data to a specific shard. Specifically, an MD5 hash function is used * to map partition keys to 128-bit integer values and to map associated * data records to shards. As a result of this hashing mechanism, all * data records with the same partition key will map to the same shard * within the stream. */ public void setPartitionKey(String partitionKey) { this.partitionKey = partitionKey; } /** * Determines which shard in the stream the data record is assigned to. * Partition keys are Unicode strings with a maximum length limit of 256 * characters for each key. Amazon Kinesis uses the partition key as * input to a hash function that maps the partition key and associated * data to a specific shard. Specifically, an MD5 hash function is used * to map partition keys to 128-bit integer values and to map associated * data records to shards. As a result of this hashing mechanism, all * data records with the same partition key will map to the same shard * within the stream. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 1 - 256
* * @param partitionKey Determines which shard in the stream the data record is assigned to. * Partition keys are Unicode strings with a maximum length limit of 256 * characters for each key. Amazon Kinesis uses the partition key as * input to a hash function that maps the partition key and associated * data to a specific shard. Specifically, an MD5 hash function is used * to map partition keys to 128-bit integer values and to map associated * data records to shards. As a result of this hashing mechanism, all * data records with the same partition key will map to the same shard * within the stream. * * @return A reference to this updated object so that method calls can be chained * together. */ public PutRecordRequest withPartitionKey(String partitionKey) { this.partitionKey = partitionKey; return this; } /** * The hash value used to explicitly determine the shard the data record * is assigned to by overriding the partition key hash. *

* Constraints:
* Pattern: 0|([1-9]\d{0,38})
* * @return The hash value used to explicitly determine the shard the data record * is assigned to by overriding the partition key hash. */ public String getExplicitHashKey() { return explicitHashKey; } /** * The hash value used to explicitly determine the shard the data record * is assigned to by overriding the partition key hash. *

* Constraints:
* Pattern: 0|([1-9]\d{0,38})
* * @param explicitHashKey The hash value used to explicitly determine the shard the data record * is assigned to by overriding the partition key hash. */ public void setExplicitHashKey(String explicitHashKey) { this.explicitHashKey = explicitHashKey; } /** * The hash value used to explicitly determine the shard the data record * is assigned to by overriding the partition key hash. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Pattern: 0|([1-9]\d{0,38})
* * @param explicitHashKey The hash value used to explicitly determine the shard the data record * is assigned to by overriding the partition key hash. * * @return A reference to this updated object so that method calls can be chained * together. */ public PutRecordRequest withExplicitHashKey(String explicitHashKey) { this.explicitHashKey = explicitHashKey; return this; } /** * Guarantees strictly increasing sequence numbers, for puts from the * same client and to the same partition key. Usage: set the * SequenceNumberForOrdering of record n to the * sequence number of record n-1 (as returned in the result when * putting record n-1). If this parameter is not set, records will * be coarsely ordered based on arrival time. *

* Constraints:
* Pattern: 0|([1-9]\d{0,128})
* * @return Guarantees strictly increasing sequence numbers, for puts from the * same client and to the same partition key. Usage: set the * SequenceNumberForOrdering of record n to the * sequence number of record n-1 (as returned in the result when * putting record n-1). If this parameter is not set, records will * be coarsely ordered based on arrival time. */ public String getSequenceNumberForOrdering() { return sequenceNumberForOrdering; } /** * Guarantees strictly increasing sequence numbers, for puts from the * same client and to the same partition key. Usage: set the * SequenceNumberForOrdering of record n to the * sequence number of record n-1 (as returned in the result when * putting record n-1). If this parameter is not set, records will * be coarsely ordered based on arrival time. *

* Constraints:
* Pattern: 0|([1-9]\d{0,128})
* * @param sequenceNumberForOrdering Guarantees strictly increasing sequence numbers, for puts from the * same client and to the same partition key. Usage: set the * SequenceNumberForOrdering of record n to the * sequence number of record n-1 (as returned in the result when * putting record n-1). If this parameter is not set, records will * be coarsely ordered based on arrival time. */ public void setSequenceNumberForOrdering(String sequenceNumberForOrdering) { this.sequenceNumberForOrdering = sequenceNumberForOrdering; } /** * Guarantees strictly increasing sequence numbers, for puts from the * same client and to the same partition key. Usage: set the * SequenceNumberForOrdering of record n to the * sequence number of record n-1 (as returned in the result when * putting record n-1). If this parameter is not set, records will * be coarsely ordered based on arrival time. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Pattern: 0|([1-9]\d{0,128})
* * @param sequenceNumberForOrdering Guarantees strictly increasing sequence numbers, for puts from the * same client and to the same partition key. Usage: set the * SequenceNumberForOrdering of record n to the * sequence number of record n-1 (as returned in the result when * putting record n-1). If this parameter is not set, records will * be coarsely ordered based on arrival time. * * @return A reference to this updated object so that method calls can be chained * together. */ public PutRecordRequest withSequenceNumberForOrdering(String sequenceNumberForOrdering) { this.sequenceNumberForOrdering = sequenceNumberForOrdering; return this; } /** * Returns a string representation of this object; useful for testing and * debugging. * * @return A string representation of this object. * * @see java.lang.Object#toString() */ @Override public String toString() { StringBuilder sb = new StringBuilder(); sb.append("{"); if (getStreamName() != null) sb.append("StreamName: " + getStreamName() + ","); if (getData() != null) sb.append("Data: " + getData() + ","); if (getPartitionKey() != null) sb.append("PartitionKey: " + getPartitionKey() + ","); if (getExplicitHashKey() != null) sb.append("ExplicitHashKey: " + getExplicitHashKey() + ","); if (getSequenceNumberForOrdering() != null) sb.append("SequenceNumberForOrdering: " + getSequenceNumberForOrdering() ); sb.append("}"); return sb.toString(); } @Override public int hashCode() { final int prime = 31; int hashCode = 1; hashCode = prime * hashCode + ((getStreamName() == null) ? 0 : getStreamName().hashCode()); hashCode = prime * hashCode + ((getData() == null) ? 0 : getData().hashCode()); hashCode = prime * hashCode + ((getPartitionKey() == null) ? 0 : getPartitionKey().hashCode()); hashCode = prime * hashCode + ((getExplicitHashKey() == null) ? 0 : getExplicitHashKey().hashCode()); hashCode = prime * hashCode + ((getSequenceNumberForOrdering() == null) ? 0 : getSequenceNumberForOrdering().hashCode()); return hashCode; } @Override public boolean equals(Object obj) { if (this == obj) return true; if (obj == null) return false; if (obj instanceof PutRecordRequest == false) return false; PutRecordRequest other = (PutRecordRequest)obj; if (other.getStreamName() == null ^ this.getStreamName() == null) return false; if (other.getStreamName() != null && other.getStreamName().equals(this.getStreamName()) == false) return false; if (other.getData() == null ^ this.getData() == null) return false; if (other.getData() != null && other.getData().equals(this.getData()) == false) return false; if (other.getPartitionKey() == null ^ this.getPartitionKey() == null) return false; if (other.getPartitionKey() != null && other.getPartitionKey().equals(this.getPartitionKey()) == false) return false; if (other.getExplicitHashKey() == null ^ this.getExplicitHashKey() == null) return false; if (other.getExplicitHashKey() != null && other.getExplicitHashKey().equals(this.getExplicitHashKey()) == false) return false; if (other.getSequenceNumberForOrdering() == null ^ this.getSequenceNumberForOrdering() == null) return false; if (other.getSequenceNumberForOrdering() != null && other.getSequenceNumberForOrdering().equals(this.getSequenceNumberForOrdering()) == false) return false; return true; } @Override public PutRecordRequest clone() { return (PutRecordRequest) super.clone(); } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy