All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.amazonaws.services.kinesis.model.PutRecordRequest Maven / Gradle / Ivy

Go to download

The AWS Java SDK for Amazon Kinesis module holds the client classes that are used for communicating with Amazon Kinesis Service

There is a newer version: 1.12.778
Show newest version
/*
 * Copyright 2019-2024 Amazon.com, Inc. or its affiliates. All Rights Reserved.
 * 
 * Licensed under the Apache License, Version 2.0 (the "License"). You may not use this file except in compliance with
 * the License. A copy of the License is located at
 * 
 * http://aws.amazon.com/apache2.0
 * 
 * or in the "license" file accompanying this file. This file is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
 * CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions
 * and limitations under the License.
 */
package com.amazonaws.services.kinesis.model;

import java.io.Serializable;
import javax.annotation.Generated;

import com.amazonaws.AmazonWebServiceRequest;

/**
 * 

* Represents the input for PutRecord. *

* * @see AWS API * Documentation */ @Generated("com.amazonaws:aws-java-sdk-code-generator") public class PutRecordRequest extends com.amazonaws.AmazonWebServiceRequest implements Serializable, Cloneable { /** *

* The name of the stream to put the data record into. *

*/ private String streamName; /** *

* The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data blob * (the payload before base64-encoding) is added to the partition key size, the total size must not exceed the * maximum record size (1 MiB). *

*/ private java.nio.ByteBuffer data; /** *

* Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings with a * maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition key as input * to a hash function that maps the partition key and associated data to a specific shard. Specifically, an MD5 hash * function is used to map partition keys to 128-bit integer values and to map associated data records to shards. As * a result of this hashing mechanism, all data records with the same partition key map to the same shard within the * stream. *

*/ private String partitionKey; /** *

* The hash value used to explicitly determine the shard the data record is assigned to by overriding the partition * key hash. *

*/ private String explicitHashKey; /** *

* Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition key. * Usage: set the SequenceNumberForOrdering of record n to the sequence number of record * n-1 (as returned in the result when putting record n-1). If this parameter is not set, records are * coarsely ordered based on arrival time. *

*/ private String sequenceNumberForOrdering; /** *

* The ARN of the stream. *

*/ private String streamARN; /** *

* The name of the stream to put the data record into. *

* * @param streamName * The name of the stream to put the data record into. */ public void setStreamName(String streamName) { this.streamName = streamName; } /** *

* The name of the stream to put the data record into. *

* * @return The name of the stream to put the data record into. */ public String getStreamName() { return this.streamName; } /** *

* The name of the stream to put the data record into. *

* * @param streamName * The name of the stream to put the data record into. * @return Returns a reference to this object so that method calls can be chained together. */ public PutRecordRequest withStreamName(String streamName) { setStreamName(streamName); return this; } /** *

* The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data blob * (the payload before base64-encoding) is added to the partition key size, the total size must not exceed the * maximum record size (1 MiB). *

*

* The AWS SDK for Java performs a Base64 encoding on this field before sending this request to the AWS service. * Users of the SDK should not perform Base64 encoding on this field. *

*

* Warning: ByteBuffers returned by the SDK are mutable. Changes to the content or position of the byte buffer will * be seen by all objects that have a reference to this object. It is recommended to call ByteBuffer.duplicate() or * ByteBuffer.asReadOnlyBuffer() before using or reading from the buffer. This behavior will be changed in a future * major version of the SDK. *

* * @param data * The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data * blob (the payload before base64-encoding) is added to the partition key size, the total size must not * exceed the maximum record size (1 MiB). */ public void setData(java.nio.ByteBuffer data) { this.data = data; } /** *

* The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data blob * (the payload before base64-encoding) is added to the partition key size, the total size must not exceed the * maximum record size (1 MiB). *

*

* {@code ByteBuffer}s are stateful. Calling their {@code get} methods changes their {@code position}. We recommend * using {@link java.nio.ByteBuffer#asReadOnlyBuffer()} to create a read-only view of the buffer with an independent * {@code position}, and calling {@code get} methods on this rather than directly on the returned {@code ByteBuffer}. * Doing so will ensure that anyone else using the {@code ByteBuffer} will not be affected by changes to the * {@code position}. *

* * @return The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data * blob (the payload before base64-encoding) is added to the partition key size, the total size must not * exceed the maximum record size (1 MiB). */ public java.nio.ByteBuffer getData() { return this.data; } /** *

* The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data blob * (the payload before base64-encoding) is added to the partition key size, the total size must not exceed the * maximum record size (1 MiB). *

*

* The AWS SDK for Java performs a Base64 encoding on this field before sending this request to the AWS service. * Users of the SDK should not perform Base64 encoding on this field. *

*

* Warning: ByteBuffers returned by the SDK are mutable. Changes to the content or position of the byte buffer will * be seen by all objects that have a reference to this object. It is recommended to call ByteBuffer.duplicate() or * ByteBuffer.asReadOnlyBuffer() before using or reading from the buffer. This behavior will be changed in a future * major version of the SDK. *

* * @param data * The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data * blob (the payload before base64-encoding) is added to the partition key size, the total size must not * exceed the maximum record size (1 MiB). * @return Returns a reference to this object so that method calls can be chained together. */ public PutRecordRequest withData(java.nio.ByteBuffer data) { setData(data); return this; } /** *

* Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings with a * maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition key as input * to a hash function that maps the partition key and associated data to a specific shard. Specifically, an MD5 hash * function is used to map partition keys to 128-bit integer values and to map associated data records to shards. As * a result of this hashing mechanism, all data records with the same partition key map to the same shard within the * stream. *

* * @param partitionKey * Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings * with a maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition * key as input to a hash function that maps the partition key and associated data to a specific shard. * Specifically, an MD5 hash function is used to map partition keys to 128-bit integer values and to map * associated data records to shards. As a result of this hashing mechanism, all data records with the same * partition key map to the same shard within the stream. */ public void setPartitionKey(String partitionKey) { this.partitionKey = partitionKey; } /** *

* Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings with a * maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition key as input * to a hash function that maps the partition key and associated data to a specific shard. Specifically, an MD5 hash * function is used to map partition keys to 128-bit integer values and to map associated data records to shards. As * a result of this hashing mechanism, all data records with the same partition key map to the same shard within the * stream. *

* * @return Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings * with a maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the * partition key as input to a hash function that maps the partition key and associated data to a specific * shard. Specifically, an MD5 hash function is used to map partition keys to 128-bit integer values and to * map associated data records to shards. As a result of this hashing mechanism, all data records with the * same partition key map to the same shard within the stream. */ public String getPartitionKey() { return this.partitionKey; } /** *

* Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings with a * maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition key as input * to a hash function that maps the partition key and associated data to a specific shard. Specifically, an MD5 hash * function is used to map partition keys to 128-bit integer values and to map associated data records to shards. As * a result of this hashing mechanism, all data records with the same partition key map to the same shard within the * stream. *

* * @param partitionKey * Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings * with a maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition * key as input to a hash function that maps the partition key and associated data to a specific shard. * Specifically, an MD5 hash function is used to map partition keys to 128-bit integer values and to map * associated data records to shards. As a result of this hashing mechanism, all data records with the same * partition key map to the same shard within the stream. * @return Returns a reference to this object so that method calls can be chained together. */ public PutRecordRequest withPartitionKey(String partitionKey) { setPartitionKey(partitionKey); return this; } /** *

* The hash value used to explicitly determine the shard the data record is assigned to by overriding the partition * key hash. *

* * @param explicitHashKey * The hash value used to explicitly determine the shard the data record is assigned to by overriding the * partition key hash. */ public void setExplicitHashKey(String explicitHashKey) { this.explicitHashKey = explicitHashKey; } /** *

* The hash value used to explicitly determine the shard the data record is assigned to by overriding the partition * key hash. *

* * @return The hash value used to explicitly determine the shard the data record is assigned to by overriding the * partition key hash. */ public String getExplicitHashKey() { return this.explicitHashKey; } /** *

* The hash value used to explicitly determine the shard the data record is assigned to by overriding the partition * key hash. *

* * @param explicitHashKey * The hash value used to explicitly determine the shard the data record is assigned to by overriding the * partition key hash. * @return Returns a reference to this object so that method calls can be chained together. */ public PutRecordRequest withExplicitHashKey(String explicitHashKey) { setExplicitHashKey(explicitHashKey); return this; } /** *

* Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition key. * Usage: set the SequenceNumberForOrdering of record n to the sequence number of record * n-1 (as returned in the result when putting record n-1). If this parameter is not set, records are * coarsely ordered based on arrival time. *

* * @param sequenceNumberForOrdering * Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition * key. Usage: set the SequenceNumberForOrdering of record n to the sequence number of * record n-1 (as returned in the result when putting record n-1). If this parameter is not * set, records are coarsely ordered based on arrival time. */ public void setSequenceNumberForOrdering(String sequenceNumberForOrdering) { this.sequenceNumberForOrdering = sequenceNumberForOrdering; } /** *

* Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition key. * Usage: set the SequenceNumberForOrdering of record n to the sequence number of record * n-1 (as returned in the result when putting record n-1). If this parameter is not set, records are * coarsely ordered based on arrival time. *

* * @return Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition * key. Usage: set the SequenceNumberForOrdering of record n to the sequence number of * record n-1 (as returned in the result when putting record n-1). If this parameter is not * set, records are coarsely ordered based on arrival time. */ public String getSequenceNumberForOrdering() { return this.sequenceNumberForOrdering; } /** *

* Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition key. * Usage: set the SequenceNumberForOrdering of record n to the sequence number of record * n-1 (as returned in the result when putting record n-1). If this parameter is not set, records are * coarsely ordered based on arrival time. *

* * @param sequenceNumberForOrdering * Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition * key. Usage: set the SequenceNumberForOrdering of record n to the sequence number of * record n-1 (as returned in the result when putting record n-1). If this parameter is not * set, records are coarsely ordered based on arrival time. * @return Returns a reference to this object so that method calls can be chained together. */ public PutRecordRequest withSequenceNumberForOrdering(String sequenceNumberForOrdering) { setSequenceNumberForOrdering(sequenceNumberForOrdering); return this; } /** *

* The ARN of the stream. *

* * @param streamARN * The ARN of the stream. */ public void setStreamARN(String streamARN) { this.streamARN = streamARN; } /** *

* The ARN of the stream. *

* * @return The ARN of the stream. */ public String getStreamARN() { return this.streamARN; } /** *

* The ARN of the stream. *

* * @param streamARN * The ARN of the stream. * @return Returns a reference to this object so that method calls can be chained together. */ public PutRecordRequest withStreamARN(String streamARN) { setStreamARN(streamARN); return this; } /** * Returns a string representation of this object. This is useful for testing and debugging. Sensitive data will be * redacted from this string using a placeholder value. * * @return A string representation of this object. * * @see java.lang.Object#toString() */ @Override public String toString() { StringBuilder sb = new StringBuilder(); sb.append("{"); if (getStreamName() != null) sb.append("StreamName: ").append(getStreamName()).append(","); if (getData() != null) sb.append("Data: ").append(getData()).append(","); if (getPartitionKey() != null) sb.append("PartitionKey: ").append(getPartitionKey()).append(","); if (getExplicitHashKey() != null) sb.append("ExplicitHashKey: ").append(getExplicitHashKey()).append(","); if (getSequenceNumberForOrdering() != null) sb.append("SequenceNumberForOrdering: ").append(getSequenceNumberForOrdering()).append(","); if (getStreamARN() != null) sb.append("StreamARN: ").append(getStreamARN()); sb.append("}"); return sb.toString(); } @Override public boolean equals(Object obj) { if (this == obj) return true; if (obj == null) return false; if (obj instanceof PutRecordRequest == false) return false; PutRecordRequest other = (PutRecordRequest) obj; if (other.getStreamName() == null ^ this.getStreamName() == null) return false; if (other.getStreamName() != null && other.getStreamName().equals(this.getStreamName()) == false) return false; if (other.getData() == null ^ this.getData() == null) return false; if (other.getData() != null && other.getData().equals(this.getData()) == false) return false; if (other.getPartitionKey() == null ^ this.getPartitionKey() == null) return false; if (other.getPartitionKey() != null && other.getPartitionKey().equals(this.getPartitionKey()) == false) return false; if (other.getExplicitHashKey() == null ^ this.getExplicitHashKey() == null) return false; if (other.getExplicitHashKey() != null && other.getExplicitHashKey().equals(this.getExplicitHashKey()) == false) return false; if (other.getSequenceNumberForOrdering() == null ^ this.getSequenceNumberForOrdering() == null) return false; if (other.getSequenceNumberForOrdering() != null && other.getSequenceNumberForOrdering().equals(this.getSequenceNumberForOrdering()) == false) return false; if (other.getStreamARN() == null ^ this.getStreamARN() == null) return false; if (other.getStreamARN() != null && other.getStreamARN().equals(this.getStreamARN()) == false) return false; return true; } @Override public int hashCode() { final int prime = 31; int hashCode = 1; hashCode = prime * hashCode + ((getStreamName() == null) ? 0 : getStreamName().hashCode()); hashCode = prime * hashCode + ((getData() == null) ? 0 : getData().hashCode()); hashCode = prime * hashCode + ((getPartitionKey() == null) ? 0 : getPartitionKey().hashCode()); hashCode = prime * hashCode + ((getExplicitHashKey() == null) ? 0 : getExplicitHashKey().hashCode()); hashCode = prime * hashCode + ((getSequenceNumberForOrdering() == null) ? 0 : getSequenceNumberForOrdering().hashCode()); hashCode = prime * hashCode + ((getStreamARN() == null) ? 0 : getStreamARN().hashCode()); return hashCode; } @Override public PutRecordRequest clone() { return (PutRecordRequest) super.clone(); } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy