All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.amazonaws.services.kinesis.model.PutRecordsRequestEntry Maven / Gradle / Ivy

/*
 * Copyright 2011-2016 Amazon.com, Inc. or its affiliates. All Rights Reserved.
 * 
 * Licensed under the Apache License, Version 2.0 (the "License"). You may not
 * use this file except in compliance with the License. A copy of the License is
 * located at
 * 
 * http://aws.amazon.com/apache2.0
 * 
 * or in the "license" file accompanying this file. This file is distributed on
 * an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
 * express or implied. See the License for the specific language governing
 * permissions and limitations under the License.
 */
package com.amazonaws.services.kinesis.model;

import java.io.Serializable;

/**
 * 

* Represents the output for PutRecords. *

*/ public class PutRecordsRequestEntry implements Serializable, Cloneable { /** *

* The data blob to put into the record, which is base64-encoded when the * blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total size must * not exceed the maximum record size (1 MB). *

*/ private java.nio.ByteBuffer data; /** *

* The hash value used to determine explicitly the shard that the data * record is assigned to by overriding the partition key hash. *

*/ private String explicitHashKey; /** *

* Determines which shard in the stream the data record is assigned to. * Partition keys are Unicode strings with a maximum length limit of 256 * characters for each key. Amazon Kinesis uses the partition key as input * to a hash function that maps the partition key and associated data to a * specific shard. Specifically, an MD5 hash function is used to map * partition keys to 128-bit integer values and to map associated data * records to shards. As a result of this hashing mechanism, all data * records with the same partition key map to the same shard within the * stream. *

*/ private String partitionKey; /** *

* The data blob to put into the record, which is base64-encoded when the * blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total size must * not exceed the maximum record size (1 MB). *

*

* AWS SDK for Java performs a Base64 encoding on this field before sending * this request to AWS service by default. Users of the SDK should not * perform Base64 encoding on this field. *

*

* Warning: ByteBuffers returned by the SDK are mutable. Changes to the * content or position of the byte buffer will be seen by all objects that * have a reference to this object. It is recommended to call * ByteBuffer.duplicate() or ByteBuffer.asReadOnlyBuffer() before using or * reading from the buffer. This behavior will be changed in a future major * version of the SDK. *

* * @param data * The data blob to put into the record, which is base64-encoded when * the blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total * size must not exceed the maximum record size (1 MB). */ public void setData(java.nio.ByteBuffer data) { this.data = data; } /** *

* The data blob to put into the record, which is base64-encoded when the * blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total size must * not exceed the maximum record size (1 MB). *

*

* {@code ByteBuffer}s are stateful. Calling their {@code get} methods * changes their {@code position}. We recommend using * {@link java.nio.ByteBuffer#asReadOnlyBuffer()} to create a read-only view * of the buffer with an independent {@code position}, and calling * {@code get} methods on this rather than directly on the returned * {@code ByteBuffer}. Doing so will ensure that anyone else using the * {@code ByteBuffer} will not be affected by changes to the {@code position} * . *

* * @return The data blob to put into the record, which is base64-encoded * when the blob is serialized. When the data blob (the payload * before base64-encoding) is added to the partition key size, the * total size must not exceed the maximum record size (1 MB). */ public java.nio.ByteBuffer getData() { return this.data; } /** *

* The data blob to put into the record, which is base64-encoded when the * blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total size must * not exceed the maximum record size (1 MB). *

* * @param data * The data blob to put into the record, which is base64-encoded when * the blob is serialized. When the data blob (the payload before * base64-encoding) is added to the partition key size, the total * size must not exceed the maximum record size (1 MB). * @return Returns a reference to this object so that method calls can be * chained together. */ public PutRecordsRequestEntry withData(java.nio.ByteBuffer data) { setData(data); return this; } /** *

* The hash value used to determine explicitly the shard that the data * record is assigned to by overriding the partition key hash. *

* * @param explicitHashKey * The hash value used to determine explicitly the shard that the * data record is assigned to by overriding the partition key hash. */ public void setExplicitHashKey(String explicitHashKey) { this.explicitHashKey = explicitHashKey; } /** *

* The hash value used to determine explicitly the shard that the data * record is assigned to by overriding the partition key hash. *

* * @return The hash value used to determine explicitly the shard that the * data record is assigned to by overriding the partition key hash. */ public String getExplicitHashKey() { return this.explicitHashKey; } /** *

* The hash value used to determine explicitly the shard that the data * record is assigned to by overriding the partition key hash. *

* * @param explicitHashKey * The hash value used to determine explicitly the shard that the * data record is assigned to by overriding the partition key hash. * @return Returns a reference to this object so that method calls can be * chained together. */ public PutRecordsRequestEntry withExplicitHashKey(String explicitHashKey) { setExplicitHashKey(explicitHashKey); return this; } /** *

* Determines which shard in the stream the data record is assigned to. * Partition keys are Unicode strings with a maximum length limit of 256 * characters for each key. Amazon Kinesis uses the partition key as input * to a hash function that maps the partition key and associated data to a * specific shard. Specifically, an MD5 hash function is used to map * partition keys to 128-bit integer values and to map associated data * records to shards. As a result of this hashing mechanism, all data * records with the same partition key map to the same shard within the * stream. *

* * @param partitionKey * Determines which shard in the stream the data record is assigned * to. Partition keys are Unicode strings with a maximum length limit * of 256 characters for each key. Amazon Kinesis uses the partition * key as input to a hash function that maps the partition key and * associated data to a specific shard. Specifically, an MD5 hash * function is used to map partition keys to 128-bit integer values * and to map associated data records to shards. As a result of this * hashing mechanism, all data records with the same partition key * map to the same shard within the stream. */ public void setPartitionKey(String partitionKey) { this.partitionKey = partitionKey; } /** *

* Determines which shard in the stream the data record is assigned to. * Partition keys are Unicode strings with a maximum length limit of 256 * characters for each key. Amazon Kinesis uses the partition key as input * to a hash function that maps the partition key and associated data to a * specific shard. Specifically, an MD5 hash function is used to map * partition keys to 128-bit integer values and to map associated data * records to shards. As a result of this hashing mechanism, all data * records with the same partition key map to the same shard within the * stream. *

* * @return Determines which shard in the stream the data record is assigned * to. Partition keys are Unicode strings with a maximum length * limit of 256 characters for each key. Amazon Kinesis uses the * partition key as input to a hash function that maps the partition * key and associated data to a specific shard. Specifically, an MD5 * hash function is used to map partition keys to 128-bit integer * values and to map associated data records to shards. As a result * of this hashing mechanism, all data records with the same * partition key map to the same shard within the stream. */ public String getPartitionKey() { return this.partitionKey; } /** *

* Determines which shard in the stream the data record is assigned to. * Partition keys are Unicode strings with a maximum length limit of 256 * characters for each key. Amazon Kinesis uses the partition key as input * to a hash function that maps the partition key and associated data to a * specific shard. Specifically, an MD5 hash function is used to map * partition keys to 128-bit integer values and to map associated data * records to shards. As a result of this hashing mechanism, all data * records with the same partition key map to the same shard within the * stream. *

* * @param partitionKey * Determines which shard in the stream the data record is assigned * to. Partition keys are Unicode strings with a maximum length limit * of 256 characters for each key. Amazon Kinesis uses the partition * key as input to a hash function that maps the partition key and * associated data to a specific shard. Specifically, an MD5 hash * function is used to map partition keys to 128-bit integer values * and to map associated data records to shards. As a result of this * hashing mechanism, all data records with the same partition key * map to the same shard within the stream. * @return Returns a reference to this object so that method calls can be * chained together. */ public PutRecordsRequestEntry withPartitionKey(String partitionKey) { setPartitionKey(partitionKey); return this; } /** * Returns a string representation of this object; useful for testing and * debugging. * * @return A string representation of this object. * * @see java.lang.Object#toString() */ @Override public String toString() { StringBuilder sb = new StringBuilder(); sb.append("{"); if (getData() != null) sb.append("Data: " + getData() + ","); if (getExplicitHashKey() != null) sb.append("ExplicitHashKey: " + getExplicitHashKey() + ","); if (getPartitionKey() != null) sb.append("PartitionKey: " + getPartitionKey()); sb.append("}"); return sb.toString(); } @Override public boolean equals(Object obj) { if (this == obj) return true; if (obj == null) return false; if (obj instanceof PutRecordsRequestEntry == false) return false; PutRecordsRequestEntry other = (PutRecordsRequestEntry) obj; if (other.getData() == null ^ this.getData() == null) return false; if (other.getData() != null && other.getData().equals(this.getData()) == false) return false; if (other.getExplicitHashKey() == null ^ this.getExplicitHashKey() == null) return false; if (other.getExplicitHashKey() != null && other.getExplicitHashKey().equals(this.getExplicitHashKey()) == false) return false; if (other.getPartitionKey() == null ^ this.getPartitionKey() == null) return false; if (other.getPartitionKey() != null && other.getPartitionKey().equals(this.getPartitionKey()) == false) return false; return true; } @Override public int hashCode() { final int prime = 31; int hashCode = 1; hashCode = prime * hashCode + ((getData() == null) ? 0 : getData().hashCode()); hashCode = prime * hashCode + ((getExplicitHashKey() == null) ? 0 : getExplicitHashKey() .hashCode()); hashCode = prime * hashCode + ((getPartitionKey() == null) ? 0 : getPartitionKey() .hashCode()); return hashCode; } @Override public PutRecordsRequestEntry clone() { try { return (PutRecordsRequestEntry) super.clone(); } catch (CloneNotSupportedException e) { throw new IllegalStateException( "Got a CloneNotSupportedException from Object.clone() " + "even though we're Cloneable!", e); } } }




© 2015 - 2025 Weber Informatics LLC | Privacy Policy