All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.amazonaws.services.appflow.model.SalesforceDestinationProperties Maven / Gradle / Ivy

Go to download

The AWS Java SDK for Amazon Appflow module holds the client classes that are used for communicating with Amazon Appflow Service

There is a newer version: 1.12.778
Show newest version
/*
 * Copyright 2019-2024 Amazon.com, Inc. or its affiliates. All Rights Reserved.
 * 
 * Licensed under the Apache License, Version 2.0 (the "License"). You may not use this file except in compliance with
 * the License. A copy of the License is located at
 * 
 * http://aws.amazon.com/apache2.0
 * 
 * or in the "license" file accompanying this file. This file is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
 * CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions
 * and limitations under the License.
 */
package com.amazonaws.services.appflow.model;

import java.io.Serializable;
import javax.annotation.Generated;
import com.amazonaws.protocol.StructuredPojo;
import com.amazonaws.protocol.ProtocolMarshaller;

/**
 * 

* The properties that are applied when Salesforce is being used as a destination. *

* * @see AWS API Documentation */ @Generated("com.amazonaws:aws-java-sdk-code-generator") public class SalesforceDestinationProperties implements Serializable, Cloneable, StructuredPojo { /** *

* The object specified in the Salesforce flow destination. *

*/ private String object; /** *

* The name of the field that Amazon AppFlow uses as an ID when performing a write operation such as update or * delete. *

*/ private java.util.List idFieldNames; /** *

* The settings that determine how Amazon AppFlow handles an error when placing data in the Salesforce destination. * For example, this setting would determine if the flow should fail after one insertion error, or continue and * attempt to insert every record regardless of the initial failure. ErrorHandlingConfig is a part of * the destination connector details. *

*/ private ErrorHandlingConfig errorHandlingConfig; /** *

* This specifies the type of write operation to be performed in Salesforce. When the value is UPSERT, * then idFieldNames is required. *

*/ private String writeOperationType; /** *

* Specifies which Salesforce API is used by Amazon AppFlow when your flow transfers data to Salesforce. *

*
*
AUTOMATIC
*
*

* The default. Amazon AppFlow selects which API to use based on the number of records that your flow transfers to * Salesforce. If your flow transfers fewer than 1,000 records, Amazon AppFlow uses Salesforce REST API. If your * flow transfers 1,000 records or more, Amazon AppFlow uses Salesforce Bulk API 2.0. *

*

* Each of these Salesforce APIs structures data differently. If Amazon AppFlow selects the API automatically, be * aware that, for recurring flows, the data output might vary from one flow run to the next. For example, if a flow * runs daily, it might use REST API on one day to transfer 900 records, and it might use Bulk API 2.0 on the next * day to transfer 1,100 records. For each of these flow runs, the respective Salesforce API formats the data * differently. Some of the differences include how dates are formatted and null values are represented. Also, Bulk * API 2.0 doesn't transfer Salesforce compound fields. *

*

* By choosing this option, you optimize flow performance for both small and large data transfers, but the tradeoff * is inconsistent formatting in the output. *

*
*
BULKV2
*
*

* Amazon AppFlow uses only Salesforce Bulk API 2.0. This API runs asynchronous data transfers, and it's optimal for * large sets of data. By choosing this option, you ensure that your flow writes consistent output, but you optimize * performance only for large data transfers. *

*

* Note that Bulk API 2.0 does not transfer Salesforce compound fields. *

*
*
REST_SYNC
*
*

* Amazon AppFlow uses only Salesforce REST API. By choosing this option, you ensure that your flow writes * consistent output, but you decrease performance for large data transfers that are better suited for Bulk API 2.0. * In some cases, if your flow attempts to transfer a vary large set of data, it might fail with a timed out error. *

*
*
*/ private String dataTransferApi; /** *

* The object specified in the Salesforce flow destination. *

* * @param object * The object specified in the Salesforce flow destination. */ public void setObject(String object) { this.object = object; } /** *

* The object specified in the Salesforce flow destination. *

* * @return The object specified in the Salesforce flow destination. */ public String getObject() { return this.object; } /** *

* The object specified in the Salesforce flow destination. *

* * @param object * The object specified in the Salesforce flow destination. * @return Returns a reference to this object so that method calls can be chained together. */ public SalesforceDestinationProperties withObject(String object) { setObject(object); return this; } /** *

* The name of the field that Amazon AppFlow uses as an ID when performing a write operation such as update or * delete. *

* * @return The name of the field that Amazon AppFlow uses as an ID when performing a write operation such as update * or delete. */ public java.util.List getIdFieldNames() { return idFieldNames; } /** *

* The name of the field that Amazon AppFlow uses as an ID when performing a write operation such as update or * delete. *

* * @param idFieldNames * The name of the field that Amazon AppFlow uses as an ID when performing a write operation such as update * or delete. */ public void setIdFieldNames(java.util.Collection idFieldNames) { if (idFieldNames == null) { this.idFieldNames = null; return; } this.idFieldNames = new java.util.ArrayList(idFieldNames); } /** *

* The name of the field that Amazon AppFlow uses as an ID when performing a write operation such as update or * delete. *

*

* NOTE: This method appends the values to the existing list (if any). Use * {@link #setIdFieldNames(java.util.Collection)} or {@link #withIdFieldNames(java.util.Collection)} if you want to * override the existing values. *

* * @param idFieldNames * The name of the field that Amazon AppFlow uses as an ID when performing a write operation such as update * or delete. * @return Returns a reference to this object so that method calls can be chained together. */ public SalesforceDestinationProperties withIdFieldNames(String... idFieldNames) { if (this.idFieldNames == null) { setIdFieldNames(new java.util.ArrayList(idFieldNames.length)); } for (String ele : idFieldNames) { this.idFieldNames.add(ele); } return this; } /** *

* The name of the field that Amazon AppFlow uses as an ID when performing a write operation such as update or * delete. *

* * @param idFieldNames * The name of the field that Amazon AppFlow uses as an ID when performing a write operation such as update * or delete. * @return Returns a reference to this object so that method calls can be chained together. */ public SalesforceDestinationProperties withIdFieldNames(java.util.Collection idFieldNames) { setIdFieldNames(idFieldNames); return this; } /** *

* The settings that determine how Amazon AppFlow handles an error when placing data in the Salesforce destination. * For example, this setting would determine if the flow should fail after one insertion error, or continue and * attempt to insert every record regardless of the initial failure. ErrorHandlingConfig is a part of * the destination connector details. *

* * @param errorHandlingConfig * The settings that determine how Amazon AppFlow handles an error when placing data in the Salesforce * destination. For example, this setting would determine if the flow should fail after one insertion error, * or continue and attempt to insert every record regardless of the initial failure. * ErrorHandlingConfig is a part of the destination connector details. */ public void setErrorHandlingConfig(ErrorHandlingConfig errorHandlingConfig) { this.errorHandlingConfig = errorHandlingConfig; } /** *

* The settings that determine how Amazon AppFlow handles an error when placing data in the Salesforce destination. * For example, this setting would determine if the flow should fail after one insertion error, or continue and * attempt to insert every record regardless of the initial failure. ErrorHandlingConfig is a part of * the destination connector details. *

* * @return The settings that determine how Amazon AppFlow handles an error when placing data in the Salesforce * destination. For example, this setting would determine if the flow should fail after one insertion error, * or continue and attempt to insert every record regardless of the initial failure. * ErrorHandlingConfig is a part of the destination connector details. */ public ErrorHandlingConfig getErrorHandlingConfig() { return this.errorHandlingConfig; } /** *

* The settings that determine how Amazon AppFlow handles an error when placing data in the Salesforce destination. * For example, this setting would determine if the flow should fail after one insertion error, or continue and * attempt to insert every record regardless of the initial failure. ErrorHandlingConfig is a part of * the destination connector details. *

* * @param errorHandlingConfig * The settings that determine how Amazon AppFlow handles an error when placing data in the Salesforce * destination. For example, this setting would determine if the flow should fail after one insertion error, * or continue and attempt to insert every record regardless of the initial failure. * ErrorHandlingConfig is a part of the destination connector details. * @return Returns a reference to this object so that method calls can be chained together. */ public SalesforceDestinationProperties withErrorHandlingConfig(ErrorHandlingConfig errorHandlingConfig) { setErrorHandlingConfig(errorHandlingConfig); return this; } /** *

* This specifies the type of write operation to be performed in Salesforce. When the value is UPSERT, * then idFieldNames is required. *

* * @param writeOperationType * This specifies the type of write operation to be performed in Salesforce. When the value is * UPSERT, then idFieldNames is required. * @see WriteOperationType */ public void setWriteOperationType(String writeOperationType) { this.writeOperationType = writeOperationType; } /** *

* This specifies the type of write operation to be performed in Salesforce. When the value is UPSERT, * then idFieldNames is required. *

* * @return This specifies the type of write operation to be performed in Salesforce. When the value is * UPSERT, then idFieldNames is required. * @see WriteOperationType */ public String getWriteOperationType() { return this.writeOperationType; } /** *

* This specifies the type of write operation to be performed in Salesforce. When the value is UPSERT, * then idFieldNames is required. *

* * @param writeOperationType * This specifies the type of write operation to be performed in Salesforce. When the value is * UPSERT, then idFieldNames is required. * @return Returns a reference to this object so that method calls can be chained together. * @see WriteOperationType */ public SalesforceDestinationProperties withWriteOperationType(String writeOperationType) { setWriteOperationType(writeOperationType); return this; } /** *

* This specifies the type of write operation to be performed in Salesforce. When the value is UPSERT, * then idFieldNames is required. *

* * @param writeOperationType * This specifies the type of write operation to be performed in Salesforce. When the value is * UPSERT, then idFieldNames is required. * @return Returns a reference to this object so that method calls can be chained together. * @see WriteOperationType */ public SalesforceDestinationProperties withWriteOperationType(WriteOperationType writeOperationType) { this.writeOperationType = writeOperationType.toString(); return this; } /** *

* Specifies which Salesforce API is used by Amazon AppFlow when your flow transfers data to Salesforce. *

*
*
AUTOMATIC
*
*

* The default. Amazon AppFlow selects which API to use based on the number of records that your flow transfers to * Salesforce. If your flow transfers fewer than 1,000 records, Amazon AppFlow uses Salesforce REST API. If your * flow transfers 1,000 records or more, Amazon AppFlow uses Salesforce Bulk API 2.0. *

*

* Each of these Salesforce APIs structures data differently. If Amazon AppFlow selects the API automatically, be * aware that, for recurring flows, the data output might vary from one flow run to the next. For example, if a flow * runs daily, it might use REST API on one day to transfer 900 records, and it might use Bulk API 2.0 on the next * day to transfer 1,100 records. For each of these flow runs, the respective Salesforce API formats the data * differently. Some of the differences include how dates are formatted and null values are represented. Also, Bulk * API 2.0 doesn't transfer Salesforce compound fields. *

*

* By choosing this option, you optimize flow performance for both small and large data transfers, but the tradeoff * is inconsistent formatting in the output. *

*
*
BULKV2
*
*

* Amazon AppFlow uses only Salesforce Bulk API 2.0. This API runs asynchronous data transfers, and it's optimal for * large sets of data. By choosing this option, you ensure that your flow writes consistent output, but you optimize * performance only for large data transfers. *

*

* Note that Bulk API 2.0 does not transfer Salesforce compound fields. *

*
*
REST_SYNC
*
*

* Amazon AppFlow uses only Salesforce REST API. By choosing this option, you ensure that your flow writes * consistent output, but you decrease performance for large data transfers that are better suited for Bulk API 2.0. * In some cases, if your flow attempts to transfer a vary large set of data, it might fail with a timed out error. *

*
*
* * @param dataTransferApi * Specifies which Salesforce API is used by Amazon AppFlow when your flow transfers data to Salesforce.

*
*
AUTOMATIC
*
*

* The default. Amazon AppFlow selects which API to use based on the number of records that your flow * transfers to Salesforce. If your flow transfers fewer than 1,000 records, Amazon AppFlow uses Salesforce * REST API. If your flow transfers 1,000 records or more, Amazon AppFlow uses Salesforce Bulk API 2.0. *

*

* Each of these Salesforce APIs structures data differently. If Amazon AppFlow selects the API * automatically, be aware that, for recurring flows, the data output might vary from one flow run to the * next. For example, if a flow runs daily, it might use REST API on one day to transfer 900 records, and it * might use Bulk API 2.0 on the next day to transfer 1,100 records. For each of these flow runs, the * respective Salesforce API formats the data differently. Some of the differences include how dates are * formatted and null values are represented. Also, Bulk API 2.0 doesn't transfer Salesforce compound fields. *

*

* By choosing this option, you optimize flow performance for both small and large data transfers, but the * tradeoff is inconsistent formatting in the output. *

*
*
BULKV2
*
*

* Amazon AppFlow uses only Salesforce Bulk API 2.0. This API runs asynchronous data transfers, and it's * optimal for large sets of data. By choosing this option, you ensure that your flow writes consistent * output, but you optimize performance only for large data transfers. *

*

* Note that Bulk API 2.0 does not transfer Salesforce compound fields. *

*
*
REST_SYNC
*
*

* Amazon AppFlow uses only Salesforce REST API. By choosing this option, you ensure that your flow writes * consistent output, but you decrease performance for large data transfers that are better suited for Bulk * API 2.0. In some cases, if your flow attempts to transfer a vary large set of data, it might fail with a * timed out error. *

*
* @see SalesforceDataTransferApi */ public void setDataTransferApi(String dataTransferApi) { this.dataTransferApi = dataTransferApi; } /** *

* Specifies which Salesforce API is used by Amazon AppFlow when your flow transfers data to Salesforce. *

*
*
AUTOMATIC
*
*

* The default. Amazon AppFlow selects which API to use based on the number of records that your flow transfers to * Salesforce. If your flow transfers fewer than 1,000 records, Amazon AppFlow uses Salesforce REST API. If your * flow transfers 1,000 records or more, Amazon AppFlow uses Salesforce Bulk API 2.0. *

*

* Each of these Salesforce APIs structures data differently. If Amazon AppFlow selects the API automatically, be * aware that, for recurring flows, the data output might vary from one flow run to the next. For example, if a flow * runs daily, it might use REST API on one day to transfer 900 records, and it might use Bulk API 2.0 on the next * day to transfer 1,100 records. For each of these flow runs, the respective Salesforce API formats the data * differently. Some of the differences include how dates are formatted and null values are represented. Also, Bulk * API 2.0 doesn't transfer Salesforce compound fields. *

*

* By choosing this option, you optimize flow performance for both small and large data transfers, but the tradeoff * is inconsistent formatting in the output. *

*
*
BULKV2
*
*

* Amazon AppFlow uses only Salesforce Bulk API 2.0. This API runs asynchronous data transfers, and it's optimal for * large sets of data. By choosing this option, you ensure that your flow writes consistent output, but you optimize * performance only for large data transfers. *

*

* Note that Bulk API 2.0 does not transfer Salesforce compound fields. *

*
*
REST_SYNC
*
*

* Amazon AppFlow uses only Salesforce REST API. By choosing this option, you ensure that your flow writes * consistent output, but you decrease performance for large data transfers that are better suited for Bulk API 2.0. * In some cases, if your flow attempts to transfer a vary large set of data, it might fail with a timed out error. *

*
*
* * @return Specifies which Salesforce API is used by Amazon AppFlow when your flow transfers data to Salesforce.

*
*
AUTOMATIC
*
*

* The default. Amazon AppFlow selects which API to use based on the number of records that your flow * transfers to Salesforce. If your flow transfers fewer than 1,000 records, Amazon AppFlow uses Salesforce * REST API. If your flow transfers 1,000 records or more, Amazon AppFlow uses Salesforce Bulk API 2.0. *

*

* Each of these Salesforce APIs structures data differently. If Amazon AppFlow selects the API * automatically, be aware that, for recurring flows, the data output might vary from one flow run to the * next. For example, if a flow runs daily, it might use REST API on one day to transfer 900 records, and it * might use Bulk API 2.0 on the next day to transfer 1,100 records. For each of these flow runs, the * respective Salesforce API formats the data differently. Some of the differences include how dates are * formatted and null values are represented. Also, Bulk API 2.0 doesn't transfer Salesforce compound * fields. *

*

* By choosing this option, you optimize flow performance for both small and large data transfers, but the * tradeoff is inconsistent formatting in the output. *

*
*
BULKV2
*
*

* Amazon AppFlow uses only Salesforce Bulk API 2.0. This API runs asynchronous data transfers, and it's * optimal for large sets of data. By choosing this option, you ensure that your flow writes consistent * output, but you optimize performance only for large data transfers. *

*

* Note that Bulk API 2.0 does not transfer Salesforce compound fields. *

*
*
REST_SYNC
*
*

* Amazon AppFlow uses only Salesforce REST API. By choosing this option, you ensure that your flow writes * consistent output, but you decrease performance for large data transfers that are better suited for Bulk * API 2.0. In some cases, if your flow attempts to transfer a vary large set of data, it might fail with a * timed out error. *

*
* @see SalesforceDataTransferApi */ public String getDataTransferApi() { return this.dataTransferApi; } /** *

* Specifies which Salesforce API is used by Amazon AppFlow when your flow transfers data to Salesforce. *

*
*
AUTOMATIC
*
*

* The default. Amazon AppFlow selects which API to use based on the number of records that your flow transfers to * Salesforce. If your flow transfers fewer than 1,000 records, Amazon AppFlow uses Salesforce REST API. If your * flow transfers 1,000 records or more, Amazon AppFlow uses Salesforce Bulk API 2.0. *

*

* Each of these Salesforce APIs structures data differently. If Amazon AppFlow selects the API automatically, be * aware that, for recurring flows, the data output might vary from one flow run to the next. For example, if a flow * runs daily, it might use REST API on one day to transfer 900 records, and it might use Bulk API 2.0 on the next * day to transfer 1,100 records. For each of these flow runs, the respective Salesforce API formats the data * differently. Some of the differences include how dates are formatted and null values are represented. Also, Bulk * API 2.0 doesn't transfer Salesforce compound fields. *

*

* By choosing this option, you optimize flow performance for both small and large data transfers, but the tradeoff * is inconsistent formatting in the output. *

*
*
BULKV2
*
*

* Amazon AppFlow uses only Salesforce Bulk API 2.0. This API runs asynchronous data transfers, and it's optimal for * large sets of data. By choosing this option, you ensure that your flow writes consistent output, but you optimize * performance only for large data transfers. *

*

* Note that Bulk API 2.0 does not transfer Salesforce compound fields. *

*
*
REST_SYNC
*
*

* Amazon AppFlow uses only Salesforce REST API. By choosing this option, you ensure that your flow writes * consistent output, but you decrease performance for large data transfers that are better suited for Bulk API 2.0. * In some cases, if your flow attempts to transfer a vary large set of data, it might fail with a timed out error. *

*
*
* * @param dataTransferApi * Specifies which Salesforce API is used by Amazon AppFlow when your flow transfers data to Salesforce.

*
*
AUTOMATIC
*
*

* The default. Amazon AppFlow selects which API to use based on the number of records that your flow * transfers to Salesforce. If your flow transfers fewer than 1,000 records, Amazon AppFlow uses Salesforce * REST API. If your flow transfers 1,000 records or more, Amazon AppFlow uses Salesforce Bulk API 2.0. *

*

* Each of these Salesforce APIs structures data differently. If Amazon AppFlow selects the API * automatically, be aware that, for recurring flows, the data output might vary from one flow run to the * next. For example, if a flow runs daily, it might use REST API on one day to transfer 900 records, and it * might use Bulk API 2.0 on the next day to transfer 1,100 records. For each of these flow runs, the * respective Salesforce API formats the data differently. Some of the differences include how dates are * formatted and null values are represented. Also, Bulk API 2.0 doesn't transfer Salesforce compound fields. *

*

* By choosing this option, you optimize flow performance for both small and large data transfers, but the * tradeoff is inconsistent formatting in the output. *

*
*
BULKV2
*
*

* Amazon AppFlow uses only Salesforce Bulk API 2.0. This API runs asynchronous data transfers, and it's * optimal for large sets of data. By choosing this option, you ensure that your flow writes consistent * output, but you optimize performance only for large data transfers. *

*

* Note that Bulk API 2.0 does not transfer Salesforce compound fields. *

*
*
REST_SYNC
*
*

* Amazon AppFlow uses only Salesforce REST API. By choosing this option, you ensure that your flow writes * consistent output, but you decrease performance for large data transfers that are better suited for Bulk * API 2.0. In some cases, if your flow attempts to transfer a vary large set of data, it might fail with a * timed out error. *

*
* @return Returns a reference to this object so that method calls can be chained together. * @see SalesforceDataTransferApi */ public SalesforceDestinationProperties withDataTransferApi(String dataTransferApi) { setDataTransferApi(dataTransferApi); return this; } /** *

* Specifies which Salesforce API is used by Amazon AppFlow when your flow transfers data to Salesforce. *

*
*
AUTOMATIC
*
*

* The default. Amazon AppFlow selects which API to use based on the number of records that your flow transfers to * Salesforce. If your flow transfers fewer than 1,000 records, Amazon AppFlow uses Salesforce REST API. If your * flow transfers 1,000 records or more, Amazon AppFlow uses Salesforce Bulk API 2.0. *

*

* Each of these Salesforce APIs structures data differently. If Amazon AppFlow selects the API automatically, be * aware that, for recurring flows, the data output might vary from one flow run to the next. For example, if a flow * runs daily, it might use REST API on one day to transfer 900 records, and it might use Bulk API 2.0 on the next * day to transfer 1,100 records. For each of these flow runs, the respective Salesforce API formats the data * differently. Some of the differences include how dates are formatted and null values are represented. Also, Bulk * API 2.0 doesn't transfer Salesforce compound fields. *

*

* By choosing this option, you optimize flow performance for both small and large data transfers, but the tradeoff * is inconsistent formatting in the output. *

*
*
BULKV2
*
*

* Amazon AppFlow uses only Salesforce Bulk API 2.0. This API runs asynchronous data transfers, and it's optimal for * large sets of data. By choosing this option, you ensure that your flow writes consistent output, but you optimize * performance only for large data transfers. *

*

* Note that Bulk API 2.0 does not transfer Salesforce compound fields. *

*
*
REST_SYNC
*
*

* Amazon AppFlow uses only Salesforce REST API. By choosing this option, you ensure that your flow writes * consistent output, but you decrease performance for large data transfers that are better suited for Bulk API 2.0. * In some cases, if your flow attempts to transfer a vary large set of data, it might fail with a timed out error. *

*
*
* * @param dataTransferApi * Specifies which Salesforce API is used by Amazon AppFlow when your flow transfers data to Salesforce.

*
*
AUTOMATIC
*
*

* The default. Amazon AppFlow selects which API to use based on the number of records that your flow * transfers to Salesforce. If your flow transfers fewer than 1,000 records, Amazon AppFlow uses Salesforce * REST API. If your flow transfers 1,000 records or more, Amazon AppFlow uses Salesforce Bulk API 2.0. *

*

* Each of these Salesforce APIs structures data differently. If Amazon AppFlow selects the API * automatically, be aware that, for recurring flows, the data output might vary from one flow run to the * next. For example, if a flow runs daily, it might use REST API on one day to transfer 900 records, and it * might use Bulk API 2.0 on the next day to transfer 1,100 records. For each of these flow runs, the * respective Salesforce API formats the data differently. Some of the differences include how dates are * formatted and null values are represented. Also, Bulk API 2.0 doesn't transfer Salesforce compound fields. *

*

* By choosing this option, you optimize flow performance for both small and large data transfers, but the * tradeoff is inconsistent formatting in the output. *

*
*
BULKV2
*
*

* Amazon AppFlow uses only Salesforce Bulk API 2.0. This API runs asynchronous data transfers, and it's * optimal for large sets of data. By choosing this option, you ensure that your flow writes consistent * output, but you optimize performance only for large data transfers. *

*

* Note that Bulk API 2.0 does not transfer Salesforce compound fields. *

*
*
REST_SYNC
*
*

* Amazon AppFlow uses only Salesforce REST API. By choosing this option, you ensure that your flow writes * consistent output, but you decrease performance for large data transfers that are better suited for Bulk * API 2.0. In some cases, if your flow attempts to transfer a vary large set of data, it might fail with a * timed out error. *

*
* @return Returns a reference to this object so that method calls can be chained together. * @see SalesforceDataTransferApi */ public SalesforceDestinationProperties withDataTransferApi(SalesforceDataTransferApi dataTransferApi) { this.dataTransferApi = dataTransferApi.toString(); return this; } /** * Returns a string representation of this object. This is useful for testing and debugging. Sensitive data will be * redacted from this string using a placeholder value. * * @return A string representation of this object. * * @see java.lang.Object#toString() */ @Override public String toString() { StringBuilder sb = new StringBuilder(); sb.append("{"); if (getObject() != null) sb.append("Object: ").append(getObject()).append(","); if (getIdFieldNames() != null) sb.append("IdFieldNames: ").append(getIdFieldNames()).append(","); if (getErrorHandlingConfig() != null) sb.append("ErrorHandlingConfig: ").append(getErrorHandlingConfig()).append(","); if (getWriteOperationType() != null) sb.append("WriteOperationType: ").append(getWriteOperationType()).append(","); if (getDataTransferApi() != null) sb.append("DataTransferApi: ").append(getDataTransferApi()); sb.append("}"); return sb.toString(); } @Override public boolean equals(Object obj) { if (this == obj) return true; if (obj == null) return false; if (obj instanceof SalesforceDestinationProperties == false) return false; SalesforceDestinationProperties other = (SalesforceDestinationProperties) obj; if (other.getObject() == null ^ this.getObject() == null) return false; if (other.getObject() != null && other.getObject().equals(this.getObject()) == false) return false; if (other.getIdFieldNames() == null ^ this.getIdFieldNames() == null) return false; if (other.getIdFieldNames() != null && other.getIdFieldNames().equals(this.getIdFieldNames()) == false) return false; if (other.getErrorHandlingConfig() == null ^ this.getErrorHandlingConfig() == null) return false; if (other.getErrorHandlingConfig() != null && other.getErrorHandlingConfig().equals(this.getErrorHandlingConfig()) == false) return false; if (other.getWriteOperationType() == null ^ this.getWriteOperationType() == null) return false; if (other.getWriteOperationType() != null && other.getWriteOperationType().equals(this.getWriteOperationType()) == false) return false; if (other.getDataTransferApi() == null ^ this.getDataTransferApi() == null) return false; if (other.getDataTransferApi() != null && other.getDataTransferApi().equals(this.getDataTransferApi()) == false) return false; return true; } @Override public int hashCode() { final int prime = 31; int hashCode = 1; hashCode = prime * hashCode + ((getObject() == null) ? 0 : getObject().hashCode()); hashCode = prime * hashCode + ((getIdFieldNames() == null) ? 0 : getIdFieldNames().hashCode()); hashCode = prime * hashCode + ((getErrorHandlingConfig() == null) ? 0 : getErrorHandlingConfig().hashCode()); hashCode = prime * hashCode + ((getWriteOperationType() == null) ? 0 : getWriteOperationType().hashCode()); hashCode = prime * hashCode + ((getDataTransferApi() == null) ? 0 : getDataTransferApi().hashCode()); return hashCode; } @Override public SalesforceDestinationProperties clone() { try { return (SalesforceDestinationProperties) super.clone(); } catch (CloneNotSupportedException e) { throw new IllegalStateException("Got a CloneNotSupportedException from Object.clone() " + "even though we're Cloneable!", e); } } @com.amazonaws.annotation.SdkInternalApi @Override public void marshall(ProtocolMarshaller protocolMarshaller) { com.amazonaws.services.appflow.model.transform.SalesforceDestinationPropertiesMarshaller.getInstance().marshall(this, protocolMarshaller); } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy