All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.amazonaws.services.elasticmapreduce.model.JobFlowInstancesConfig Maven / Gradle / Ivy

Go to download

The Amazon Web Services SDK for Java provides Java APIs for building software on AWS' cost-effective, scalable, and reliable infrastructure products. The AWS Java SDK allows developers to code against APIs for all of Amazon's infrastructure web services (Amazon S3, Amazon EC2, Amazon SQS, Amazon Relational Database Service, Amazon AutoScaling, etc).

The newest version!
/*
 * Copyright 2010-2014 Amazon.com, Inc. or its affiliates. All Rights Reserved.
 * 
 * Licensed under the Apache License, Version 2.0 (the "License").
 * You may not use this file except in compliance with the License.
 * A copy of the License is located at
 * 
 *  http://aws.amazon.com/apache2.0
 * 
 * or in the "license" file accompanying this file. This file is distributed
 * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
 * express or implied. See the License for the specific language governing
 * permissions and limitations under the License.
 */
package com.amazonaws.services.elasticmapreduce.model;

import java.io.Serializable;

/**
 * 

* A description of the Amazon EC2 instance running the job flow. *

*/ public class JobFlowInstancesConfig implements Serializable { /** * The EC2 instance type of the master node. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
*/ private String masterInstanceType; /** * The EC2 instance type of the slave nodes. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
*/ private String slaveInstanceType; /** * The number of Amazon EC2 instances used to execute the job flow. */ private Integer instanceCount; /** * Configuration for the job flow's instance groups. */ private com.amazonaws.internal.ListWithAutoConstructFlag instanceGroups; /** * The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
*/ private String ec2KeyName; /** * The Availability Zone the job flow will run in. */ private PlacementType placement; /** * Specifies whether the job flow should terminate after completing all * steps. */ private Boolean keepJobFlowAliveWhenNoSteps; /** * Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. */ private Boolean terminationProtected; /** * The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
*/ private String hadoopVersion; /** * To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
*/ private String ec2SubnetId; /** * Default constructor for a new JobFlowInstancesConfig object. Callers should use the * setter or fluent setter (with...) methods to initialize this object after creating it. */ public JobFlowInstancesConfig() {} /** * The EC2 instance type of the master node. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @return The EC2 instance type of the master node. */ public String getMasterInstanceType() { return masterInstanceType; } /** * The EC2 instance type of the master node. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param masterInstanceType The EC2 instance type of the master node. */ public void setMasterInstanceType(String masterInstanceType) { this.masterInstanceType = masterInstanceType; } /** * The EC2 instance type of the master node. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param masterInstanceType The EC2 instance type of the master node. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withMasterInstanceType(String masterInstanceType) { this.masterInstanceType = masterInstanceType; return this; } /** * The EC2 instance type of the slave nodes. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @return The EC2 instance type of the slave nodes. */ public String getSlaveInstanceType() { return slaveInstanceType; } /** * The EC2 instance type of the slave nodes. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param slaveInstanceType The EC2 instance type of the slave nodes. */ public void setSlaveInstanceType(String slaveInstanceType) { this.slaveInstanceType = slaveInstanceType; } /** * The EC2 instance type of the slave nodes. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param slaveInstanceType The EC2 instance type of the slave nodes. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withSlaveInstanceType(String slaveInstanceType) { this.slaveInstanceType = slaveInstanceType; return this; } /** * The number of Amazon EC2 instances used to execute the job flow. * * @return The number of Amazon EC2 instances used to execute the job flow. */ public Integer getInstanceCount() { return instanceCount; } /** * The number of Amazon EC2 instances used to execute the job flow. * * @param instanceCount The number of Amazon EC2 instances used to execute the job flow. */ public void setInstanceCount(Integer instanceCount) { this.instanceCount = instanceCount; } /** * The number of Amazon EC2 instances used to execute the job flow. *

* Returns a reference to this object so that method calls can be chained together. * * @param instanceCount The number of Amazon EC2 instances used to execute the job flow. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withInstanceCount(Integer instanceCount) { this.instanceCount = instanceCount; return this; } /** * Configuration for the job flow's instance groups. * * @return Configuration for the job flow's instance groups. */ public java.util.List getInstanceGroups() { if (instanceGroups == null) { instanceGroups = new com.amazonaws.internal.ListWithAutoConstructFlag(); instanceGroups.setAutoConstruct(true); } return instanceGroups; } /** * Configuration for the job flow's instance groups. * * @param instanceGroups Configuration for the job flow's instance groups. */ public void setInstanceGroups(java.util.Collection instanceGroups) { if (instanceGroups == null) { this.instanceGroups = null; return; } com.amazonaws.internal.ListWithAutoConstructFlag instanceGroupsCopy = new com.amazonaws.internal.ListWithAutoConstructFlag(instanceGroups.size()); instanceGroupsCopy.addAll(instanceGroups); this.instanceGroups = instanceGroupsCopy; } /** * Configuration for the job flow's instance groups. *

* Returns a reference to this object so that method calls can be chained together. * * @param instanceGroups Configuration for the job flow's instance groups. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withInstanceGroups(InstanceGroupConfig... instanceGroups) { if (getInstanceGroups() == null) setInstanceGroups(new java.util.ArrayList(instanceGroups.length)); for (InstanceGroupConfig value : instanceGroups) { getInstanceGroups().add(value); } return this; } /** * Configuration for the job flow's instance groups. *

* Returns a reference to this object so that method calls can be chained together. * * @param instanceGroups Configuration for the job flow's instance groups. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withInstanceGroups(java.util.Collection instanceGroups) { if (instanceGroups == null) { this.instanceGroups = null; } else { com.amazonaws.internal.ListWithAutoConstructFlag instanceGroupsCopy = new com.amazonaws.internal.ListWithAutoConstructFlag(instanceGroups.size()); instanceGroupsCopy.addAll(instanceGroups); this.instanceGroups = instanceGroupsCopy; } return this; } /** * The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @return The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." */ public String getEc2KeyName() { return ec2KeyName; } /** * The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param ec2KeyName The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." */ public void setEc2KeyName(String ec2KeyName) { this.ec2KeyName = ec2KeyName; } /** * The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param ec2KeyName The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withEc2KeyName(String ec2KeyName) { this.ec2KeyName = ec2KeyName; return this; } /** * The Availability Zone the job flow will run in. * * @return The Availability Zone the job flow will run in. */ public PlacementType getPlacement() { return placement; } /** * The Availability Zone the job flow will run in. * * @param placement The Availability Zone the job flow will run in. */ public void setPlacement(PlacementType placement) { this.placement = placement; } /** * The Availability Zone the job flow will run in. *

* Returns a reference to this object so that method calls can be chained together. * * @param placement The Availability Zone the job flow will run in. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withPlacement(PlacementType placement) { this.placement = placement; return this; } /** * Specifies whether the job flow should terminate after completing all * steps. * * @return Specifies whether the job flow should terminate after completing all * steps. */ public Boolean isKeepJobFlowAliveWhenNoSteps() { return keepJobFlowAliveWhenNoSteps; } /** * Specifies whether the job flow should terminate after completing all * steps. * * @param keepJobFlowAliveWhenNoSteps Specifies whether the job flow should terminate after completing all * steps. */ public void setKeepJobFlowAliveWhenNoSteps(Boolean keepJobFlowAliveWhenNoSteps) { this.keepJobFlowAliveWhenNoSteps = keepJobFlowAliveWhenNoSteps; } /** * Specifies whether the job flow should terminate after completing all * steps. *

* Returns a reference to this object so that method calls can be chained together. * * @param keepJobFlowAliveWhenNoSteps Specifies whether the job flow should terminate after completing all * steps. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withKeepJobFlowAliveWhenNoSteps(Boolean keepJobFlowAliveWhenNoSteps) { this.keepJobFlowAliveWhenNoSteps = keepJobFlowAliveWhenNoSteps; return this; } /** * Specifies whether the job flow should terminate after completing all * steps. * * @return Specifies whether the job flow should terminate after completing all * steps. */ public Boolean getKeepJobFlowAliveWhenNoSteps() { return keepJobFlowAliveWhenNoSteps; } /** * Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. * * @return Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. */ public Boolean isTerminationProtected() { return terminationProtected; } /** * Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. * * @param terminationProtected Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. */ public void setTerminationProtected(Boolean terminationProtected) { this.terminationProtected = terminationProtected; } /** * Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. *

* Returns a reference to this object so that method calls can be chained together. * * @param terminationProtected Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withTerminationProtected(Boolean terminationProtected) { this.terminationProtected = terminationProtected; return this; } /** * Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. * * @return Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. */ public Boolean getTerminationProtected() { return terminationProtected; } /** * The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @return The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. */ public String getHadoopVersion() { return hadoopVersion; } /** * The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param hadoopVersion The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. */ public void setHadoopVersion(String hadoopVersion) { this.hadoopVersion = hadoopVersion; } /** * The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param hadoopVersion The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withHadoopVersion(String hadoopVersion) { this.hadoopVersion = hadoopVersion; return this; } /** * To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @return To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. */ public String getEc2SubnetId() { return ec2SubnetId; } /** * To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param ec2SubnetId To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. */ public void setEc2SubnetId(String ec2SubnetId) { this.ec2SubnetId = ec2SubnetId; } /** * To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param ec2SubnetId To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withEc2SubnetId(String ec2SubnetId) { this.ec2SubnetId = ec2SubnetId; return this; } /** * Returns a string representation of this object; useful for testing and * debugging. * * @return A string representation of this object. * * @see java.lang.Object#toString() */ @Override public String toString() { StringBuilder sb = new StringBuilder(); sb.append("{"); if (getMasterInstanceType() != null) sb.append("MasterInstanceType: " + getMasterInstanceType() + ","); if (getSlaveInstanceType() != null) sb.append("SlaveInstanceType: " + getSlaveInstanceType() + ","); if (getInstanceCount() != null) sb.append("InstanceCount: " + getInstanceCount() + ","); if (getInstanceGroups() != null) sb.append("InstanceGroups: " + getInstanceGroups() + ","); if (getEc2KeyName() != null) sb.append("Ec2KeyName: " + getEc2KeyName() + ","); if (getPlacement() != null) sb.append("Placement: " + getPlacement() + ","); if (isKeepJobFlowAliveWhenNoSteps() != null) sb.append("KeepJobFlowAliveWhenNoSteps: " + isKeepJobFlowAliveWhenNoSteps() + ","); if (isTerminationProtected() != null) sb.append("TerminationProtected: " + isTerminationProtected() + ","); if (getHadoopVersion() != null) sb.append("HadoopVersion: " + getHadoopVersion() + ","); if (getEc2SubnetId() != null) sb.append("Ec2SubnetId: " + getEc2SubnetId() ); sb.append("}"); return sb.toString(); } @Override public int hashCode() { final int prime = 31; int hashCode = 1; hashCode = prime * hashCode + ((getMasterInstanceType() == null) ? 0 : getMasterInstanceType().hashCode()); hashCode = prime * hashCode + ((getSlaveInstanceType() == null) ? 0 : getSlaveInstanceType().hashCode()); hashCode = prime * hashCode + ((getInstanceCount() == null) ? 0 : getInstanceCount().hashCode()); hashCode = prime * hashCode + ((getInstanceGroups() == null) ? 0 : getInstanceGroups().hashCode()); hashCode = prime * hashCode + ((getEc2KeyName() == null) ? 0 : getEc2KeyName().hashCode()); hashCode = prime * hashCode + ((getPlacement() == null) ? 0 : getPlacement().hashCode()); hashCode = prime * hashCode + ((isKeepJobFlowAliveWhenNoSteps() == null) ? 0 : isKeepJobFlowAliveWhenNoSteps().hashCode()); hashCode = prime * hashCode + ((isTerminationProtected() == null) ? 0 : isTerminationProtected().hashCode()); hashCode = prime * hashCode + ((getHadoopVersion() == null) ? 0 : getHadoopVersion().hashCode()); hashCode = prime * hashCode + ((getEc2SubnetId() == null) ? 0 : getEc2SubnetId().hashCode()); return hashCode; } @Override public boolean equals(Object obj) { if (this == obj) return true; if (obj == null) return false; if (obj instanceof JobFlowInstancesConfig == false) return false; JobFlowInstancesConfig other = (JobFlowInstancesConfig)obj; if (other.getMasterInstanceType() == null ^ this.getMasterInstanceType() == null) return false; if (other.getMasterInstanceType() != null && other.getMasterInstanceType().equals(this.getMasterInstanceType()) == false) return false; if (other.getSlaveInstanceType() == null ^ this.getSlaveInstanceType() == null) return false; if (other.getSlaveInstanceType() != null && other.getSlaveInstanceType().equals(this.getSlaveInstanceType()) == false) return false; if (other.getInstanceCount() == null ^ this.getInstanceCount() == null) return false; if (other.getInstanceCount() != null && other.getInstanceCount().equals(this.getInstanceCount()) == false) return false; if (other.getInstanceGroups() == null ^ this.getInstanceGroups() == null) return false; if (other.getInstanceGroups() != null && other.getInstanceGroups().equals(this.getInstanceGroups()) == false) return false; if (other.getEc2KeyName() == null ^ this.getEc2KeyName() == null) return false; if (other.getEc2KeyName() != null && other.getEc2KeyName().equals(this.getEc2KeyName()) == false) return false; if (other.getPlacement() == null ^ this.getPlacement() == null) return false; if (other.getPlacement() != null && other.getPlacement().equals(this.getPlacement()) == false) return false; if (other.isKeepJobFlowAliveWhenNoSteps() == null ^ this.isKeepJobFlowAliveWhenNoSteps() == null) return false; if (other.isKeepJobFlowAliveWhenNoSteps() != null && other.isKeepJobFlowAliveWhenNoSteps().equals(this.isKeepJobFlowAliveWhenNoSteps()) == false) return false; if (other.isTerminationProtected() == null ^ this.isTerminationProtected() == null) return false; if (other.isTerminationProtected() != null && other.isTerminationProtected().equals(this.isTerminationProtected()) == false) return false; if (other.getHadoopVersion() == null ^ this.getHadoopVersion() == null) return false; if (other.getHadoopVersion() != null && other.getHadoopVersion().equals(this.getHadoopVersion()) == false) return false; if (other.getEc2SubnetId() == null ^ this.getEc2SubnetId() == null) return false; if (other.getEc2SubnetId() != null && other.getEc2SubnetId().equals(this.getEc2SubnetId()) == false) return false; return true; } }





© 2015 - 2025 Weber Informatics LLC | Privacy Policy