All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.amazonaws.services.elasticmapreduce.model.JobFlowInstancesConfig Maven / Gradle / Ivy

/*
 * Copyright 2010-2014 Amazon.com, Inc. or its affiliates. All Rights Reserved.
 * 
 * Licensed under the Apache License, Version 2.0 (the "License").
 * You may not use this file except in compliance with the License.
 * A copy of the License is located at
 * 
 *  http://aws.amazon.com/apache2.0
 * 
 * or in the "license" file accompanying this file. This file is distributed
 * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
 * express or implied. See the License for the specific language governing
 * permissions and limitations under the License.
 */
package com.amazonaws.services.elasticmapreduce.model;

import java.io.Serializable;

/**
 * 

* A description of the Amazon EC2 instance running the job flow. A valid * JobFlowInstancesConfig must contain at least InstanceGroups, which is * the recommended configuration. However, a valid alternative is to have * MasterInstanceType, SlaveInstanceType, and InstanceCount (all three * must be present). *

*/ public class JobFlowInstancesConfig implements Serializable { /** * The EC2 instance type of the master node. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
*/ private String masterInstanceType; /** * The EC2 instance type of the slave nodes. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
*/ private String slaveInstanceType; /** * The number of Amazon EC2 instances used to execute the job flow. */ private Integer instanceCount; /** * Configuration for the job flow's instance groups. */ private com.amazonaws.internal.ListWithAutoConstructFlag instanceGroups; /** * The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
*/ private String ec2KeyName; /** * The Availability Zone the job flow will run in. */ private PlacementType placement; /** * Specifies whether the job flow should terminate after completing all * steps. */ private Boolean keepJobFlowAliveWhenNoSteps; /** * Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. */ private Boolean terminationProtected; /** * The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
*/ private String hadoopVersion; /** * To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
*/ private String ec2SubnetId; /** * Default constructor for a new JobFlowInstancesConfig object. Callers should use the * setter or fluent setter (with...) methods to initialize this object after creating it. */ public JobFlowInstancesConfig() {} /** * The EC2 instance type of the master node. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @return The EC2 instance type of the master node. */ public String getMasterInstanceType() { return masterInstanceType; } /** * The EC2 instance type of the master node. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param masterInstanceType The EC2 instance type of the master node. */ public void setMasterInstanceType(String masterInstanceType) { this.masterInstanceType = masterInstanceType; } /** * The EC2 instance type of the master node. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param masterInstanceType The EC2 instance type of the master node. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withMasterInstanceType(String masterInstanceType) { this.masterInstanceType = masterInstanceType; return this; } /** * The EC2 instance type of the slave nodes. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @return The EC2 instance type of the slave nodes. */ public String getSlaveInstanceType() { return slaveInstanceType; } /** * The EC2 instance type of the slave nodes. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param slaveInstanceType The EC2 instance type of the slave nodes. */ public void setSlaveInstanceType(String slaveInstanceType) { this.slaveInstanceType = slaveInstanceType; } /** * The EC2 instance type of the slave nodes. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 1 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param slaveInstanceType The EC2 instance type of the slave nodes. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withSlaveInstanceType(String slaveInstanceType) { this.slaveInstanceType = slaveInstanceType; return this; } /** * The number of Amazon EC2 instances used to execute the job flow. * * @return The number of Amazon EC2 instances used to execute the job flow. */ public Integer getInstanceCount() { return instanceCount; } /** * The number of Amazon EC2 instances used to execute the job flow. * * @param instanceCount The number of Amazon EC2 instances used to execute the job flow. */ public void setInstanceCount(Integer instanceCount) { this.instanceCount = instanceCount; } /** * The number of Amazon EC2 instances used to execute the job flow. *

* Returns a reference to this object so that method calls can be chained together. * * @param instanceCount The number of Amazon EC2 instances used to execute the job flow. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withInstanceCount(Integer instanceCount) { this.instanceCount = instanceCount; return this; } /** * Configuration for the job flow's instance groups. * * @return Configuration for the job flow's instance groups. */ public java.util.List getInstanceGroups() { if (instanceGroups == null) { instanceGroups = new com.amazonaws.internal.ListWithAutoConstructFlag(); instanceGroups.setAutoConstruct(true); } return instanceGroups; } /** * Configuration for the job flow's instance groups. * * @param instanceGroups Configuration for the job flow's instance groups. */ public void setInstanceGroups(java.util.Collection instanceGroups) { if (instanceGroups == null) { this.instanceGroups = null; return; } com.amazonaws.internal.ListWithAutoConstructFlag instanceGroupsCopy = new com.amazonaws.internal.ListWithAutoConstructFlag(instanceGroups.size()); instanceGroupsCopy.addAll(instanceGroups); this.instanceGroups = instanceGroupsCopy; } /** * Configuration for the job flow's instance groups. *

* Returns a reference to this object so that method calls can be chained together. * * @param instanceGroups Configuration for the job flow's instance groups. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withInstanceGroups(InstanceGroupConfig... instanceGroups) { if (getInstanceGroups() == null) setInstanceGroups(new java.util.ArrayList(instanceGroups.length)); for (InstanceGroupConfig value : instanceGroups) { getInstanceGroups().add(value); } return this; } /** * Configuration for the job flow's instance groups. *

* Returns a reference to this object so that method calls can be chained together. * * @param instanceGroups Configuration for the job flow's instance groups. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withInstanceGroups(java.util.Collection instanceGroups) { if (instanceGroups == null) { this.instanceGroups = null; } else { com.amazonaws.internal.ListWithAutoConstructFlag instanceGroupsCopy = new com.amazonaws.internal.ListWithAutoConstructFlag(instanceGroups.size()); instanceGroupsCopy.addAll(instanceGroups); this.instanceGroups = instanceGroupsCopy; } return this; } /** * The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @return The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." */ public String getEc2KeyName() { return ec2KeyName; } /** * The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param ec2KeyName The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." */ public void setEc2KeyName(String ec2KeyName) { this.ec2KeyName = ec2KeyName; } /** * The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param ec2KeyName The name of the Amazon EC2 key pair that can be used to ssh to the * master node as the user called "hadoop." * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withEc2KeyName(String ec2KeyName) { this.ec2KeyName = ec2KeyName; return this; } /** * The Availability Zone the job flow will run in. * * @return The Availability Zone the job flow will run in. */ public PlacementType getPlacement() { return placement; } /** * The Availability Zone the job flow will run in. * * @param placement The Availability Zone the job flow will run in. */ public void setPlacement(PlacementType placement) { this.placement = placement; } /** * The Availability Zone the job flow will run in. *

* Returns a reference to this object so that method calls can be chained together. * * @param placement The Availability Zone the job flow will run in. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withPlacement(PlacementType placement) { this.placement = placement; return this; } /** * Specifies whether the job flow should terminate after completing all * steps. * * @return Specifies whether the job flow should terminate after completing all * steps. */ public Boolean isKeepJobFlowAliveWhenNoSteps() { return keepJobFlowAliveWhenNoSteps; } /** * Specifies whether the job flow should terminate after completing all * steps. * * @param keepJobFlowAliveWhenNoSteps Specifies whether the job flow should terminate after completing all * steps. */ public void setKeepJobFlowAliveWhenNoSteps(Boolean keepJobFlowAliveWhenNoSteps) { this.keepJobFlowAliveWhenNoSteps = keepJobFlowAliveWhenNoSteps; } /** * Specifies whether the job flow should terminate after completing all * steps. *

* Returns a reference to this object so that method calls can be chained together. * * @param keepJobFlowAliveWhenNoSteps Specifies whether the job flow should terminate after completing all * steps. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withKeepJobFlowAliveWhenNoSteps(Boolean keepJobFlowAliveWhenNoSteps) { this.keepJobFlowAliveWhenNoSteps = keepJobFlowAliveWhenNoSteps; return this; } /** * Specifies whether the job flow should terminate after completing all * steps. * * @return Specifies whether the job flow should terminate after completing all * steps. */ public Boolean getKeepJobFlowAliveWhenNoSteps() { return keepJobFlowAliveWhenNoSteps; } /** * Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. * * @return Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. */ public Boolean isTerminationProtected() { return terminationProtected; } /** * Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. * * @param terminationProtected Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. */ public void setTerminationProtected(Boolean terminationProtected) { this.terminationProtected = terminationProtected; } /** * Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. *

* Returns a reference to this object so that method calls can be chained together. * * @param terminationProtected Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withTerminationProtected(Boolean terminationProtected) { this.terminationProtected = terminationProtected; return this; } /** * Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. * * @return Specifies whether to lock the job flow to prevent the Amazon EC2 * instances from being terminated by API call, user intervention, or in * the event of a job flow error. */ public Boolean getTerminationProtected() { return terminationProtected; } /** * The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @return The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. */ public String getHadoopVersion() { return hadoopVersion; } /** * The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param hadoopVersion The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. */ public void setHadoopVersion(String hadoopVersion) { this.hadoopVersion = hadoopVersion; } /** * The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param hadoopVersion The Hadoop version for the job flow. Valid inputs are "0.18", "0.20", * or "0.20.205". If you do not set this value, the default of 0.18 is * used, unless the AmiVersion parameter is set in the RunJobFlow call, * in which case the default version of Hadoop for that AMI version is * used. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withHadoopVersion(String hadoopVersion) { this.hadoopVersion = hadoopVersion; return this; } /** * To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @return To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. */ public String getEc2SubnetId() { return ec2SubnetId; } /** * To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param ec2SubnetId To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. */ public void setEc2SubnetId(String ec2SubnetId) { this.ec2SubnetId = ec2SubnetId; } /** * To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. *

* Returns a reference to this object so that method calls can be chained together. *

* Constraints:
* Length: 0 - 256
* Pattern: [\u0020-\uD7FF\uE000-\uFFFD\uD800\uDC00-\uDBFF\uDFFF\r\n\t]*
* * @param ec2SubnetId To launch the job flow in Amazon Virtual Private Cloud (Amazon VPC), * set this parameter to the identifier of the Amazon VPC subnet where * you want the job flow to launch. If you do not specify this value, the * job flow is launched in the normal Amazon Web Services cloud, outside * of an Amazon VPC.

Amazon VPC currently does not support cluster * compute quadruple extra large (cc1.4xlarge) instances. Thus you cannot * specify the cc1.4xlarge instance type for nodes of a job flow launched * in a Amazon VPC. * * @return A reference to this updated object so that method calls can be chained * together. */ public JobFlowInstancesConfig withEc2SubnetId(String ec2SubnetId) { this.ec2SubnetId = ec2SubnetId; return this; } /** * Returns a string representation of this object; useful for testing and * debugging. * * @return A string representation of this object. * * @see java.lang.Object#toString() */ @Override public String toString() { StringBuilder sb = new StringBuilder(); sb.append("{"); if (getMasterInstanceType() != null) sb.append("MasterInstanceType: " + getMasterInstanceType() + ","); if (getSlaveInstanceType() != null) sb.append("SlaveInstanceType: " + getSlaveInstanceType() + ","); if (getInstanceCount() != null) sb.append("InstanceCount: " + getInstanceCount() + ","); if (getInstanceGroups() != null) sb.append("InstanceGroups: " + getInstanceGroups() + ","); if (getEc2KeyName() != null) sb.append("Ec2KeyName: " + getEc2KeyName() + ","); if (getPlacement() != null) sb.append("Placement: " + getPlacement() + ","); if (isKeepJobFlowAliveWhenNoSteps() != null) sb.append("KeepJobFlowAliveWhenNoSteps: " + isKeepJobFlowAliveWhenNoSteps() + ","); if (isTerminationProtected() != null) sb.append("TerminationProtected: " + isTerminationProtected() + ","); if (getHadoopVersion() != null) sb.append("HadoopVersion: " + getHadoopVersion() + ","); if (getEc2SubnetId() != null) sb.append("Ec2SubnetId: " + getEc2SubnetId() ); sb.append("}"); return sb.toString(); } @Override public int hashCode() { final int prime = 31; int hashCode = 1; hashCode = prime * hashCode + ((getMasterInstanceType() == null) ? 0 : getMasterInstanceType().hashCode()); hashCode = prime * hashCode + ((getSlaveInstanceType() == null) ? 0 : getSlaveInstanceType().hashCode()); hashCode = prime * hashCode + ((getInstanceCount() == null) ? 0 : getInstanceCount().hashCode()); hashCode = prime * hashCode + ((getInstanceGroups() == null) ? 0 : getInstanceGroups().hashCode()); hashCode = prime * hashCode + ((getEc2KeyName() == null) ? 0 : getEc2KeyName().hashCode()); hashCode = prime * hashCode + ((getPlacement() == null) ? 0 : getPlacement().hashCode()); hashCode = prime * hashCode + ((isKeepJobFlowAliveWhenNoSteps() == null) ? 0 : isKeepJobFlowAliveWhenNoSteps().hashCode()); hashCode = prime * hashCode + ((isTerminationProtected() == null) ? 0 : isTerminationProtected().hashCode()); hashCode = prime * hashCode + ((getHadoopVersion() == null) ? 0 : getHadoopVersion().hashCode()); hashCode = prime * hashCode + ((getEc2SubnetId() == null) ? 0 : getEc2SubnetId().hashCode()); return hashCode; } @Override public boolean equals(Object obj) { if (this == obj) return true; if (obj == null) return false; if (obj instanceof JobFlowInstancesConfig == false) return false; JobFlowInstancesConfig other = (JobFlowInstancesConfig)obj; if (other.getMasterInstanceType() == null ^ this.getMasterInstanceType() == null) return false; if (other.getMasterInstanceType() != null && other.getMasterInstanceType().equals(this.getMasterInstanceType()) == false) return false; if (other.getSlaveInstanceType() == null ^ this.getSlaveInstanceType() == null) return false; if (other.getSlaveInstanceType() != null && other.getSlaveInstanceType().equals(this.getSlaveInstanceType()) == false) return false; if (other.getInstanceCount() == null ^ this.getInstanceCount() == null) return false; if (other.getInstanceCount() != null && other.getInstanceCount().equals(this.getInstanceCount()) == false) return false; if (other.getInstanceGroups() == null ^ this.getInstanceGroups() == null) return false; if (other.getInstanceGroups() != null && other.getInstanceGroups().equals(this.getInstanceGroups()) == false) return false; if (other.getEc2KeyName() == null ^ this.getEc2KeyName() == null) return false; if (other.getEc2KeyName() != null && other.getEc2KeyName().equals(this.getEc2KeyName()) == false) return false; if (other.getPlacement() == null ^ this.getPlacement() == null) return false; if (other.getPlacement() != null && other.getPlacement().equals(this.getPlacement()) == false) return false; if (other.isKeepJobFlowAliveWhenNoSteps() == null ^ this.isKeepJobFlowAliveWhenNoSteps() == null) return false; if (other.isKeepJobFlowAliveWhenNoSteps() != null && other.isKeepJobFlowAliveWhenNoSteps().equals(this.isKeepJobFlowAliveWhenNoSteps()) == false) return false; if (other.isTerminationProtected() == null ^ this.isTerminationProtected() == null) return false; if (other.isTerminationProtected() != null && other.isTerminationProtected().equals(this.isTerminationProtected()) == false) return false; if (other.getHadoopVersion() == null ^ this.getHadoopVersion() == null) return false; if (other.getHadoopVersion() != null && other.getHadoopVersion().equals(this.getHadoopVersion()) == false) return false; if (other.getEc2SubnetId() == null ^ this.getEc2SubnetId() == null) return false; if (other.getEc2SubnetId() != null && other.getEc2SubnetId().equals(this.getEc2SubnetId()) == false) return false; return true; } }





© 2015 - 2025 Weber Informatics LLC | Privacy Policy