
com.databricks.sdk.service.jobs.RepairRun Maven / Gradle / Ivy
// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
package com.databricks.sdk.service.jobs;
import com.databricks.sdk.support.Generated;
import com.databricks.sdk.support.ToStringer;
import com.fasterxml.jackson.annotation.JsonProperty;
import java.util.Collection;
import java.util.Map;
import java.util.Objects;
@Generated
public class RepairRun {
/**
* An array of commands to execute for jobs with the dbt task, for example `"dbt_commands": ["dbt
* deps", "dbt seed", "dbt run"]`
*/
@JsonProperty("dbt_commands")
private Collection dbtCommands;
/**
* A list of parameters for jobs with Spark JAR tasks, for example `\"jar_params\": [\"john doe\",
* \"35\"]`. The parameters are used to invoke the main function of the main class specified in
* the Spark JAR task. If not specified upon `run-now`, it defaults to an empty list. jar_params
* cannot be specified in conjunction with notebook_params. The JSON representation of this field
* (for example `{\"jar_params\":[\"john doe\",\"35\"]}`) cannot exceed 10,000 bytes.
*
* Use [Task parameter variables](/jobs.html"#parameter-variables") to set parameters
* containing information about job runs.
*/
@JsonProperty("jar_params")
private Collection jarParams;
/**
* The ID of the latest repair. This parameter is not required when repairing a run for the first
* time, but must be provided on subsequent requests to repair the same run.
*/
@JsonProperty("latest_repair_id")
private Long latestRepairId;
/**
* A map from keys to values for jobs with notebook task, for example `\"notebook_params\":
* {\"name\": \"john doe\", \"age\": \"35\"}`. The map is passed to the notebook and is accessible
* through the [dbutils.widgets.get] function.
*
* If not specified upon `run-now`, the triggered run uses the job’s base parameters.
*
*
notebook_params cannot be specified in conjunction with jar_params.
*
*
Use [Task parameter variables] to set parameters containing information about job runs.
*
*
The JSON representation of this field (for example `{\"notebook_params\":{\"name\":\"john
* doe\",\"age\":\"35\"}}`) cannot exceed 10,000 bytes.
*
*
[Task parameter variables]: https://docs.databricks.com/jobs.html#parameter-variables
* [dbutils.widgets.get]: https://docs.databricks.com/dev-tools/databricks-utils.html
*/
@JsonProperty("notebook_params")
private Map notebookParams;
/** */
@JsonProperty("pipeline_params")
private PipelineParams pipelineParams;
/**
* A map from keys to values for jobs with Python wheel task, for example `"python_named_params":
* {"name": "task", "data": "dbfs:/path/to/data.json"}`.
*/
@JsonProperty("python_named_params")
private Map pythonNamedParams;
/**
* A list of parameters for jobs with Python tasks, for example `\"python_params\": [\"john doe\",
* \"35\"]`. The parameters are passed to Python file as command-line parameters. If specified
* upon `run-now`, it would overwrite the parameters specified in job setting. The JSON
* representation of this field (for example `{\"python_params\":[\"john doe\",\"35\"]}`) cannot
* exceed 10,000 bytes.
*
* Use [Task parameter variables] to set parameters containing information about job runs.
*
*
Important
*
*
These parameters accept only Latin characters (ASCII character set). Using non-ASCII
* characters returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese
* kanjis, and emojis.
*
*
[Task parameter variables]: https://docs.databricks.com/jobs.html#parameter-variables
*/
@JsonProperty("python_params")
private Collection pythonParams;
/**
* If true, repair all failed tasks. Only one of `rerun_tasks` or `rerun_all_failed_tasks` can be
* used.
*/
@JsonProperty("rerun_all_failed_tasks")
private Boolean rerunAllFailedTasks;
/**
* If true, repair all tasks that depend on the tasks in `rerun_tasks`, even if they were
* previously successful. Can be also used in combination with `rerun_all_failed_tasks`.
*/
@JsonProperty("rerun_dependent_tasks")
private Boolean rerunDependentTasks;
/** The task keys of the task runs to repair. */
@JsonProperty("rerun_tasks")
private Collection rerunTasks;
/** The job run ID of the run to repair. The run must not be in progress. */
@JsonProperty("run_id")
private Long runId;
/**
* A list of parameters for jobs with spark submit task, for example `\"spark_submit_params\":
* [\"--class\", \"org.apache.spark.examples.SparkPi\"]`. The parameters are passed to
* spark-submit script as command-line parameters. If specified upon `run-now`, it would overwrite
* the parameters specified in job setting. The JSON representation of this field (for example
* `{\"python_params\":[\"john doe\",\"35\"]}`) cannot exceed 10,000 bytes.
*
* Use [Task parameter variables] to set parameters containing information about job runs
*
*
Important
*
*
These parameters accept only Latin characters (ASCII character set). Using non-ASCII
* characters returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese
* kanjis, and emojis.
*
*
[Task parameter variables]: https://docs.databricks.com/jobs.html#parameter-variables
*/
@JsonProperty("spark_submit_params")
private Collection sparkSubmitParams;
/**
* A map from keys to values for jobs with SQL task, for example `"sql_params": {"name": "john
* doe", "age": "35"}`. The SQL alert task does not support custom parameters.
*/
@JsonProperty("sql_params")
private Map sqlParams;
public RepairRun setDbtCommands(Collection dbtCommands) {
this.dbtCommands = dbtCommands;
return this;
}
public Collection getDbtCommands() {
return dbtCommands;
}
public RepairRun setJarParams(Collection jarParams) {
this.jarParams = jarParams;
return this;
}
public Collection getJarParams() {
return jarParams;
}
public RepairRun setLatestRepairId(Long latestRepairId) {
this.latestRepairId = latestRepairId;
return this;
}
public Long getLatestRepairId() {
return latestRepairId;
}
public RepairRun setNotebookParams(Map notebookParams) {
this.notebookParams = notebookParams;
return this;
}
public Map getNotebookParams() {
return notebookParams;
}
public RepairRun setPipelineParams(PipelineParams pipelineParams) {
this.pipelineParams = pipelineParams;
return this;
}
public PipelineParams getPipelineParams() {
return pipelineParams;
}
public RepairRun setPythonNamedParams(Map pythonNamedParams) {
this.pythonNamedParams = pythonNamedParams;
return this;
}
public Map getPythonNamedParams() {
return pythonNamedParams;
}
public RepairRun setPythonParams(Collection pythonParams) {
this.pythonParams = pythonParams;
return this;
}
public Collection getPythonParams() {
return pythonParams;
}
public RepairRun setRerunAllFailedTasks(Boolean rerunAllFailedTasks) {
this.rerunAllFailedTasks = rerunAllFailedTasks;
return this;
}
public Boolean getRerunAllFailedTasks() {
return rerunAllFailedTasks;
}
public RepairRun setRerunDependentTasks(Boolean rerunDependentTasks) {
this.rerunDependentTasks = rerunDependentTasks;
return this;
}
public Boolean getRerunDependentTasks() {
return rerunDependentTasks;
}
public RepairRun setRerunTasks(Collection rerunTasks) {
this.rerunTasks = rerunTasks;
return this;
}
public Collection getRerunTasks() {
return rerunTasks;
}
public RepairRun setRunId(Long runId) {
this.runId = runId;
return this;
}
public Long getRunId() {
return runId;
}
public RepairRun setSparkSubmitParams(Collection sparkSubmitParams) {
this.sparkSubmitParams = sparkSubmitParams;
return this;
}
public Collection getSparkSubmitParams() {
return sparkSubmitParams;
}
public RepairRun setSqlParams(Map sqlParams) {
this.sqlParams = sqlParams;
return this;
}
public Map getSqlParams() {
return sqlParams;
}
@Override
public boolean equals(Object o) {
if (this == o) return true;
if (o == null || getClass() != o.getClass()) return false;
RepairRun that = (RepairRun) o;
return Objects.equals(dbtCommands, that.dbtCommands)
&& Objects.equals(jarParams, that.jarParams)
&& Objects.equals(latestRepairId, that.latestRepairId)
&& Objects.equals(notebookParams, that.notebookParams)
&& Objects.equals(pipelineParams, that.pipelineParams)
&& Objects.equals(pythonNamedParams, that.pythonNamedParams)
&& Objects.equals(pythonParams, that.pythonParams)
&& Objects.equals(rerunAllFailedTasks, that.rerunAllFailedTasks)
&& Objects.equals(rerunDependentTasks, that.rerunDependentTasks)
&& Objects.equals(rerunTasks, that.rerunTasks)
&& Objects.equals(runId, that.runId)
&& Objects.equals(sparkSubmitParams, that.sparkSubmitParams)
&& Objects.equals(sqlParams, that.sqlParams);
}
@Override
public int hashCode() {
return Objects.hash(
dbtCommands,
jarParams,
latestRepairId,
notebookParams,
pipelineParams,
pythonNamedParams,
pythonParams,
rerunAllFailedTasks,
rerunDependentTasks,
rerunTasks,
runId,
sparkSubmitParams,
sqlParams);
}
@Override
public String toString() {
return new ToStringer(RepairRun.class)
.add("dbtCommands", dbtCommands)
.add("jarParams", jarParams)
.add("latestRepairId", latestRepairId)
.add("notebookParams", notebookParams)
.add("pipelineParams", pipelineParams)
.add("pythonNamedParams", pythonNamedParams)
.add("pythonParams", pythonParams)
.add("rerunAllFailedTasks", rerunAllFailedTasks)
.add("rerunDependentTasks", rerunDependentTasks)
.add("rerunTasks", rerunTasks)
.add("runId", runId)
.add("sparkSubmitParams", sparkSubmitParams)
.add("sqlParams", sqlParams)
.toString();
}
}