All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.databricks.sdk.service.jobs.RepairRun Maven / Gradle / Ivy

// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.

package com.databricks.sdk.service.jobs;

import com.databricks.sdk.support.Generated;
import com.databricks.sdk.support.ToStringer;
import com.fasterxml.jackson.annotation.JsonProperty;
import java.util.Collection;
import java.util.Map;
import java.util.Objects;

@Generated
public class RepairRun {
  /**
   * An array of commands to execute for jobs with the dbt task, for example `"dbt_commands": ["dbt
   * deps", "dbt seed", "dbt run"]`
   */
  @JsonProperty("dbt_commands")
  private Collection dbtCommands;

  /**
   * A list of parameters for jobs with Spark JAR tasks, for example `\"jar_params\": [\"john doe\",
   * \"35\"]`. The parameters are used to invoke the main function of the main class specified in
   * the Spark JAR task. If not specified upon `run-now`, it defaults to an empty list. jar_params
   * cannot be specified in conjunction with notebook_params. The JSON representation of this field
   * (for example `{\"jar_params\":[\"john doe\",\"35\"]}`) cannot exceed 10,000 bytes.
   *
   * 

Use [Task parameter variables](/jobs.html"#parameter-variables") to set parameters * containing information about job runs. */ @JsonProperty("jar_params") private Collection jarParams; /** * The ID of the latest repair. This parameter is not required when repairing a run for the first * time, but must be provided on subsequent requests to repair the same run. */ @JsonProperty("latest_repair_id") private Long latestRepairId; /** * A map from keys to values for jobs with notebook task, for example `\"notebook_params\": * {\"name\": \"john doe\", \"age\": \"35\"}`. The map is passed to the notebook and is accessible * through the [dbutils.widgets.get] function. * *

If not specified upon `run-now`, the triggered run uses the job’s base parameters. * *

notebook_params cannot be specified in conjunction with jar_params. * *

Use [Task parameter variables] to set parameters containing information about job runs. * *

The JSON representation of this field (for example `{\"notebook_params\":{\"name\":\"john * doe\",\"age\":\"35\"}}`) cannot exceed 10,000 bytes. * *

[Task parameter variables]: https://docs.databricks.com/jobs.html#parameter-variables * [dbutils.widgets.get]: https://docs.databricks.com/dev-tools/databricks-utils.html */ @JsonProperty("notebook_params") private Map notebookParams; /** */ @JsonProperty("pipeline_params") private PipelineParams pipelineParams; /** * A map from keys to values for jobs with Python wheel task, for example `"python_named_params": * {"name": "task", "data": "dbfs:/path/to/data.json"}`. */ @JsonProperty("python_named_params") private Map pythonNamedParams; /** * A list of parameters for jobs with Python tasks, for example `\"python_params\": [\"john doe\", * \"35\"]`. The parameters are passed to Python file as command-line parameters. If specified * upon `run-now`, it would overwrite the parameters specified in job setting. The JSON * representation of this field (for example `{\"python_params\":[\"john doe\",\"35\"]}`) cannot * exceed 10,000 bytes. * *

Use [Task parameter variables] to set parameters containing information about job runs. * *

Important * *

These parameters accept only Latin characters (ASCII character set). Using non-ASCII * characters returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese * kanjis, and emojis. * *

[Task parameter variables]: https://docs.databricks.com/jobs.html#parameter-variables */ @JsonProperty("python_params") private Collection pythonParams; /** * If true, repair all failed tasks. Only one of `rerun_tasks` or `rerun_all_failed_tasks` can be * used. */ @JsonProperty("rerun_all_failed_tasks") private Boolean rerunAllFailedTasks; /** * If true, repair all tasks that depend on the tasks in `rerun_tasks`, even if they were * previously successful. Can be also used in combination with `rerun_all_failed_tasks`. */ @JsonProperty("rerun_dependent_tasks") private Boolean rerunDependentTasks; /** The task keys of the task runs to repair. */ @JsonProperty("rerun_tasks") private Collection rerunTasks; /** The job run ID of the run to repair. The run must not be in progress. */ @JsonProperty("run_id") private Long runId; /** * A list of parameters for jobs with spark submit task, for example `\"spark_submit_params\": * [\"--class\", \"org.apache.spark.examples.SparkPi\"]`. The parameters are passed to * spark-submit script as command-line parameters. If specified upon `run-now`, it would overwrite * the parameters specified in job setting. The JSON representation of this field (for example * `{\"python_params\":[\"john doe\",\"35\"]}`) cannot exceed 10,000 bytes. * *

Use [Task parameter variables] to set parameters containing information about job runs * *

Important * *

These parameters accept only Latin characters (ASCII character set). Using non-ASCII * characters returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese * kanjis, and emojis. * *

[Task parameter variables]: https://docs.databricks.com/jobs.html#parameter-variables */ @JsonProperty("spark_submit_params") private Collection sparkSubmitParams; /** * A map from keys to values for jobs with SQL task, for example `"sql_params": {"name": "john * doe", "age": "35"}`. The SQL alert task does not support custom parameters. */ @JsonProperty("sql_params") private Map sqlParams; public RepairRun setDbtCommands(Collection dbtCommands) { this.dbtCommands = dbtCommands; return this; } public Collection getDbtCommands() { return dbtCommands; } public RepairRun setJarParams(Collection jarParams) { this.jarParams = jarParams; return this; } public Collection getJarParams() { return jarParams; } public RepairRun setLatestRepairId(Long latestRepairId) { this.latestRepairId = latestRepairId; return this; } public Long getLatestRepairId() { return latestRepairId; } public RepairRun setNotebookParams(Map notebookParams) { this.notebookParams = notebookParams; return this; } public Map getNotebookParams() { return notebookParams; } public RepairRun setPipelineParams(PipelineParams pipelineParams) { this.pipelineParams = pipelineParams; return this; } public PipelineParams getPipelineParams() { return pipelineParams; } public RepairRun setPythonNamedParams(Map pythonNamedParams) { this.pythonNamedParams = pythonNamedParams; return this; } public Map getPythonNamedParams() { return pythonNamedParams; } public RepairRun setPythonParams(Collection pythonParams) { this.pythonParams = pythonParams; return this; } public Collection getPythonParams() { return pythonParams; } public RepairRun setRerunAllFailedTasks(Boolean rerunAllFailedTasks) { this.rerunAllFailedTasks = rerunAllFailedTasks; return this; } public Boolean getRerunAllFailedTasks() { return rerunAllFailedTasks; } public RepairRun setRerunDependentTasks(Boolean rerunDependentTasks) { this.rerunDependentTasks = rerunDependentTasks; return this; } public Boolean getRerunDependentTasks() { return rerunDependentTasks; } public RepairRun setRerunTasks(Collection rerunTasks) { this.rerunTasks = rerunTasks; return this; } public Collection getRerunTasks() { return rerunTasks; } public RepairRun setRunId(Long runId) { this.runId = runId; return this; } public Long getRunId() { return runId; } public RepairRun setSparkSubmitParams(Collection sparkSubmitParams) { this.sparkSubmitParams = sparkSubmitParams; return this; } public Collection getSparkSubmitParams() { return sparkSubmitParams; } public RepairRun setSqlParams(Map sqlParams) { this.sqlParams = sqlParams; return this; } public Map getSqlParams() { return sqlParams; } @Override public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; RepairRun that = (RepairRun) o; return Objects.equals(dbtCommands, that.dbtCommands) && Objects.equals(jarParams, that.jarParams) && Objects.equals(latestRepairId, that.latestRepairId) && Objects.equals(notebookParams, that.notebookParams) && Objects.equals(pipelineParams, that.pipelineParams) && Objects.equals(pythonNamedParams, that.pythonNamedParams) && Objects.equals(pythonParams, that.pythonParams) && Objects.equals(rerunAllFailedTasks, that.rerunAllFailedTasks) && Objects.equals(rerunDependentTasks, that.rerunDependentTasks) && Objects.equals(rerunTasks, that.rerunTasks) && Objects.equals(runId, that.runId) && Objects.equals(sparkSubmitParams, that.sparkSubmitParams) && Objects.equals(sqlParams, that.sqlParams); } @Override public int hashCode() { return Objects.hash( dbtCommands, jarParams, latestRepairId, notebookParams, pipelineParams, pythonNamedParams, pythonParams, rerunAllFailedTasks, rerunDependentTasks, rerunTasks, runId, sparkSubmitParams, sqlParams); } @Override public String toString() { return new ToStringer(RepairRun.class) .add("dbtCommands", dbtCommands) .add("jarParams", jarParams) .add("latestRepairId", latestRepairId) .add("notebookParams", notebookParams) .add("pipelineParams", pipelineParams) .add("pythonNamedParams", pythonNamedParams) .add("pythonParams", pythonParams) .add("rerunAllFailedTasks", rerunAllFailedTasks) .add("rerunDependentTasks", rerunDependentTasks) .add("rerunTasks", rerunTasks) .add("runId", runId) .add("sparkSubmitParams", sparkSubmitParams) .add("sqlParams", sqlParams) .toString(); } }





© 2015 - 2025 Weber Informatics LLC | Privacy Policy