All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.databricks.sdk.service.jobs.RunParameters Maven / Gradle / Ivy

There is a newer version: 0.35.0
Show newest version
// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.

package com.databricks.sdk.service.jobs;

import com.databricks.sdk.support.Generated;
import com.databricks.sdk.support.ToStringer;
import com.fasterxml.jackson.annotation.JsonProperty;
import java.util.Collection;
import java.util.Map;
import java.util.Objects;

@Generated
public class RunParameters {
  /**
   * An array of commands to execute for jobs with the dbt task, for example `"dbt_commands": ["dbt
   * deps", "dbt seed", "dbt deps", "dbt seed", "dbt run"]`
   */
  @JsonProperty("dbt_commands")
  private Collection dbtCommands;

  /**
   * A list of parameters for jobs with Spark JAR tasks, for example `"jar_params": ["john doe",
   * "35"]`. The parameters are used to invoke the main function of the main class specified in the
   * Spark JAR task. If not specified upon `run-now`, it defaults to an empty list. jar_params
   * cannot be specified in conjunction with notebook_params. The JSON representation of this field
   * (for example `{"jar_params":["john doe","35"]}`) cannot exceed 10,000 bytes.
   *
   * 

Use [Task parameter variables](/jobs.html\"#parameter-variables\") to set parameters * containing information about job runs. */ @JsonProperty("jar_params") private Collection jarParams; /** * A map from keys to values for jobs with notebook task, for example `"notebook_params": {"name": * "john doe", "age": "35"}`. The map is passed to the notebook and is accessible through the * [dbutils.widgets.get] function. * *

If not specified upon `run-now`, the triggered run uses the job’s base parameters. * *

notebook_params cannot be specified in conjunction with jar_params. * *

Use [Task parameter variables] to set parameters containing information about job runs. * *

The JSON representation of this field (for example `{"notebook_params":{"name":"john * doe","age":"35"}}`) cannot exceed 10,000 bytes. * *

[Task parameter variables]: https://docs.databricks.com/jobs.html#parameter-variables * [dbutils.widgets.get]: https://docs.databricks.com/dev-tools/databricks-utils.html */ @JsonProperty("notebook_params") private Map notebookParams; /** */ @JsonProperty("pipeline_params") private PipelineParams pipelineParams; /** */ @JsonProperty("python_named_params") private Map pythonNamedParams; /** * A list of parameters for jobs with Python tasks, for example `"python_params": ["john doe", * "35"]`. The parameters are passed to Python file as command-line parameters. If specified upon * `run-now`, it would overwrite the parameters specified in job setting. The JSON representation * of this field (for example `{"python_params":["john doe","35"]}`) cannot exceed 10,000 bytes. * *

Use [Task parameter variables] to set parameters containing information about job runs. * *

Important * *

These parameters accept only Latin characters (ASCII character set). Using non-ASCII * characters returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese * kanjis, and emojis. * *

[Task parameter variables]: https://docs.databricks.com/jobs.html#parameter-variables */ @JsonProperty("python_params") private Collection pythonParams; /** * A list of parameters for jobs with spark submit task, for example `"spark_submit_params": * ["--class", "org.apache.spark.examples.SparkPi"]`. The parameters are passed to spark-submit * script as command-line parameters. If specified upon `run-now`, it would overwrite the * parameters specified in job setting. The JSON representation of this field (for example * `{"python_params":["john doe","35"]}`) cannot exceed 10,000 bytes. * *

Use [Task parameter variables] to set parameters containing information about job runs * *

Important * *

These parameters accept only Latin characters (ASCII character set). Using non-ASCII * characters returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese * kanjis, and emojis. * *

[Task parameter variables]: https://docs.databricks.com/jobs.html#parameter-variables */ @JsonProperty("spark_submit_params") private Collection sparkSubmitParams; /** * A map from keys to values for jobs with SQL task, for example `"sql_params": {"name": "john * doe", "age": "35"}`. The SQL alert task does not support custom parameters. */ @JsonProperty("sql_params") private Map sqlParams; public RunParameters setDbtCommands(Collection dbtCommands) { this.dbtCommands = dbtCommands; return this; } public Collection getDbtCommands() { return dbtCommands; } public RunParameters setJarParams(Collection jarParams) { this.jarParams = jarParams; return this; } public Collection getJarParams() { return jarParams; } public RunParameters setNotebookParams(Map notebookParams) { this.notebookParams = notebookParams; return this; } public Map getNotebookParams() { return notebookParams; } public RunParameters setPipelineParams(PipelineParams pipelineParams) { this.pipelineParams = pipelineParams; return this; } public PipelineParams getPipelineParams() { return pipelineParams; } public RunParameters setPythonNamedParams(Map pythonNamedParams) { this.pythonNamedParams = pythonNamedParams; return this; } public Map getPythonNamedParams() { return pythonNamedParams; } public RunParameters setPythonParams(Collection pythonParams) { this.pythonParams = pythonParams; return this; } public Collection getPythonParams() { return pythonParams; } public RunParameters setSparkSubmitParams(Collection sparkSubmitParams) { this.sparkSubmitParams = sparkSubmitParams; return this; } public Collection getSparkSubmitParams() { return sparkSubmitParams; } public RunParameters setSqlParams(Map sqlParams) { this.sqlParams = sqlParams; return this; } public Map getSqlParams() { return sqlParams; } @Override public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; RunParameters that = (RunParameters) o; return Objects.equals(dbtCommands, that.dbtCommands) && Objects.equals(jarParams, that.jarParams) && Objects.equals(notebookParams, that.notebookParams) && Objects.equals(pipelineParams, that.pipelineParams) && Objects.equals(pythonNamedParams, that.pythonNamedParams) && Objects.equals(pythonParams, that.pythonParams) && Objects.equals(sparkSubmitParams, that.sparkSubmitParams) && Objects.equals(sqlParams, that.sqlParams); } @Override public int hashCode() { return Objects.hash( dbtCommands, jarParams, notebookParams, pipelineParams, pythonNamedParams, pythonParams, sparkSubmitParams, sqlParams); } @Override public String toString() { return new ToStringer(RunParameters.class) .add("dbtCommands", dbtCommands) .add("jarParams", jarParams) .add("notebookParams", notebookParams) .add("pipelineParams", pipelineParams) .add("pythonNamedParams", pythonNamedParams) .add("pythonParams", pythonParams) .add("sparkSubmitParams", sparkSubmitParams) .add("sqlParams", sqlParams) .toString(); } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy