All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.databricks.sdk.service.serving.ServedModelOutput Maven / Gradle / Ivy

// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.

package com.databricks.sdk.service.serving;

import com.databricks.sdk.support.Generated;
import com.databricks.sdk.support.ToStringer;
import com.fasterxml.jackson.annotation.JsonProperty;
import java.util.Map;
import java.util.Objects;

@Generated
public class ServedModelOutput {
  /** The creation timestamp of the served model in Unix time. */
  @JsonProperty("creation_timestamp")
  private Long creationTimestamp;

  /** The email of the user who created the served model. */
  @JsonProperty("creator")
  private String creator;

  /**
   * An object containing a set of optional, user-specified environment variable key-value pairs
   * used for serving this model. Note: this is an experimental feature and subject to change.
   * Example model environment variables that refer to Databricks secrets: `{"OPENAI_API_KEY":
   * "{{secrets/my_scope/my_key}}", "DATABRICKS_TOKEN": "{{secrets/my_scope2/my_key2}}"}`
   */
  @JsonProperty("environment_vars")
  private Map environmentVars;

  /** ARN of the instance profile that the served model will use to access AWS resources. */
  @JsonProperty("instance_profile_arn")
  private String instanceProfileArn;

  /**
   * The name of the model in Databricks Model Registry or the full name of the model in Unity
   * Catalog.
   */
  @JsonProperty("model_name")
  private String modelName;

  /** The version of the model in Databricks Model Registry or Unity Catalog to be served. */
  @JsonProperty("model_version")
  private String modelVersion;

  /** The name of the served model. */
  @JsonProperty("name")
  private String name;

  /** Whether the compute resources for the Served Model should scale down to zero. */
  @JsonProperty("scale_to_zero_enabled")
  private Boolean scaleToZeroEnabled;

  /** Information corresponding to the state of the Served Model. */
  @JsonProperty("state")
  private ServedModelState state;

  /**
   * The workload size of the served model. The workload size corresponds to a range of provisioned
   * concurrency that the compute will autoscale between. A single unit of provisioned concurrency
   * can process one request at a time. Valid workload sizes are "Small" (4 - 4 provisioned
   * concurrency), "Medium" (8 - 16 provisioned concurrency), and "Large" (16 - 64 provisioned
   * concurrency). If scale-to-zero is enabled, the lower bound of the provisioned concurrency for
   * each workload size will be 0.
   */
  @JsonProperty("workload_size")
  private String workloadSize;

  /**
   * The workload type of the served model. The workload type selects which type of compute to use
   * in the endpoint. The default value for this parameter is "CPU". For deep learning workloads,
   * GPU acceleration is available by selecting workload types like GPU_SMALL and others. See the
   * available [GPU types].
   *
   * 

[GPU types]: * https://docs.databricks.com/machine-learning/model-serving/create-manage-serving-endpoints.html#gpu-workload-types */ @JsonProperty("workload_type") private String workloadType; public ServedModelOutput setCreationTimestamp(Long creationTimestamp) { this.creationTimestamp = creationTimestamp; return this; } public Long getCreationTimestamp() { return creationTimestamp; } public ServedModelOutput setCreator(String creator) { this.creator = creator; return this; } public String getCreator() { return creator; } public ServedModelOutput setEnvironmentVars(Map environmentVars) { this.environmentVars = environmentVars; return this; } public Map getEnvironmentVars() { return environmentVars; } public ServedModelOutput setInstanceProfileArn(String instanceProfileArn) { this.instanceProfileArn = instanceProfileArn; return this; } public String getInstanceProfileArn() { return instanceProfileArn; } public ServedModelOutput setModelName(String modelName) { this.modelName = modelName; return this; } public String getModelName() { return modelName; } public ServedModelOutput setModelVersion(String modelVersion) { this.modelVersion = modelVersion; return this; } public String getModelVersion() { return modelVersion; } public ServedModelOutput setName(String name) { this.name = name; return this; } public String getName() { return name; } public ServedModelOutput setScaleToZeroEnabled(Boolean scaleToZeroEnabled) { this.scaleToZeroEnabled = scaleToZeroEnabled; return this; } public Boolean getScaleToZeroEnabled() { return scaleToZeroEnabled; } public ServedModelOutput setState(ServedModelState state) { this.state = state; return this; } public ServedModelState getState() { return state; } public ServedModelOutput setWorkloadSize(String workloadSize) { this.workloadSize = workloadSize; return this; } public String getWorkloadSize() { return workloadSize; } public ServedModelOutput setWorkloadType(String workloadType) { this.workloadType = workloadType; return this; } public String getWorkloadType() { return workloadType; } @Override public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; ServedModelOutput that = (ServedModelOutput) o; return Objects.equals(creationTimestamp, that.creationTimestamp) && Objects.equals(creator, that.creator) && Objects.equals(environmentVars, that.environmentVars) && Objects.equals(instanceProfileArn, that.instanceProfileArn) && Objects.equals(modelName, that.modelName) && Objects.equals(modelVersion, that.modelVersion) && Objects.equals(name, that.name) && Objects.equals(scaleToZeroEnabled, that.scaleToZeroEnabled) && Objects.equals(state, that.state) && Objects.equals(workloadSize, that.workloadSize) && Objects.equals(workloadType, that.workloadType); } @Override public int hashCode() { return Objects.hash( creationTimestamp, creator, environmentVars, instanceProfileArn, modelName, modelVersion, name, scaleToZeroEnabled, state, workloadSize, workloadType); } @Override public String toString() { return new ToStringer(ServedModelOutput.class) .add("creationTimestamp", creationTimestamp) .add("creator", creator) .add("environmentVars", environmentVars) .add("instanceProfileArn", instanceProfileArn) .add("modelName", modelName) .add("modelVersion", modelVersion) .add("name", name) .add("scaleToZeroEnabled", scaleToZeroEnabled) .add("state", state) .add("workloadSize", workloadSize) .add("workloadType", workloadType) .toString(); } }





© 2015 - 2025 Weber Informatics LLC | Privacy Policy