com.databricks.sdk.service.serving.ServedModelOutput Maven / Gradle / Ivy
// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
package com.databricks.sdk.service.serving;
import com.databricks.sdk.support.Generated;
import com.databricks.sdk.support.ToStringer;
import com.fasterxml.jackson.annotation.JsonProperty;
import java.util.Map;
import java.util.Objects;
@Generated
public class ServedModelOutput {
/** The creation timestamp of the served model in Unix time. */
@JsonProperty("creation_timestamp")
private Long creationTimestamp;
/** The email of the user who created the served model. */
@JsonProperty("creator")
private String creator;
/**
* An object containing a set of optional, user-specified environment variable key-value pairs
* used for serving this model. Note: this is an experimental feature and subject to change.
* Example model environment variables that refer to Databricks secrets: `{"OPENAI_API_KEY":
* "{{secrets/my_scope/my_key}}", "DATABRICKS_TOKEN": "{{secrets/my_scope2/my_key2}}"}`
*/
@JsonProperty("environment_vars")
private Map environmentVars;
/** ARN of the instance profile that the served model will use to access AWS resources. */
@JsonProperty("instance_profile_arn")
private String instanceProfileArn;
/**
* The name of the model in Databricks Model Registry or the full name of the model in Unity
* Catalog.
*/
@JsonProperty("model_name")
private String modelName;
/** The version of the model in Databricks Model Registry or Unity Catalog to be served. */
@JsonProperty("model_version")
private String modelVersion;
/** The name of the served model. */
@JsonProperty("name")
private String name;
/** Whether the compute resources for the Served Model should scale down to zero. */
@JsonProperty("scale_to_zero_enabled")
private Boolean scaleToZeroEnabled;
/** Information corresponding to the state of the Served Model. */
@JsonProperty("state")
private ServedModelState state;
/**
* The workload size of the served model. The workload size corresponds to a range of provisioned
* concurrency that the compute will autoscale between. A single unit of provisioned concurrency
* can process one request at a time. Valid workload sizes are "Small" (4 - 4 provisioned
* concurrency), "Medium" (8 - 16 provisioned concurrency), and "Large" (16 - 64 provisioned
* concurrency). If scale-to-zero is enabled, the lower bound of the provisioned concurrency for
* each workload size will be 0.
*/
@JsonProperty("workload_size")
private String workloadSize;
/**
* The workload type of the served model. The workload type selects which type of compute to use
* in the endpoint. The default value for this parameter is "CPU". For deep learning workloads,
* GPU acceleration is available by selecting workload types like GPU_SMALL and others. See the
* available [GPU types].
*
* [GPU types]:
* https://docs.databricks.com/machine-learning/model-serving/create-manage-serving-endpoints.html#gpu-workload-types
*/
@JsonProperty("workload_type")
private String workloadType;
public ServedModelOutput setCreationTimestamp(Long creationTimestamp) {
this.creationTimestamp = creationTimestamp;
return this;
}
public Long getCreationTimestamp() {
return creationTimestamp;
}
public ServedModelOutput setCreator(String creator) {
this.creator = creator;
return this;
}
public String getCreator() {
return creator;
}
public ServedModelOutput setEnvironmentVars(Map environmentVars) {
this.environmentVars = environmentVars;
return this;
}
public Map getEnvironmentVars() {
return environmentVars;
}
public ServedModelOutput setInstanceProfileArn(String instanceProfileArn) {
this.instanceProfileArn = instanceProfileArn;
return this;
}
public String getInstanceProfileArn() {
return instanceProfileArn;
}
public ServedModelOutput setModelName(String modelName) {
this.modelName = modelName;
return this;
}
public String getModelName() {
return modelName;
}
public ServedModelOutput setModelVersion(String modelVersion) {
this.modelVersion = modelVersion;
return this;
}
public String getModelVersion() {
return modelVersion;
}
public ServedModelOutput setName(String name) {
this.name = name;
return this;
}
public String getName() {
return name;
}
public ServedModelOutput setScaleToZeroEnabled(Boolean scaleToZeroEnabled) {
this.scaleToZeroEnabled = scaleToZeroEnabled;
return this;
}
public Boolean getScaleToZeroEnabled() {
return scaleToZeroEnabled;
}
public ServedModelOutput setState(ServedModelState state) {
this.state = state;
return this;
}
public ServedModelState getState() {
return state;
}
public ServedModelOutput setWorkloadSize(String workloadSize) {
this.workloadSize = workloadSize;
return this;
}
public String getWorkloadSize() {
return workloadSize;
}
public ServedModelOutput setWorkloadType(String workloadType) {
this.workloadType = workloadType;
return this;
}
public String getWorkloadType() {
return workloadType;
}
@Override
public boolean equals(Object o) {
if (this == o) return true;
if (o == null || getClass() != o.getClass()) return false;
ServedModelOutput that = (ServedModelOutput) o;
return Objects.equals(creationTimestamp, that.creationTimestamp)
&& Objects.equals(creator, that.creator)
&& Objects.equals(environmentVars, that.environmentVars)
&& Objects.equals(instanceProfileArn, that.instanceProfileArn)
&& Objects.equals(modelName, that.modelName)
&& Objects.equals(modelVersion, that.modelVersion)
&& Objects.equals(name, that.name)
&& Objects.equals(scaleToZeroEnabled, that.scaleToZeroEnabled)
&& Objects.equals(state, that.state)
&& Objects.equals(workloadSize, that.workloadSize)
&& Objects.equals(workloadType, that.workloadType);
}
@Override
public int hashCode() {
return Objects.hash(
creationTimestamp,
creator,
environmentVars,
instanceProfileArn,
modelName,
modelVersion,
name,
scaleToZeroEnabled,
state,
workloadSize,
workloadType);
}
@Override
public String toString() {
return new ToStringer(ServedModelOutput.class)
.add("creationTimestamp", creationTimestamp)
.add("creator", creator)
.add("environmentVars", environmentVars)
.add("instanceProfileArn", instanceProfileArn)
.add("modelName", modelName)
.add("modelVersion", modelVersion)
.add("name", name)
.add("scaleToZeroEnabled", scaleToZeroEnabled)
.add("state", state)
.add("workloadSize", workloadSize)
.add("workloadType", workloadType)
.toString();
}
}