
co.elastic.clients.elasticsearch.inference.PutOpenaiRequest Maven / Gradle / Ivy
Show all versions of org.apache.servicemix.bundles.elasticsearch-java
/*
* Licensed to Elasticsearch B.V. under one or more contributor
* license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright
* ownership. Elasticsearch B.V. licenses this file to you under
* the Apache License, Version 2.0 (the "License"); you may
* not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package co.elastic.clients.elasticsearch.inference;
import co.elastic.clients.elasticsearch._types.ErrorResponse;
import co.elastic.clients.elasticsearch._types.RequestBase;
import co.elastic.clients.json.JsonpDeserializable;
import co.elastic.clients.json.JsonpDeserializer;
import co.elastic.clients.json.JsonpMapper;
import co.elastic.clients.json.JsonpSerializable;
import co.elastic.clients.json.ObjectBuilderDeserializer;
import co.elastic.clients.json.ObjectDeserializer;
import co.elastic.clients.transport.Endpoint;
import co.elastic.clients.transport.endpoints.SimpleEndpoint;
import co.elastic.clients.util.ApiTypeHelper;
import co.elastic.clients.util.ObjectBuilder;
import jakarta.json.stream.JsonGenerator;
import java.lang.String;
import java.util.Collections;
import java.util.HashMap;
import java.util.Map;
import java.util.Objects;
import java.util.function.Function;
import javax.annotation.Nullable;
//----------------------------------------------------------------
// THIS CODE IS GENERATED. MANUAL EDITS WILL BE LOST.
//----------------------------------------------------------------
//
// This code is generated from the Elasticsearch API specification
// at https://github.com/elastic/elasticsearch-specification
//
// Manual updates to this file will be lost when the code is
// re-generated.
//
// If you find a property that is missing or wrongly typed, please
// open an issue or a PR on the API specification repository.
//
//----------------------------------------------------------------
// typedef: inference.put_openai.Request
/**
* Create an OpenAI inference endpoint.
*
* Create an inference endpoint to perform an inference task with the
* openai
service or openai
compatible APIs.
*
* @see API
* specification
*/
@JsonpDeserializable
public class PutOpenaiRequest extends RequestBase implements JsonpSerializable {
@Nullable
private final InferenceChunkingSettings chunkingSettings;
private final String openaiInferenceId;
private final OpenAIServiceType service;
private final OpenAIServiceSettings serviceSettings;
@Nullable
private final OpenAITaskSettings taskSettings;
private final OpenAITaskType taskType;
// ---------------------------------------------------------------------------------------------
private PutOpenaiRequest(Builder builder) {
this.chunkingSettings = builder.chunkingSettings;
this.openaiInferenceId = ApiTypeHelper.requireNonNull(builder.openaiInferenceId, this, "openaiInferenceId");
this.service = ApiTypeHelper.requireNonNull(builder.service, this, "service");
this.serviceSettings = ApiTypeHelper.requireNonNull(builder.serviceSettings, this, "serviceSettings");
this.taskSettings = builder.taskSettings;
this.taskType = ApiTypeHelper.requireNonNull(builder.taskType, this, "taskType");
}
public static PutOpenaiRequest of(Function> fn) {
return fn.apply(new Builder()).build();
}
/**
* The chunking configuration object.
*
* API name: {@code chunking_settings}
*/
@Nullable
public final InferenceChunkingSettings chunkingSettings() {
return this.chunkingSettings;
}
/**
* Required - The unique identifier of the inference endpoint.
*
* API name: {@code openai_inference_id}
*/
public final String openaiInferenceId() {
return this.openaiInferenceId;
}
/**
* Required - The type of service supported for the specified task type. In this
* case, openai
.
*
* API name: {@code service}
*/
public final OpenAIServiceType service() {
return this.service;
}
/**
* Required - Settings used to install the inference model. These settings are
* specific to the openai
service.
*
* API name: {@code service_settings}
*/
public final OpenAIServiceSettings serviceSettings() {
return this.serviceSettings;
}
/**
* Settings to configure the inference task. These settings are specific to the
* task type you specified.
*
* API name: {@code task_settings}
*/
@Nullable
public final OpenAITaskSettings taskSettings() {
return this.taskSettings;
}
/**
* Required - The type of the inference task that the model will perform. NOTE:
* The chat_completion
task type only supports streaming and only
* through the _stream API.
*
* API name: {@code task_type}
*/
public final OpenAITaskType taskType() {
return this.taskType;
}
/**
* Serialize this object to JSON.
*/
public void serialize(JsonGenerator generator, JsonpMapper mapper) {
generator.writeStartObject();
serializeInternal(generator, mapper);
generator.writeEnd();
}
protected void serializeInternal(JsonGenerator generator, JsonpMapper mapper) {
if (this.chunkingSettings != null) {
generator.writeKey("chunking_settings");
this.chunkingSettings.serialize(generator, mapper);
}
generator.writeKey("service");
this.service.serialize(generator, mapper);
generator.writeKey("service_settings");
this.serviceSettings.serialize(generator, mapper);
if (this.taskSettings != null) {
generator.writeKey("task_settings");
this.taskSettings.serialize(generator, mapper);
}
}
// ---------------------------------------------------------------------------------------------
/**
* Builder for {@link PutOpenaiRequest}.
*/
public static class Builder extends RequestBase.AbstractBuilder
implements
ObjectBuilder {
@Nullable
private InferenceChunkingSettings chunkingSettings;
private String openaiInferenceId;
private OpenAIServiceType service;
private OpenAIServiceSettings serviceSettings;
@Nullable
private OpenAITaskSettings taskSettings;
private OpenAITaskType taskType;
/**
* The chunking configuration object.
*
* API name: {@code chunking_settings}
*/
public final Builder chunkingSettings(@Nullable InferenceChunkingSettings value) {
this.chunkingSettings = value;
return this;
}
/**
* The chunking configuration object.
*
* API name: {@code chunking_settings}
*/
public final Builder chunkingSettings(
Function> fn) {
return this.chunkingSettings(fn.apply(new InferenceChunkingSettings.Builder()).build());
}
/**
* Required - The unique identifier of the inference endpoint.
*
* API name: {@code openai_inference_id}
*/
public final Builder openaiInferenceId(String value) {
this.openaiInferenceId = value;
return this;
}
/**
* Required - The type of service supported for the specified task type. In this
* case, openai
.
*
* API name: {@code service}
*/
public final Builder service(OpenAIServiceType value) {
this.service = value;
return this;
}
/**
* Required - Settings used to install the inference model. These settings are
* specific to the openai
service.
*
* API name: {@code service_settings}
*/
public final Builder serviceSettings(OpenAIServiceSettings value) {
this.serviceSettings = value;
return this;
}
/**
* Required - Settings used to install the inference model. These settings are
* specific to the openai
service.
*
* API name: {@code service_settings}
*/
public final Builder serviceSettings(
Function> fn) {
return this.serviceSettings(fn.apply(new OpenAIServiceSettings.Builder()).build());
}
/**
* Settings to configure the inference task. These settings are specific to the
* task type you specified.
*
* API name: {@code task_settings}
*/
public final Builder taskSettings(@Nullable OpenAITaskSettings value) {
this.taskSettings = value;
return this;
}
/**
* Settings to configure the inference task. These settings are specific to the
* task type you specified.
*
* API name: {@code task_settings}
*/
public final Builder taskSettings(Function> fn) {
return this.taskSettings(fn.apply(new OpenAITaskSettings.Builder()).build());
}
/**
* Required - The type of the inference task that the model will perform. NOTE:
* The chat_completion
task type only supports streaming and only
* through the _stream API.
*
* API name: {@code task_type}
*/
public final Builder taskType(OpenAITaskType value) {
this.taskType = value;
return this;
}
@Override
protected Builder self() {
return this;
}
/**
* Builds a {@link PutOpenaiRequest}.
*
* @throws NullPointerException
* if some of the required fields are null.
*/
public PutOpenaiRequest build() {
_checkSingleUse();
return new PutOpenaiRequest(this);
}
}
// ---------------------------------------------------------------------------------------------
/**
* Json deserializer for {@link PutOpenaiRequest}
*/
public static final JsonpDeserializer _DESERIALIZER = ObjectBuilderDeserializer.lazy(Builder::new,
PutOpenaiRequest::setupPutOpenaiRequestDeserializer);
protected static void setupPutOpenaiRequestDeserializer(ObjectDeserializer op) {
op.add(Builder::chunkingSettings, InferenceChunkingSettings._DESERIALIZER, "chunking_settings");
op.add(Builder::service, OpenAIServiceType._DESERIALIZER, "service");
op.add(Builder::serviceSettings, OpenAIServiceSettings._DESERIALIZER, "service_settings");
op.add(Builder::taskSettings, OpenAITaskSettings._DESERIALIZER, "task_settings");
}
// ---------------------------------------------------------------------------------------------
/**
* Endpoint "{@code inference.put_openai}".
*/
public static final Endpoint _ENDPOINT = new SimpleEndpoint<>(
"es/inference.put_openai",
// Request method
request -> {
return "PUT";
},
// Request path
request -> {
final int _openaiInferenceId = 1 << 0;
final int _taskType = 1 << 1;
int propsSet = 0;
propsSet |= _openaiInferenceId;
propsSet |= _taskType;
if (propsSet == (_taskType | _openaiInferenceId)) {
StringBuilder buf = new StringBuilder();
buf.append("/_inference");
buf.append("/");
SimpleEndpoint.pathEncode(request.taskType.jsonValue(), buf);
buf.append("/");
SimpleEndpoint.pathEncode(request.openaiInferenceId, buf);
return buf.toString();
}
throw SimpleEndpoint.noPathTemplateFound("path");
},
// Path parameters
request -> {
Map params = new HashMap<>();
final int _openaiInferenceId = 1 << 0;
final int _taskType = 1 << 1;
int propsSet = 0;
propsSet |= _openaiInferenceId;
propsSet |= _taskType;
if (propsSet == (_taskType | _openaiInferenceId)) {
params.put("taskType", request.taskType.jsonValue());
params.put("openaiInferenceId", request.openaiInferenceId);
}
return params;
},
// Request parameters
request -> {
return Collections.emptyMap();
}, SimpleEndpoint.emptyMap(), true, PutOpenaiResponse._DESERIALIZER);
}