commonMain.com.xebia.functional.openai.generated.model.CreateChatCompletionResponseChoicesInner.kt Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of xef-openai-client Show documentation
Show all versions of xef-openai-client Show documentation
Building applications with LLMs through composability in Kotlin
/**
*
* Please note:
* This class is auto generated by OpenAPI Generator (https://openapi-generator.tech).
* Do not edit this file manually.
*
*/
@file:Suppress(
"ArrayInDataClass",
"EnumEntryName",
"RemoveRedundantQualifierName",
"UnusedImport"
)
package com.xebia.functional.openai.generated.model
import com.xebia.functional.openai.generated.model.ChatCompletionResponseMessage
import com.xebia.functional.openai.generated.model.CreateChatCompletionResponseChoicesInnerLogprobs
import kotlinx.serialization.Serializable
import kotlinx.serialization.SerialName
import kotlinx.serialization.Contextual
import kotlin.js.JsName
import kotlinx.serialization.json.*
/**
*
*
* @param finishReason The reason the model stopped generating tokens. This will be `stop` if the model hit a natural stop point or a provided stop sequence, `length` if the maximum number of tokens specified in the request was reached, `content_filter` if content was omitted due to a flag from our content filters, `tool_calls` if the model called a tool, or `function_call` (deprecated) if the model called a function.
* @param index The index of the choice in the list of choices.
* @param message
* @param logprobs
*/
@Serializable
data class CreateChatCompletionResponseChoicesInner (
/* The reason the model stopped generating tokens. This will be `stop` if the model hit a natural stop point or a provided stop sequence, `length` if the maximum number of tokens specified in the request was reached, `content_filter` if content was omitted due to a flag from our content filters, `tool_calls` if the model called a tool, or `function_call` (deprecated) if the model called a function. */
@SerialName(value = "finish_reason") val finishReason: CreateChatCompletionResponseChoicesInner.FinishReason,
/* The index of the choice in the list of choices. */
@SerialName(value = "index") val index: kotlin.Int,
@SerialName(value = "message") val message: ChatCompletionResponseMessage,
@SerialName(value = "logprobs") val logprobs: CreateChatCompletionResponseChoicesInnerLogprobs?
) {
/**
* The reason the model stopped generating tokens. This will be `stop` if the model hit a natural stop point or a provided stop sequence, `length` if the maximum number of tokens specified in the request was reached, `content_filter` if content was omitted due to a flag from our content filters, `tool_calls` if the model called a tool, or `function_call` (deprecated) if the model called a function.
*
* Values: stop,length,tool_calls,content_filter,function_call
*/
@Serializable
enum class FinishReason(val value: kotlin.String) {
@SerialName(value = "stop") stop("stop"),
@SerialName(value = "length") @JsName("length_type") length("length"),
@SerialName(value = "tool_calls") tool_calls("tool_calls"),
@SerialName(value = "content_filter") content_filter("content_filter"),
@SerialName(value = "function_call") function_call("function_call");
}
}