
tri.ai.gemini.GeminiTextChat.kt Maven / Gradle / Ivy
/*-
* #%L
* tri.promptfx:promptkt
* %%
* Copyright (C) 2023 - 2024 Johns Hopkins University Applied Physics Laboratory
* %%
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
* #L%
*/
package tri.ai.gemini
import tri.ai.core.TextChat
import tri.ai.core.TextChatMessage
import tri.ai.core.TextChatRole
import tri.ai.gemini.GeminiModelIndex.GEMINI_PRO
import tri.ai.prompt.trace.AiExecInfo
import tri.ai.prompt.trace.AiModelInfo
import tri.ai.prompt.trace.AiOutputInfo
import tri.ai.prompt.trace.AiPromptTrace
/** Text chat with Gemini models. */
class GeminiTextChat(override val modelId: String = GEMINI_PRO, val client: GeminiClient = GeminiClient.INSTANCE) :
TextChat {
override fun toString() = "$modelId (Gemini)"
override suspend fun chat(messages: List, tokens: Int?, stop: List?, requestJson: Boolean?, numResponses: Int?): AiPromptTrace {
val modelInfo = AiModelInfo.info(modelId, tokens = tokens, stop = stop, requestJson = requestJson, numResponses = numResponses)
val t0 = System.currentTimeMillis()
val resp = client.generateContent(
messages, modelId,
GenerationConfig(
maxOutputTokens = tokens ?: 500,
stopSequences = stop,
responseMimeType = if (requestJson == true) "application/json" else null
)
)
return resp.trace(modelInfo, t0)
}
companion object {
/** Create trace for chat message response, with given model info and start query time. */
internal fun GenerateContentResponse.trace(modelInfo: AiModelInfo, t0: Long): AiPromptTrace {
val err = error
return if (err != null) {
AiPromptTrace.error(modelInfo, err.message, duration = System.currentTimeMillis() - t0)
} else {
val firstCandidate = candidates!!.first()
val role = when (firstCandidate.content.role) {
"user" -> TextChatRole.User
"model" -> TextChatRole.Assistant
else -> error("Invalid role: ${firstCandidate.content.role}")
}
val msgs = firstCandidate.content.parts.map { TextChatMessage(role, it.text) }
AiPromptTrace(
null,
modelInfo,
AiExecInfo(responseTimeMillis = System.currentTimeMillis() - t0),
AiOutputInfo(msgs)
)
}
}
}
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy