All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.agentsflex.llm.spark.SparkLlm Maven / Gradle / Ivy

/*
 *  Copyright (c) 2023-2025, Agents-Flex ([email protected]).
 *  

* Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at *

* http://www.apache.org/licenses/LICENSE-2.0 *

* Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package com.agentsflex.llm.spark; import com.agentsflex.core.document.Document; import com.agentsflex.core.llm.*; import com.agentsflex.core.llm.client.BaseLlmClientListener; import com.agentsflex.core.llm.client.HttpClient; import com.agentsflex.core.llm.client.LlmClient; import com.agentsflex.core.llm.client.LlmClientListener; import com.agentsflex.core.llm.client.impl.WebSocketClient; import com.agentsflex.core.llm.embedding.EmbeddingOptions; import com.agentsflex.core.llm.response.AbstractBaseMessageResponse; import com.agentsflex.core.llm.response.AiMessageResponse; import com.agentsflex.core.llm.response.FunctionMessageResponse; import com.agentsflex.core.message.AiMessage; import com.agentsflex.core.message.FunctionMessage; import com.agentsflex.core.parser.AiMessageParser; import com.agentsflex.core.parser.FunctionMessageParser; import com.agentsflex.core.prompt.Prompt; import com.agentsflex.core.store.VectorData; import com.agentsflex.core.util.StringUtil; import com.alibaba.fastjson.JSONPath; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import java.nio.ByteBuffer; import java.nio.ByteOrder; import java.util.Base64; import java.util.concurrent.CountDownLatch; public class SparkLlm extends BaseLlm { private static final Logger logger = LoggerFactory.getLogger(SparkLlm.class); public AiMessageParser aiMessageParser = SparkLlmUtil.getAiMessageParser(); public FunctionMessageParser functionMessageParser = SparkLlmUtil.getFunctionMessageParser(); private final HttpClient httpClient = new HttpClient(); public SparkLlm(SparkLlmConfig config) { super(config); } @Override public VectorData embed(Document document, EmbeddingOptions options) { String payload = SparkLlmUtil.embedPayload(config, document); String resp = httpClient.post(SparkLlmUtil.createEmbedURL(config), null, payload); if (StringUtil.noText(resp)) { return null; } Integer code = JSONPath.read(resp, "$.header.code", Integer.class); if (code == null || code != 0) { logger.error(resp); return null; } String text = JSONPath.read(resp, "$.payload.feature.text", String.class); if (StringUtil.noText(text)) { return null; } byte[] buffer = Base64.getDecoder().decode(text); double[] vector = new double[buffer.length / 4]; for (int i = 0; i < vector.length; i++) { int n = i * 4; vector[i] = ByteBuffer.wrap(buffer, n, 4).order(ByteOrder.LITTLE_ENDIAN).getFloat(); } VectorData vectorData = new VectorData(); vectorData.setVector(vector); return vectorData; } @SuppressWarnings("unchecked") @Override public > R chat(Prompt prompt, ChatOptions options) { CountDownLatch latch = new CountDownLatch(1); Throwable[] failureThrowable = new Throwable[1]; AbstractBaseMessageResponse[] messageResponse = {null}; waitResponse(prompt, options, messageResponse, latch, failureThrowable); AbstractBaseMessageResponse response = messageResponse[0]; if (response == null){ return null; } Throwable fialureThrowable = failureThrowable[0]; if (null == response.getMessage() || fialureThrowable != null) { response.setError(true); if (fialureThrowable != null) { response.setErrorMessage(fialureThrowable.getMessage()); } } else { response.setError(false); } return (R) response; } private > void waitResponse(Prompt prompt , ChatOptions options , AbstractBaseMessageResponse[] messageResponse , CountDownLatch latch , Throwable[] failureThrowable) { chatStream(prompt, new StreamResponseListener() { @Override public void onMessage(ChatContext context, R response) { if (response.getMessage() instanceof FunctionMessage) { messageResponse[0] = (FunctionMessageResponse) response; } else { AiMessage aiMessage = new AiMessage(); aiMessage.setContent(response.getMessage().getFullContent()); messageResponse[0] = new AiMessageResponse(aiMessage); } } @Override public void onStop(ChatContext context) { StreamResponseListener.super.onStop(context); latch.countDown(); } @Override public void onFailure(ChatContext context, Throwable throwable) { logger.error(throwable.toString(), throwable); failureThrowable[0] = throwable; } }, options); try { latch.await(); } catch (InterruptedException e) { throw new RuntimeException(e); } } @Override public > void chatStream(Prompt prompt, StreamResponseListener listener, ChatOptions options) { LlmClient llmClient = new WebSocketClient(); String url = SparkLlmUtil.createURL(config); String payload = SparkLlmUtil.promptToPayload(prompt, config, options); LlmClientListener clientListener = new BaseLlmClientListener(this, llmClient, listener, prompt, aiMessageParser, functionMessageParser); llmClient.start(url, null, payload, clientListener, config); } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy