io.quarkiverse.langchain4j.huggingface.runtime.config.EmbeddingModelConfig.jdp Maven / Gradle / Ivy
io.quarkiverse.langchain4j.huggingface.runtime.config.EmbeddingModelConfig.inferenceEndpointUrl=The URL of the inference endpoint for the embedding.\n\nWhen using Hugging Face with the inference API, the URL is\n{@code https\://api-inference.huggingface.co/pipeline/feature-extraction/},\nfor example\n{@code https\://api-inference.huggingface.co/pipeline/feature-extraction/sentence-transformers/all-mpnet-base-v2}.\n\nWhen using a deployed inference endpoint, the URL is the URL of the endpoint.\nWhen using a local hugging face model, the URL is the URL of the local model.
io.quarkiverse.langchain4j.huggingface.runtime.config.EmbeddingModelConfig.waitForModel=If the model is not ready, wait for it instead of receiving 503. It limits the number of requests required to get your\ninference done. It is advised to only set this flag to true after receiving a 503 error as it will limit hanging in your\napplication to known places