Configuration property fixed at build time - All other configuration properties are overridable at runtime
LangChain4j Easy RAG |
Type |
Default |
|---|---|---|
Path to the directory containing the documents to be ingested. This is either an absolute or relative path in the filesystem. A relative path is resolved against the current working directory at runtime. Environment variable: |
string |
required |
Does Environment variable: |
tooltip:filesystem[The |
tooltip:filesystem[The {@link #path()} represents a filesystem reference] |
Matcher used for filtering which files from the directory should be ingested. This uses the Environment variable: |
string |
|
Whether to recursively ingest documents from subdirectories. Environment variable: |
boolean |
|
Maximum segment size when splitting documents, in tokens. Environment variable: |
int |
|
Maximum overlap (in tokens) when splitting documents. Environment variable: |
int |
|
Maximum number of results to return when querying the retrieval augmentor. Environment variable: |
int |
|
The minimum score for results to return when querying the retrieval augmentor. Environment variable: |
double |
|
The strategy to decide whether document ingestion into the store should happen at startup or not. The default is ON. Changing to OFF generally only makes sense if running against a persistent embedding store that was already populated. When set to MANUAL, it is expected that the application will inject and call the Environment variable: |
|
|
Whether or not to reuse embeddings. Defaults to Environment variable: |
boolean |
|
The file path to load/save embeddings, assuming Defaults to Environment variable: |
string |
|
LangChain4j Model Context Protocol client |
Type |
Default |
Whether the MCP extension should automatically generate a ToolProvider that is wired up to all the configured MCP clients. The default is true if at least one MCP client is configured, false otherwise. Environment variable: |
boolean |
|
File containing the MCP servers configuration in the Claude Desktop format. This configuration can only be used to configure This file is read at build time which means that which MCP servers the client will use, is determined at build time. However, specific configuration of each MCP server can be overridden at runtime. Environment variable: |
string |
|
Whether the MCP extension should automatically register a health check for configured MCP clients. The default is true if at least one MCP client is configured, false otherwise. Environment variable: |
boolean |
|
Whether resources should be exposed as MCP tools. Environment variable: |
boolean |
|
Type |
Default |
|
Transport type Environment variable: |
|
|
The URL of the SSE endpoint. This only applies to MCP clients using the HTTP transport. Environment variable: |
string |
|
The command to execute to spawn the MCP server process. This only applies to MCP clients using the STDIO transport. Environment variable: |
list of string |
|
Environment variables for the spawned MCP server process. This only applies to MCP clients using the STDIO transport. Environment variable: |
Map<String,String> |
|
Whether to log requests Environment variable: |
boolean |
|
Whether to log responses Environment variable: |
boolean |
|
Whether to prefer MicroProfile health checks. Applies to MCP HTTP clients only. If this property is enabled, an HTTP GET call is made to an MCP Server MicroProfile Health endpoint. MicroProfile Health endpoint URL is calculated by extracting a base URL that has no path component from the Default MCP Client health check that opens a Streamable HTTP or SSE transport channel is attempted when a MicroProfile health check returns an HTTP 404 or other error status. Environment variable: |
boolean |
|
Relative path of an MCP Server MicroProfile Health endpoint. This property is effective only when the MicroProfile Health endpoint URL is calculated by extracting the base URL that has no path component from the Environment variable: |
string |
|
Timeout for tool executions performed by the MCP client Environment variable: |
|
|
Timeout for resource-related operations (retrieving a list of resources as well as the actual contents of resources). Environment variable: |
|
|
Timeout for pinging the MCP server process to check if it’s still alive. If a ping times out, the client’s health check will start failing. Environment variable: |
|
|
The initial list of MCP roots that the client can present to the server. The list can be later updated programmatically during runtime. The list is formatted as key-value pairs separated by commas. For example: workspace1=/path/to/workspace1,workspace2=/path/to/workspace2 Environment variable: |
list of string |
|
The name of the TLS configuration (bucket) used for client authentication in the TLS registry. This does not have any effect when the stdio transport is used. Environment variable: |
string |
|
Whether to cache the tool list obtained from the MCP server. When set to true (the default), the tool list is cached until the server notifies of changes or the cache is manually evicted. When false, the client always fetches a fresh tool list from the server. This is useful when using MCP servers that don’t support tool list change notifications. Environment variable: |
boolean |
|
Type |
Default |
|
The base URL of the MCP registry, without the API version segment. The default value points at the official registry (https://registry.modelcontextprotocol.io). Environment variable: |
string |
|
Whether to log requests Environment variable: |
boolean |
|
Whether to log responses Environment variable: |
boolean |
|
The name of the TLS configuration (bucket) that this MCP client registry will use. Environment variable: |
string |
|
The read timeout for the MCP registry’s underlying http client Environment variable: |
|
|
The connect timeout for the MCP registry’s underlying http client Environment variable: |
|
|
LangChain4j Neo4j embedding store |
Type |
Default |
Dimension of the embeddings that will be stored in the Neo4j store. Environment variable: |
int |
required |
Label for the created nodes. Environment variable: |
string |
|
Name of the property to store the embedding vectors. Environment variable: |
string |
|
Name of the property to store embedding IDs. Environment variable: |
string |
|
Prefix to be added to the metadata keys. By default, no prefix is used. Environment variable: |
string |
|
Name of the property to store the embedding text. Environment variable: |
string |
|
Name of the index to be created for vector search. Environment variable: |
string |
|
Name of the database to connect to. Environment variable: |
string |
|
The query to use when retrieving embeddings. This query has to return the following columns:
Environment variable: |
string |
|
LangChain4j Pinecone embedding store |
Type |
Default |
The API key to Pinecone. Environment variable: |
string |
required |
Environment name, e.g. gcp-starter or northamerica-northeast1-gcp. Environment variable: |
string |
required |
ID of the project. Environment variable: |
string |
required |
Name of the index within the project. If the index doesn’t exist, it will be created. Environment variable: |
string |
required |
Dimension of the embeddings in the index. This is required only in case that the index doesn’t exist yet and needs to be created. Environment variable: |
int |
|
The type of the pod to use. This is only used if the index doesn’t exist yet and needs to be created. The format: One of Environment variable: |
string |
|
The timeout duration for the index to become ready. Only relevant if the index doesn’t exist yet and needs to be created. If not specified, 1 minute will be used. Environment variable: |
||
The namespace. Environment variable: |
string |
|
The name of the field that contains the text segment. Environment variable: |
string |
|
The timeout duration for the Pinecone client. If not specified, 5 seconds will be used. Environment variable: |
||
LangChain4j Tavily Web Search Engine |
Type |
Default |
Base URL of the Tavily API Environment variable: |
string |
|
API key for the Tavily API Environment variable: |
string |
required |
Maximum number of results to return Environment variable: |
int |
|
The timeout duration for Tavily requests. Environment variable: |
|
|
Whether requests to Tavily should be logged Environment variable: |
boolean |
|
Whether responses from Tavily should be logged Environment variable: |
boolean |
|
The search depth to use. This can be "basic" or "advanced". Basic is the default. Environment variable: |
|
|
Include a short answer to original query. Default is false. Environment variable: |
boolean |
|
Include the cleaned and parsed HTML content of each search result. Default is false. Environment variable: |
boolean |
|
A list of domains to specifically include in the search results. By default all domains are included. Environment variable: |
list of string |
|
A list of domains to specifically exclude from the search results. By default no domains are excluded. Environment variable: |
list of string |
|
Quarkus LangChain4j - AI Gemini |
Type |
Default |
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
The api key Environment variable: |
string |
|
Publisher of model Environment variable: |
string |
|
Meant to be used for testing only in order to override the base URL used by the client Environment variable: |
string |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Whether the Vertex AI client should log requests Environment variable: |
boolean |
|
Whether the Vertex AI client should log responses Environment variable: |
boolean |
|
Timeout for requests to gemini APIs Environment variable: |
|
|
The id of the model to use. Environment variable: |
string |
|
The temperature is used for sampling during response generation, which occurs when topP and topK are applied. Temperature controls the degree of randomness in token selection. Lower temperatures are good for prompts that require a less open-ended or creative response, while higher temperatures can lead to more diverse or creative results. A temperature of 0 means that the highest probability tokens are always selected. In this case, responses for a given prompt are mostly deterministic, but a small amount of variation is still possible. If the model returns a response that’s too generic, too short, or the model gives a fallback response, try increasing the temperature. Environment variable: |
double |
|
Maximum number of tokens that can be generated in the response. A token is approximately four characters. 100 tokens correspond to roughly 60-80 words. Specify a lower value for shorter responses and a higher value for potentially longer responses. Environment variable: |
int |
|
Top-P changes how the model selects tokens for output. Tokens are selected from the most (see top-K) to least probable until the sum of their probabilities equals the top-P value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-P value is 0.5, then the model will select either A or B as the next token by using temperature and excludes C as a candidate. Specify a lower value for less random responses and a higher value for more random responses. Range: 0.0 - 1.0 Default for gemini-2.5-flash: 0.95 Environment variable: |
double |
|
Top-K changes how the model selects tokens for output. A top-K of 1 means the next selected token is the most probable among all tokens in the model’s vocabulary (also called greedy decoding), while a top-K of 3 means that the next token is selected from among the three most probable tokens by using temperature. For each token selection step, the top-K tokens with the highest probabilities are sampled. Then tokens are further filtered based on top-P with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Range: 1-40 gemini-2.5-flash doesn’t support topK Environment variable: |
int |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Global timeout for requests to gemini APIs Environment variable: |
|
|
Controls whether thought summaries are enabled. Thought summaries are synthesized versions of the model’s raw thoughts and offer insights into the model’s internal reasoning process. Environment variable: |
boolean |
|
The thinkingBudget parameter guides the model on the number of thinking tokens to use when generating a response. A higher token count generally allows for more detailed reasoning, which can be beneficial for tackling more complex tasks. If latency is more important, use a lower budget or disable thinking by setting thinkingBudget to 0. Setting the thinkingBudget to -1 turns on dynamic thinking, meaning the model will adjust the budget based on the complexity of the request. The thinkingBudget is only supported in Gemini 2.5 Flash, 2.5 Pro, and 2.5 Flash-Lite. Depending on the prompt, the model might overflow or underflow the token budget. See Gemini API docs for more details. Environment variable: |
long |
|
The id of the model to use. Environment variable: |
string |
|
Reduced dimension for the output embedding Environment variable: |
int |
|
Optional task type for which the embeddings will be used. Can only be set for models/embedding-001 Possible values: TASK_TYPE_UNSPECIFIED, RETRIEVAL_QUERY, RETRIEVAL_DOCUMENT, SEMANTIC_SIMILARITY, CLASSIFICATION, CLUSTERING, QUESTION_ANSWERING, FACT_VERIFICATION Environment variable: |
string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Global timeout for requests to gemini APIs Environment variable: |
|
|
Type |
Default |
|
The api key Environment variable: |
string |
|
Publisher of model Environment variable: |
string |
|
Meant to be used for testing only in order to override the base URL used by the client Environment variable: |
string |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Whether the Vertex AI client should log requests Environment variable: |
boolean |
|
Whether the Vertex AI client should log responses Environment variable: |
boolean |
|
Timeout for requests to gemini APIs Environment variable: |
|
|
The id of the model to use. Environment variable: |
string |
|
The temperature is used for sampling during response generation, which occurs when topP and topK are applied. Temperature controls the degree of randomness in token selection. Lower temperatures are good for prompts that require a less open-ended or creative response, while higher temperatures can lead to more diverse or creative results. A temperature of 0 means that the highest probability tokens are always selected. In this case, responses for a given prompt are mostly deterministic, but a small amount of variation is still possible. If the model returns a response that’s too generic, too short, or the model gives a fallback response, try increasing the temperature. Environment variable: |
double |
|
Maximum number of tokens that can be generated in the response. A token is approximately four characters. 100 tokens correspond to roughly 60-80 words. Specify a lower value for shorter responses and a higher value for potentially longer responses. Environment variable: |
int |
|
Top-P changes how the model selects tokens for output. Tokens are selected from the most (see top-K) to least probable until the sum of their probabilities equals the top-P value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-P value is 0.5, then the model will select either A or B as the next token by using temperature and excludes C as a candidate. Specify a lower value for less random responses and a higher value for more random responses. Range: 0.0 - 1.0 Default for gemini-2.5-flash: 0.95 Environment variable: |
double |
|
Top-K changes how the model selects tokens for output. A top-K of 1 means the next selected token is the most probable among all tokens in the model’s vocabulary (also called greedy decoding), while a top-K of 3 means that the next token is selected from among the three most probable tokens by using temperature. For each token selection step, the top-K tokens with the highest probabilities are sampled. Then tokens are further filtered based on top-P with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Range: 1-40 gemini-2.5-flash doesn’t support topK Environment variable: |
int |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Global timeout for requests to gemini APIs Environment variable: |
|
|
Controls whether thought summaries are enabled. Thought summaries are synthesized versions of the model’s raw thoughts and offer insights into the model’s internal reasoning process. Environment variable: |
boolean |
|
The thinkingBudget parameter guides the model on the number of thinking tokens to use when generating a response. A higher token count generally allows for more detailed reasoning, which can be beneficial for tackling more complex tasks. If latency is more important, use a lower budget or disable thinking by setting thinkingBudget to 0. Setting the thinkingBudget to -1 turns on dynamic thinking, meaning the model will adjust the budget based on the complexity of the request. The thinkingBudget is only supported in Gemini 2.5 Flash, 2.5 Pro, and 2.5 Flash-Lite. Depending on the prompt, the model might overflow or underflow the token budget. See Gemini API docs for more details. Environment variable: |
long |
|
The id of the model to use. Environment variable: |
string |
|
Reduced dimension for the output embedding Environment variable: |
int |
|
Optional task type for which the embeddings will be used. Can only be set for models/embedding-001 Possible values: TASK_TYPE_UNSPECIFIED, RETRIEVAL_QUERY, RETRIEVAL_DOCUMENT, SEMANTIC_SIMILARITY, CLASSIFICATION, CLUSTERING, QUESTION_ANSWERING, FACT_VERIFICATION Environment variable: |
string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Global timeout for requests to gemini APIs Environment variable: |
|
|
Quarkus LangChain4j - Anthropic |
Type |
Default |
Whether the model should be enabled Environment variable: |
boolean |
|
Base URL of the Anthropic API Environment variable: |
string |
|
Anthropic API key Environment variable: |
string |
|
The Anthropic version Environment variable: |
string |
|
Timeout for Anthropic calls Environment variable: |
|
|
Whether the Anthropic client should log requests Environment variable: |
boolean |
|
Whether the Anthropic client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Model name to use Environment variable: |
string |
|
What sampling temperature to use, between 0.0 and 1.0. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. It is generally recommended to set this or the Environment variable: |
double |
|
The maximum number of tokens to generate in the completion. The token count of your prompt plus Environment variable: |
int |
|
Double (0.0-1.0). Nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. It is generally recommended to set this or the Environment variable: |
double |
|
Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative Environment variable: |
int |
|
The maximum number of times to retry. 1 means exactly one attempt, with retrying disabled. Environment variable: |
int |
|
The custom text sequences that will cause the model to stop generating Environment variable: |
list of string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Cache system messages to reduce costs for repeated prompts. Requires minimum 1024 tokens (Claude Opus/Sonnet) or 2048-4096 tokens (Haiku). Supported models: Claude Opus 4.1, Sonnet 4.5, Haiku 4.5, and later models. Environment variable: |
boolean |
|
Cache tool definitions to reduce costs. Requires minimum 1024 tokens (Claude Opus/Sonnet) or 2048-4096 tokens (Haiku). Supported models: Claude Opus 4.1, Sonnet 4.5, Haiku 4.5, and later models. Environment variable: |
boolean |
|
The thinking type to enable Claude’s reasoning process Environment variable: |
string |
|
The token budget for the model’s thinking process Environment variable: |
int |
|
Whether thinking results should be returned in the response Environment variable: |
boolean |
|
Whether previously stored thinking should be sent in follow-up requests Environment variable: |
boolean |
|
Enable interleaved thinking for Claude 4 models, allowing reasoning between tool calls. Requires Claude 4 model (e.g., claude-opus-4-20250514) and thinking.type: enabled. Environment variable: |
boolean |
|
Type |
Default |
|
Base URL of the Anthropic API Environment variable: |
string |
|
Anthropic API key Environment variable: |
string |
|
The Anthropic version Environment variable: |
string |
|
Timeout for Anthropic calls Environment variable: |
|
|
Whether the Anthropic client should log requests Environment variable: |
boolean |
|
Whether the Anthropic client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Model name to use Environment variable: |
string |
|
What sampling temperature to use, between 0.0 and 1.0. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. It is generally recommended to set this or the Environment variable: |
double |
|
The maximum number of tokens to generate in the completion. The token count of your prompt plus Environment variable: |
int |
|
Double (0.0-1.0). Nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. It is generally recommended to set this or the Environment variable: |
double |
|
Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative Environment variable: |
int |
|
The maximum number of times to retry. 1 means exactly one attempt, with retrying disabled. Environment variable: |
int |
|
The custom text sequences that will cause the model to stop generating Environment variable: |
list of string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Cache system messages to reduce costs for repeated prompts. Requires minimum 1024 tokens (Claude Opus/Sonnet) or 2048-4096 tokens (Haiku). Supported models: Claude Opus 4.1, Sonnet 4.5, Haiku 4.5, and later models. Environment variable: |
boolean |
|
Cache tool definitions to reduce costs. Requires minimum 1024 tokens (Claude Opus/Sonnet) or 2048-4096 tokens (Haiku). Supported models: Claude Opus 4.1, Sonnet 4.5, Haiku 4.5, and later models. Environment variable: |
boolean |
|
The thinking type to enable Claude’s reasoning process Environment variable: |
string |
|
The token budget for the model’s thinking process Environment variable: |
int |
|
Whether thinking results should be returned in the response Environment variable: |
boolean |
|
Whether previously stored thinking should be sent in follow-up requests Environment variable: |
boolean |
|
Enable interleaved thinking for Claude 4 models, allowing reasoning between tool calls. Requires Claude 4 model (e.g., claude-opus-4-20250514) and thinking.type: enabled. Environment variable: |
boolean |
|
Quarkus LangChain4j - Azure OpenAI |
Type |
Default |
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
The name of your Azure OpenAI Resource. You’re required to first deploy a model before you can make calls. This and Environment variable: |
string |
|
The domain name of your Azure OpenAI Resource. You’re required to first deploy a model before you can make calls. This and Environment variable: |
string |
|
The name of your model deployment. You’re required to first deploy a model before you can make calls. This and Environment variable: |
string |
|
The endpoint for the Azure OpenAI resource. If not specified, then Environment variable: |
string |
|
The Azure AD token to use for this operation. If present, then the requests towards OpenAI will include this in the Authorization header. Note that this property overrides the functionality of Environment variable: |
string |
|
The API version to use for this operation. This follows the YYYY-MM-DD format Environment variable: |
string |
|
Azure OpenAI API key Environment variable: |
string |
|
Timeout for OpenAI calls Environment variable: |
|
|
The maximum number of times to retry. 1 means exactly one attempt, with retrying disabled. Environment variable: |
int |
|
Whether the OpenAI client should log requests Environment variable: |
boolean |
|
Whether the OpenAI client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
The Proxy type Environment variable: |
string |
|
The Proxy host Environment variable: |
string |
|
The Proxy port Environment variable: |
int |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
The Azure AD token to use for this operation. If present, then the requests towards OpenAI will include this in the Authorization header. Note that this property overrides the functionality of Environment variable: |
string |
|
The API version to use for this operation. This follows the YYYY-MM-DD format Environment variable: |
string |
|
Azure OpenAI API key Environment variable: |
string |
|
What sampling temperature to use, with values between 0 and 2. Higher values means the model will take more risks. A value of 0.9 is good for more creative applications, while 0 (argmax sampling) is good for ones with a well-defined answer. It is recommended to alter this or topP, but not both. Environment variable: |
double |
|
An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with topP probability mass. 0.1 means only the tokens comprising the top 10% probability mass are considered. It is recommended to alter this or temperature, but not both. Environment variable: |
double |
|
If specified, our system will make the best effort to sample deterministically, such that repeated requests with the same seed and parameters should return the same result. Determinism isn’t guaranteed. Support for reproducible output was first added in API version 2023-12-01-preview Environment variable: |
int |
|
The maximum number of tokens to generate in the completion. The token count of your prompt plus max_tokens can’t exceed the model’s context length. Most models have a context length of 2048 tokens (except for the newest models, which support 4096). Environment variable: |
int |
|
Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model’s likelihood to talk about new topics. Environment variable: |
double |
|
Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model’s likelihood to repeat the same line verbatim. Environment variable: |
double |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
The response format the model should use. Some models are not compatible with some response formats, make sure to review OpenAI documentation. Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
The Azure AD token to use for this operation. If present, then the requests towards OpenAI will include this in the Authorization header. Note that this property overrides the functionality of Environment variable: |
string |
|
The API version to use for this operation. This follows the YYYY-MM-DD format Environment variable: |
string |
|
Azure OpenAI API key Environment variable: |
string |
|
Whether embedding model requests should be logged Environment variable: |
boolean |
|
Whether embedding model responses should be logged Environment variable: |
boolean |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
The Azure AD token to use for this operation. If present, then the requests towards OpenAI will include this in the Authorization header. Note that this property overrides the functionality of Environment variable: |
string |
|
The API version to use for this operation. This follows the YYYY-MM-DD format Environment variable: |
string |
|
Azure OpenAI API key Environment variable: |
string |
|
Model name to use Environment variable: |
string |
|
Configure whether the generated images will be saved to disk. By default, persisting is disabled, but it is implicitly enabled when Environment variable: |
boolean |
|
The path where the generated images will be persisted to disk. This only applies of Environment variable: |
path |
|
The format in which the generated images are returned. Must be one of Environment variable: |
string |
|
The size of the generated images. Must be one of Must be one of Environment variable: |
string |
|
The quality of the image that will be generated.
This param is only supported for when the model is Environment variable: |
string |
|
The number of images to generate. Must be between 1 and 10. When the model is dall-e-3, only n=1 is supported. Environment variable: |
int |
|
The style of the generated images. Must be one of This param is only supported for when the model is Environment variable: |
string |
|
A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. Environment variable: |
string |
|
Whether image model requests should be logged Environment variable: |
boolean |
|
Whether image model responses should be logged Environment variable: |
boolean |
|
Type |
Default |
|
The name of your Azure OpenAI Resource. You’re required to first deploy a model before you can make calls. This and Environment variable: |
string |
|
The domain name of your Azure OpenAI Resource. You’re required to first deploy a model before you can make calls. This and Environment variable: |
string |
|
The name of your model deployment. You’re required to first deploy a model before you can make calls. This and Environment variable: |
string |
|
The endpoint for the Azure OpenAI resource. If not specified, then Environment variable: |
string |
|
The Azure AD token to use for this operation. If present, then the requests towards OpenAI will include this in the Authorization header. Note that this property overrides the functionality of Environment variable: |
string |
|
The API version to use for this operation. This follows the YYYY-MM-DD format Environment variable: |
string |
|
Azure OpenAI API key Environment variable: |
string |
|
Timeout for OpenAI calls Environment variable: |
|
|
The maximum number of times to retry. 1 means exactly one attempt, with retrying disabled. Environment variable: |
int |
|
Whether the OpenAI client should log requests Environment variable: |
boolean |
|
Whether the OpenAI client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
The Proxy type Environment variable: |
string |
|
The Proxy host Environment variable: |
string |
|
The Proxy port Environment variable: |
int |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
The Azure AD token to use for this operation. If present, then the requests towards OpenAI will include this in the Authorization header. Note that this property overrides the functionality of Environment variable: |
string |
|
The API version to use for this operation. This follows the YYYY-MM-DD format Environment variable: |
string |
|
Azure OpenAI API key Environment variable: |
string |
|
What sampling temperature to use, with values between 0 and 2. Higher values means the model will take more risks. A value of 0.9 is good for more creative applications, while 0 (argmax sampling) is good for ones with a well-defined answer. It is recommended to alter this or topP, but not both. Environment variable: |
double |
|
An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with topP probability mass. 0.1 means only the tokens comprising the top 10% probability mass are considered. It is recommended to alter this or temperature, but not both. Environment variable: |
double |
|
If specified, our system will make the best effort to sample deterministically, such that repeated requests with the same seed and parameters should return the same result. Determinism isn’t guaranteed. Support for reproducible output was first added in API version 2023-12-01-preview Environment variable: |
int |
|
The maximum number of tokens to generate in the completion. The token count of your prompt plus max_tokens can’t exceed the model’s context length. Most models have a context length of 2048 tokens (except for the newest models, which support 4096). Environment variable: |
int |
|
Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model’s likelihood to talk about new topics. Environment variable: |
double |
|
Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model’s likelihood to repeat the same line verbatim. Environment variable: |
double |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
The response format the model should use. Some models are not compatible with some response formats, make sure to review OpenAI documentation. Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
The Azure AD token to use for this operation. If present, then the requests towards OpenAI will include this in the Authorization header. Note that this property overrides the functionality of Environment variable: |
string |
|
The API version to use for this operation. This follows the YYYY-MM-DD format Environment variable: |
string |
|
Azure OpenAI API key Environment variable: |
string |
|
Whether embedding model requests should be logged Environment variable: |
boolean |
|
Whether embedding model responses should be logged Environment variable: |
boolean |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
This property will override the Environment variable: |
string |
|
The Azure AD token to use for this operation. If present, then the requests towards OpenAI will include this in the Authorization header. Note that this property overrides the functionality of Environment variable: |
string |
|
The API version to use for this operation. This follows the YYYY-MM-DD format Environment variable: |
string |
|
Azure OpenAI API key Environment variable: |
string |
|
Model name to use Environment variable: |
string |
|
Configure whether the generated images will be saved to disk. By default, persisting is disabled, but it is implicitly enabled when Environment variable: |
boolean |
|
The path where the generated images will be persisted to disk. This only applies of Environment variable: |
path |
|
The format in which the generated images are returned. Must be one of Environment variable: |
string |
|
The size of the generated images. Must be one of Must be one of Environment variable: |
string |
|
The quality of the image that will be generated.
This param is only supported for when the model is Environment variable: |
string |
|
The number of images to generate. Must be between 1 and 10. When the model is dall-e-3, only n=1 is supported. Environment variable: |
int |
|
The style of the generated images. Must be one of This param is only supported for when the model is Environment variable: |
string |
|
A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. Environment variable: |
string |
|
Whether image model requests should be logged Environment variable: |
boolean |
|
Whether image model responses should be logged Environment variable: |
boolean |
|
Quarkus LangChain4j - Bedrock |
Type |
Default |
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Whether the Bedrock client should log requests Environment variable: |
boolean |
|
Whether the Bedrock client should log responses Environment variable: |
boolean |
|
Whether chat model body in request and response should be logged Environment variable: |
boolean |
|
Region used by the bedrock runtime api. See Regions Supported. Environment variable: |
string |
|
Override the endpoint used by the bedrock client Environment variable: |
string |
|
Specify a custom credentials provider to use for the bedrock client. Identified by bean name. Environment variable: |
string |
|
The maximum number retries the aws sdk client will attempt. Environment variable: |
int |
|
Configure the amount of time to allow the client to complete the execution of an API call. This timeout covers the entire client execution except for marshalling. This includes request handler execution, all HTTP requests including retries, unmarshalling, etc. This value should always be positive, if present. Environment variable: |
|
|
Connect Timeout for Bedrock calls Environment variable: |
|
|
Read Timeout for Bedrock calls Environment variable: |
|
|
A string value in the form of Environment variable: |
string |
|
Proxy username, equivalent to the http.proxy or https.proxy JVM settings. Environment variable: |
string |
|
Proxy password, equivalent to the http.proxyPassword or https.proxyPassword JVM settings. Environment variable: |
string |
|
Hosts to access without proxy, similar to the http.nonProxyHosts or https.nonProxyHosts JVM settings. Please note that unlike the JVM settings, this property is empty by default. Environment variable: |
string |
|
If true, the REST clients will not provide additional contextual information (like REST client class and method names) when exception occurs during a client invocation. Environment variable: |
boolean |
|
The time in ms for which a connection remains unused in the connection pool before being evicted and closed. A timeout of Environment variable: |
int |
|
The size of the connection pool for this client. Environment variable: |
int |
|
If set to false disables the keep alive completely. Environment variable: |
boolean |
|
The class name of the host name verifier. The class must have a public no-argument constructor. Environment variable: |
string |
|
Set whether hostname verification is enabled. Default is enabled. This setting should not be disabled in production as it makes the client vulnerable to MITM attacks. Environment variable: |
boolean |
|
The trust store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The trust store password. Environment variable: |
string |
|
The type of the trust store. Defaults to "JKS". Environment variable: |
string |
|
The key store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The key store password. Environment variable: |
string |
|
The type of the key store. Defaults to "JKS". Environment variable: |
string |
|
The name of the TLS configuration to use. If not set and the default TLS configuration is configured ( Environment variable: |
string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Whether chat model body in request and response should be logged Environment variable: |
boolean |
|
Region used by the bedrock runtime api. See Regions Supported. Environment variable: |
string |
|
Override the endpoint used by the bedrock client Environment variable: |
string |
|
Specify a custom credentials provider to use for the bedrock client. Identified by bean name. Environment variable: |
string |
|
The maximum number retries the aws sdk client will attempt. Environment variable: |
int |
|
Configure the amount of time to allow the client to complete the execution of an API call. This timeout covers the entire client execution except for marshalling. This includes request handler execution, all HTTP requests including retries, unmarshalling, etc. This value should always be positive, if present. Environment variable: |
|
|
Model id to use. See Models Supported Environment variable: |
string |
|
What sampling temperature to use, between 0.0 and 1.0. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. It is generally recommended to set this or the Environment variable: |
double |
|
The maximum number of tokens to generate in the completion. The token count of your prompt plus Environment variable: |
int |
|
Double (0.0-1.0). Nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. It is generally recommended to set this or the Environment variable: |
double |
|
Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative Environment variable: |
int |
|
The custom text sequences that will cause the model to stop generating Environment variable: |
list of string |
|
Connect Timeout for Bedrock calls Environment variable: |
|
|
Read Timeout for Bedrock calls Environment variable: |
|
|
A string value in the form of Environment variable: |
string |
|
Proxy username, equivalent to the http.proxy or https.proxy JVM settings. Environment variable: |
string |
|
Proxy password, equivalent to the http.proxyPassword or https.proxyPassword JVM settings. Environment variable: |
string |
|
Hosts to access without proxy, similar to the http.nonProxyHosts or https.nonProxyHosts JVM settings. Please note that unlike the JVM settings, this property is empty by default. Environment variable: |
string |
|
If true, the REST clients will not provide additional contextual information (like REST client class and method names) when exception occurs during a client invocation. Environment variable: |
boolean |
|
The time in ms for which a connection remains unused in the connection pool before being evicted and closed. A timeout of Environment variable: |
int |
|
The size of the connection pool for this client. Environment variable: |
int |
|
If set to false disables the keep alive completely. Environment variable: |
boolean |
|
The class name of the host name verifier. The class must have a public no-argument constructor. Environment variable: |
string |
|
Set whether hostname verification is enabled. Default is enabled. This setting should not be disabled in production as it makes the client vulnerable to MITM attacks. Environment variable: |
boolean |
|
The trust store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The trust store password. Environment variable: |
string |
|
The type of the trust store. Defaults to "JKS". Environment variable: |
string |
|
The key store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The key store password. Environment variable: |
string |
|
The type of the key store. Defaults to "JKS". Environment variable: |
string |
|
The name of the TLS configuration to use. If not set and the default TLS configuration is configured ( Environment variable: |
string |
|
Whether embedding model requests should be logged Environment variable: |
boolean |
|
Whether embedding model responses should be logged Environment variable: |
boolean |
|
Whether embedding model body in request and response should be logged Environment variable: |
boolean |
|
Region used by the bedrock runtime api. See Regions Supported. Environment variable: |
string |
|
Override the endpoint used by the bedrock client Environment variable: |
string |
|
Specify a custom credentials provider to use for the bedrock client. Identified by bean name. Environment variable: |
string |
|
The maximum number retries the aws sdk client will attempt. Environment variable: |
int |
|
Configure the amount of time to allow the client to complete the execution of an API call. This timeout covers the entire client execution except for marshalling. This includes request handler execution, all HTTP requests including retries, unmarshalling, etc. This value should always be positive, if present. Environment variable: |
|
|
Model name to use Environment variable: |
string |
|
The number of dimensions the output embedding should have Environment variable: |
int |
|
Flag indicating whether to normalize the output embedding Environment variable: |
boolean |
|
Prepends special tokens to differentiate each type from one another. You should not mix different types together, except when mixing types for search and retrieval. In this case, embed your corpus with the search_document type and embedded queries with type search_query type. Environment variable: |
string |
|
Specifies how the API handles inputs longer than the maximum token length Environment variable: |
string |
|
Connect Timeout for Bedrock calls Environment variable: |
|
|
Read Timeout for Bedrock calls Environment variable: |
|
|
A string value in the form of Environment variable: |
string |
|
Proxy username, equivalent to the http.proxy or https.proxy JVM settings. Environment variable: |
string |
|
Proxy password, equivalent to the http.proxyPassword or https.proxyPassword JVM settings. Environment variable: |
string |
|
Hosts to access without proxy, similar to the http.nonProxyHosts or https.nonProxyHosts JVM settings. Please note that unlike the JVM settings, this property is empty by default. Environment variable: |
string |
|
If true, the REST clients will not provide additional contextual information (like REST client class and method names) when exception occurs during a client invocation. Environment variable: |
boolean |
|
The time in ms for which a connection remains unused in the connection pool before being evicted and closed. A timeout of Environment variable: |
int |
|
The size of the connection pool for this client. Environment variable: |
int |
|
If set to false disables the keep alive completely. Environment variable: |
boolean |
|
The class name of the host name verifier. The class must have a public no-argument constructor. Environment variable: |
string |
|
Set whether hostname verification is enabled. Default is enabled. This setting should not be disabled in production as it makes the client vulnerable to MITM attacks. Environment variable: |
boolean |
|
The trust store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The trust store password. Environment variable: |
string |
|
The type of the trust store. Defaults to "JKS". Environment variable: |
string |
|
The key store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The key store password. Environment variable: |
string |
|
The type of the key store. Defaults to "JKS". Environment variable: |
string |
|
The name of the TLS configuration to use. If not set and the default TLS configuration is configured ( Environment variable: |
string |
|
Type |
Default |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Whether the Bedrock client should log requests Environment variable: |
boolean |
|
Whether the Bedrock client should log responses Environment variable: |
boolean |
|
Whether chat model body in request and response should be logged Environment variable: |
boolean |
|
Region used by the bedrock runtime api. See Regions Supported. Environment variable: |
string |
|
Override the endpoint used by the bedrock client Environment variable: |
string |
|
Specify a custom credentials provider to use for the bedrock client. Identified by bean name. Environment variable: |
string |
|
The maximum number retries the aws sdk client will attempt. Environment variable: |
int |
|
Configure the amount of time to allow the client to complete the execution of an API call. This timeout covers the entire client execution except for marshalling. This includes request handler execution, all HTTP requests including retries, unmarshalling, etc. This value should always be positive, if present. Environment variable: |
|
|
Connect Timeout for Bedrock calls Environment variable: |
|
|
Read Timeout for Bedrock calls Environment variable: |
|
|
A string value in the form of Environment variable: |
string |
|
Proxy username, equivalent to the http.proxy or https.proxy JVM settings. Environment variable: |
string |
|
Proxy password, equivalent to the http.proxyPassword or https.proxyPassword JVM settings. Environment variable: |
string |
|
Hosts to access without proxy, similar to the http.nonProxyHosts or https.nonProxyHosts JVM settings. Please note that unlike the JVM settings, this property is empty by default. Environment variable: |
string |
|
If true, the REST clients will not provide additional contextual information (like REST client class and method names) when exception occurs during a client invocation. Environment variable: |
boolean |
|
The time in ms for which a connection remains unused in the connection pool before being evicted and closed. A timeout of Environment variable: |
int |
|
The size of the connection pool for this client. Environment variable: |
int |
|
If set to false disables the keep alive completely. Environment variable: |
boolean |
|
The class name of the host name verifier. The class must have a public no-argument constructor. Environment variable: |
string |
|
Set whether hostname verification is enabled. Default is enabled. This setting should not be disabled in production as it makes the client vulnerable to MITM attacks. Environment variable: |
boolean |
|
The trust store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The trust store password. Environment variable: |
string |
|
The type of the trust store. Defaults to "JKS". Environment variable: |
string |
|
The key store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The key store password. Environment variable: |
string |
|
The type of the key store. Defaults to "JKS". Environment variable: |
string |
|
The name of the TLS configuration to use. If not set and the default TLS configuration is configured ( Environment variable: |
string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Whether chat model body in request and response should be logged Environment variable: |
boolean |
|
Region used by the bedrock runtime api. See Regions Supported. Environment variable: |
string |
|
Override the endpoint used by the bedrock client Environment variable: |
string |
|
Specify a custom credentials provider to use for the bedrock client. Identified by bean name. Environment variable: |
string |
|
The maximum number retries the aws sdk client will attempt. Environment variable: |
int |
|
Configure the amount of time to allow the client to complete the execution of an API call. This timeout covers the entire client execution except for marshalling. This includes request handler execution, all HTTP requests including retries, unmarshalling, etc. This value should always be positive, if present. Environment variable: |
|
|
Model id to use. See Models Supported Environment variable: |
string |
|
What sampling temperature to use, between 0.0 and 1.0. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. It is generally recommended to set this or the Environment variable: |
double |
|
The maximum number of tokens to generate in the completion. The token count of your prompt plus Environment variable: |
int |
|
Double (0.0-1.0). Nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. It is generally recommended to set this or the Environment variable: |
double |
|
Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative Environment variable: |
int |
|
The custom text sequences that will cause the model to stop generating Environment variable: |
list of string |
|
Connect Timeout for Bedrock calls Environment variable: |
|
|
Read Timeout for Bedrock calls Environment variable: |
|
|
A string value in the form of Environment variable: |
string |
|
Proxy username, equivalent to the http.proxy or https.proxy JVM settings. Environment variable: |
string |
|
Proxy password, equivalent to the http.proxyPassword or https.proxyPassword JVM settings. Environment variable: |
string |
|
Hosts to access without proxy, similar to the http.nonProxyHosts or https.nonProxyHosts JVM settings. Please note that unlike the JVM settings, this property is empty by default. Environment variable: |
string |
|
If true, the REST clients will not provide additional contextual information (like REST client class and method names) when exception occurs during a client invocation. Environment variable: |
boolean |
|
The time in ms for which a connection remains unused in the connection pool before being evicted and closed. A timeout of Environment variable: |
int |
|
The size of the connection pool for this client. Environment variable: |
int |
|
If set to false disables the keep alive completely. Environment variable: |
boolean |
|
The class name of the host name verifier. The class must have a public no-argument constructor. Environment variable: |
string |
|
Set whether hostname verification is enabled. Default is enabled. This setting should not be disabled in production as it makes the client vulnerable to MITM attacks. Environment variable: |
boolean |
|
The trust store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The trust store password. Environment variable: |
string |
|
The type of the trust store. Defaults to "JKS". Environment variable: |
string |
|
The key store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The key store password. Environment variable: |
string |
|
The type of the key store. Defaults to "JKS". Environment variable: |
string |
|
The name of the TLS configuration to use. If not set and the default TLS configuration is configured ( Environment variable: |
string |
|
Whether embedding model requests should be logged Environment variable: |
boolean |
|
Whether embedding model responses should be logged Environment variable: |
boolean |
|
Whether embedding model body in request and response should be logged Environment variable: |
boolean |
|
Region used by the bedrock runtime api. See Regions Supported. Environment variable: |
string |
|
Override the endpoint used by the bedrock client Environment variable: |
string |
|
Specify a custom credentials provider to use for the bedrock client. Identified by bean name. Environment variable: |
string |
|
The maximum number retries the aws sdk client will attempt. Environment variable: |
int |
|
Configure the amount of time to allow the client to complete the execution of an API call. This timeout covers the entire client execution except for marshalling. This includes request handler execution, all HTTP requests including retries, unmarshalling, etc. This value should always be positive, if present. Environment variable: |
|
|
Model name to use Environment variable: |
string |
|
The number of dimensions the output embedding should have Environment variable: |
int |
|
Flag indicating whether to normalize the output embedding Environment variable: |
boolean |
|
Prepends special tokens to differentiate each type from one another. You should not mix different types together, except when mixing types for search and retrieval. In this case, embed your corpus with the search_document type and embedded queries with type search_query type. Environment variable: |
string |
|
Specifies how the API handles inputs longer than the maximum token length Environment variable: |
string |
|
Connect Timeout for Bedrock calls Environment variable: |
|
|
Read Timeout for Bedrock calls Environment variable: |
|
|
A string value in the form of Environment variable: |
string |
|
Proxy username, equivalent to the http.proxy or https.proxy JVM settings. Environment variable: |
string |
|
Proxy password, equivalent to the http.proxyPassword or https.proxyPassword JVM settings. Environment variable: |
string |
|
Hosts to access without proxy, similar to the http.nonProxyHosts or https.nonProxyHosts JVM settings. Please note that unlike the JVM settings, this property is empty by default. Environment variable: |
string |
|
If true, the REST clients will not provide additional contextual information (like REST client class and method names) when exception occurs during a client invocation. Environment variable: |
boolean |
|
The time in ms for which a connection remains unused in the connection pool before being evicted and closed. A timeout of Environment variable: |
int |
|
The size of the connection pool for this client. Environment variable: |
int |
|
If set to false disables the keep alive completely. Environment variable: |
boolean |
|
The class name of the host name verifier. The class must have a public no-argument constructor. Environment variable: |
string |
|
Set whether hostname verification is enabled. Default is enabled. This setting should not be disabled in production as it makes the client vulnerable to MITM attacks. Environment variable: |
boolean |
|
The trust store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The trust store password. Environment variable: |
string |
|
The type of the trust store. Defaults to "JKS". Environment variable: |
string |
|
The key store location. Can point to either a classpath resource or a file. Environment variable: |
string |
|
The key store password. Environment variable: |
string |
|
The type of the key store. Defaults to "JKS". Environment variable: |
string |
|
The name of the TLS configuration to use. If not set and the default TLS configuration is configured ( Environment variable: |
string |
|
Quarkus LangChain4j - Chroma |
Type |
Default |
If DevServices has been explicitly enabled or disabled. DevServices is generally enabled by default, unless there is an existing configuration present. When DevServices is enabled Quarkus will attempt to automatically configure and start a database when running in Dev or Test mode and when Docker is running. Environment variable: |
boolean |
|
The container image name to use, for container based DevServices providers. Environment variable: |
string |
|
Optional fixed port the dev service will listen to. If not defined, the port will be chosen randomly. Environment variable: |
int |
|
Indicates if the Chroma server managed by Quarkus Dev Services is shared. When shared, Quarkus looks for running containers using label-based service discovery. If a matching container is found, it is used, and so a second one is not started. Otherwise, Dev Services for Chroma starts a new container. The discovery uses the Container sharing is only used in dev mode. Environment variable: |
boolean |
|
The value of the This property is used when you need multiple shared Chroma servers. Environment variable: |
string |
|
Environment variables that are passed to the container. Environment variable: |
Map<String,String> |
|
URL where the Chroma database is listening for requests Environment variable: |
string |
required |
The collection name. Environment variable: |
string |
|
The timeout duration for the Chroma client. If not specified, 5 seconds will be used. Environment variable: |
||
Whether requests to Chroma should be logged Environment variable: |
boolean |
|
Whether responses from Chroma should be logged Environment variable: |
boolean |
|
The Chroma API version to use. V1 is deprecated (Chroma 0.x) and its support will be removed in the future. Please use Chroma 1.x which uses the V2 API. Environment variable: |
|
|
Quarkus LangChain4j - Cohere |
Type |
Default |
Whether the scoring model should be enabled. Environment variable: |
boolean |
|
Base URL of the Cohere API. Environment variable: |
string |
|
Cohere API key. Environment variable: |
string |
|
Timeout for Cohere calls. Environment variable: |
|
|
Reranking model to use. The current list of supported models can be found in the Cohere docs Environment variable: |
string |
|
Timeout for Cohere calls Environment variable: |
|
|
Whether embedding model requests should be logged. Environment variable: |
boolean |
|
Whether embedding model responses should be logged. Environment variable: |
boolean |
|
The maximum number of times to retry. 1 means exactly one attempt, with retrying disabled. Environment variable: |
int |
|
Type |
Default |
|
Base URL of the Cohere API. Environment variable: |
string |
|
Cohere API key. Environment variable: |
string |
|
Timeout for Cohere calls. Environment variable: |
|
|
Reranking model to use. The current list of supported models can be found in the Cohere docs Environment variable: |
string |
|
Timeout for Cohere calls Environment variable: |
|
|
Whether embedding model requests should be logged. Environment variable: |
boolean |
|
Whether embedding model responses should be logged. Environment variable: |
boolean |
|
The maximum number of times to retry. 1 means exactly one attempt, with retrying disabled. Environment variable: |
int |
|
Quarkus LangChain4j - Core |
Type |
Default |
Configure the type of The extension provides a default bean that configures If Users can choose to provide their own Environment variable: |
|
|
If DevServices has been explicitly enabled or disabled. DevServices is generally enabled by default, unless there is an existing configuration present. When DevServices is enabled Quarkus will attempt to automatically serve a model if there are any matching ones. Environment variable: |
boolean |
|
The default port where the inference server listens for requests Environment variable: |
int |
|
Instructs Ollama to preload a model in order to get faster response times Environment variable: |
boolean |
|
Configuration property to enable or disable the use of the {response schema} placeholder in the @SystemMessage/@UserMessage. Environment variable: |
boolean |
|
The maximum number of messages the configured Environment variable: |
int |
|
The maximum number of tokens the configured Environment variable: |
int |
|
Whether clients should log requests Environment variable: |
boolean |
|
Whether clients should log responses Environment variable: |
boolean |
|
Global timeout for requests to LLM APIs Environment variable: |
|
|
Global temperature for LLM APIs Environment variable: |
double |
|
Configures the maximum number of retries for the guardrail. Sets it to 0 to disable retries. Environment variable: |
int |
|
If enabled, the prompt is included on the generated spans Environment variable: |
boolean |
|
If enabled, the completion is included on the generated spans Environment variable: |
boolean |
|
If enabled, tool call arguments are included on the generated spans Environment variable: |
boolean |
|
If enabled, tool call results are included on the generated spans Environment variable: |
boolean |
|
Type |
Default |
|
The model provider to use Environment variable: |
string |
|
The model provider to use Environment variable: |
string |
|
The model provider to use Environment variable: |
string |
|
The model provider to use Environment variable: |
string |
|
The model provider to use Environment variable: |
string |
|
The model provider to use Environment variable: |
string |
|
The model provider to use Environment variable: |
string |
|
The model provider to use Environment variable: |
string |
|
The model provider to use Environment variable: |
string |
|
The model provider to use Environment variable: |
string |
|
Quarkus LangChain4j - GPULlama3 |
Type |
Default |
Determines whether the necessary GPULlama3 models are downloaded and included in the jar at build time. Currently, this option is only valid for Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Model name to use Environment variable: |
string |
|
Quantization of the model to use Environment variable: |
string |
|
Location on the file-system which serves as a cache for the models Environment variable: |
path |
|
What sampling temperature to use, between 0.0 and 1.0. Environment variable: |
double |
|
What sampling topP to use, between 0.0 and 1.0. Environment variable: |
double |
|
What seed value to use. Environment variable: |
int |
|
The maximum number of tokens to generate in the completion. Environment variable: |
int |
|
Whether to enable the integration. Set to Environment variable: |
boolean |
|
Whether GPULlama3 should log requests Environment variable: |
boolean |
|
Whether GPULlama3 client should log responses Environment variable: |
boolean |
|
Type |
Default |
|
Model name to use Environment variable: |
string |
|
Quantization of the model to use Environment variable: |
string |
|
What sampling temperature to use, between 0.0 and 1.0. Environment variable: |
double |
|
What sampling topP to use, between 0.0 and 1.0. Environment variable: |
double |
|
What seed value to use. Environment variable: |
int |
|
The maximum number of tokens to generate in the completion. Environment variable: |
int |
|
Whether to enable the integration. Set to Environment variable: |
boolean |
|
Whether GPULlama3 should log requests Environment variable: |
boolean |
|
Whether GPULlama3 client should log responses Environment variable: |
boolean |
|
Quarkus LangChain4j - Hugging Face |
Type |
Default |
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
HuggingFace API key Environment variable: |
string |
|
Timeout for HuggingFace calls Environment variable: |
|
|
The URL of the inference endpoint for the chat model. When using Hugging Face with the inference API, the URL is When using a deployed inference endpoint, the URL is the URL of the endpoint. When using a local hugging face model, the URL is the URL of the local model. Environment variable: |
|
|
Float (0.0-100.0). The temperature of the sampling operation. 1 means regular sampling, 0 means always take the highest score, 100.0 is getting closer to uniform probability Environment variable: |
double |
|
Int (0-250). The amount of new tokens to be generated, this does not include the input length it is a estimate of the size of generated text you want. Each new tokens slows down the request, so look for balance between response times and length of text generated Environment variable: |
int |
|
If set to Environment variable: |
boolean |
|
If the model is not ready, wait for it instead of receiving 503. It limits the number of requests required to get your inference done. It is advised to only set this flag to true after receiving a 503 error as it will limit hanging in your application to known places Environment variable: |
boolean |
|
Whether or not to use sampling ; use greedy decoding otherwise. Environment variable: |
boolean |
|
The number of highest probability vocabulary tokens to keep for top-k-filtering. Environment variable: |
int |
|
If set to less than Environment variable: |
double |
|
The parameter for repetition penalty. 1.0 means no penalty. See this paper for more details. Environment variable: |
double |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
The URL of the inference endpoint for the embedding. When using Hugging Face with the inference API, the URL is When using a deployed inference endpoint, the URL is the URL of the endpoint. When using a local hugging face model, the URL is the URL of the local model. Environment variable: |
|
|
If the model is not ready, wait for it instead of receiving 503. It limits the number of requests required to get your inference done. It is advised to only set this flag to true after receiving a 503 error as it will limit hanging in your application to known places Environment variable: |
boolean |
|
Whether the HuggingFace client should log requests Environment variable: |
boolean |
|
Whether the HuggingFace client should log responses Environment variable: |
boolean |
|
Whether or not to enable the integration. Defaults to Environment variable: |
boolean |
|
Type |
Default |
|
HuggingFace API key Environment variable: |
string |
|
Timeout for HuggingFace calls Environment variable: |
|
|
The URL of the inference endpoint for the chat model. When using Hugging Face with the inference API, the URL is When using a deployed inference endpoint, the URL is the URL of the endpoint. When using a local hugging face model, the URL is the URL of the local model. Environment variable: |
|
|
Float (0.0-100.0). The temperature of the sampling operation. 1 means regular sampling, 0 means always take the highest score, 100.0 is getting closer to uniform probability Environment variable: |
double |
|
Int (0-250). The amount of new tokens to be generated, this does not include the input length it is a estimate of the size of generated text you want. Each new tokens slows down the request, so look for balance between response times and length of text generated Environment variable: |
int |
|
If set to Environment variable: |
boolean |
|
If the model is not ready, wait for it instead of receiving 503. It limits the number of requests required to get your inference done. It is advised to only set this flag to true after receiving a 503 error as it will limit hanging in your application to known places Environment variable: |
boolean |
|
Whether or not to use sampling ; use greedy decoding otherwise. Environment variable: |
boolean |
|
The number of highest probability vocabulary tokens to keep for top-k-filtering. Environment variable: |
int |
|
If set to less than Environment variable: |
double |
|
The parameter for repetition penalty. 1.0 means no penalty. See this paper for more details. Environment variable: |
double |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
The URL of the inference endpoint for the embedding. When using Hugging Face with the inference API, the URL is When using a deployed inference endpoint, the URL is the URL of the endpoint. When using a local hugging face model, the URL is the URL of the local model. Environment variable: |
|
|
If the model is not ready, wait for it instead of receiving 503. It limits the number of requests required to get your inference done. It is advised to only set this flag to true after receiving a 503 error as it will limit hanging in your application to known places Environment variable: |
boolean |
|
Whether the HuggingFace client should log requests Environment variable: |
boolean |
|
Whether the HuggingFace client should log responses Environment variable: |
boolean |
|
Whether or not to enable the integration. Defaults to Environment variable: |
boolean |
|
Quarkus LangChain4j - Infinispan embedding store |
Type |
Default |
The name of the Infinispan client to use. These clients are configured by means of the Environment variable: |
string |
|
The dimension of the embedding vectors. This has to be the same as the dimension of vectors produced by the embedding model that you use. For example, AllMiniLmL6V2QuantizedEmbeddingModel produces vectors of dimension 384. OpenAI’s text-embedding-ada-002 produces vectors of dimension 1536. Environment variable: |
long |
required |
Name of the cache that will be used in Infinispan when searching for related embeddings. If this cache doesn’t exist, it will be created. Environment variable: |
string |
|
The maximum distance. The most distance between vectors is how close or far apart two embeddings are. Environment variable: |
int |
|
Quarkus LangChain4j - Jlama |
Type |
Default |
Determines whether the necessary Jlama models are downloaded and included in the jar at build time. Currently, this option is only valid for Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Model name to use Environment variable: |
string |
|
Model name to use Environment variable: |
string |
|
Location on the file-system which serves as a cache for the models Environment variable: |
path |
|
What sampling temperature to use, between 0.0 and 1.0. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. It is generally recommended to set this or the Environment variable: |
double |
|
The maximum number of tokens to generate in the completion. The token count of your prompt plus Environment variable: |
int |
|
Whether to enable the integration. Set to Environment variable: |
boolean |
|
Whether Jlama should log requests Environment variable: |
boolean |
|
Whether Jlama client should log responses Environment variable: |
boolean |
|
Type |
Default |
|
Model name to use Environment variable: |
string |
|
Model name to use Environment variable: |
string |
|
What sampling temperature to use, between 0.0 and 1.0. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. It is generally recommended to set this or the Environment variable: |
double |
|
The maximum number of tokens to generate in the completion. The token count of your prompt plus Environment variable: |
int |
|
Whether to enable the integration. Set to Environment variable: |
boolean |
|
Whether Jlama should log requests Environment variable: |
boolean |
|
Whether Jlama client should log responses Environment variable: |
boolean |
|
Quarkus LangChain4j - Llama3 - Java |
Type |
Default |
Determines whether the necessary Jlama models are downloaded and included in the jar at build time. Currently, this option is only valid for Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Model name to use Environment variable: |
string |
|
Quantization of the model to use Environment variable: |
string |
|
Llama3.java supports AOT model preloading, enabling 0-overhead, instant inference, with minimal TTFT (time-to-first-token). A specialized, larger binary will be generated, with no parsing overhead for that particular model. It can still run other models, although incurring the usual parsing overhead. Environment variable: |
boolean |
|
Location on the file-system which serves as a cache for the models Environment variable: |
path |
|
Temperature in [0,inf] Environment variable: |
double |
|
Number of steps to run for < 0 = limited by context length Environment variable: |
int |
|
Whether to enable the integration. Set to Environment variable: |
boolean |
|
Whether Jlama should log requests Environment variable: |
boolean |
|
Whether Jlama client should log responses Environment variable: |
boolean |
|
Type |
Default |
|
Model name to use Environment variable: |
string |
|
Quantization of the model to use Environment variable: |
string |
|
Llama3.java supports AOT model preloading, enabling 0-overhead, instant inference, with minimal TTFT (time-to-first-token). A specialized, larger binary will be generated, with no parsing overhead for that particular model. It can still run other models, although incurring the usual parsing overhead. Environment variable: |
boolean |
|
Temperature in [0,inf] Environment variable: |
double |
|
Number of steps to run for < 0 = limited by context length Environment variable: |
int |
|
Whether to enable the integration. Set to Environment variable: |
boolean |
|
Whether Jlama should log requests Environment variable: |
boolean |
|
Whether Jlama client should log responses Environment variable: |
boolean |
|
Quarkus Langchain4j - Memory Store - MongoDB |
Type |
Default |
The name of the MongoDB client to use. These clients are configured by means of the Environment variable: |
string |
|
The name of the database to use. Environment variable: |
string |
|
The name of the collection to use. Environment variable: |
string |
|
Quarkus Langchain4j - Memory Store - Redis |
Type |
Default |
The name of the Redis client to use. These clients are configured by means of the Environment variable: |
string |
|
Quarkus LangChain4j - Milvus embedding store |
Type |
Default |
Whether Dev Services for Milvus are enabled or not. Environment variable: |
boolean |
|
Container image for Milvus. Environment variable: |
string |
|
Optional fixed port the Milvus dev service will listen to. If not defined, the port will be chosen randomly. Environment variable: |
int |
|
Indicates if the Dev Service containers managed by Quarkus for Milvus are shared. Environment variable: |
boolean |
|
Service label to apply to created Dev Services containers. Environment variable: |
string |
|
The URL of the Milvus server. Environment variable: |
string |
required |
The port of the Milvus server. Environment variable: |
int |
required |
The authentication token for the Milvus server. Environment variable: |
string |
|
The username for the Milvus server. Environment variable: |
string |
|
The password for the Milvus server. Environment variable: |
string |
|
The timeout duration for the Milvus client. If not specified, 5 seconds will be used. Environment variable: |
||
Name of the database. Environment variable: |
string |
|
Create the collection if it does not exist yet. Environment variable: |
boolean |
|
Name of the collection. Environment variable: |
string |
|
Dimension of the vectors. Only applicable when the collection yet has to be created. Environment variable: |
int |
|
Name of the field that contains the ID of the vector. Environment variable: |
string |
|
Name of the field that contains the text from which the vector was calculated. Environment variable: |
string |
|
Name of the field that contains JSON metadata associated with the text. Environment variable: |
string |
|
Name of the field to store the vector in. Environment variable: |
string |
|
Description of the collection. Environment variable: |
string |
|
The index type to use for the collection. Environment variable: |
|
|
The metric type to use for searching. Environment variable: |
|
|
The consistency level. Environment variable: |
|
|
Quarkus LangChain4j - Mistral AI |
Type |
Default |
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Base URL of Mistral API Environment variable: |
string |
|
Mistral API key Environment variable: |
string |
|
Timeout for Mistral calls Environment variable: |
|
|
Model name to use Environment variable: |
string |
|
What sampling temperature to use, between 0.0 and 1.0. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. It is generally recommended to set this or the Environment variable: |
double |
|
The maximum number of tokens to generate in the completion. The token count of your prompt plus Environment variable: |
int |
|
Double (0.0-1.0). Nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. It is generally recommended to set this or the Environment variable: |
double |
|
Whether to inject a safety prompt before all conversations Environment variable: |
boolean |
|
The seed to use for random sampling. If set, different calls will generate deterministic results. Environment variable: |
int |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Model name to use Environment variable: |
string |
|
Whether embedding model requests should be logged Environment variable: |
boolean |
|
Whether embedding model responses should be logged Environment variable: |
boolean |
|
Model name to use Environment variable: |
string |
|
Whether moderation model requests should be logged Environment variable: |
boolean |
|
Whether moderation model responses should be logged Environment variable: |
boolean |
|
Whether the Mistral client should log requests Environment variable: |
boolean |
|
Whether the Mistral client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Type |
Default |
|
Base URL of Mistral API Environment variable: |
string |
|
Mistral API key Environment variable: |
string |
|
Timeout for Mistral calls Environment variable: |
|
|
Model name to use Environment variable: |
string |
|
What sampling temperature to use, between 0.0 and 1.0. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. It is generally recommended to set this or the Environment variable: |
double |
|
The maximum number of tokens to generate in the completion. The token count of your prompt plus Environment variable: |
int |
|
Double (0.0-1.0). Nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. It is generally recommended to set this or the Environment variable: |
double |
|
Whether to inject a safety prompt before all conversations Environment variable: |
boolean |
|
The seed to use for random sampling. If set, different calls will generate deterministic results. Environment variable: |
int |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Model name to use Environment variable: |
string |
|
Whether embedding model requests should be logged Environment variable: |
boolean |
|
Whether embedding model responses should be logged Environment variable: |
boolean |
|
Model name to use Environment variable: |
string |
|
Whether moderation model requests should be logged Environment variable: |
boolean |
|
Whether moderation model responses should be logged Environment variable: |
boolean |
|
Whether the Mistral client should log requests Environment variable: |
boolean |
|
Whether the Mistral client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Quarkus LangChain4j - Ollama |
Type |
Default |
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
If Dev Services for Ollama has been explicitly enabled or disabled. Dev Services are generally enabled by default, unless there is an existing configuration present. Environment variable: |
boolean |
|
The Ollama container image to use. Environment variable: |
string |
|
Model to use Environment variable: |
string |
|
Model to use. According to Ollama docs, the default value is Environment variable: |
string |
|
Base URL where the Ollama serving is running Environment variable: |
string |
|
If set, the named TLS configuration with the configured name will be applied to the REST Client Environment variable: |
string |
|
Timeout for Ollama calls Environment variable: |
|
|
Whether the Ollama client should log requests Environment variable: |
boolean |
|
Whether the Ollama client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
The temperature of the model. Increasing the temperature will make the model answer with more variability. A lower temperature will make the model answer more conservatively. Environment variable: |
double |
|
Maximum number of tokens to predict when generating text Environment variable: |
int |
|
Sets the stop sequences to use. When this pattern is encountered the LLM will stop generating text and return Environment variable: |
list of string |
|
Works together with top-k. A higher value (e.g., 0.95) will lead to more diverse text, while a lower value (e.g., 0.5) will generate more focused and conservative text Environment variable: |
double |
|
Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative Environment variable: |
int |
|
With a static number the result is always the same. With a random number the result varies Example:
Environment variable: |
int |
|
The format to return a response in. Format can be Environment variable: |
string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
The temperature of the model. Increasing the temperature will make the model answer with more variability. A lower temperature will make the model answer more conservatively. Environment variable: |
double |
|
Maximum number of tokens to predict when generating text Environment variable: |
int |
|
Sets the stop sequences to use. When this pattern is encountered the LLM will stop generating text and return Environment variable: |
list of string |
|
Works together with top-k. A higher value (e.g., 0.95) will lead to more diverse text, while a lower value (e.g., 0.5) will generate more focused and conservative text Environment variable: |
double |
|
Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative Environment variable: |
int |
|
Whether embedding model requests should be logged Environment variable: |
boolean |
|
Whether embedding model responses should be logged Environment variable: |
boolean |
|
Type |
Default |
|
Model to use Environment variable: |
string |
|
Model to use. According to Ollama docs, the default value is Environment variable: |
string |
|
Base URL where the Ollama serving is running Environment variable: |
string |
|
If set, the named TLS configuration with the configured name will be applied to the REST Client Environment variable: |
string |
|
Timeout for Ollama calls Environment variable: |
|
|
Whether the Ollama client should log requests Environment variable: |
boolean |
|
Whether the Ollama client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
The temperature of the model. Increasing the temperature will make the model answer with more variability. A lower temperature will make the model answer more conservatively. Environment variable: |
double |
|
Maximum number of tokens to predict when generating text Environment variable: |
int |
|
Sets the stop sequences to use. When this pattern is encountered the LLM will stop generating text and return Environment variable: |
list of string |
|
Works together with top-k. A higher value (e.g., 0.95) will lead to more diverse text, while a lower value (e.g., 0.5) will generate more focused and conservative text Environment variable: |
double |
|
Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative Environment variable: |
int |
|
With a static number the result is always the same. With a random number the result varies Example:
Environment variable: |
int |
|
The format to return a response in. Format can be Environment variable: |
string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
The temperature of the model. Increasing the temperature will make the model answer with more variability. A lower temperature will make the model answer more conservatively. Environment variable: |
double |
|
Maximum number of tokens to predict when generating text Environment variable: |
int |
|
Sets the stop sequences to use. When this pattern is encountered the LLM will stop generating text and return Environment variable: |
list of string |
|
Works together with top-k. A higher value (e.g., 0.95) will lead to more diverse text, while a lower value (e.g., 0.5) will generate more focused and conservative text Environment variable: |
double |
|
Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative Environment variable: |
int |
|
Whether embedding model requests should be logged Environment variable: |
boolean |
|
Whether embedding model responses should be logged Environment variable: |
boolean |
|
Quarkus LangChain4j - OpenAI |
Type |
Default |
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
Base URL of OpenAI API Environment variable: |
string |
|
If set, the named TLS configuration with the configured name will be applied to the REST Client Environment variable: |
string |
|
OpenAI API key Environment variable: |
string |
|
OpenAI Organization ID (https://platform.openai.com/docs/api-reference/organization-optional) Environment variable: |
string |
|
Timeout for OpenAI calls Environment variable: |
|
|
The maximum number of times to retry. 1 means exactly one attempt, with retrying disabled. Environment variable: |
int |
|
Whether the OpenAI client should log requests Environment variable: |
boolean |
|
Whether the OpenAI client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
The Proxy type Environment variable: |
string |
|
The Proxy host Environment variable: |
string |
|
The Proxy port Environment variable: |
int |
|
Model name to use Environment variable: |
string |
|
What sampling temperature to use, with values between 0 and 2. Higher values means the model will take more risks. A value of 0.9 is good for more creative applications, while 0 (argmax sampling) is good for ones with a well-defined answer. It is recommended to alter this or topP, but not both. Environment variable: |
double |
|
An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with topP probability mass. 0.1 means only the tokens comprising the top 10% probability mass are considered. It is recommended to alter this or temperature, but not both. Environment variable: |
double |
|
An upper bound for the number of tokens that can be generated for a completion, including visible output tokens and reasoning tokens. Environment variable: |
int |
|
Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model’s likelihood to talk about new topics. Environment variable: |
double |
|
Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model’s likelihood to repeat the same line verbatim. Environment variable: |
double |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
The response format the model should use. Some models are not compatible with some response formats, make sure to review OpenAI documentation. Environment variable: |
string |
|
Whether responses follow JSON Schema for Structured Outputs Environment variable: |
boolean |
|
The list of stop words to use. Environment variable: |
list of string |
|
Constrains effort on reasoning for reasoning models. Currently supported values are Note: The Environment variable: |
string |
|
Specifies the processing type used for serving the request. If set to When the service tier parameter is set, the response body will include the Environment variable: |
string |
|
Model name to use Environment variable: |
string |
|
Whether embedding model requests should be logged Environment variable: |
boolean |
|
Whether embedding model responses should be logged Environment variable: |
boolean |
|
A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. Environment variable: |
string |
|
Model name to use Environment variable: |
string |
|
Whether moderation model requests should be logged Environment variable: |
boolean |
|
Whether moderation model responses should be logged Environment variable: |
boolean |
|
Model name to use Environment variable: |
string |
|
Configure whether the generated images will be saved to disk. By default, persisting is disabled, but it is implicitly enabled when Environment variable: |
boolean |
|
The path where the generated images will be persisted to disk. This only applies of Environment variable: |
path |
|
The format in which the generated images are returned. Must be one of Environment variable: |
string |
|
The size of the generated images. Must be one of Must be one of Environment variable: |
string |
|
The quality of the image that will be generated.
This param is only supported for when the model is Environment variable: |
string |
|
The number of images to generate. Must be between 1 and 10. When the model is dall-e-3, only n=1 is supported. Environment variable: |
int |
|
The style of the generated images. Must be one of This param is only supported for when the model is Environment variable: |
string |
|
A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. Environment variable: |
string |
|
Whether image model requests should be logged Environment variable: |
boolean |
|
Whether image model responses should be logged Environment variable: |
boolean |
|
Type |
Default |
|
Base URL of OpenAI API Environment variable: |
string |
|
If set, the named TLS configuration with the configured name will be applied to the REST Client Environment variable: |
string |
|
OpenAI API key Environment variable: |
string |
|
OpenAI Organization ID (https://platform.openai.com/docs/api-reference/organization-optional) Environment variable: |
string |
|
Timeout for OpenAI calls Environment variable: |
|
|
The maximum number of times to retry. 1 means exactly one attempt, with retrying disabled. Environment variable: |
int |
|
Whether the OpenAI client should log requests Environment variable: |
boolean |
|
Whether the OpenAI client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
The Proxy type Environment variable: |
string |
|
The Proxy host Environment variable: |
string |
|
The Proxy port Environment variable: |
int |
|
Model name to use Environment variable: |
string |
|
What sampling temperature to use, with values between 0 and 2. Higher values means the model will take more risks. A value of 0.9 is good for more creative applications, while 0 (argmax sampling) is good for ones with a well-defined answer. It is recommended to alter this or topP, but not both. Environment variable: |
double |
|
An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with topP probability mass. 0.1 means only the tokens comprising the top 10% probability mass are considered. It is recommended to alter this or temperature, but not both. Environment variable: |
double |
|
An upper bound for the number of tokens that can be generated for a completion, including visible output tokens and reasoning tokens. Environment variable: |
int |
|
Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model’s likelihood to talk about new topics. Environment variable: |
double |
|
Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model’s likelihood to repeat the same line verbatim. Environment variable: |
double |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
The response format the model should use. Some models are not compatible with some response formats, make sure to review OpenAI documentation. Environment variable: |
string |
|
Whether responses follow JSON Schema for Structured Outputs Environment variable: |
boolean |
|
The list of stop words to use. Environment variable: |
list of string |
|
Constrains effort on reasoning for reasoning models. Currently supported values are Note: The Environment variable: |
string |
|
Specifies the processing type used for serving the request. If set to When the service tier parameter is set, the response body will include the Environment variable: |
string |
|
Model name to use Environment variable: |
string |
|
Whether embedding model requests should be logged Environment variable: |
boolean |
|
Whether embedding model responses should be logged Environment variable: |
boolean |
|
A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. Environment variable: |
string |
|
Model name to use Environment variable: |
string |
|
Whether moderation model requests should be logged Environment variable: |
boolean |
|
Whether moderation model responses should be logged Environment variable: |
boolean |
|
Model name to use Environment variable: |
string |
|
Configure whether the generated images will be saved to disk. By default, persisting is disabled, but it is implicitly enabled when Environment variable: |
boolean |
|
The path where the generated images will be persisted to disk. This only applies of Environment variable: |
path |
|
The format in which the generated images are returned. Must be one of Environment variable: |
string |
|
The size of the generated images. Must be one of Must be one of Environment variable: |
string |
|
The quality of the image that will be generated.
This param is only supported for when the model is Environment variable: |
string |
|
The number of images to generate. Must be between 1 and 10. When the model is dall-e-3, only n=1 is supported. Environment variable: |
int |
|
The style of the generated images. Must be one of This param is only supported for when the model is Environment variable: |
string |
|
A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. Environment variable: |
string |
|
Whether image model requests should be logged Environment variable: |
boolean |
|
Whether image model responses should be logged Environment variable: |
boolean |
|
Quarkus LangChain4j - OpenId Connect (OIDC) Client McpClientAuthProvider |
Type |
Default |
Whether the OIDC Client McpClientAuthProvider should be enabled Environment variable: |
boolean |
|
Quarkus LangChain4j - OpenId Connect (OIDC) McpAuthProvider |
Type |
Default |
Whether the OIDC McpClientAuthProvider should be enabled Environment variable: |
boolean |
|
Quarkus LangChain4j - OpenId Connect (OIDC) ModelAuthProvider |
Type |
Default |
Whether the OIDC ModelAuthProvider should be enabled Environment variable: |
boolean |
|
Quarkus LangChain4j - OpenShift AI |
Type |
Default |
Whether the model should be enabled Environment variable: |
boolean |
|
Base URL where OpenShift AI serving is running, such as Environment variable: |
|
|
Timeout for OpenShift AI calls Environment variable: |
|
|
Whether the OpenShift AI client should log requests Environment variable: |
boolean |
|
Whether the OpenShift AI client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Model to use Environment variable: |
string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Type |
Default |
|
Base URL where OpenShift AI serving is running, such as Environment variable: |
|
|
Timeout for OpenShift AI calls Environment variable: |
|
|
Whether the OpenShift AI client should log requests Environment variable: |
boolean |
|
Whether the OpenShift AI client should log responses Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Model to use Environment variable: |
string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Quarkus LangChain4j - pgvector |
Type |
Default |
The name of the configured Postgres datasource to use for this store. If not set, the default datasource from the Agroal extension will be used. Environment variable: |
string |
|
The table name for storing embeddings Environment variable: |
string |
|
The dimension of the embedding vectors. This has to be the same as the dimension of vectors produced by the embedding model that you use. For example, AllMiniLmL6V2QuantizedEmbeddingModel produces vectors of dimension 384. OpenAI’s text-embedding-ada-002 produces vectors of dimension 1536. Environment variable: |
int |
required |
Use index or not Environment variable: |
boolean |
|
index size Environment variable: |
int |
|
Whether the table should be created if not already existing. Environment variable: |
boolean |
|
Whether the table should be dropped prior to being created. Environment variable: |
boolean |
|
Whether the PG extension should be created on Start. By Default, if it’s dev or test environment the value is overridden to true Environment variable: |
boolean |
|
Metadata type:
Default value: COMBINED_JSON Environment variable: |
|
|
Metadata Definition: SQL definition of metadata field(s). By default, "metadata JSON NULL" configured. This is only suitable if using the JSON metadata type. If using JSONB metadata type, this should in most cases be set to If using COLUMNS metadata type, this should be a list of columns, one column for each desired metadata field. Example: condominium_id uuid null, user uuid null Environment variable: |
list of string |
|
Metadata Indexes, list of fields to use as index. For instance:
Environment variable: |
list of string |
|
Index Type:
Environment variable: |
string |
|
Quarkus LangChain4j - Qdrant embedding store |
Type |
Default |
Whether Dev Services for Qdrant are enabled or not. Environment variable: |
boolean |
|
Container image for Qdrant. Environment variable: |
string |
|
Optional fixed port the Qdrant dev service will listen to. If not defined, the port will be chosen randomly. Environment variable: |
int |
|
Indicates if the Dev Service containers managed by Quarkus for Qdrant are shared. Environment variable: |
boolean |
|
Service label to apply to created Dev Services containers. Environment variable: |
string |
|
Distance function used for comparing vectors Environment variable: |
|
required |
Size of the vectors Environment variable: |
long |
|
The URL of the Qdrant server. Environment variable: |
string |
required |
The gRPC port of the Qdrant server. Defaults to 6334 Environment variable: |
int |
|
The Qdrant API key to authenticate with. Environment variable: |
string |
|
Whether to use TLS(HTTPS). Defaults to false. Environment variable: |
boolean |
|
The field name of the text segment in the payload. Defaults to "text_segment" Environment variable: |
string |
|
The name of the collection. Environment variable: |
string |
required |
Quarkus LangChain4j - Redis embedding store |
Type |
Default |
The name of the Redis client to use. These clients are configured by means of the Environment variable: |
string |
|
The dimension of the embedding vectors. This has to be the same as the dimension of vectors produced by the embedding model that you use. For example, AllMiniLmL6V2QuantizedEmbeddingModel produces vectors of dimension 384. OpenAI’s text-embedding-ada-002 produces vectors of dimension 1536. Environment variable: |
long |
required |
Name of the index that will be used in Redis when searching for related embeddings. If this index doesn’t exist, it will be created. Environment variable: |
string |
|
Names of fields that will store textual metadata associated with embeddings. NOTE: Filtering based on textual metadata fields is not supported at the moment. Environment variable: |
list of string |
|
Names of fields that will store numeric metadata associated with embeddings. Environment variable: |
list of string |
|
Metric used to compute the distance between two vectors. Environment variable: |
|
|
Name of the key that will be used to store the embedding vector. Environment variable: |
string |
|
Name of the key that will be used to store the embedded text. Environment variable: |
string |
|
Prefix to be applied to all keys by the embedding store. Embeddings are stored in Redis under a key that is the concatenation of this prefix and the embedding ID. If the configured prefix does not ends with Environment variable: |
string |
|
Algorithm used to index the embedding vectors. Environment variable: |
|
|
Quarkus LangChain4j - Vertex AI |
Type |
Default |
Whether the model should be enabled Environment variable: |
boolean |
|
The unique identifier of the project Environment variable: |
string |
|
GCP location Environment variable: |
string |
|
Publisher of model Environment variable: |
string |
|
Meant to be used for testing only in order to override the base URL used by the client Environment variable: |
string |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
The Proxy type Environment variable: |
string |
|
The Proxy host Environment variable: |
string |
|
The Proxy port Environment variable: |
int |
|
Whether the Vertex AI client should log requests Environment variable: |
boolean |
|
Whether the Vertex AI client should log responses Environment variable: |
boolean |
|
The id of the model to use Environment variable: |
string |
|
The temperature is used for sampling during response generation, which occurs when topP and topK are applied. Temperature controls the degree of randomness in token selection. Lower temperatures are good for prompts that require a less open-ended or creative response, while higher temperatures can lead to more diverse or creative results. A temperature of 0 means that the highest probability tokens are always selected. In this case, responses for a given prompt are mostly deterministic, but a small amount of variation is still possible. If the model returns a response that’s too generic, too short, or the model gives a fallback response, try increasing the temperature. Environment variable: |
double |
|
Maximum number of tokens that can be generated in the response. A token is approximately four characters. 100 tokens correspond to roughly 60-80 words. Specify a lower value for shorter responses and a higher value for potentially longer responses. Environment variable: |
int |
|
Top-P changes how the model selects tokens for output. Tokens are selected from the most (see top-K) to least probable until the sum of their probabilities equals the top-P value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-P value is 0.5, then the model will select either A or B as the next token by using temperature and excludes C as a candidate. Specify a lower value for less random responses and a higher value for more random responses. Environment variable: |
double |
|
Top-K changes how the model selects tokens for output. A top-K of 1 means the next selected token is the most probable among all tokens in the model’s vocabulary (also called greedy decoding), while a top-K of 3 means that the next token is selected from among the three most probable tokens by using temperature. For each token selection step, the top-K tokens with the highest probabilities are sampled. Then tokens are further filtered based on top-P with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Environment variable: |
int |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Type |
Default |
|
The unique identifier of the project Environment variable: |
string |
|
GCP location Environment variable: |
string |
|
Publisher of model Environment variable: |
string |
|
Meant to be used for testing only in order to override the base URL used by the client Environment variable: |
string |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
The Proxy type Environment variable: |
string |
|
The Proxy host Environment variable: |
string |
|
The Proxy port Environment variable: |
int |
|
Whether the Vertex AI client should log requests Environment variable: |
boolean |
|
Whether the Vertex AI client should log responses Environment variable: |
boolean |
|
The id of the model to use Environment variable: |
string |
|
The temperature is used for sampling during response generation, which occurs when topP and topK are applied. Temperature controls the degree of randomness in token selection. Lower temperatures are good for prompts that require a less open-ended or creative response, while higher temperatures can lead to more diverse or creative results. A temperature of 0 means that the highest probability tokens are always selected. In this case, responses for a given prompt are mostly deterministic, but a small amount of variation is still possible. If the model returns a response that’s too generic, too short, or the model gives a fallback response, try increasing the temperature. Environment variable: |
double |
|
Maximum number of tokens that can be generated in the response. A token is approximately four characters. 100 tokens correspond to roughly 60-80 words. Specify a lower value for shorter responses and a higher value for potentially longer responses. Environment variable: |
int |
|
Top-P changes how the model selects tokens for output. Tokens are selected from the most (see top-K) to least probable until the sum of their probabilities equals the top-P value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-P value is 0.5, then the model will select either A or B as the next token by using temperature and excludes C as a candidate. Specify a lower value for less random responses and a higher value for more random responses. Environment variable: |
double |
|
Top-K changes how the model selects tokens for output. A top-K of 1 means the next selected token is the most probable among all tokens in the model’s vocabulary (also called greedy decoding), while a top-K of 3 means that the next token is selected from among the three most probable tokens by using temperature. For each token selection step, the top-K tokens with the highest probabilities are sampled. Then tokens are further filtered based on top-P with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Environment variable: |
int |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Quarkus LangChain4j - Vertex AI Gemini |
Type |
Default |
Whether the model should be enabled Environment variable: |
boolean |
|
Whether the model should be enabled Environment variable: |
boolean |
|
The unique identifier of the project Environment variable: |
string |
|
GCP location Environment variable: |
string |
|
Publisher of model Environment variable: |
string |
|
Meant to be used for testing only in order to override the base URL used by the client Environment variable: |
string |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
The Proxy type Environment variable: |
string |
|
The Proxy host Environment variable: |
string |
|
The Proxy port Environment variable: |
int |
|
Whether the Vertex AI client should log requests Environment variable: |
boolean |
|
Whether the Vertex AI client should log responses Environment variable: |
boolean |
|
Timeout for requests to gemini APIs Environment variable: |
|
|
The id of the model to use. Environment variable: |
string |
|
The temperature is used for sampling during response generation, which occurs when topP and topK are applied. Temperature controls the degree of randomness in token selection. Lower temperatures are good for prompts that require a less open-ended or creative response, while higher temperatures can lead to more diverse or creative results. A temperature of 0 means that the highest probability tokens are always selected. In this case, responses for a given prompt are mostly deterministic, but a small amount of variation is still possible. If the model returns a response that’s too generic, too short, or the model gives a fallback response, try increasing the temperature. Range for gemini-2.5-flash: 0.0 - 2.0 Default for gemini-2.5-flash: 1.0 Environment variable: |
double |
|
Maximum number of tokens that can be generated in the response. A token is approximately four characters. 100 tokens correspond to roughly 60-80 words. Specify a lower value for shorter responses and a higher value for potentially longer responses. Environment variable: |
int |
|
Top-P changes how the model selects tokens for output. Tokens are selected from the most (see top-K) to least probable until the sum of their probabilities equals the top-P value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-P value is 0.5, then the model will select either A or B as the next token by using temperature and excludes C as a candidate. Specify a lower value for less random responses and a higher value for more random responses. Range: 0.0 - 1.0 Default for gemini-2.5-flash: 0.95 Environment variable: |
double |
|
Top-K changes how the model selects tokens for output. A top-K of 1 means the next selected token is the most probable among all tokens in the model’s vocabulary (also called greedy decoding), while a top-K of 3 means that the next token is selected from among the three most probable tokens by using temperature. For each token selection step, the top-K tokens with the highest probabilities are sampled. Then tokens are further filtered based on top-P with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Range: 1-40 gemini-2.5-flash doesn’t support topK Environment variable: |
int |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Global timeout for requests to gemini APIs Environment variable: |
|
|
The id of the model to use. Environment variable: |
string |
|
Reduced dimension for the output embedding Environment variable: |
int |
|
Optional task type for which the embeddings will be used. Can only be set for models/embedding-001 Possible values: TASK_TYPE_UNSPECIFIED, RETRIEVAL_QUERY, RETRIEVAL_DOCUMENT, SEMANTIC_SIMILARITY, CLASSIFICATION, CLUSTERING, QUESTION_ANSWERING, FACT_VERIFICATION Environment variable: |
string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Global timeout for requests to gemini APIs Environment variable: |
|
|
Type |
Default |
|
The unique identifier of the project Environment variable: |
string |
|
GCP location Environment variable: |
string |
|
Publisher of model Environment variable: |
string |
|
Meant to be used for testing only in order to override the base URL used by the client Environment variable: |
string |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
The Proxy type Environment variable: |
string |
|
The Proxy host Environment variable: |
string |
|
The Proxy port Environment variable: |
int |
|
Whether the Vertex AI client should log requests Environment variable: |
boolean |
|
Whether the Vertex AI client should log responses Environment variable: |
boolean |
|
Timeout for requests to gemini APIs Environment variable: |
|
|
The id of the model to use. Environment variable: |
string |
|
The temperature is used for sampling during response generation, which occurs when topP and topK are applied. Temperature controls the degree of randomness in token selection. Lower temperatures are good for prompts that require a less open-ended or creative response, while higher temperatures can lead to more diverse or creative results. A temperature of 0 means that the highest probability tokens are always selected. In this case, responses for a given prompt are mostly deterministic, but a small amount of variation is still possible. If the model returns a response that’s too generic, too short, or the model gives a fallback response, try increasing the temperature. Range for gemini-2.5-flash: 0.0 - 2.0 Default for gemini-2.5-flash: 1.0 Environment variable: |
double |
|
Maximum number of tokens that can be generated in the response. A token is approximately four characters. 100 tokens correspond to roughly 60-80 words. Specify a lower value for shorter responses and a higher value for potentially longer responses. Environment variable: |
int |
|
Top-P changes how the model selects tokens for output. Tokens are selected from the most (see top-K) to least probable until the sum of their probabilities equals the top-P value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-P value is 0.5, then the model will select either A or B as the next token by using temperature and excludes C as a candidate. Specify a lower value for less random responses and a higher value for more random responses. Range: 0.0 - 1.0 Default for gemini-2.5-flash: 0.95 Environment variable: |
double |
|
Top-K changes how the model selects tokens for output. A top-K of 1 means the next selected token is the most probable among all tokens in the model’s vocabulary (also called greedy decoding), while a top-K of 3 means that the next token is selected from among the three most probable tokens by using temperature. For each token selection step, the top-K tokens with the highest probabilities are sampled. Then tokens are further filtered based on top-P with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Range: 1-40 gemini-2.5-flash doesn’t support topK Environment variable: |
int |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Global timeout for requests to gemini APIs Environment variable: |
|
|
The id of the model to use. Environment variable: |
string |
|
Reduced dimension for the output embedding Environment variable: |
int |
|
Optional task type for which the embeddings will be used. Can only be set for models/embedding-001 Possible values: TASK_TYPE_UNSPECIFIED, RETRIEVAL_QUERY, RETRIEVAL_DOCUMENT, SEMANTIC_SIMILARITY, CLASSIFICATION, CLUSTERING, QUESTION_ANSWERING, FACT_VERIFICATION Environment variable: |
string |
|
Whether chat model requests should be logged Environment variable: |
boolean |
|
Whether chat model responses should be logged Environment variable: |
boolean |
|
Global timeout for requests to gemini APIs Environment variable: |
|
|
Quarkus LangChain4j - Watsonx |
Type |
Default |
Whether the model should be enabled. Environment variable: |
boolean |
|
Whether the embedding model should be enabled. Environment variable: |
boolean |
|
Whether the scoring model should be enabled. Environment variable: |
boolean |
|
Specifies the mode of interaction with the LLM model. This property allows you to choose between two modes of operation:
Allowable values: Environment variable: |
string |
|
Specifies the base URL of the watsonx.ai API. A list of all available URLs is provided in the IBM Watsonx.ai documentation at the this link. Environment variable: |
string |
|
IBM Cloud API key. Environment variable: |
string |
|
Timeout for watsonx.ai calls. Environment variable: |
|
|
The version date for the API of the form YYYY-MM-DD. Environment variable: |
string |
|
The space that contains the resource. Either Environment variable: |
string |
|
The project that contains the resource. Either Environment variable: |
string |
|
Whether the watsonx.ai client should log requests. Environment variable: |
boolean |
|
Whether the watsonx.ai client should log responses. Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Base URL of the IAM Authentication API. Environment variable: |
|
|
Timeout for IAM authentication calls. Environment variable: |
|
|
Grant type for the IAM Authentication API. Environment variable: |
string |
|
Base URL of the Cloud Object Storage API. Environment variable: |
string |
required |
The ID of the connection asset that contains the credentials required to access the data. Environment variable: |
string |
required |
The name of the bucket containing the input document. Environment variable: |
string |
required |
The ID of the connection asset used to store the extracted results. Environment variable: |
string |
required |
The name of the bucket where the output files will be written. Environment variable: |
string |
required |
Whether the Cloud Object Storage client should log requests. Environment variable: |
boolean |
|
Whether the Cloud Object Storage client should log responses. Environment variable: |
boolean |
|
Specifies the model to use for the chat completion. A list of all available models is provided in the IBM watsonx.ai documentation at the this link. To use a model, locate the Environment variable: |
string |
|
Specifies how the model should choose which tool to call during a request. This value can be:
If Setting this value influences the tool-calling behavior of the model when no specific tool is required. Environment variable: |
|
|
Specifies the name of a specific tool that the model must call. When set, the model will be forced to call the specified tool. The name must exactly match one of the available tools defined for the service. Environment variable: |
string |
|
Positive values penalize new tokens based on their existing frequency in the generated text, reducing the likelihood of the model repeating the same lines verbatim. Possible values: Environment variable: |
double |
|
Specifies whether to return the log probabilities of the output tokens. If set to Environment variable: |
boolean |
|
An integer specifying the number of most likely tokens to return at each token position, each with an associated log probability. The option Possible values: Environment variable: |
int |
|
The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model’s context length. Set to 0 for the model’s configured max generated tokens. Environment variable: |
int |
|
Specifies how many chat completion choices to generate for each input message. Environment variable: |
int |
|
Applies a penalty to new tokens based on whether they already appear in the generated text so far, encouraging the model to introduce new topics rather than repeat itself. Possible values: Environment variable: |
double |
|
Random number generator seed to use in sampling mode for experimental repeatability. Environment variable: |
int |
|
Defines one or more stop sequences that will cause the model to stop generating further tokens if any of them are encountered in the output. This allows control over where the model should end its response. If a stop sequence is encountered before the minimum number of tokens has been generated, it will be ignored. Possible values: Environment variable: |
list of string |
|
Specifies the sampling temperature to use in the generation process. Higher values (e.g. Possible values: Environment variable: |
double |
|
An alternative to sampling with Possible values: Environment variable: |
double |
|
Specifies the desired format for the model’s output. Allowable values: Environment variable: |
string |
|
Whether chat model requests should be logged. Environment variable: |
boolean |
|
Whether chat model responses should be logged. Environment variable: |
boolean |
|
The id of the model to be used. All available models are listed in the IBM Watsonx.ai documentation at the link: following link. To use a model, locate the Environment variable: |
string |
|
Represents the strategy used for picking the tokens during generation of the output text. During text generation when parameter value is set to Allowable values: Environment variable: |
string |
|
Represents the factor of exponential decay. Larger values correspond to more aggressive decay. Possible values: Environment variable: |
double |
|
A number of generated tokens after which this should take effect. Possible values: Environment variable: |
int |
|
The maximum number of new tokens to be generated. The maximum supported value for this field depends on the model being used. How the "token" is defined depends on the tokenizer and vocabulary size, which in turn depends on the model. Often the tokens are a mix of full words and sub-words. Depending on the users plan, and on the model being used, there may be an enforced maximum number of new tokens. Possible values: Environment variable: |
int |
|
If stop sequences are given, they are ignored until minimum tokens are generated. Possible values: Environment variable: |
int |
|
Random number generator seed to use in sampling mode for experimental repeatability. Possible values: Environment variable: |
int |
|
Stop sequences are one or more strings which will cause the text generation to stop if/when they are produced as part of the output. Stop sequences encountered prior to the minimum number of tokens being generated will be ignored. Possible values: Environment variable: |
list of string |
|
A value used to modify the next-token probabilities in Possible values: Environment variable: |
double |
|
The number of highest probability vocabulary tokens to keep for top-k-filtering. Only applies for Possible values: Environment variable: |
int |
|
Similar to Possible values: Environment variable: |
double |
|
Represents the penalty for penalizing tokens that have already been generated or belong to the context. The value Possible values: Environment variable: |
double |
|
Represents the maximum number of input tokens accepted. This can be used to avoid requests failing due to input being longer than configured limits. If the text is truncated, then it truncates the start of the input (on the left), so the end of the input will remain the same. If this value exceeds the maximum sequence length (refer to the documentation to find this value for the model) then the call will fail if the total number of tokens exceeds the maximum sequence length. Zero means don’t truncate. Possible values: Environment variable: |
int |
|
Pass Environment variable: |
boolean |
|
Whether chat model requests should be logged. Environment variable: |
boolean |
|
Whether chat model responses should be logged. Environment variable: |
boolean |
|
Delimiter used to concatenate the ChatMessage elements into a single string. By setting this property, you can define your preferred way of concatenating messages to ensure that the prompt is structured in the correct way. Environment variable: |
string |
|
Specifies the ID of the model to be used. A list of all available models is provided in the IBM watsonx.ai documentation at the this link. To use a model, locate the Environment variable: |
string |
|
Specifies the maximum number of input tokens accepted. This can be used to prevent requests from failing due to input exceeding the configured token limits. If the input exceeds the specified token limit, the input will be truncated from the end (right side), ensuring that the start of the input remains intact. If the provided value exceeds the model’s maximum sequence length (refer to the documentation for the model’s maximum sequence length), the request will fail if the total number of tokens exceeds the maximum limit. Environment variable: |
int |
|
Whether embedding model requests should be logged. Environment variable: |
boolean |
|
Whether embedding model responses should be logged. Environment variable: |
boolean |
|
The id of the model to be used. All available models are listed in the IBM Watsonx.ai documentation at the link: following link. To use a model, locate the Environment variable: |
string |
|
Specifies the maximum number of input tokens accepted. This helps to avoid requests failing due to input exceeding the configured token limits. If the input exceeds the specified token limit, the text will be truncated from the end (right side), ensuring that the start of the input remains intact. If the provided value exceeds the model’s maximum sequence length (refer to the documentation for the model’s maximum sequence length), the request will fail if the total number of tokens exceeds the maximum limit. Environment variable: |
int |
|
Whether embedding model requests should be logged. Environment variable: |
boolean |
|
Whether embedding model responses should be logged. Environment variable: |
boolean |
|
Base URL for the built-in service. All available URLs are listed in the IBM Watsonx.ai documentation at the following link. Note: If empty, the URL is automatically calculated based on the Environment variable: |
string |
|
IBM Cloud API key. If empty, the api key inherits the value from the Environment variable: |
string |
|
Timeout for built-in tools APIs. If empty, the api key inherits the value from the Environment variable: |
|
|
Whether the built-in rest client should log requests. Environment variable: |
boolean |
|
Whether the built-in rest client should log responses. Environment variable: |
boolean |
|
Maximum number of search results. Possible values: Environment variable: |
int |
|
Type |
Default |
|
Specifies the mode of interaction with the LLM model. This property allows you to choose between two modes of operation:
Allowable values: Environment variable: |
string |
|
Specifies the base URL of the watsonx.ai API. A list of all available URLs is provided in the IBM Watsonx.ai documentation at the this link. Environment variable: |
string |
|
IBM Cloud API key. Environment variable: |
string |
|
Timeout for watsonx.ai calls. Environment variable: |
|
|
The version date for the API of the form YYYY-MM-DD. Environment variable: |
string |
|
The space that contains the resource. Either Environment variable: |
string |
|
The project that contains the resource. Either Environment variable: |
string |
|
Whether the watsonx.ai client should log requests. Environment variable: |
boolean |
|
Whether the watsonx.ai client should log responses. Environment variable: |
boolean |
|
Whether to enable the integration. Defaults to Environment variable: |
boolean |
|
Base URL of the IAM Authentication API. Environment variable: |
|
|
Timeout for IAM authentication calls. Environment variable: |
|
|
Grant type for the IAM Authentication API. Environment variable: |
string |
|
Base URL of the Cloud Object Storage API. Environment variable: |
string |
required |
The ID of the connection asset that contains the credentials required to access the data. Environment variable: |
string |
required |
The name of the bucket containing the input document. Environment variable: |
string |
required |
The ID of the connection asset used to store the extracted results. Environment variable: |
string |
required |
The name of the bucket where the output files will be written. Environment variable: |
string |
required |
Whether the Cloud Object Storage client should log requests. Environment variable: |
boolean |
|
Whether the Cloud Object Storage client should log responses. Environment variable: |
boolean |
|
Specifies the model to use for the chat completion. A list of all available models is provided in the IBM watsonx.ai documentation at the this link. To use a model, locate the Environment variable: |
string |
|
Specifies how the model should choose which tool to call during a request. This value can be:
If Setting this value influences the tool-calling behavior of the model when no specific tool is required. Environment variable: |
|
|
Specifies the name of a specific tool that the model must call. When set, the model will be forced to call the specified tool. The name must exactly match one of the available tools defined for the service. Environment variable: |
string |
|
Positive values penalize new tokens based on their existing frequency in the generated text, reducing the likelihood of the model repeating the same lines verbatim. Possible values: Environment variable: |
double |
|
Specifies whether to return the log probabilities of the output tokens. If set to Environment variable: |
boolean |
|
An integer specifying the number of most likely tokens to return at each token position, each with an associated log probability. The option Possible values: Environment variable: |
int |
|
The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model’s context length. Set to 0 for the model’s configured max generated tokens. Environment variable: |
int |
|
Specifies how many chat completion choices to generate for each input message. Environment variable: |
int |
|
Applies a penalty to new tokens based on whether they already appear in the generated text so far, encouraging the model to introduce new topics rather than repeat itself. Possible values: Environment variable: |
double |
|
Random number generator seed to use in sampling mode for experimental repeatability. Environment variable: |
int |
|
Defines one or more stop sequences that will cause the model to stop generating further tokens if any of them are encountered in the output. This allows control over where the model should end its response. If a stop sequence is encountered before the minimum number of tokens has been generated, it will be ignored. Possible values: Environment variable: |
list of string |
|
Specifies the sampling temperature to use in the generation process. Higher values (e.g. Possible values: Environment variable: |
double |
|
An alternative to sampling with Possible values: Environment variable: |
double |
|
Specifies the desired format for the model’s output. Allowable values: Environment variable: |
string |
|
Whether chat model requests should be logged. Environment variable: |
boolean |
|
Whether chat model responses should be logged. Environment variable: |
boolean |
|
The id of the model to be used. All available models are listed in the IBM Watsonx.ai documentation at the link: following link. To use a model, locate the Environment variable: |
string |
|
Represents the strategy used for picking the tokens during generation of the output text. During text generation when parameter value is set to Allowable values: Environment variable: |
string |
|
Represents the factor of exponential decay. Larger values correspond to more aggressive decay. Possible values: Environment variable: |
double |
|
A number of generated tokens after which this should take effect. Possible values: Environment variable: |
int |
|
The maximum number of new tokens to be generated. The maximum supported value for this field depends on the model being used. How the "token" is defined depends on the tokenizer and vocabulary size, which in turn depends on the model. Often the tokens are a mix of full words and sub-words. Depending on the users plan, and on the model being used, there may be an enforced maximum number of new tokens. Possible values: Environment variable: |
int |
|
If stop sequences are given, they are ignored until minimum tokens are generated. Possible values: Environment variable: |
int |
|
Random number generator seed to use in sampling mode for experimental repeatability. Possible values: Environment variable: |
int |
|
Stop sequences are one or more strings which will cause the text generation to stop if/when they are produced as part of the output. Stop sequences encountered prior to the minimum number of tokens being generated will be ignored. Possible values: Environment variable: |
list of string |
|
A value used to modify the next-token probabilities in Possible values: Environment variable: |
double |
|
The number of highest probability vocabulary tokens to keep for top-k-filtering. Only applies for Possible values: Environment variable: |
int |
|
Similar to Possible values: Environment variable: |
double |
|
Represents the penalty for penalizing tokens that have already been generated or belong to the context. The value Possible values: Environment variable: |
double |
|
Represents the maximum number of input tokens accepted. This can be used to avoid requests failing due to input being longer than configured limits. If the text is truncated, then it truncates the start of the input (on the left), so the end of the input will remain the same. If this value exceeds the maximum sequence length (refer to the documentation to find this value for the model) then the call will fail if the total number of tokens exceeds the maximum sequence length. Zero means don’t truncate. Possible values: Environment variable: |
int |
|
Pass Environment variable: |
boolean |
|
Whether chat model requests should be logged. Environment variable: |
boolean |
|
Whether chat model responses should be logged. Environment variable: |
boolean |
|
Delimiter used to concatenate the ChatMessage elements into a single string. By setting this property, you can define your preferred way of concatenating messages to ensure that the prompt is structured in the correct way. Environment variable: |
string |
|
Specifies the ID of the model to be used. A list of all available models is provided in the IBM watsonx.ai documentation at the this link. To use a model, locate the Environment variable: |
string |
|
Specifies the maximum number of input tokens accepted. This can be used to prevent requests from failing due to input exceeding the configured token limits. If the input exceeds the specified token limit, the input will be truncated from the end (right side), ensuring that the start of the input remains intact. If the provided value exceeds the model’s maximum sequence length (refer to the documentation for the model’s maximum sequence length), the request will fail if the total number of tokens exceeds the maximum limit. Environment variable: |
int |
|
Whether embedding model requests should be logged. Environment variable: |
boolean |
|
Whether embedding model responses should be logged. Environment variable: |
boolean |
|
The id of the model to be used. All available models are listed in the IBM Watsonx.ai documentation at the link: following link. To use a model, locate the Environment variable: |
string |
|
Specifies the maximum number of input tokens accepted. This helps to avoid requests failing due to input exceeding the configured token limits. If the input exceeds the specified token limit, the text will be truncated from the end (right side), ensuring that the start of the input remains intact. If the provided value exceeds the model’s maximum sequence length (refer to the documentation for the model’s maximum sequence length), the request will fail if the total number of tokens exceeds the maximum limit. Environment variable: |
int |
|
Whether embedding model requests should be logged. Environment variable: |
boolean |
|
Whether embedding model responses should be logged. Environment variable: |
boolean |
|
Quarkus LangChain4j - Weaviate |
Type |
Default |
If DevServices has been explicitly enabled or disabled. DevServices is generally enabled by default, unless there is an existing configuration present. When DevServices is enabled Quarkus will attempt to automatically configure and start a database when running in Dev or Test mode and when Docker is running. Environment variable: |
boolean |
|
The container image name to use, for container based DevServices providers. If you want to use Redis Stack modules (bloom, graph, search…), use: Environment variable: |
string |
|
Optional fixed port the dev service will listen to. If not defined, the port will be chosen randomly. Environment variable: |
int |
|
Indicates if the Redis server managed by Quarkus Dev Services is shared. When shared, Quarkus looks for running containers using label-based service discovery. If a matching container is found, it is used, and so a second one is not started. Otherwise, Dev Services for Redis starts a new container. The discovery uses the Container sharing is only used in dev mode. Environment variable: |
boolean |
|
The value of the This property is used when you need multiple shared Weaviate servers. Environment variable: |
string |
|
Environment variables that are passed to the container. Environment variable: |
Map<String,String> |
|
The Weaviate API key to authenticate with. Environment variable: |
string |
|
The scheme, e.g. "https" of cluster URL. Find it under Details of your Weaviate cluster. Environment variable: |
string |
|
The URL of the Weaviate server. Environment variable: |
string |
|
The gRPC port of the Weaviate server. Defaults to 8080 Environment variable: |
int |
|
The gRPC port of the Weaviate server. Defaults to 50051 Environment variable: |
int |
|
The gRPC connection is secured. Environment variable: |
boolean |
|
Use gRPC instead of http for batch inserts only. Will still be used for search. Environment variable: |
boolean |
|
The object class you want to store, e.g. "MyGreatClass". Must start from an uppercase letter. Environment variable: |
string |
|
The name of the field that contains the text of a Environment variable: |
string |
|
If true (default), then Environment variable: |
boolean |
|
Consistency level: ONE, QUORUM (default) or ALL. Environment variable: |
|
|
Metadata keys that should be persisted. The default in Weaviate [], however it is required to specify at least one for the EmbeddingStore to work. Thus, we use "tags" as default Environment variable: |
list of string |
|
The name of the field where Environment variable: |
string |
|
|
About the Duration format
To write duration values, use the standard You can also use a simplified format, starting with a number:
In other cases, the simplified format is translated to the
|
|
About the MemorySize format
A size configuration option recognizes strings in this format (shown as a regular expression): If no suffix is given, assume bytes. |