Podman AI Lab Chat Models
Podman AI Lab is a local AI development environment provided as an extension to Podman Desktop. It simplifies experimentation with AI by providing a curated catalog of models and recipes, all runnable on your local machine via containerized infrastructure.
Podman AI Lab includes an inference server that is compatible with the OpenAI REST API, making it usable with the quarkus-langchain4j-openai
extension.
Prerequisites
To use Podman AI Lab with Quarkus:
-
Install Podman Desktop.
-
From the Podman Desktop UI, install the Podman AI Lab extension from the Extensions catalog. See this guide for step-by-step instructions.
-
Launch Podman AI Lab and download a model (such as
granite-7b
) through the UI. -
Start the inference server by selecting the model and clicking Run.
Dependency
Add the OpenAI extension to your project:
<dependency>
<groupId>io.quarkiverse.langchain4j</groupId>
<artifactId>quarkus-langchain4j-openai</artifactId>
<version>1.1.0.CR2</version>
</dependency>
Even though this is not an actual OpenAI deployment, the API is compatible and works seamlessly with this extension.
Only the chat model API is supported at the moment. Other OpenAI APIs (like embeddings or moderation) are not available through Podman AI Lab. |
Configuration
Once the inference server is running (e.g., on port 44079
), configure your application like so:
quarkus.langchain4j.openai.base-url=http://localhost:44079/v1
# Responses may be slightly slower than cloud models
quarkus.langchain4j.openai.timeout=60s
You can verify the port from the Podman AI Lab UI under the Running Inference Server section. |
Podman AI Lab currently supports only a single running model at a time.
This means that any |