Models
The top-level models section declares AI models that are used by your Compose application. These models are typically pulled as OCI artifacts, run by a model runner, and exposed as an API that your service containers can consume.
Services can only access models when explicitly granted by a models attribute within the services top-level element.
Examples
Example 1
services:
app:
image: app
models:
- ai_model
models:
ai_model:
model: ai/modelIn this basic example:
- The app service uses the
ai_model. - The
ai_modelis defined as an OCI artifact (ai/model) that is pulled and served by the model runner. - Docker Compose injects connection info, for example
AI_MODEL_URL, into the container.
Example 2
services:
app:
image: app
models:
my_model:
endpoint_var: MODEL_URL
models:
my_model:
model: ai/model
context_size: 1024
runtime_flags:
- "--a-flag"
- "--another-flag=42"In this advanced setup:
- The service app references
my_modelusing the long syntax. - Compose injects the model runner's URL as the environment variable
MODEL_URL.
Attributes
model(required): The OCI artifact identifier for the model. This is what Compose pulls and runs via the model runner.context_size: Defines the maximum token context size for the model.runtime_flags: A list of raw command-line flags passed to the inference engine when the model is started.
Additional resources
For more examples and information on using model, see
Use AI models in Compose

Formed in 2009, the Archive Team (not to be confused with the archive.org Archive-It Team) is a rogue archivist collective dedicated to saving copies of rapidly dying or deleted websites for the sake of history and digital heritage. The group is 100% composed of volunteers and interested parties, and has expanded into a large amount of related projects for saving online and digital history.
