llama.cpp
Local or self-hosted servers where the model boundary stays operator-owned. This page covers the built-inllamacpp provider kind directly.
At A Glance
| Field | Value |
|---|---|
| Built-in kind | llamacpp |
| Provider group | Local And Self-Hosted Providers |
| Protocol family | openai_chat_completions |
| Feature family | openai_compatible |
| Auth scheme | bearer |
| Credential envs | none required by default |
| Aliases | llama.cpp, llama_cpp |
| Default base URL | http://127.0.0.1:8080 |
| Request endpoint | http://127.0.0.1:8080/v1/chat/completions |
| Models endpoint | http://127.0.0.1:8080/v1/models |
Minimal Config
Verify It
list-models is unreliable for this account or region, pin an explicit provider.model or add preferred_models instead of leaving recovery implicit.
Auth And Routing Contract
| Contract | Value |
|---|---|
| Auth optional | yes |
| Model probe auth optional | yes |
| Default API key env | none |
| OAuth env | none |
| Primary request route | http://127.0.0.1:8080/v1/chat/completions |
| Primary model-catalog route | http://127.0.0.1:8080/v1/models |
Operator Notes
- Authentication is optional by default on this profile, which is useful for local or self-hosted servers during first setup.
Related Docs
- Continue to Providers And Models when you still need the broader chooser page.
- Continue to Provider Guides for the full built-in provider matrix.
- Continue to Provider Recipes for representative rollout recipes.
- Continue to Configuration Patterns for the shared provider-profile shape.