Upstream docs: https://docs.ollama.com/
Everything not listed in this document should behave the same as upstream Ollama. If a feature, setting, or behavior is not mentioned here, the upstream documentation is accurate and fully applicable.
Ollama makes it easy to get up and running with self-hosted, open source large language models (LLMs). It supports a wide range of models from the Ollama library.
- Image and Container Runtime
- Volume and Data Layout
- Installation and First-Run Flow
- Configuration Management
- Network Access and Interfaces
- Actions (StartOS UI)
- Dependencies
- Backups and Restore
- Health Checks
- Limitations and Differences
- What Is Unchanged from Upstream
- Contributing
- Quick Reference for AI Consumers
| Property | Value |
|---|---|
| Image | ollama/ollama (upstream unmodified) |
| Architectures | x86_64, aarch64 |
| Entrypoint | Default upstream entrypoint |
| Volume | Mount Point | Purpose |
|---|---|---|
main |
/root/.ollama |
All Ollama data (models, blobs, manifests) |
Key directories on the main volume:
models/-- downloaded model files (blobs and manifests)
| Step | Upstream | StartOS |
|---|---|---|
| Installation | curl -fsSL https://ollama.com/install.sh | sh |
Install from marketplace or sideload .s9pk |
| Start service | ollama serve |
Automatic via StartOS |
| Pull models | ollama pull <model> |
Use API or a connected UI (e.g. Open WebUI) |
Key difference: On StartOS, the Ollama API is exposed as a network interface. Use a compatible client or UI service (such as Open WebUI) to interact with it.
Ollama on StartOS runs with default upstream configuration. No user-configurable settings are currently exposed through StartOS actions.
Configuration NOT exposed on StartOS:
OLLAMA_HOST-- fixed:0.0.0.0:11434OLLAMA_MODELS-- fixed:/root/.ollamaOLLAMA_NUM_PARALLEL-- uses upstream defaultOLLAMA_MAX_LOADED_MODELS-- uses upstream default- GPU/CUDA settings -- uses upstream auto-detection
- Proxy settings
| Interface | Port | Protocol | Type | Purpose |
|---|---|---|---|---|
| Ollama API | 11434 | HTTP | API | Model inference and management API |
Access methods (StartOS 0.4.0):
- LAN IP with unique port
<hostname>.localwith unique port- Tor
.onionaddress - Custom domains (if configured)
API endpoints (subset):
| Endpoint | Method | Purpose |
|---|---|---|
/api/generate |
POST | Generate text completion |
/api/chat |
POST | Chat completion |
/api/pull |
POST | Download a model |
/api/tags |
GET | List local models |
/api/show |
POST | Show model info |
/api/delete |
DELETE | Remove a model |
None. Ollama is managed entirely through its API.
None. Ollama is a standalone application.
Included in backup:
mainvolume -- all Ollama data including downloaded models
Restore behavior:
- All models and data are restored
- No reconfiguration needed
Note: Backups can be very large depending on the number and size of downloaded models. A single 7B parameter model is typically 4-5 GB.
| Check | Method | Grace Period |
|---|---|---|
| Ollama API | Port listening on 11434 | Default |
Messages:
- Success: "Your Ollama API is ready"
- Error: "Error launching your Ollama API"
- No GPU passthrough -- StartOS does not currently support GPU passthrough, so inference runs on CPU only
- No exposed configuration -- environment variables and runtime settings cannot be changed through StartOS
- Large storage requirements -- models are stored on-device and can consume significant disk space (4-5 GB per 7B model)
- Memory requirements -- 8 GB RAM minimum for 7B models, 16 GB for 13B, 32 GB for 33B+
- CPU-only performance -- without GPU acceleration, inference is significantly slower than GPU-accelerated setups
- Full Ollama API compatibility
- All supported model formats (GGUF, Safetensors via conversion)
- Model pulling from Ollama library
- Chat and generate endpoints
- Concurrent request handling
- Model loading and unloading
- Modelfile support for custom models
- Embedding generation
- All client library compatibility (Python, JavaScript, Go, etc.)
See CONTRIBUTING.md for build instructions and development workflow.
package_id: ollama
image: ollama/ollama
architectures:
- x86_64
- aarch64
volumes:
main: /root/.ollama
ports:
api: 11434
dependencies: none
startos_managed_env_vars: []
actions: none