Containerized local LLM stack for the Framework Desktop / Strix Halo,
plus the OpenCode harness on the Mac side.
- pyinfra/framework/: pyinfra deploy targeting the box
- llama.cpp (Vulkan), vLLM (ROCm), Ollama (ROCm with HSA override
for gfx1151), OpenWebUI
- Beszel (host + container + AMD GPU dashboard via sysfs)
- OpenLIT (LLM fleet metrics)
- Phoenix (per-trace agent waterfall)
- OpenHands (autonomous agent in a Docker sandbox)
- opencode/: OpenCode config + Phoenix bridge plugin (OTel exporter)
- install.sh deploys to ~/.config/opencode/
- StrixHaloSetup.md / StrixHaloMemory.md / Roadmap.md / TODO.md:
documentation and planning
- testing/qwen3-coder-30b/: small evaluation harness
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
26 lines
998 B
YAML
26 lines
998 B
YAML
# OpenWebUI — ChatGPT-like web UI in front of Ollama. Pre-configured to
|
|
# use the host's Ollama instance and the project's SearXNG for web
|
|
# search. Default port 3000.
|
|
#
|
|
# Persistent state (users, conversations, uploaded docs, RAG vector
|
|
# index) lives at /srv/docker/openwebui/data so backups touch one path.
|
|
services:
|
|
openwebui:
|
|
image: ghcr.io/open-webui/open-webui:main
|
|
container_name: openwebui
|
|
restart: unless-stopped
|
|
ports:
|
|
- "3000:8080"
|
|
extra_hosts:
|
|
# Lets the container reach Ollama on the host's :11434 without
|
|
# needing to share Docker networks.
|
|
- "host.docker.internal:host-gateway"
|
|
environment:
|
|
- OLLAMA_BASE_URL=http://host.docker.internal:11434
|
|
# Built-in web search via the project's SearXNG instance.
|
|
- ENABLE_RAG_WEB_SEARCH=true
|
|
- RAG_WEB_SEARCH_ENGINE=searxng
|
|
- SEARXNG_QUERY_URL=https://searxng.n0n.io/search?q=<query>&format=json
|
|
volumes:
|
|
- /srv/docker/openwebui/data:/app/backend/data
|