Containerized local LLM stack for the Framework Desktop / Strix Halo,
plus the OpenCode harness on the Mac side.
- pyinfra/framework/: pyinfra deploy targeting the box
- llama.cpp (Vulkan), vLLM (ROCm), Ollama (ROCm with HSA override
for gfx1151), OpenWebUI
- Beszel (host + container + AMD GPU dashboard via sysfs)
- OpenLIT (LLM fleet metrics)
- Phoenix (per-trace agent waterfall)
- OpenHands (autonomous agent in a Docker sandbox)
- opencode/: OpenCode config + Phoenix bridge plugin (OTel exporter)
- install.sh deploys to ~/.config/opencode/
- StrixHaloSetup.md / StrixHaloMemory.md / Roadmap.md / TODO.md:
documentation and planning
- testing/qwen3-coder-30b/: small evaluation harness
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
42 lines
950 B
JSON
42 lines
950 B
JSON
{
|
|
"$schema": "https://opencode.ai/config.json",
|
|
"experimental": {
|
|
"openTelemetry": true
|
|
},
|
|
"plugin": ["./.opencode/plugin/phoenix-bridge.js"],
|
|
"provider": {
|
|
"framework": {
|
|
"npm": "@ai-sdk/openai-compatible",
|
|
"name": "Framework Desktop (Strix Halo)",
|
|
"options": {
|
|
"baseURL": "http://framework:11434/v1"
|
|
},
|
|
"models": {
|
|
"qwen3-coder:30b": {
|
|
"name": "Qwen3 Coder 30B (local)",
|
|
"limit": {
|
|
"context": 131072,
|
|
"output": 16384
|
|
}
|
|
}
|
|
}
|
|
}
|
|
},
|
|
"mcp": {
|
|
"playwright": {
|
|
"type": "local",
|
|
"command": ["npx", "-y", "@playwright/mcp@latest"],
|
|
"enabled": true
|
|
},
|
|
"searxng": {
|
|
"type": "local",
|
|
"command": ["npx", "-y", "mcp-searxng"],
|
|
"enabled": true,
|
|
"environment": {
|
|
"SEARXNG_URL": "https://searxng.n0n.io"
|
|
}
|
|
}
|
|
},
|
|
"model": "framework/qwen3-coder:30b"
|
|
}
|