fix(cloud): mount llm.cloud.yaml over llm.yaml; restrict to vllm+ollama only
Some checks failed
CI / test (push) Failing after 17s

Remove claude_code, github_copilot, and anthropic from all cloud fallback
orders — cloud accounts must not route through personal/dev LLM backends.
vllm_research and ollama_research are the only permitted research backends.
llm.cloud.yaml is now bind-mounted at /app/config/llm.yaml in compose.cloud.yml.
This commit is contained in:
pyr0ball 2026-04-01 19:59:01 -07:00
parent 5b296b3e01
commit 13cd4c0d8a
2 changed files with 63 additions and 0 deletions

View file

@ -21,6 +21,7 @@ services:
- "8505:8501" - "8505:8501"
volumes: volumes:
- /devl/menagerie-data:/devl/menagerie-data # per-user data trees - /devl/menagerie-data:/devl/menagerie-data # per-user data trees
- ./config/llm.cloud.yaml:/app/config/llm.yaml:ro # cloud-safe backends only (no claude_code/copilot/anthropic)
environment: environment:
- CLOUD_MODE=true - CLOUD_MODE=true
- CLOUD_DATA_ROOT=/devl/menagerie-data - CLOUD_DATA_ROOT=/devl/menagerie-data

62
config/llm.cloud.yaml Normal file
View file

@ -0,0 +1,62 @@
backends:
anthropic:
api_key_env: ANTHROPIC_API_KEY
enabled: false
model: claude-sonnet-4-6
supports_images: true
type: anthropic
claude_code:
api_key: any
base_url: http://localhost:3009/v1
enabled: false
model: claude-code-terminal
supports_images: true
type: openai_compat
github_copilot:
api_key: any
base_url: http://localhost:3010/v1
enabled: false
model: gpt-4o
supports_images: false
type: openai_compat
ollama:
api_key: ollama
base_url: http://host.docker.internal:11434/v1
enabled: true
model: llama3.1:8b # generic — no personal fine-tunes in cloud
supports_images: false
type: openai_compat
ollama_research:
api_key: ollama
base_url: http://host.docker.internal:11434/v1
enabled: true
model: llama3.1:8b
supports_images: false
type: openai_compat
vision_service:
base_url: http://host.docker.internal:8002
enabled: true
supports_images: true
type: vision_service
vllm:
api_key: ''
base_url: http://host.docker.internal:8000/v1
enabled: true
model: __auto__
supports_images: false
type: openai_compat
vllm_research:
api_key: ''
base_url: http://host.docker.internal:8000/v1
enabled: true
model: __auto__
supports_images: false
type: openai_compat
fallback_order:
- vllm
- ollama
research_fallback_order:
- vllm_research
- ollama_research
vision_fallback_order:
- vision_service