fix(cloud): mount llm.cloud.yaml over llm.yaml; restrict to vllm+ollama only
Some checks failed
CI / test (push) Failing after 17s
Some checks failed
CI / test (push) Failing after 17s
Remove claude_code, github_copilot, and anthropic from all cloud fallback orders — cloud accounts must not route through personal/dev LLM backends. vllm_research and ollama_research are the only permitted research backends. llm.cloud.yaml is now bind-mounted at /app/config/llm.yaml in compose.cloud.yml.
This commit is contained in:
parent
5b296b3e01
commit
13cd4c0d8a
2 changed files with 63 additions and 0 deletions
|
|
@ -21,6 +21,7 @@ services:
|
||||||
- "8505:8501"
|
- "8505:8501"
|
||||||
volumes:
|
volumes:
|
||||||
- /devl/menagerie-data:/devl/menagerie-data # per-user data trees
|
- /devl/menagerie-data:/devl/menagerie-data # per-user data trees
|
||||||
|
- ./config/llm.cloud.yaml:/app/config/llm.yaml:ro # cloud-safe backends only (no claude_code/copilot/anthropic)
|
||||||
environment:
|
environment:
|
||||||
- CLOUD_MODE=true
|
- CLOUD_MODE=true
|
||||||
- CLOUD_DATA_ROOT=/devl/menagerie-data
|
- CLOUD_DATA_ROOT=/devl/menagerie-data
|
||||||
|
|
|
||||||
62
config/llm.cloud.yaml
Normal file
62
config/llm.cloud.yaml
Normal file
|
|
@ -0,0 +1,62 @@
|
||||||
|
backends:
|
||||||
|
anthropic:
|
||||||
|
api_key_env: ANTHROPIC_API_KEY
|
||||||
|
enabled: false
|
||||||
|
model: claude-sonnet-4-6
|
||||||
|
supports_images: true
|
||||||
|
type: anthropic
|
||||||
|
claude_code:
|
||||||
|
api_key: any
|
||||||
|
base_url: http://localhost:3009/v1
|
||||||
|
enabled: false
|
||||||
|
model: claude-code-terminal
|
||||||
|
supports_images: true
|
||||||
|
type: openai_compat
|
||||||
|
github_copilot:
|
||||||
|
api_key: any
|
||||||
|
base_url: http://localhost:3010/v1
|
||||||
|
enabled: false
|
||||||
|
model: gpt-4o
|
||||||
|
supports_images: false
|
||||||
|
type: openai_compat
|
||||||
|
ollama:
|
||||||
|
api_key: ollama
|
||||||
|
base_url: http://host.docker.internal:11434/v1
|
||||||
|
enabled: true
|
||||||
|
model: llama3.1:8b # generic — no personal fine-tunes in cloud
|
||||||
|
supports_images: false
|
||||||
|
type: openai_compat
|
||||||
|
ollama_research:
|
||||||
|
api_key: ollama
|
||||||
|
base_url: http://host.docker.internal:11434/v1
|
||||||
|
enabled: true
|
||||||
|
model: llama3.1:8b
|
||||||
|
supports_images: false
|
||||||
|
type: openai_compat
|
||||||
|
vision_service:
|
||||||
|
base_url: http://host.docker.internal:8002
|
||||||
|
enabled: true
|
||||||
|
supports_images: true
|
||||||
|
type: vision_service
|
||||||
|
vllm:
|
||||||
|
api_key: ''
|
||||||
|
base_url: http://host.docker.internal:8000/v1
|
||||||
|
enabled: true
|
||||||
|
model: __auto__
|
||||||
|
supports_images: false
|
||||||
|
type: openai_compat
|
||||||
|
vllm_research:
|
||||||
|
api_key: ''
|
||||||
|
base_url: http://host.docker.internal:8000/v1
|
||||||
|
enabled: true
|
||||||
|
model: __auto__
|
||||||
|
supports_images: false
|
||||||
|
type: openai_compat
|
||||||
|
fallback_order:
|
||||||
|
- vllm
|
||||||
|
- ollama
|
||||||
|
research_fallback_order:
|
||||||
|
- vllm_research
|
||||||
|
- ollama_research
|
||||||
|
vision_fallback_order:
|
||||||
|
- vision_service
|
||||||
Loading…
Reference in a new issue