From 13cd4c0d8a4c924979ec08847ed383fdd4c6a62a Mon Sep 17 00:00:00 2001 From: pyr0ball Date: Wed, 1 Apr 2026 19:59:01 -0700 Subject: [PATCH] fix(cloud): mount llm.cloud.yaml over llm.yaml; restrict to vllm+ollama only MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Remove claude_code, github_copilot, and anthropic from all cloud fallback orders — cloud accounts must not route through personal/dev LLM backends. vllm_research and ollama_research are the only permitted research backends. llm.cloud.yaml is now bind-mounted at /app/config/llm.yaml in compose.cloud.yml. --- compose.cloud.yml | 1 + config/llm.cloud.yaml | 62 +++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 63 insertions(+) create mode 100644 config/llm.cloud.yaml diff --git a/compose.cloud.yml b/compose.cloud.yml index 4f42880..417a6c6 100644 --- a/compose.cloud.yml +++ b/compose.cloud.yml @@ -21,6 +21,7 @@ services: - "8505:8501" volumes: - /devl/menagerie-data:/devl/menagerie-data # per-user data trees + - ./config/llm.cloud.yaml:/app/config/llm.yaml:ro # cloud-safe backends only (no claude_code/copilot/anthropic) environment: - CLOUD_MODE=true - CLOUD_DATA_ROOT=/devl/menagerie-data diff --git a/config/llm.cloud.yaml b/config/llm.cloud.yaml new file mode 100644 index 0000000..62af14f --- /dev/null +++ b/config/llm.cloud.yaml @@ -0,0 +1,62 @@ +backends: + anthropic: + api_key_env: ANTHROPIC_API_KEY + enabled: false + model: claude-sonnet-4-6 + supports_images: true + type: anthropic + claude_code: + api_key: any + base_url: http://localhost:3009/v1 + enabled: false + model: claude-code-terminal + supports_images: true + type: openai_compat + github_copilot: + api_key: any + base_url: http://localhost:3010/v1 + enabled: false + model: gpt-4o + supports_images: false + type: openai_compat + ollama: + api_key: ollama + base_url: http://host.docker.internal:11434/v1 + enabled: true + model: llama3.1:8b # generic — no personal fine-tunes in cloud + supports_images: false + type: openai_compat + ollama_research: + api_key: ollama + base_url: http://host.docker.internal:11434/v1 + enabled: true + model: llama3.1:8b + supports_images: false + type: openai_compat + vision_service: + base_url: http://host.docker.internal:8002 + enabled: true + supports_images: true + type: vision_service + vllm: + api_key: '' + base_url: http://host.docker.internal:8000/v1 + enabled: true + model: __auto__ + supports_images: false + type: openai_compat + vllm_research: + api_key: '' + base_url: http://host.docker.internal:8000/v1 + enabled: true + model: __auto__ + supports_images: false + type: openai_compat +fallback_order: +- vllm +- ollama +research_fallback_order: +- vllm_research +- ollama_research +vision_fallback_order: +- vision_service