From 7c9dcd262091c720aa36907db951ae7ffbe3e160 Mon Sep 17 00:00:00 2001 From: pyr0ball Date: Thu, 2 Apr 2026 12:20:41 -0700 Subject: [PATCH] config(llm): add cf_orch block to vllm backend --- config/llm.yaml | 8 +++++++- 1 file changed, 7 insertions(+), 1 deletion(-) diff --git a/config/llm.yaml b/config/llm.yaml index 0f08746..0ef6abc 100644 --- a/config/llm.yaml +++ b/config/llm.yaml @@ -30,7 +30,7 @@ backends: api_key: ollama base_url: http://host.docker.internal:11434/v1 enabled: true - model: llama3.2:3b + model: llama3.1:8b supports_images: false type: openai_compat vision_service: @@ -45,6 +45,12 @@ backends: model: __auto__ supports_images: false type: openai_compat + cf_orch: + service: vllm + model_candidates: + - Ouro-2.6B-Thinking + - Ouro-1.4B + ttl_s: 300 vllm_research: api_key: '' base_url: http://host.docker.internal:8000/v1