From c5e2dc975f0ae56ef204d82b15ce587819f49b1d Mon Sep 17 00:00:00 2001 From: pyr0ball Date: Sat, 4 Apr 2026 18:37:58 -0700 Subject: [PATCH] chore: add LLM env-var config + CF coordinator vars to .env.example --- .env.example | 14 ++++++++++++++ 1 file changed, 14 insertions(+) diff --git a/.env.example b/.env.example index 9763220..61d12b2 100644 --- a/.env.example +++ b/.env.example @@ -19,6 +19,14 @@ VLLM_MAX_MODEL_LEN=4096 # increase to 8192 for Thinking models with VLLM_GPU_MEM_UTIL=0.75 # lower to 0.6 if sharing GPU with other services OLLAMA_DEFAULT_MODEL=llama3.2:3b +# ── LLM env-var auto-config (alternative to config/llm.yaml) ───────────────── +# Set any of these to configure LLM backends without needing a config/llm.yaml. +# Priority: Anthropic > OpenAI-compat > Ollama (always tried as local fallback). +OLLAMA_HOST=http://localhost:11434 # Ollama host; override if on a different machine +OLLAMA_MODEL=llama3.2:3b # model to request from Ollama +OPENAI_MODEL=gpt-4o-mini # model override for OpenAI-compat backend +ANTHROPIC_MODEL=claude-haiku-4-5-20251001 # model override for Anthropic backend + # API keys (required for remote profile) ANTHROPIC_API_KEY= OPENAI_COMPAT_URL= @@ -31,6 +39,12 @@ FORGEJO_API_URL=https://git.opensourcesolarpunk.com/api/v1 # GITHUB_TOKEN= # future — enable when public mirror is active # GITHUB_REPO= # future +# ── CF-hosted coordinator (Paid+ tier) ─────────────────────────────────────── +# Set CF_LICENSE_KEY to authenticate with the hosted coordinator. +# Leave both blank for local self-hosted cf-orch or bare-metal inference. +CF_LICENSE_KEY= +CF_ORCH_URL=https://orch.circuitforge.tech + # Cloud multi-tenancy (compose.cloud.yml only — do not set for local installs) CLOUD_MODE=false CLOUD_DATA_ROOT=/devl/menagerie-data