From e86c07c59ea286e4620f5f11d67c97592311147a Mon Sep 17 00:00:00 2001 From: pyr0ball Date: Tue, 24 Feb 2026 19:14:33 -0800 Subject: [PATCH] test: add ollama_research URL assertion to llm config generation test --- tests/test_llm_config_generation.py | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/tests/test_llm_config_generation.py b/tests/test_llm_config_generation.py index ba778df..5e6bb69 100644 --- a/tests/test_llm_config_generation.py +++ b/tests/test_llm_config_generation.py @@ -19,8 +19,9 @@ def test_urls_applied_to_llm_yaml(tmp_path): })) llm_yaml = tmp_path / "llm.yaml" llm_yaml.write_text(yaml.dump({"backends": { - "ollama": {"base_url": "http://old:11434/v1", "type": "openai_compat"}, - "vllm": {"base_url": "http://old:8000/v1", "type": "openai_compat"}, + "ollama": {"base_url": "http://old:11434/v1", "type": "openai_compat"}, + "ollama_research": {"base_url": "http://old:11434/v1", "type": "openai_compat"}, + "vllm": {"base_url": "http://old:8000/v1", "type": "openai_compat"}, }})) profile = UserProfile(user_yaml) @@ -28,6 +29,7 @@ def test_urls_applied_to_llm_yaml(tmp_path): result = yaml.safe_load(llm_yaml.read_text()) assert result["backends"]["ollama"]["base_url"] == "http://myserver:11434/v1" + assert result["backends"]["ollama_research"]["base_url"] == "http://myserver:11434/v1" assert result["backends"]["vllm"]["base_url"] == "http://localhost:8000/v1"