diff --git a/Makefile b/Makefile index dcb770a..4998e2e 100644 --- a/Makefile +++ b/Makefile @@ -55,7 +55,7 @@ prepare-training: ## Scan docs_dir for cover letters and build training JSONL finetune: ## Fine-tune your personal cover letter model (run prepare-training first) @echo "Starting fine-tune (30-90 min on GPU, much longer on CPU)..." - $(COMPOSE) $(COMPOSE_FILES) --profile $(PROFILE) run --rm finetune + $(COMPOSE) $(COMPOSE_FILES) -f compose.gpu.yml --profile finetune run --rm finetune clean: ## Remove containers, images, and data volumes (DESTRUCTIVE) @echo "WARNING: This will delete all Peregrine containers and data." diff --git a/compose.yml b/compose.yml index eac74f4..d2b7b08 100644 --- a/compose.yml +++ b/compose.yml @@ -92,8 +92,5 @@ services: - OLLAMA_URL=http://ollama:11434 - OLLAMA_MODELS_MOUNT=/ollama-models - OLLAMA_MODELS_OLLAMA_PATH=/root/.ollama - depends_on: - ollama: - condition: service_started - profiles: [cpu, single-gpu, dual-gpu] + profiles: [finetune] restart: "no" diff --git a/scripts/preflight.py b/scripts/preflight.py index c4c6367..7c57790 100644 --- a/scripts/preflight.py +++ b/scripts/preflight.py @@ -48,7 +48,7 @@ _SERVICES: dict[str, tuple[str, int, str, bool, bool]] = { "searxng": ("searxng_port", 8888, "SEARXNG_PORT", True, True), "vllm": ("vllm_port", 8000, "VLLM_PORT", True, True), "vision": ("vision_port", 8002, "VISION_PORT", True, True), - "ollama": ("ollama_port", 11434, "OLLAMA_PORT", False, True), + "ollama": ("ollama_port", 11434, "OLLAMA_PORT", True, True), } # LLM yaml backend keys → url suffix, keyed by service name