feat: LLM queue optimizer complete — closes #2
Resource-aware batch scheduler for LLM tasks: - scripts/task_scheduler.py (new): TaskScheduler singleton with VRAM-aware batch scheduling, durability, thread-safe singleton, memory safety - scripts/task_runner.py: submit_task() routes LLM types through scheduler - scripts/db.py: reset_running_tasks() for durable restart behavior - app/app.py: _startup() preserves queued tasks on restart - config/llm.yaml.example: scheduler VRAM budget config documented - tests/test_task_scheduler.py (new): 24 tests covering all behaviors Pre-existing failure: test_generate_calls_llm_router (issue #12, unrelated)
This commit is contained in:
parent
95378c106e
commit
27d4b0e732