mirror of
https://github.com/dogkeeper886/ollama37.git
synced 2025-12-17 19:27:00 +00:00
Tesla K80 needs ~60-180s to load model into VRAM on first inference. Add warmup step with 5-minute timeout to preload model before subsequent inference tests run. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>