mirror of
https://github.com/dogkeeper886/ollama37.git
synced 2025-12-10 07:46:59 +00:00
2.0 KiB
2.0 KiB
Ollama37 🚀
Tesla K80 Compatible Ollama Fork
Run modern LLMs on NVIDIA Tesla K80 and other CUDA Compute Capability 3.7 GPUs. While official Ollama dropped legacy GPU support, Ollama37 keeps your Tesla K80 hardware functional with the latest models and features.
Key Features
- ⚡ Tesla K80 Support - Full compatibility with CUDA Compute Capability 3.7
- 🛠️ Optimized Build - CUDA 11 toolchain for maximum legacy GPU compatibility
Quick Start
Docker (Recommended)
# Pull and run
docker pull dogkeeper886/ollama37
docker run --runtime=nvidia --gpus all -p 11434:11434 dogkeeper886/ollama37
Docker Compose
version: "3.8"
services:
ollama:
image: dogkeeper886/ollama37:latest
container_name: ollama37
runtime: nvidia
deploy:
resources:
reservations:
devices:
- driver: nvidia
count: all
capabilities: [gpu]
ports:
- "11434:11434"
volumes:
- ollama-data:/root/.ollama
environment:
- OLLAMA_HOST=0.0.0.0:11434
- NVIDIA_VISIBLE_DEVICES=all
- NVIDIA_DRIVER_CAPABILITIES=compute,utility
restart: unless-stopped
healthcheck:
test: ["CMD", "/usr/local/bin/ollama", "list"]
interval: 30s
timeout: 10s
retries: 3
start_period: 5s
volumes:
ollama-data:
name: ollama-data
docker-compose up -d
Usage
Run Your First Model
# Download and run a model
ollama pull gemma3
ollama run gemma3 "Why is the sky blue?"
# Interactive chat
ollama run gemma3
Contributing
Found an issue or want to contribute? Check our GitHub issues or submit Tesla K80-specific bug reports and compatibility fixes.
License
Same license as upstream Ollama. See LICENSE file for details.
CLI Commands
ollama list # List models
ollama show llama3.2 # Model info
ollama ps # Running models
ollama stop llama3.2 # Stop model
ollama serve # Start server