mirror of
https://github.com/dogkeeper886/ollama37.git
synced 2025-12-09 23:37:06 +00:00
This commit represents a complete rework after pulling the latest changes from official ollama/ollama repository and re-applying Tesla K80 compatibility patches. ## Key Changes ### CUDA Compute Capability 3.7 Support (Tesla K80) - Added sm_37 (compute 3.7) to CMAKE_CUDA_ARCHITECTURES in CMakeLists.txt - Updated CMakePresets.json to include compute 3.7 in "CUDA 11" preset - Using 37-virtual (PTX with JIT compilation) for maximum compatibility ### Legacy Toolchain Compatibility - **NVIDIA Driver**: 470.256.02 (last version supporting Kepler/K80) - **CUDA Version**: 11.4.4 (last CUDA 11.x supporting compute 3.7) - **GCC Version**: 10.5.0 (required by CUDA 11.4 host_config.h) ### CPU Architecture Trade-offs Due to GCC 10.5 limitation, sacrificed newer CPU optimizations: - Alderlake CPU variant enabled WITHOUT AVX_VNNI (requires GCC 11+) - Still supports: SSE4.2, AVX, F16C, AVX2, BMI2, FMA - Performance impact: ~3-7% on newer CPUs (acceptable for K80 compatibility) ### Build System Updates - Modified ml/backend/ggml/ggml/src/ggml-cuda/CMakeLists.txt for compute 3.7 - Added -Wno-deprecated-gpu-targets flag to suppress warnings - Updated ml/backend/ggml/ggml/src/CMakeLists.txt for Alderlake without AVX_VNNI ### Upstream Sync Merged latest llama.cpp changes including: - Enhanced KV cache management with ISWA and hybrid memory support - Improved multi-modal support (mtmd framework) - New model architectures (Gemma3, Llama4, Qwen3, etc.) - GPU backend improvements for CUDA, Metal, and ROCm - Updated quantization support and GGUF format handling ### Documentation - Updated CLAUDE.md with comprehensive build instructions - Documented toolchain constraints and CPU architecture trade-offs - Removed outdated CI/CD workflows (tesla-k80-*.yml) - Cleaned up temporary development artifacts ## Rationale This fork maintains Tesla K80 GPU support (compute 3.7) which was dropped in official Ollama due to legacy driver/CUDA requirements. The toolchain constraint creates a deadlock: - K80 → Driver 470 → CUDA 11.4 → GCC 10 → No AVX_VNNI We accept the loss of cutting-edge CPU optimizations to enable running modern LLMs on legacy but still capable Tesla K80 hardware (12GB VRAM per GPU). 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
45 lines
1.8 KiB
Plaintext
45 lines
1.8 KiB
Plaintext
---
|
|
title: macOS
|
|
---
|
|
|
|
## System Requirements
|
|
|
|
* MacOS Sonoma (v14) or newer
|
|
* Apple M series (CPU and GPU support) or x86 (CPU only)
|
|
|
|
|
|
## Filesystem Requirements
|
|
|
|
The preferred method of installation is to mount the `ollama.dmg` and drag-and-drop the Ollama application to the system-wide `Applications` folder. Upon startup, the Ollama app will verify the `ollama` CLI is present in your PATH, and if not detected, will prompt for permission to create a link in `/usr/local/bin`
|
|
|
|
Once you've installed Ollama, you'll need additional space for storing the Large Language models, which can be tens to hundreds of GB in size. If your home directory doesn't have enough space, you can change where the binaries are installed, and where the models are stored.
|
|
|
|
### Changing Install Location
|
|
|
|
To install the Ollama application somewhere other than `Applications`, place the Ollama application in the desired location, and ensure the CLI `Ollama.app/Contents/Resources/ollama` or a sym-link to the CLI can be found in your path. Upon first start decline the "Move to Applications?" request.
|
|
|
|
|
|
## Troubleshooting
|
|
|
|
Ollama on MacOS stores files in a few different locations.
|
|
- `~/.ollama` contains models and configuration
|
|
- `~/.ollama/logs` contains logs
|
|
- *app.log* contains most recent logs from the GUI application
|
|
- *server.log* contains the most recent server logs
|
|
- `<install location>/Ollama.app/Contents/Resources/ollama` the CLI binary
|
|
|
|
## Uninstall
|
|
|
|
To fully remove Ollama from your system, remove the following files and folders:
|
|
|
|
```
|
|
sudo rm -rf /Applications/Ollama.app
|
|
sudo rm /usr/local/bin/ollama
|
|
rm -rf "~/Library/Application Support/Ollama"
|
|
rm -rf "~/Library/Saved Application State/com.electron.ollama.savedState"
|
|
rm -rf ~/Library/Caches/com.electron.ollama/
|
|
rm -rf ~/Library/Caches/ollama
|
|
rm -rf ~/Library/WebKit/com.electron.ollama
|
|
rm -rf ~/.ollama
|
|
```
|