llm: Remove GGML_CUDA_NO_PEER_COPY for ROCm (#7174)

This workaround logic in llama.cpp is causing crashes for users with less system memory than VRAM.
This commit is contained in:
Daniel Hiltgen
2024-10-12 09:56:49 -07:00
committed by GitHub
parent 7fe3902552
commit c3d321d405
2 changed files with 1 additions and 2 deletions

View File

@@ -340,7 +340,6 @@ function build_rocm() {
"-DCMAKE_C_COMPILER=clang.exe",
"-DCMAKE_CXX_COMPILER=clang++.exe",
"-DGGML_HIPBLAS=on",
"-DGGML_CUDA_NO_PEER_COPY=on",
"-DHIP_PLATFORM=amd",
"-DGGML_AVX=on",
"-DGGML_AVX2=off",