Open
Description
Git commit
commit 3f3769b (HEAD -> master, origin/master, origin/HEAD)
Operating systems
Windows
GGML backends
CUDA
Problem description & steps to reproduce
After replacing 3090 with 5070 I see compilation error:
nvcc fatal : Unsupported gpu architecture 'compute_120'
I found this: ggml-org/whisper.cpp#3030
Adding -DCMAKE_CUDA_ARCHITECTURES="86" solved my llama.cpp compilation problem.
Should this fix be merged also into llama.cpp?
First Bad Commit
No response
Compile command
cmake -DGGML_CUDA=ON -DLLAMA_CURL=OFF ..
cmake --build . --config Release -j 30
Relevant log output
nvcc fatal : Unsupported gpu architecture 'compute_120'