Compile llama.cpp with CUDA 12 #13403
Replies: 4 comments
-
Those are cuda compute capabilities, basically device architectures, not cuda versions. See https://developer.nvidia.com/cuda-gpus |
Beta Was this translation helpful? Give feedback.
-
@0cc4m OK. How do I set this parameter correctly so that RTX4000 and RTX5000 can be used? |
Beta Was this translation helpful? Give feedback.
-
Dont modify CMAKE_CUDA_ARCHITECTURES the llama.cpp CUDA backend should run on newer GPUs without explicitly setting the architectures. |
Beta Was this translation helpful? Give feedback.
-
@CommanderLake I will try, thanks. I'll come back with feedback. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I am building llamacpp with CUDA 12 support (RTX5000). How can I add support for RTX4000 and RXT5000 using
cmake -B build -DGGML_CUDA=ON -DCMAKE_CUDA_ARCHITECTURES=“86;89”
? The range of acceptable values for theDCMAKE_CUDA_ARCHITECTURES
parameter is not specified in documentation. Should I specify CUDA 12 as120
or as12
? For example:Beta Was this translation helpful? Give feedback.
All reactions