Files
llama.cpp/.github/workflows
Oliver Simons 1da013c66e Build with CCCL 3.2 for CUDA backends
Gives best perf for backend-sampling on CUDA. Flag can be removed once
CCCL 3.2 is bundled within CTK and that CTK version is used in llama.cpp
2025-12-19 16:10:51 +01:00
..
2025-12-05 19:44:17 +08:00