Files
llama.cpp/tests/test-backend-sampler.cpp
Daniel Bevenius 311c1a347f sampling : ensure at most one output token per seq
This commit adds a check in the batch allocator to ensure that when
backend sampling is enabled, at most one output token is specified per
sequence.
2025-11-18 16:06:23 +01:00

30 KiB