Files
llama.cpp/src
Daniel Bevenius 311c1a347f sampling : ensure at most one output token per seq
This commit adds a check in the batch allocator to ensure that when
backend sampling is enabled, at most one output token is specified per
sequence.
2025-11-18 16:06:23 +01:00
..
2025-09-05 17:32:39 -06:00
2025-09-05 17:32:39 -06:00