Back to feed

b7852

Jan 28, 2026
Meta/llama.cppCLIvb7852

sampling : remove sampling branching in output_reserve (#18811)

  • sampling : remove sampling branching in output_reserve

This commit updates output_reserve in llama-context.cpp to always allocate sampling buffers regardless of whether sampling is needed for the current batch.

The motivation for this is to avoid reallocations and branching based on the sampling requirements of the batch.

macOS/iOS:

Linux:

Windows:

openEuler: