Back to feed

b7972

Feb 8, 2026
Meta/llama.cppCLIvb7972

CUDA: Fix non-contig rope (#19338)

  • Rename variables + fix rope_neox

Seems memory layout is shared with Vulkan so we can port fix from https://github.com/ggml-org/llama.cpp/pull/19299

  • Fix rope_multi

  • Fix rope_vision

  • Fix rope_norm

  • Rename ne* to ne0* for consistent variable naming

  • cont : consistent stride names


Co-authored-by: Georgi Gerganov ggerganov@gmail.com

macOS/iOS:

Linux:

Windows:

openEuler: