Back to feed

b7990

Feb 10, 2026
Meta/llama.cppCLIvb7990

models : support qwen3.5 series (#19468)

  • support qwen3.5 series

  • remove deepstack for now, and some code clean

  • code clean

  • add FULL_ATTENTION_INTERVAL metadata

  • code clean

  • reorder v heads for linear attention to avoid expensive interleaved repeat

macOS/iOS:

Linux:

Windows:

openEuler: