Files
llama.cpp/ggml/include
Georgi Gerganov 9067487c44 ggml : fix FA mask dim 2 and 3 (#14505)
* ggml : fix FA mask dim 2 and 3

ggml-ci

* backends : unsupport batched FA in CUDA and Vulkan

ggml-ci

* vulkan : disable FA for mask->ne[2] != 1
2025-07-03 10:46:57 +03:00
..
2025-06-27 16:41:40 +03:00
2025-07-03 10:46:57 +03:00