mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-09-05 06:45:21 -04:00
* CUDA: add bf16 and f32 support to cublas_mul_mat_batched * Review: add type traits and make function more generic * Review: make check more explicit, add back comments, and fix formatting * Review: fix formatting, remove useless type conversion, fix naming for bools