mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-07-23 03:08:08 +00:00
CUDA: refactor mmq, dmmv, mmvq (#7716)
* CUDA: refactor mmq, dmmv, mmvq * fix out-of-bounds write * struct for qk, qr, qi * fix cmake build * mmq_type_traits
This commit is contained in:
5
ggml-cuda/template-instances/mmq-instance-q8_0.cu
Normal file
5
ggml-cuda/template-instances/mmq-instance-q8_0.cu
Normal file
@ -0,0 +1,5 @@
|
||||
// This file has been autogenerated by generate_cu_files.py, do not edit manually.
|
||||
|
||||
#include "../mmq.cuh"
|
||||
|
||||
DECL_MMQ_CASE(GGML_TYPE_Q8_0);
|
Reference in New Issue
Block a user