mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-08-18 22:20:16 -04:00
Speedup the AVX-512 implementation of ggml_vec_dot_q4_0() (#933)
This commit is contained in:
2
ggml.h
2
ggml.h
@@ -808,6 +808,8 @@ size_t ggml_quantize_q4_1(const float * src, void * dst, int n, int k, int64_t *
|
||||
int ggml_cpu_has_avx(void);
|
||||
int ggml_cpu_has_avx2(void);
|
||||
int ggml_cpu_has_avx512(void);
|
||||
int ggml_cpu_has_avx512_vbmi(void);
|
||||
int ggml_cpu_has_avx512_vnni(void);
|
||||
int ggml_cpu_has_fma(void);
|
||||
int ggml_cpu_has_neon(void);
|
||||
int ggml_cpu_has_arm_fma(void);
|
||||
|
Reference in New Issue
Block a user