mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-07-27 03:33:46 -04:00
ggml-cpu: fp16 load ensured to hit
Signed-off-by: Aaron Teo <aaron.teo1@ibm.com>
This commit is contained in:
@ -967,7 +967,6 @@ static inline float32x4_t __lzs_f16cx4_load(const ggml_fp16_t * x) {
|
|||||||
#ifdef __NNPA__
|
#ifdef __NNPA__
|
||||||
uint16x8_t v_x = vec_xl(0, (const ggml_fp16_t *)x);
|
uint16x8_t v_x = vec_xl(0, (const ggml_fp16_t *)x);
|
||||||
uint16x8_t nnpa_dlf16 = vec_convert_from_fp16(v_x, 0);
|
uint16x8_t nnpa_dlf16 = vec_convert_from_fp16(v_x, 0);
|
||||||
raise(SIGINT); // TODO: Ensure it is called
|
|
||||||
return vec_extend_to_fp32_hi(nnpa_dlf16, 0);
|
return vec_extend_to_fp32_hi(nnpa_dlf16, 0);
|
||||||
#else
|
#else
|
||||||
float tmp[4];
|
float tmp[4];
|
||||||
|
Reference in New Issue
Block a user