Files
llama.cpp/examples
Georgi Gerganov ba69bbc84c imatrix : offload to GPU support (#4957)
* backend : add eval callback

ggml-ci

* backend : group nodes in a single compute when user don't need them

* backend : clean-up the implementation

ggml-ci

* simple : do not perform tensor data copy if not needed

* simple : fix

* imatrix : offload to GPU support

* imatrix : fix ggml_mul_mat_id hanlding

ggml-ci

* ci : add imatrix test

ggml-ci

* ci : rearrange output

ggml-ci
2024-01-17 18:46:30 +02:00
..
2024-01-14 09:45:56 +02:00
2023-12-21 23:08:14 +02:00
2023-03-29 20:21:09 +03:00
2023-08-30 09:29:32 +03:00
2024-01-13 20:45:45 +02:00
2023-08-08 14:44:48 +03:00