mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-08-26 10:09:41 -04:00
metal : parallel command buffer encoding (#1860)
* metal : parallel command buffer encoding * metal : determine number of command buffers based on gf->n_threads
This commit is contained in:
@@ -55,6 +55,7 @@ void ggml_metal_set_tensor(struct ggml_metal_context * ctx, struct ggml_tensor *
|
||||
void ggml_metal_get_tensor(struct ggml_metal_context * ctx, struct ggml_tensor * t);
|
||||
|
||||
// same as ggml_graph_compute but uses Metal
|
||||
// creates gf->n_threads command buffers in parallel
|
||||
void ggml_metal_graph_compute(struct ggml_metal_context * ctx, struct ggml_cgraph * gf);
|
||||
|
||||
#ifdef __cplusplus
|
||||
|
Reference in New Issue
Block a user