mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-08-24 08:49:15 -04:00
Fixed mmap prefetch for GPU offloading (#2529)
This commit is contained in:
@@ -219,7 +219,7 @@ struct llama_mmap {
|
||||
// prefetch/readahead impairs performance on NUMA systems
|
||||
if (numa) { prefetch = 0; }
|
||||
#ifdef __linux__
|
||||
if (prefetch) { flags |= MAP_POPULATE; }
|
||||
if (prefetch >= file->size) { flags |= MAP_POPULATE; }
|
||||
#endif
|
||||
addr = mmap(NULL, file->size, PROT_READ, flags, fd, 0);
|
||||
if (addr == MAP_FAILED) {
|
||||
|
Reference in New Issue
Block a user