mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-08-12 03:21:10 -04:00
llama : bump max layers from 256 to 512 (#8530)
* llama : bump max layers from 256 to 512 * llama : replace asserts with exceptions
This commit is contained in:
@@ -40,7 +40,7 @@
|
||||
#define LLAMA_FILE_MAGIC_GGSQ 0x67677371u // 'ggsq'
|
||||
|
||||
#define LLAMA_SESSION_MAGIC LLAMA_FILE_MAGIC_GGSN
|
||||
#define LLAMA_SESSION_VERSION 6
|
||||
#define LLAMA_SESSION_VERSION 7
|
||||
|
||||
#define LLAMA_STATE_SEQ_MAGIC LLAMA_FILE_MAGIC_GGSQ
|
||||
#define LLAMA_STATE_SEQ_VERSION 1
|
||||
|
Reference in New Issue
Block a user