Default Branch

8846aace49 · model : gemma3n text-only (#14400) · Updated 2025-06-26 17:34:02 +00:00

Branches

bc82fc2ed8 · llama-bench : add time-to-first-byte stat · Updated 2024-10-18 13:40:02 +00:00    tqcq

1821
1

2d3fc54ac6 · add amx kernel for gemm · Updated 2024-10-18 03:35:49 +00:00    tqcq

1831
1

630bce5a7f · ggml : fix possible buffer use after free in sched reserve · Updated 2024-10-17 22:21:54 +00:00    tqcq

1829
1

17b3a3e8cc · llama : minor llama_grammar refactoring · Updated 2024-10-17 09:23:27 +00:00    tqcq

1857
4

a34fc0dd86 · ci : reduce severity of unused Pyright ignore comments · Updated 2024-09-30 17:59:40 +00:00    tqcq

1912
1

114ab6347e · sampling : fix off-by-one in tail-free sampling · Updated 2024-09-23 08:44:55 +00:00    tqcq

1956
1

6e873e561a · llama : make llm_tokenizer more private · Updated 2024-09-20 08:41:51 +00:00    tqcq

1975
2

6b0248c29a · Update ggml/src/ggml.c · Updated 2024-09-18 16:00:26 +00:00    tqcq

1980
2

a6a8f8d09c · Update docs/backend/SYCL.md · Updated 2024-09-17 08:25:43 +00:00    tqcq

2012
2

cc1c017191 · naming : normalize the name of callback-related identifiers · Updated 2024-09-16 06:11:42 +00:00    tqcq

2000
1

73ef3f769c · Update llama-server-intel.Dockerfile · Updated 2024-09-15 15:21:46 +00:00    tqcq

2005
3

fb8f142554 · one more CMAKE_CXX_FLAGS fix (#9471) · Updated 2024-09-13 13:13:07 +00:00    tqcq

2014
5

d7c042d1ae · ggml : make n_threads_cur atomic_int · Updated 2024-09-11 18:12:11 +00:00    tqcq

2030
1

f9968f661d · ggml : update comments [no ci] · Updated 2024-09-11 10:16:39 +00:00    tqcq

2043
5

cfbf33a705 · ggml : style changes + fix 512-bit nb loop check · Updated 2024-09-09 09:50:35 +00:00    tqcq

2104
4

c3e2bb6dcf · rpc : fix nkvo · Updated 2024-09-07 01:24:47 +00:00    tqcq

2085
1

b979fc97ba · cmake : use ggml-metal.metal from source dir to build default.metallib · Updated 2024-09-05 16:17:56 +00:00    tqcq

2094
1

75b3a09602 · test-backend-ops : add TQ1_0 and TQ2_0 comments for later · Updated 2024-09-04 19:00:21 +00:00    tqcq

2096
33

f648ca2cee · llama : add llama_sampling API + move grammar in libllama · Updated 2024-09-03 07:31:54 +00:00    tqcq

2103
1

40fa68cb46 · readme : add API change notice · Updated 2024-09-02 15:32:24 +00:00    tqcq

2112
3