You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+9-3
Original file line number
Diff line number
Diff line change
@@ -9,11 +9,17 @@
9
9
10
10
Inference of [LLaMA](https://arxiv.org/abs/2302.13971) model in pure C/C++
11
11
12
-
### 🚧 Incoming breaking change + refactoring:
12
+
### Hot topics
13
13
14
-
See PR https://github.com/ggerganov/llama.cpp/pull/2398 for more info.
14
+
A new file format has been introduced: [GGUF](https://github.com/ggerganov/llama.cpp/pull/2398)
15
15
16
-
To devs: avoid making big changes to `llama.h` / `llama.cpp` until merged
16
+
Last revision compatible with the old format: [dadbed9](https://github.com/ggerganov/llama.cpp/commit/dadbed99e65252d79f81101a392d0d6497b86caa)
17
+
18
+
### Current `master` should be considered in Beta - expect some issues for a few days!
19
+
20
+
### Be prepared to re-convert and / or re-quantize your GGUF models while this notice is up!
21
+
22
+
### Issues with non-GGUF models will be considered with low priority!
0 commit comments