You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+2-3
Original file line number
Diff line number
Diff line change
@@ -20,12 +20,11 @@ The main goal of `llama.cpp` is to run the LLaMA model using 4-bit integer quant
20
20
- Apple silicon first-class citizen - optimized via ARM NEON and Accelerate framework
21
21
- AVX2 support for x86 architectures
22
22
- Mixed F16 / F32 precision
23
-
- 4-bit integer quantization support
23
+
- 4 & 8 bit integer quantization support
24
24
- Runs on the CPU
25
25
26
26
The original implementation of `llama.cpp` was [hacked in an evening](https://github.com/ggerganov/llama.cpp/issues/33#issuecomment-1465108022).
27
-
Since then, the project has improved significantly thanks to many contributions. This project is for educational purposes and serves
28
-
as the main playground for developing new features for the [ggml](https://github.com/ggerganov/ggml) library.
27
+
Since then, the project has improved significantly thanks to many contributions. This project is for educational purposes and serves as the main playground for developing new features for the [ggml](https://github.com/ggerganov/ggml) library.
0 commit comments