Skip to content

Commit 4b0eff3

Browse files
docs : Quantum -> Quantized (#8666)
* docfix: imatrix readme, quantum models -> quantized models. * docfix: server readme: quantum models -> quantized models.
1 parent 8a4bad5 commit 4b0eff3

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

examples/imatrix/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# llama.cpp/examples/imatrix
22

3-
Compute an importance matrix for a model and given text dataset. Can be used during quantization to enchance the quality of the quantum models.
3+
Compute an importance matrix for a model and given text dataset. Can be used during quantization to enchance the quality of the quantized models.
44
More information is available here: https://github.com/ggerganov/llama.cpp/pull/4861
55

66
## Usage

examples/server/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ Fast, lightweight, pure C/C++ HTTP server based on [httplib](https://github.com/
55
Set of LLM REST APIs and a simple web front end to interact with llama.cpp.
66

77
**Features:**
8-
* LLM inference of F16 and quantum models on GPU and CPU
8+
* LLM inference of F16 and quantized models on GPU and CPU
99
* [OpenAI API](https://github.com/openai/openai-openapi) compatible chat completions and embeddings routes
1010
* Parallel decoding with multi-user support
1111
* Continuous batching

0 commit comments

Comments
 (0)