Add link to llama-imatrix
Browse files
README.md
CHANGED
@@ -144,6 +144,8 @@ is unfortunately very frequent).
|
|
144 |
|
145 |
## What do I need to do to compute imatrix files for large models?
|
146 |
|
|
|
|
|
147 |
### Hardware
|
148 |
|
149 |
* RAM: A lot of RAM is required to compute imatrix files. Example: 512 GB is just enough to compute 405B imatrix quants in Q8.
|
|
|
144 |
|
145 |
## What do I need to do to compute imatrix files for large models?
|
146 |
|
147 |
+
Use [`llama-imatrix`](https://github.com/ggml-org/llama.cpp/blob/master/examples/imatrix/README.md) to compute imatrix files.
|
148 |
+
|
149 |
### Hardware
|
150 |
|
151 |
* RAM: A lot of RAM is required to compute imatrix files. Example: 512 GB is just enough to compute 405B imatrix quants in Q8.
|