treehugg3 commited on
Commit
1efd43b
·
verified ·
1 Parent(s): 8507cc0

Add link to llama-imatrix

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -144,6 +144,8 @@ is unfortunately very frequent).
144
 
145
  ## What do I need to do to compute imatrix files for large models?
146
 
 
 
147
  ### Hardware
148
 
149
  * RAM: A lot of RAM is required to compute imatrix files. Example: 512 GB is just enough to compute 405B imatrix quants in Q8.
 
144
 
145
  ## What do I need to do to compute imatrix files for large models?
146
 
147
+ Use [`llama-imatrix`](https://github.com/ggml-org/llama.cpp/blob/master/examples/imatrix/README.md) to compute imatrix files.
148
+
149
  ### Hardware
150
 
151
  * RAM: A lot of RAM is required to compute imatrix files. Example: 512 GB is just enough to compute 405B imatrix quants in Q8.