CISCai commited on
Commit
7e257ad
·
verified ·
1 Parent(s): a609734

Added links to full context YaRN-enabled GGUFs

Browse files
Files changed (1) hide show
  1. README.md +13 -11
README.md CHANGED
@@ -32,6 +32,8 @@ Quantization was done with an importance matrix that was trained for ~1M tokens
32
 
33
  Fill-in-Middle token metadata has been added, see [example](#simple-llama-cpp-python-example-fill-in-middle-code).
34
 
 
 
35
  <!-- description end -->
36
 
37
 
@@ -85,17 +87,17 @@ Refer to the Provided Files table below to see what files use which methods, and
85
 
86
  | Name | Quant method | Bits | Size | Max RAM required | Use case |
87
  | ---- | ---- | ---- | ---- | ---- | ----- |
88
- | [Qwen2.5-Coder-7B-Instruct.IQ1_S.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ1_S.gguf) | IQ1_S | 1 | 1.8 GB| 2.0 GB | smallest, significant quality loss |
89
- | [Qwen2.5-Coder-7B-Instruct.IQ1_M.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ1_M.gguf) | IQ1_M | 1 | 1.9 GB| 2.1 GB | very small, significant quality loss |
90
- | [Qwen2.5-Coder-7B-Instruct.IQ2_XXS.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ2_XXS.gguf) | IQ2_XXS | 2 | 2.1 GB| 2.3 GB | very small, high quality loss |
91
- | [Qwen2.5-Coder-7B-Instruct.IQ2_XS.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ2_XS.gguf) | IQ2_XS | 2 | 2.3 GB| 2.5 GB | very small, high quality loss |
92
- | [Qwen2.5-Coder-7B-Instruct.IQ2_S.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ2_S.gguf) | IQ2_S | 2 | 2.4 GB| 2.6 GB | small, substantial quality loss |
93
- | [Qwen2.5-Coder-7B-Instruct.IQ2_M.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ2_M.gguf) | IQ2_M | 2 | 2.6 GB| 2.8 GB | small, greater quality loss |
94
- | [Qwen2.5-Coder-7B-Instruct.IQ3_XXS.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ3_XXS.gguf) | IQ3_XXS | 3 | 2.9 GB| 3.1 GB | very small, high quality loss |
95
- | [Qwen2.5-Coder-7B-Instruct.IQ3_XS.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ3_XS.gguf) | IQ3_XS | 3 | 3.1 GB| 3.3 GB | small, substantial quality loss |
96
- | [Qwen2.5-Coder-7B-Instruct.IQ3_S.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ3_S.gguf) | IQ3_S | 3 | 3.3 GB| 3.4 GB | small, greater quality loss |
97
- | [Qwen2.5-Coder-7B-Instruct.IQ3_M.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ3_M.gguf) | IQ3_M | 3 | 3.3 GB| 3.5 GB | medium, balanced quality - recommended |
98
- | [Qwen2.5-Coder-7B-Instruct.IQ4_XS.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ4_XS.gguf) | IQ4_XS | 4 | 3.9 GB| 4.1 GB | small, substantial quality loss |
99
 
100
  Generated importance matrix file: [Qwen2.5-Coder-7B-Instruct.imatrix.dat](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.imatrix.dat)
101
 
 
32
 
33
  Fill-in-Middle token metadata has been added, see [example](#simple-llama-cpp-python-example-fill-in-middle-code).
34
 
35
+ **Update January 6th 2025**: Added links to full context YaRN-enabled GGUFs (using [GGUF Editor](https://huggingface.co/spaces/CISCai/gguf-editor)).
36
+
37
  <!-- description end -->
38
 
39
 
 
87
 
88
  | Name | Quant method | Bits | Size | Max RAM required | Use case |
89
  | ---- | ---- | ---- | ---- | ---- | ----- |
90
+ | [Qwen2.5-Coder-7B-Instruct.IQ1_S.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ1_S.gguf) ([with YaRN](https://ciscai-gguf-editor.hf.space/download/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/Qwen2.5-Coder-7B-Instruct.IQ1_S.gguf?branch=main&add=%5B%22qwen2.context_length%22,4,131072%5D&add=%5B%22qwen2.rope.scaling.type%22,8,%22yarn%22%5D&add=%5B%22qwen2.rope.scaling.factor%22,6,4%5D&add=%5B%22qwen2.rope.scaling.original_context_length%22,4,32768%5D)) | IQ1_S | 1 | 1.8 GB| 2.0 GB | smallest, significant quality loss |
91
+ | [Qwen2.5-Coder-7B-Instruct.IQ1_M.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ1_M.gguf) ([with YaRN](https://ciscai-gguf-editor.hf.space/download/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/Qwen2.5-Coder-7B-Instruct.IQ1_M.gguf?branch=main&add=%5B%22qwen2.context_length%22,4,131072%5D&add=%5B%22qwen2.rope.scaling.type%22,8,%22yarn%22%5D&add=%5B%22qwen2.rope.scaling.factor%22,6,4%5D&add=%5B%22qwen2.rope.scaling.original_context_length%22,4,32768%5D)) | IQ1_M | 1 | 1.9 GB| 2.1 GB | very small, significant quality loss |
92
+ | [Qwen2.5-Coder-7B-Instruct.IQ2_XXS.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ2_XXS.gguf) ([with YaRN](https://ciscai-gguf-editor.hf.space/download/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/Qwen2.5-Coder-7B-Instruct.IQ2_XXS.gguf?branch=main&add=%5B%22qwen2.context_length%22,4,131072%5D&add=%5B%22qwen2.rope.scaling.type%22,8,%22yarn%22%5D&add=%5B%22qwen2.rope.scaling.factor%22,6,4%5D&add=%5B%22qwen2.rope.scaling.original_context_length%22,4,32768%5D)) | IQ2_XXS | 2 | 2.1 GB| 2.3 GB | very small, high quality loss |
93
+ | [Qwen2.5-Coder-7B-Instruct.IQ2_XS.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ2_XS.gguf) ([with YaRN](https://ciscai-gguf-editor.hf.space/download/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/Qwen2.5-Coder-7B-Instruct.IQ2_XS.gguf?branch=main&add=%5B%22qwen2.context_length%22,4,131072%5D&add=%5B%22qwen2.rope.scaling.type%22,8,%22yarn%22%5D&add=%5B%22qwen2.rope.scaling.factor%22,6,4%5D&add=%5B%22qwen2.rope.scaling.original_context_length%22,4,32768%5D)) | IQ2_XS | 2 | 2.3 GB| 2.5 GB | very small, high quality loss |
94
+ | [Qwen2.5-Coder-7B-Instruct.IQ2_S.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ2_S.gguf) ([with YaRN](https://ciscai-gguf-editor.hf.space/download/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/Qwen2.5-Coder-7B-Instruct.IQ2_S.gguf?branch=main&add=%5B%22qwen2.context_length%22,4,131072%5D&add=%5B%22qwen2.rope.scaling.type%22,8,%22yarn%22%5D&add=%5B%22qwen2.rope.scaling.factor%22,6,4%5D&add=%5B%22qwen2.rope.scaling.original_context_length%22,4,32768%5D)) | IQ2_S | 2 | 2.4 GB| 2.6 GB | small, substantial quality loss |
95
+ | [Qwen2.5-Coder-7B-Instruct.IQ2_M.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ2_M.gguf) ([with YaRN](https://ciscai-gguf-editor.hf.space/download/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/Qwen2.5-Coder-7B-Instruct.IQ2_M.gguf?branch=main&add=%5B%22qwen2.context_length%22,4,131072%5D&add=%5B%22qwen2.rope.scaling.type%22,8,%22yarn%22%5D&add=%5B%22qwen2.rope.scaling.factor%22,6,4%5D&add=%5B%22qwen2.rope.scaling.original_context_length%22,4,32768%5D)) | IQ2_M | 2 | 2.6 GB| 2.8 GB | small, greater quality loss |
96
+ | [Qwen2.5-Coder-7B-Instruct.IQ3_XXS.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ3_XXS.gguf) ([with YaRN](https://ciscai-gguf-editor.hf.space/download/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/Qwen2.5-Coder-7B-Instruct.IQ3_XXS.gguf?branch=main&add=%5B%22qwen2.context_length%22,4,131072%5D&add=%5B%22qwen2.rope.scaling.type%22,8,%22yarn%22%5D&add=%5B%22qwen2.rope.scaling.factor%22,6,4%5D&add=%5B%22qwen2.rope.scaling.original_context_length%22,4,32768%5D)) | IQ3_XXS | 3 | 2.9 GB| 3.1 GB | very small, high quality loss |
97
+ | [Qwen2.5-Coder-7B-Instruct.IQ3_XS.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ3_XS.gguf) ([with YaRN](https://ciscai-gguf-editor.hf.space/download/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/Qwen2.5-Coder-7B-Instruct.IQ3_XS.gguf?branch=main&add=%5B%22qwen2.context_length%22,4,131072%5D&add=%5B%22qwen2.rope.scaling.type%22,8,%22yarn%22%5D&add=%5B%22qwen2.rope.scaling.factor%22,6,4%5D&add=%5B%22qwen2.rope.scaling.original_context_length%22,4,32768%5D)) | IQ3_XS | 3 | 3.1 GB| 3.3 GB | small, substantial quality loss |
98
+ | [Qwen2.5-Coder-7B-Instruct.IQ3_S.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ3_S.gguf) ([with YaRN](https://ciscai-gguf-editor.hf.space/download/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/Qwen2.5-Coder-7B-Instruct.IQ3_S.gguf?branch=main&add=%5B%22qwen2.context_length%22,4,131072%5D&add=%5B%22qwen2.rope.scaling.type%22,8,%22yarn%22%5D&add=%5B%22qwen2.rope.scaling.factor%22,6,4%5D&add=%5B%22qwen2.rope.scaling.original_context_length%22,4,32768%5D)) | IQ3_S | 3 | 3.3 GB| 3.4 GB | small, greater quality loss |
99
+ | [Qwen2.5-Coder-7B-Instruct.IQ3_M.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ3_M.gguf) ([with YaRN](https://ciscai-gguf-editor.hf.space/download/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/Qwen2.5-Coder-7B-Instruct.IQ3_M.gguf?branch=main&add=%5B%22qwen2.context_length%22,4,131072%5D&add=%5B%22qwen2.rope.scaling.type%22,8,%22yarn%22%5D&add=%5B%22qwen2.rope.scaling.factor%22,6,4%5D&add=%5B%22qwen2.rope.scaling.original_context_length%22,4,32768%5D)) | IQ3_M | 3 | 3.3 GB| 3.5 GB | medium, balanced quality - recommended |
100
+ | [Qwen2.5-Coder-7B-Instruct.IQ4_XS.gguf](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.IQ4_XS.gguf) ([with YaRN](https://ciscai-gguf-editor.hf.space/download/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/Qwen2.5-Coder-7B-Instruct.IQ4_XS.gguf?branch=main&add=%5B%22qwen2.context_length%22,4,131072%5D&add=%5B%22qwen2.rope.scaling.type%22,8,%22yarn%22%5D&add=%5B%22qwen2.rope.scaling.factor%22,6,4%5D&add=%5B%22qwen2.rope.scaling.original_context_length%22,4,32768%5D)) | IQ4_XS | 4 | 3.9 GB| 4.1 GB | small, substantial quality loss |
101
 
102
  Generated importance matrix file: [Qwen2.5-Coder-7B-Instruct.imatrix.dat](https://huggingface.co/CISCai/Qwen2.5-Coder-7B-Instruct-SOTA-GGUF/blob/main/Qwen2.5-Coder-7B-Instruct.imatrix.dat)
103