bartowski commited on
Commit
c2c8956
·
verified ·
1 Parent(s): e394ec3

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +5 -8
README.md CHANGED
@@ -1,17 +1,11 @@
1
  ---
2
  quantized_by: bartowski
3
  pipeline_tag: text-generation
4
- tags:
5
- - falcon3
6
- license: other
7
- base_model: tiiuae/Falcon3-10B-Instruct
8
- license_name: falcon-llm-license
9
- license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
10
  ---
11
 
12
  ## Llamacpp imatrix Quantizations of Falcon3-10B-Instruct
13
 
14
- Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b4341">b4341</a> for quantization.
15
 
16
  Original model: https://huggingface.co/tiiuae/Falcon3-10B-Instruct
17
 
@@ -29,11 +23,14 @@ Run them in [LM Studio](https://lmstudio.ai/)
29
  <|assistant|>
30
  ```
31
 
 
 
 
 
32
  ## Download a file (not the whole branch) from below:
33
 
34
  | Filename | Quant type | File Size | Split | Description |
35
  | -------- | ---------- | --------- | ----- | ----------- |
36
- | [Falcon3-10B-Instruct-f32.gguf](https://huggingface.co/bartowski/Falcon3-10B-Instruct-GGUF/blob/main/Falcon3-10B-Instruct-f32.gguf) | f32 | 41.23GB | false | Full F32 weights. |
37
  | [Falcon3-10B-Instruct-f16.gguf](https://huggingface.co/bartowski/Falcon3-10B-Instruct-GGUF/blob/main/Falcon3-10B-Instruct-f16.gguf) | f16 | 20.62GB | false | Full F16 weights. |
38
  | [Falcon3-10B-Instruct-Q8_0.gguf](https://huggingface.co/bartowski/Falcon3-10B-Instruct-GGUF/blob/main/Falcon3-10B-Instruct-Q8_0.gguf) | Q8_0 | 10.96GB | false | Extremely high quality, generally unneeded but max available quant. |
39
  | [Falcon3-10B-Instruct-Q6_K_L.gguf](https://huggingface.co/bartowski/Falcon3-10B-Instruct-GGUF/blob/main/Falcon3-10B-Instruct-Q6_K_L.gguf) | Q6_K_L | 8.65GB | false | Uses Q8_0 for embed and output weights. Very high quality, near perfect, *recommended*. |
 
1
  ---
2
  quantized_by: bartowski
3
  pipeline_tag: text-generation
 
 
 
 
 
 
4
  ---
5
 
6
  ## Llamacpp imatrix Quantizations of Falcon3-10B-Instruct
7
 
8
+ Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b4381">b4381</a> for quantization.
9
 
10
  Original model: https://huggingface.co/tiiuae/Falcon3-10B-Instruct
11
 
 
23
  <|assistant|>
24
  ```
25
 
26
+ ## What's new:
27
+
28
+ Fix tokenizer
29
+
30
  ## Download a file (not the whole branch) from below:
31
 
32
  | Filename | Quant type | File Size | Split | Description |
33
  | -------- | ---------- | --------- | ----- | ----------- |
 
34
  | [Falcon3-10B-Instruct-f16.gguf](https://huggingface.co/bartowski/Falcon3-10B-Instruct-GGUF/blob/main/Falcon3-10B-Instruct-f16.gguf) | f16 | 20.62GB | false | Full F16 weights. |
35
  | [Falcon3-10B-Instruct-Q8_0.gguf](https://huggingface.co/bartowski/Falcon3-10B-Instruct-GGUF/blob/main/Falcon3-10B-Instruct-Q8_0.gguf) | Q8_0 | 10.96GB | false | Extremely high quality, generally unneeded but max available quant. |
36
  | [Falcon3-10B-Instruct-Q6_K_L.gguf](https://huggingface.co/bartowski/Falcon3-10B-Instruct-GGUF/blob/main/Falcon3-10B-Instruct-Q6_K_L.gguf) | Q6_K_L | 8.65GB | false | Uses Q8_0 for embed and output weights. Very high quality, near perfect, *recommended*. |