starble-dev
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -13,14 +13,13 @@ library_name: transformers
|
|
13 |
> Mistral-Nemo-12B is very sensitive to the temperature sampler, try values near **0.3** at first or else you will get some weird results. This is mentioned by MistralAI at their [Transformers](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407#transformers) section. <br>
|
14 |
> In my personal testing, Flash-Attention seems to have seem weird effects with the model as well, however there is no confirmation on this.
|
15 |
|
16 |
-
**Original Model:**
|
17 |
-
[BeaverAI/mistral-doryV2-12b](https://huggingface.co/BeaverAI/mistral-doryV2-12b)
|
18 |
|
19 |
-
**How to Use:**
|
20 |
-
[llama.cpp](https://github.com/ggerganov/llama.cpp)
|
21 |
|
22 |
-
**License:**
|
23 |
-
|
|
|
24 |
|
25 |
# Quants
|
26 |
| Name | Quant Type | Size |
|
|
|
13 |
> Mistral-Nemo-12B is very sensitive to the temperature sampler, try values near **0.3** at first or else you will get some weird results. This is mentioned by MistralAI at their [Transformers](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407#transformers) section. <br>
|
14 |
> In my personal testing, Flash-Attention seems to have seem weird effects with the model as well, however there is no confirmation on this.
|
15 |
|
16 |
+
**Original Model:** [BeaverAI/mistral-doryV2-12b](https://huggingface.co/BeaverAI/mistral-doryV2-12b)
|
|
|
17 |
|
18 |
+
**How to Use:** [llama.cpp](https://github.com/ggerganov/llama.cpp)
|
|
|
19 |
|
20 |
+
**Original Model License:** Apache 2.0
|
21 |
+
|
22 |
+
**Release Used:** [b3441](https://github.com/ggerganov/llama.cpp/releases/tag/b3441)
|
23 |
|
24 |
# Quants
|
25 |
| Name | Quant Type | Size |
|