starble-dev commited on
Commit
dc6a974
·
verified ·
1 Parent(s): 7665ca2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -6
README.md CHANGED
@@ -13,14 +13,13 @@ library_name: transformers
13
  > Mistral-Nemo-12B is very sensitive to the temperature sampler, try values near **0.3** at first or else you will get some weird results. This is mentioned by MistralAI at their [Transformers](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407#transformers) section. <br>
14
  > In my personal testing, Flash-Attention seems to have seem weird effects with the model as well, however there is no confirmation on this.
15
 
16
- **Original Model:**
17
- [BeaverAI/mistral-doryV2-12b](https://huggingface.co/BeaverAI/mistral-doryV2-12b)
18
 
19
- **How to Use:**
20
- [llama.cpp](https://github.com/ggerganov/llama.cpp)
21
 
22
- **License:**
23
- Apache 2.0
 
24
 
25
  # Quants
26
  | Name | Quant Type | Size |
 
13
  > Mistral-Nemo-12B is very sensitive to the temperature sampler, try values near **0.3** at first or else you will get some weird results. This is mentioned by MistralAI at their [Transformers](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407#transformers) section. <br>
14
  > In my personal testing, Flash-Attention seems to have seem weird effects with the model as well, however there is no confirmation on this.
15
 
16
+ **Original Model:** [BeaverAI/mistral-doryV2-12b](https://huggingface.co/BeaverAI/mistral-doryV2-12b)
 
17
 
18
+ **How to Use:** [llama.cpp](https://github.com/ggerganov/llama.cpp)
 
19
 
20
+ **Original Model License:** Apache 2.0
21
+
22
+ **Release Used:** [b3441](https://github.com/ggerganov/llama.cpp/releases/tag/b3441)
23
 
24
  # Quants
25
  | Name | Quant Type | Size |