BramVanroy
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -24,13 +24,11 @@ should probably proofread and complete it, then remove this comment. -->
|
|
24 |
|
25 |
# GEITje-ultra-sft
|
26 |
|
27 |
-
This model is a fine-tuned version of [Rijgersberg/GEITje-7B](https://huggingface.co/Rijgersberg/GEITje-7B) on
|
28 |
-
It achieves the following results on the evaluation set:
|
29 |
-
- Loss: 0.8563
|
30 |
|
31 |
## Model description
|
32 |
|
33 |
-
This model is a SFT (chat-tuned) version of [Rijgersberg/GEITje-7B](https://huggingface.co/Rijgersberg/GEITje-7B), which in turn is based on Mistral 7B and further
|
34 |
|
35 |
## Intended uses & limitations
|
36 |
|
|
|
24 |
|
25 |
# GEITje-ultra-sft
|
26 |
|
27 |
+
This model is a fine-tuned version of [Rijgersberg/GEITje-7B](https://huggingface.co/Rijgersberg/GEITje-7B) on a number of synthetic datasets including gpt-3.5-turbo and gpt-4-turbo data, multi- and single turn conversations, and code. The training set consists of around 240M tokens. The model was trained with context length 8192.
|
|
|
|
|
28 |
|
29 |
## Model description
|
30 |
|
31 |
+
This model is a SFT (chat-tuned) version of [Rijgersberg/GEITje-7B](https://huggingface.co/Rijgersberg/GEITje-7B), which in turn is based on Mistral 7B and further pretrained on Dutch data.
|
32 |
|
33 |
## Intended uses & limitations
|
34 |
|