BramVanroy commited on
Commit
744c229
·
verified ·
1 Parent(s): 0f0ac94

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -4
README.md CHANGED
@@ -24,13 +24,11 @@ should probably proofread and complete it, then remove this comment. -->
24
 
25
  # GEITje-ultra-sft
26
 
27
- This model is a fine-tuned version of [Rijgersberg/GEITje-7B](https://huggingface.co/Rijgersberg/GEITje-7B) on the BramVanroy/ultrachat_200k_dutch, the BramVanroy/stackoverflow-chat-dutch, the BramVanroy/alpaca-cleaned-dutch, the BramVanroy/dolly-15k-dutch and the BramVanroy/no_robots_dutch datasets.
28
- It achieves the following results on the evaluation set:
29
- - Loss: 0.8563
30
 
31
  ## Model description
32
 
33
- This model is a SFT (chat-tuned) version of [Rijgersberg/GEITje-7B](https://huggingface.co/Rijgersberg/GEITje-7B), which in turn is based on Mistral 7B and further finetuned.
34
 
35
  ## Intended uses & limitations
36
 
 
24
 
25
  # GEITje-ultra-sft
26
 
27
+ This model is a fine-tuned version of [Rijgersberg/GEITje-7B](https://huggingface.co/Rijgersberg/GEITje-7B) on a number of synthetic datasets including gpt-3.5-turbo and gpt-4-turbo data, multi- and single turn conversations, and code. The training set consists of around 240M tokens. The model was trained with context length 8192.
 
 
28
 
29
  ## Model description
30
 
31
+ This model is a SFT (chat-tuned) version of [Rijgersberg/GEITje-7B](https://huggingface.co/Rijgersberg/GEITje-7B), which in turn is based on Mistral 7B and further pretrained on Dutch data.
32
 
33
  ## Intended uses & limitations
34