BramVanroy
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -30,6 +30,22 @@ This model is a fine-tuned version of [Rijgersberg/GEITje-7B](https://huggingfac
|
|
30 |
> Note that this model has not been aligned with DPO or other techniques. In practice, it is therefore recommended to use the [DPO variant](https://huggingface.co/BramVanroy/GEITje-7B-ultra) of this model.
|
31 |
|
32 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
## Model description
|
34 |
|
35 |
This model is a SFT (chat-tuned) version of [Rijgersberg/GEITje-7B](https://huggingface.co/Rijgersberg/GEITje-7B), which in turn is based on Mistral 7B and further pretrained on Dutch data.
|
|
|
30 |
> Note that this model has not been aligned with DPO or other techniques. In practice, it is therefore recommended to use the [DPO variant](https://huggingface.co/BramVanroy/GEITje-7B-ultra) of this model.
|
31 |
|
32 |
|
33 |
+
## Citation
|
34 |
+
|
35 |
+
If you use GEITje 7B Ultra (SFT) or any of its derivatives or quantizations, place cite the following paper:
|
36 |
+
|
37 |
+
```bibtex
|
38 |
+
@misc{vanroy2024geitje7bultraconversational,
|
39 |
+
title={GEITje 7B Ultra: A Conversational Model for Dutch},
|
40 |
+
author={Bram Vanroy},
|
41 |
+
year={2024},
|
42 |
+
eprint={2412.04092},
|
43 |
+
archivePrefix={arXiv},
|
44 |
+
primaryClass={cs.CL},
|
45 |
+
url={https://arxiv.org/abs/2412.04092},
|
46 |
+
}
|
47 |
+
```
|
48 |
+
|
49 |
## Model description
|
50 |
|
51 |
This model is a SFT (chat-tuned) version of [Rijgersberg/GEITje-7B](https://huggingface.co/Rijgersberg/GEITje-7B), which in turn is based on Mistral 7B and further pretrained on Dutch data.
|