BramVanroy commited on
Commit
14cc9c3
·
1 Parent(s): 113d430

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -14
README.md CHANGED
@@ -1,37 +1,38 @@
1
  ---
2
- license: apache-2.0
3
- tags:
4
- - generated_from_trainer
5
  datasets:
6
  - BramVanroy/alpaca-dolly-dutch
 
 
 
7
  model-index:
8
- - name: 2e-4lr+512tbs+5ep+32a+4r-v2
9
  results: []
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
 
15
- # 2e-4lr+512tbs+5ep+32a+4r-v2
16
 
17
  This model is a fine-tuned version of [tiiuae/falcon-40b](https://huggingface.co/tiiuae/falcon-40b) on the BramVanroy/alpaca-dolly-dutch dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 0.9154
20
 
21
- ## Model description
 
22
 
23
- More information needed
24
 
25
  ## Intended uses & limitations
26
 
27
- More information needed
28
 
29
  ## Training and evaluation data
30
 
31
- More information needed
 
32
 
33
  ## Training procedure
34
 
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
@@ -47,7 +48,7 @@ The following hyperparameters were used during training:
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: linear
49
  - lr_scheduler_warmup_steps: 150
50
- - num_epochs: 5
51
 
52
  ### Training results
53
 
 
1
  ---
2
+ license: cc-by-nc-4.0
 
 
3
  datasets:
4
  - BramVanroy/alpaca-dolly-dutch
5
+ language:
6
+ - nl
7
+ inference: false
8
  model-index:
9
+ - name: falcon-7b-ft-alpaca-cleaned-dutch
10
  results: []
11
  ---
12
 
13
+ # falcon-40b-ft-alpaca-dolly-dutch
 
14
 
15
+ ## Model description
16
 
17
  This model is a fine-tuned version of [tiiuae/falcon-40b](https://huggingface.co/tiiuae/falcon-40b) on the BramVanroy/alpaca-dolly-dutch dataset.
 
 
18
 
19
+ This model is a fine-tuned version of [tiiuae/falcon-40b](https://huggingface.co/tiiuae/falcon-40b) on the [BramVanroy/alpaca-dolly-dutch](https://huggingface.co/datasets/BramVanroy/alpaca-dolly-dutch) dataset.
20
+ See the original [tiiuae/falcon-40b](https://huggingface.co/tiiuae/falcon-40b) for more information, intended use, and biases.
21
 
 
22
 
23
  ## Intended uses & limitations
24
 
25
+ This model is intended as a (poor) baseline for Dutch generative LLMs. It by no means aims to provide SOTA performance and is specifically intended for research purposes and experimentation.
26
 
27
  ## Training and evaluation data
28
 
29
+ Trained on the synthetic [BramVanroy/alpaca-dolly-dutch](https://huggingface.co/datasets/BramVanroy/alpaca-dolly-dutch) instruction dataset.
30
+ Therefore, commercial use of this model is forbidden. The model is intended for research purposes only.
31
 
32
  ## Training procedure
33
 
34
+ Trained with LoRA and merged before upload. The adapters are in the `adapters` branch.
35
+
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
 
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
  - lr_scheduler_warmup_steps: 150
51
+ - num_epochs: 5 (but with early stopping)
52
 
53
  ### Training results
54