DUNHILL commited on
Commit
e64c545
1 Parent(s): 5ac0422

Model save

Browse files
README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: dandelin/vilt-b32-mlm
5
+ tags:
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: results2
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # results2
16
+
17
+ This model is a fine-tuned version of [dandelin/vilt-b32-mlm](https://huggingface.co/dandelin/vilt-b32-mlm) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 53.7912
20
+ - Bleu Score: {'bleu': 0.0, 'precisions': [0.5786516853932584, 0.0, 0.0, 0.0], 'brevity_penalty': 0.0015814103520051996, 'length_ratio': 0.13423831070889894, 'translation_length': 712, 'reference_length': 5304}
21
+
22
+ ## Model description
23
+
24
+ More information needed
25
+
26
+ ## Intended uses & limitations
27
+
28
+ More information needed
29
+
30
+ ## Training and evaluation data
31
+
32
+ More information needed
33
+
34
+ ## Training procedure
35
+
36
+ ### Training hyperparameters
37
+
38
+ The following hyperparameters were used during training:
39
+ - learning_rate: 5e-05
40
+ - train_batch_size: 8
41
+ - eval_batch_size: 16
42
+ - seed: 42
43
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
+ - lr_scheduler_type: linear
45
+ - num_epochs: 3
46
+
47
+ ### Training results
48
+
49
+ | Training Loss | Epoch | Step | Validation Loss | Bleu Score |
50
+ |:-------------:|:-----:|:----:|:---------------:|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:|
51
+ | 611.4564 | 1.0 | 63 | 163.9706 | {'bleu': 0.0, 'precisions': [0.0, 0.0, 0.0, 0.0], 'brevity_penalty': 5.701223175160721e-08, 'length_ratio': 0.05656108597285068, 'translation_length': 300, 'reference_length': 5304} |
52
+ | 130.132 | 2.0 | 126 | 63.4101 | {'bleu': 0.0, 'precisions': [0.6240601503759399, 0.0, 0.0, 0.0], 'brevity_penalty': 0.0035294383040863748, 'length_ratio': 0.1504524886877828, 'translation_length': 798, 'reference_length': 5304} |
53
+ | 57.0285 | 3.0 | 189 | 53.7912 | {'bleu': 0.0, 'precisions': [0.5786516853932584, 0.0, 0.0, 0.0], 'brevity_penalty': 0.0015814103520051996, 'length_ratio': 0.13423831070889894, 'translation_length': 712, 'reference_length': 5304} |
54
+
55
+
56
+ ### Framework versions
57
+
58
+ - Transformers 4.44.2
59
+ - Pytorch 2.4.1+cu121
60
+ - Datasets 3.0.1
61
+ - Tokenizers 0.19.1
runs/Oct20_11-00-36_551f9879ebe4/events.out.tfevents.1729422047.551f9879ebe4.778.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e793024e12dedd4ecc3b43d91173de7ba25db4f63b5324c05637a20e8eadb9d
3
- size 264036
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44affc0fa7f916bead065ff4e76fba76d44fe794f7d8346fba5f1c8ac4c93dd4
3
+ size 264661
runs/Oct20_11-00-36_551f9879ebe4/events.out.tfevents.1729429774.551f9879ebe4.778.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e47f22a670b6ab172a340542c453faf65a7455203d1240ea891325b6fb227b38
3
+ size 359