Kudod commited on
Commit
20f922a
1 Parent(s): 43dee39

End of training

Browse files
Files changed (2) hide show
  1. README.md +76 -0
  2. model.safetensors +1 -1
README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model: FacebookAI/xlm-roberta-large
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: roberta-large-ner-ghtk-cs-6-label-new-data-3090-6Sep-1
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # roberta-large-ner-ghtk-cs-6-label-new-data-3090-6Sep-1
15
+
16
+ This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the None dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 0.1725
19
+ - Tk: {'precision': 0.8651685393258427, 'recall': 0.6637931034482759, 'f1': 0.751219512195122, 'number': 116}
20
+ - Gày: {'precision': 0.75, 'recall': 0.8823529411764706, 'f1': 0.8108108108108107, 'number': 34}
21
+ - Gày trừu tượng: {'precision': 0.9111111111111111, 'recall': 0.9241803278688525, 'f1': 0.9175991861648015, 'number': 488}
22
+ - Ã đơn: {'precision': 0.8958333333333334, 'recall': 0.8472906403940886, 'f1': 0.8708860759493672, 'number': 203}
23
+ - Đt: {'precision': 0.9220917822838848, 'recall': 0.9840546697038725, 'f1': 0.9520661157024795, 'number': 878}
24
+ - Đt trừu tượng: {'precision': 0.8170731707317073, 'recall': 0.8626609442060086, 'f1': 0.8392484342379958, 'number': 233}
25
+ - Overall Precision: 0.8979
26
+ - Overall Recall: 0.9196
27
+ - Overall F1: 0.9086
28
+ - Overall Accuracy: 0.9685
29
+
30
+ ## Model description
31
+
32
+ More information needed
33
+
34
+ ## Intended uses & limitations
35
+
36
+ More information needed
37
+
38
+ ## Training and evaluation data
39
+
40
+ More information needed
41
+
42
+ ## Training procedure
43
+
44
+ ### Training hyperparameters
45
+
46
+ The following hyperparameters were used during training:
47
+ - learning_rate: 2.5e-05
48
+ - train_batch_size: 8
49
+ - eval_batch_size: 8
50
+ - seed: 42
51
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
+ - lr_scheduler_type: linear
53
+ - num_epochs: 10
54
+
55
+ ### Training results
56
+
57
+ | Training Loss | Epoch | Step | Validation Loss | Tk | Gày | Gày trừu tượng | Ã đơn | Đt | Đt trừu tượng | Overall Precision | Overall Recall | Overall F1 | Overall Accuracy |
58
+ |:-------------:|:-----:|:----:|:---------------:|:--------------------------------------------------------------------------------------------------------:|:--------------------------------------------------------------------------------------------------------:|:--------------------------------------------------------------------------------------------------------:|:--------------------------------------------------------------------------------------------------------:|:--------------------------------------------------------------------------------------------------------:|:--------------------------------------------------------------------------------------------------------:|:-----------------:|:--------------:|:----------:|:----------------:|
59
+ | No log | 1.0 | 467 | 0.1594 | {'precision': 0.75, 'recall': 0.07758620689655173, 'f1': 0.14062500000000003, 'number': 116} | {'precision': 0.30120481927710846, 'recall': 0.7352941176470589, 'f1': 0.4273504273504274, 'number': 34} | {'precision': 0.8500986193293886, 'recall': 0.8831967213114754, 'f1': 0.8663316582914573, 'number': 488} | {'precision': 0.777292576419214, 'recall': 0.8768472906403941, 'f1': 0.8240740740740742, 'number': 203} | {'precision': 0.8541666666666666, 'recall': 0.9806378132118451, 'f1': 0.9130434782608696, 'number': 878} | {'precision': 0.755868544600939, 'recall': 0.6909871244635193, 'f1': 0.7219730941704036, 'number': 233} | 0.8114 | 0.8530 | 0.8317 | 0.9466 |
60
+ | 0.2588 | 2.0 | 934 | 0.1616 | {'precision': 0.578125, 'recall': 0.31896551724137934, 'f1': 0.41111111111111115, 'number': 116} | {'precision': 0.5918367346938775, 'recall': 0.8529411764705882, 'f1': 0.6987951807228915, 'number': 34} | {'precision': 0.8096947935368043, 'recall': 0.9241803278688525, 'f1': 0.8631578947368421, 'number': 488} | {'precision': 0.9083969465648855, 'recall': 0.5862068965517241, 'f1': 0.7125748502994012, 'number': 203} | {'precision': 0.8858307849133538, 'recall': 0.989749430523918, 'f1': 0.9349112426035503, 'number': 878} | {'precision': 0.5903614457831325, 'recall': 0.8412017167381974, 'f1': 0.6938053097345134, 'number': 233} | 0.8046 | 0.8714 | 0.8367 | 0.9465 |
61
+ | 0.1107 | 3.0 | 1401 | 0.1661 | {'precision': 0.8070175438596491, 'recall': 0.7931034482758621, 'f1': 0.8, 'number': 116} | {'precision': 0.5370370370370371, 'recall': 0.8529411764705882, 'f1': 0.6590909090909091, 'number': 34} | {'precision': 0.8352272727272727, 'recall': 0.9036885245901639, 'f1': 0.8681102362204726, 'number': 488} | {'precision': 0.9352941176470588, 'recall': 0.7832512315270936, 'f1': 0.8525469168900804, 'number': 203} | {'precision': 0.9374325782092773, 'recall': 0.989749430523918, 'f1': 0.9628808864265929, 'number': 878} | {'precision': 0.6057692307692307, 'recall': 0.8111587982832618, 'f1': 0.6935779816513762, 'number': 233} | 0.8451 | 0.9114 | 0.8770 | 0.9562 |
62
+ | 0.0835 | 4.0 | 1868 | 0.1308 | {'precision': 0.8243243243243243, 'recall': 0.5258620689655172, 'f1': 0.6421052631578947, 'number': 116} | {'precision': 0.7741935483870968, 'recall': 0.7058823529411765, 'f1': 0.7384615384615385, 'number': 34} | {'precision': 0.8868686868686869, 'recall': 0.8995901639344263, 'f1': 0.8931841302136317, 'number': 488} | {'precision': 0.8695652173913043, 'recall': 0.7881773399014779, 'f1': 0.82687338501292, 'number': 203} | {'precision': 0.9097586568730325, 'recall': 0.9874715261958997, 'f1': 0.947023484434735, 'number': 878} | {'precision': 0.684931506849315, 'recall': 0.8583690987124464, 'f1': 0.7619047619047619, 'number': 233} | 0.8630 | 0.8970 | 0.8797 | 0.9594 |
63
+ | 0.0594 | 5.0 | 2335 | 0.1368 | {'precision': 0.6888888888888889, 'recall': 0.8017241379310345, 'f1': 0.7410358565737052, 'number': 116} | {'precision': 0.75, 'recall': 0.7058823529411765, 'f1': 0.7272727272727272, 'number': 34} | {'precision': 0.9032258064516129, 'recall': 0.9180327868852459, 'f1': 0.9105691056910569, 'number': 488} | {'precision': 0.9090909090909091, 'recall': 0.7881773399014779, 'f1': 0.8443271767810027, 'number': 203} | {'precision': 0.935659760087241, 'recall': 0.9772209567198178, 'f1': 0.9559888579387186, 'number': 878} | {'precision': 0.7751937984496124, 'recall': 0.8583690987124464, 'f1': 0.814663951120163, 'number': 233} | 0.8853 | 0.9134 | 0.8991 | 0.9647 |
64
+ | 0.044 | 6.0 | 2802 | 0.1338 | {'precision': 0.9333333333333333, 'recall': 0.603448275862069, 'f1': 0.7329842931937173, 'number': 116} | {'precision': 0.6744186046511628, 'recall': 0.8529411764705882, 'f1': 0.7532467532467532, 'number': 34} | {'precision': 0.9276595744680851, 'recall': 0.8934426229508197, 'f1': 0.9102296450939458, 'number': 488} | {'precision': 0.875, 'recall': 0.8275862068965517, 'f1': 0.8506329113924052, 'number': 203} | {'precision': 0.924468085106383, 'recall': 0.989749430523918, 'f1': 0.955995599559956, 'number': 878} | {'precision': 0.7807692307692308, 'recall': 0.871244635193133, 'f1': 0.8235294117647058, 'number': 233} | 0.8965 | 0.9093 | 0.9028 | 0.9658 |
65
+ | 0.0264 | 7.0 | 3269 | 0.1551 | {'precision': 0.7857142857142857, 'recall': 0.6637931034482759, 'f1': 0.719626168224299, 'number': 116} | {'precision': 0.6041666666666666, 'recall': 0.8529411764705882, 'f1': 0.7073170731707317, 'number': 34} | {'precision': 0.888015717092338, 'recall': 0.9262295081967213, 'f1': 0.9067201604814443, 'number': 488} | {'precision': 0.848780487804878, 'recall': 0.8571428571428571, 'f1': 0.8529411764705881, 'number': 203} | {'precision': 0.924468085106383, 'recall': 0.989749430523918, 'f1': 0.955995599559956, 'number': 878} | {'precision': 0.8253275109170306, 'recall': 0.8111587982832618, 'f1': 0.8181818181818181, 'number': 233} | 0.8822 | 0.9170 | 0.8993 | 0.9641 |
66
+ | 0.02 | 8.0 | 3736 | 0.1654 | {'precision': 0.9240506329113924, 'recall': 0.6293103448275862, 'f1': 0.7487179487179487, 'number': 116} | {'precision': 0.7142857142857143, 'recall': 0.8823529411764706, 'f1': 0.7894736842105262, 'number': 34} | {'precision': 0.9001996007984032, 'recall': 0.9241803278688525, 'f1': 0.9120323559150658, 'number': 488} | {'precision': 0.9, 'recall': 0.8423645320197044, 'f1': 0.8702290076335878, 'number': 203} | {'precision': 0.9234913793103449, 'recall': 0.9760820045558086, 'f1': 0.9490586932447398, 'number': 878} | {'precision': 0.8031496062992126, 'recall': 0.8755364806866953, 'f1': 0.837782340862423, 'number': 233} | 0.8957 | 0.9150 | 0.9052 | 0.9676 |
67
+ | 0.0107 | 9.0 | 4203 | 0.1804 | {'precision': 0.8390804597701149, 'recall': 0.6293103448275862, 'f1': 0.7192118226600984, 'number': 116} | {'precision': 0.725, 'recall': 0.8529411764705882, 'f1': 0.7837837837837837, 'number': 34} | {'precision': 0.907258064516129, 'recall': 0.9221311475409836, 'f1': 0.9146341463414632, 'number': 488} | {'precision': 0.8693467336683417, 'recall': 0.8522167487684729, 'f1': 0.8606965174129352, 'number': 203} | {'precision': 0.917558886509636, 'recall': 0.9760820045558086, 'f1': 0.945916114790287, 'number': 878} | {'precision': 0.8355555555555556, 'recall': 0.8068669527896996, 'f1': 0.8209606986899564, 'number': 233} | 0.8935 | 0.9068 | 0.9001 | 0.9661 |
68
+ | 0.009 | 10.0 | 4670 | 0.1725 | {'precision': 0.8651685393258427, 'recall': 0.6637931034482759, 'f1': 0.751219512195122, 'number': 116} | {'precision': 0.75, 'recall': 0.8823529411764706, 'f1': 0.8108108108108107, 'number': 34} | {'precision': 0.9111111111111111, 'recall': 0.9241803278688525, 'f1': 0.9175991861648015, 'number': 488} | {'precision': 0.8958333333333334, 'recall': 0.8472906403940886, 'f1': 0.8708860759493672, 'number': 203} | {'precision': 0.9220917822838848, 'recall': 0.9840546697038725, 'f1': 0.9520661157024795, 'number': 878} | {'precision': 0.8170731707317073, 'recall': 0.8626609442060086, 'f1': 0.8392484342379958, 'number': 233} | 0.8979 | 0.9196 | 0.9086 | 0.9685 |
69
+
70
+
71
+ ### Framework versions
72
+
73
+ - Transformers 4.44.0
74
+ - Pytorch 2.3.1+cu121
75
+ - Datasets 2.19.1
76
+ - Tokenizers 0.19.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7bb506bf2ad7f25db49640992dc964567a661bf242967f7a2f07c0deb7d0b886
3
  size 2235440556
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f4fefef8ba2e4fa0144f3acd69d64bb4186ec52d733f7d1754c8a139ed6b8e6
3
  size 2235440556