rishavranaut
commited on
Commit
•
13cd2c6
1
Parent(s):
aceb721
End of training
Browse files- README.md +135 -103
- model.safetensors +1 -1
- runs/Oct02_19-53-09_iit-p/events.out.tfevents.1727879006.iit-p +2 -2
README.md
CHANGED
@@ -20,11 +20,11 @@ should probably proofread and complete it, then remove this comment. -->
|
|
20 |
|
21 |
This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the None dataset.
|
22 |
It achieves the following results on the evaluation set:
|
23 |
-
- Loss: 1.
|
24 |
-
- Accuracy: 0.
|
25 |
-
- Precision: 0.
|
26 |
-
- Recall: 0.
|
27 |
-
- F1 score: 0.
|
28 |
|
29 |
## Model description
|
30 |
|
@@ -49,107 +49,139 @@ The following hyperparameters were used during training:
|
|
49 |
- seed: 42
|
50 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
51 |
- lr_scheduler_type: linear
|
52 |
-
- num_epochs:
|
53 |
|
54 |
### Training results
|
55 |
|
56 |
-
| Training Loss | Epoch | Step |
|
57 |
-
|
58 |
-
| 1.2434 | 0.0314 | 200 | 0.
|
59 |
-
| 1.4323 | 0.0628 | 400 |
|
60 |
-
| 1.253 | 0.0941 | 600 |
|
61 |
-
| 1.2986 | 0.1255 | 800 |
|
62 |
-
| 1.2843 | 0.1569 | 1000 |
|
63 |
-
| 1.2245 | 0.1883 | 1200 |
|
64 |
-
| 1.2389 | 0.2197 | 1400 | 0.
|
65 |
-
| 1.132 | 0.2511 | 1600 | 0.
|
66 |
-
| 1.0836 | 0.2824 | 1800 |
|
67 |
-
| 1.2434 | 0.3138 | 2000 |
|
68 |
-
| 1.1196 | 0.3452 | 2200 |
|
69 |
-
| 1.5791 | 0.3766 | 2400 |
|
70 |
-
| 1.3035 | 0.4080 | 2600 |
|
71 |
-
| 1.1401 | 0.4394 | 2800 |
|
72 |
-
| 0.9518 | 0.4707 | 3000 |
|
73 |
-
| 1.1623 | 0.5021 | 3200 |
|
74 |
-
| 1.1316 | 0.5335 | 3400 |
|
75 |
-
| 1.2599 | 0.5649 | 3600 | 0.
|
76 |
-
| 0.9866 | 0.5963 | 3800 |
|
77 |
-
| 1.147 | 0.6276 | 4000 |
|
78 |
-
| 1.0307 | 0.6590 | 4200 |
|
79 |
-
| 1.0616 | 0.6904 | 4400 |
|
80 |
-
| 1.0724 | 0.7218 | 4600 |
|
81 |
-
| 0.9499 | 0.7532 | 4800 |
|
82 |
-
| 1.1302 | 0.7846 | 5000 |
|
83 |
-
| 1.1921 | 0.8159 | 5200 |
|
84 |
-
| 0.9532 | 0.8473 | 5400 |
|
85 |
-
| 0.783 | 0.8787 | 5600 |
|
86 |
-
| 0.9835 | 0.9101 | 5800 |
|
87 |
-
| 0.9898 | 0.9415 | 6000 |
|
88 |
-
| 0.9768 | 0.9729 | 6200 |
|
89 |
-
| 1.043 | 1.0042 | 6400 |
|
90 |
-
| 0.9531 | 1.0356 | 6600 |
|
91 |
-
| 1.0585 | 1.0670 | 6800 |
|
92 |
-
| 0.8862 | 1.0984 | 7000 |
|
93 |
-
| 0.8721 | 1.1298 | 7200 | 0.
|
94 |
-
| 0.8678 | 1.1611 | 7400 |
|
95 |
-
| 0.7617 | 1.1925 | 7600 |
|
96 |
-
| 1.0394 | 1.2239 | 7800 |
|
97 |
-
| 0.822 | 1.2553 | 8000 |
|
98 |
-
| 0.8406 | 1.2867 | 8200 |
|
99 |
-
| 0.7059 | 1.3181 | 8400 |
|
100 |
-
| 0.8649 | 1.3494 | 8600 | 0.
|
101 |
-
| 0.7142 | 1.3808 | 8800 |
|
102 |
-
| 0.9057 | 1.4122 | 9000 | 0.
|
103 |
-
| 0.9312 | 1.4436 | 9200 |
|
104 |
-
| 0.8459 | 1.4750 | 9400 |
|
105 |
-
| 0.8427 | 1.5064 | 9600 |
|
106 |
-
| 0.7245 | 1.5377 | 9800 |
|
107 |
-
| 0.6386 | 1.5691 | 10000 |
|
108 |
-
| 0.7513 | 1.6005 | 10200 | 0.
|
109 |
-
| 0.828 | 1.6319 | 10400 | 0.
|
110 |
-
| 0.8393 | 1.6633 | 10600 | 0.
|
111 |
-
| 0.8679 | 1.6946 | 10800 | 0.
|
112 |
-
| 0.6735 | 1.7260 | 11000 |
|
113 |
-
| 0.8702 | 1.7574 | 11200 | 0.
|
114 |
-
| 0.7435 | 1.7888 | 11400 |
|
115 |
-
| 0.8796 | 1.8202 | 11600 | 0.
|
116 |
-
| 0.6257 | 1.8516 | 11800 | 0.
|
117 |
-
| 0.8589 | 1.8829 | 12000 | 0.
|
118 |
-
| 0.865 | 1.9143 | 12200 |
|
119 |
-
| 0.8068 | 1.9457 | 12400 | 0.
|
120 |
-
| 0.6212 | 1.9771 | 12600 |
|
121 |
-
| 0.7657 | 2.0085 | 12800 | 0.
|
122 |
-
| 0.6631 | 2.0399 | 13000 |
|
123 |
-
| 0.3003 | 2.0712 | 13200 |
|
124 |
-
| 0.5982 | 2.1026 | 13400 |
|
125 |
-
| 0.4828 | 2.1340 | 13600 |
|
126 |
-
| 0.5463 | 2.1654 | 13800 |
|
127 |
-
| 0.5429 | 2.1968 | 14000 |
|
128 |
-
| 0.7112 | 2.2282 | 14200 |
|
129 |
-
| 0.4816 | 2.2595 | 14400 |
|
130 |
-
| 0.7882 | 2.2909 | 14600 | 0.
|
131 |
-
| 0.5265 | 2.3223 | 14800 |
|
132 |
-
| 0.6116 | 2.3537 | 15000 |
|
133 |
-
| 0.575 | 2.3851 | 15200 |
|
134 |
-
| 0.5599 | 2.4164 | 15400 |
|
135 |
-
| 0.5821 | 2.4478 | 15600 |
|
136 |
-
| 0.4777 | 2.4792 | 15800 |
|
137 |
-
| 0.432 | 2.5106 | 16000 |
|
138 |
-
| 0.4385 | 2.5420 | 16200 |
|
139 |
-
| 0.6103 | 2.5734 | 16400 |
|
140 |
-
| 0.4618 | 2.6047 | 16600 |
|
141 |
-
| 0.471 | 2.6361 | 16800 |
|
142 |
-
| 0.6206 | 2.6675 | 17000 |
|
143 |
-
| 0.5869 | 2.6989 | 17200 |
|
144 |
-
| 0.5647 | 2.7303 | 17400 |
|
145 |
-
| 0.5527 | 2.7617 | 17600 |
|
146 |
-
| 0.8013 | 2.7930 | 17800 | 0.
|
147 |
-
| 0.5232 | 2.8244 | 18000 |
|
148 |
-
| 0.617 | 2.8558 | 18200 |
|
149 |
-
| 0.5093 | 2.8872 | 18400 |
|
150 |
-
| 0.5099 | 2.9186 | 18600 |
|
151 |
-
| 0.6239 | 2.9499 | 18800 |
|
152 |
-
| 0.6592 | 2.9813 | 19000 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
153 |
|
154 |
|
155 |
### Framework versions
|
|
|
20 |
|
21 |
This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the None dataset.
|
22 |
It achieves the following results on the evaluation set:
|
23 |
+
- Loss: 1.5959
|
24 |
+
- Accuracy: 0.7812
|
25 |
+
- Precision: 0.8048
|
26 |
+
- Recall: 0.7647
|
27 |
+
- F1 score: 0.7842
|
28 |
|
29 |
## Model description
|
30 |
|
|
|
49 |
- seed: 42
|
50 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
51 |
- lr_scheduler_type: linear
|
52 |
+
- num_epochs: 5
|
53 |
|
54 |
### Training results
|
55 |
|
56 |
+
| Training Loss | Epoch | Step | Accuracy | F1 score | Precision | Recall | Validation Loss |
|
57 |
+
|:-------------:|:------:|:-----:|:--------:|:--------:|:---------:|:------:|:---------------:|
|
58 |
+
| 1.2434 | 0.0314 | 200 | 0.6059 | 0.6974 | 0.5805 | 0.8733 | 0.7545 |
|
59 |
+
| 1.4323 | 0.0628 | 400 | 0.6 | 0.6970 | 0.575 | 0.8846 | 1.7199 |
|
60 |
+
| 1.253 | 0.0941 | 600 | 0.6059 | 0.6957 | 0.5812 | 0.8665 | 1.3828 |
|
61 |
+
| 1.2986 | 0.1255 | 800 | 0.48 | 0.0 | 0.0 | 0.0 | 2.4072 |
|
62 |
+
| 1.2843 | 0.1569 | 1000 | 0.6553 | 0.5925 | 0.7690 | 0.4819 | 1.2474 |
|
63 |
+
| 1.2245 | 0.1883 | 1200 | 0.5435 | 0.2271 | 0.95 | 0.1290 | 1.3841 |
|
64 |
+
| 1.2389 | 0.2197 | 1400 | 0.6706 | 0.6923 | 0.6731 | 0.7127 | 0.8993 |
|
65 |
+
| 1.132 | 0.2511 | 1600 | 0.6271 | 0.6846 | 0.6110 | 0.7783 | 0.6845 |
|
66 |
+
| 1.0836 | 0.2824 | 1800 | 0.6824 | 0.6574 | 0.7486 | 0.5860 | 1.1694 |
|
67 |
+
| 1.2434 | 0.3138 | 2000 | 0.6788 | 0.6422 | 0.7632 | 0.5543 | 1.4787 |
|
68 |
+
| 1.1196 | 0.3452 | 2200 | 0.6694 | 0.6659 | 0.7018 | 0.6335 | 1.5004 |
|
69 |
+
| 1.5791 | 0.3766 | 2400 | 0.6376 | 0.5157 | 0.8454 | 0.3710 | 1.1289 |
|
70 |
+
| 1.3035 | 0.4080 | 2600 | 0.6859 | 0.6877 | 0.7119 | 0.6652 | 1.0136 |
|
71 |
+
| 1.1401 | 0.4394 | 2800 | 0.6753 | 0.6443 | 0.7485 | 0.5656 | 1.2340 |
|
72 |
+
| 0.9518 | 0.4707 | 3000 | 0.7024 | 0.6810 | 0.7692 | 0.6109 | 1.2197 |
|
73 |
+
| 1.1623 | 0.5021 | 3200 | 0.6788 | 0.6807 | 0.7046 | 0.6584 | 1.2827 |
|
74 |
+
| 1.1316 | 0.5335 | 3400 | 0.6659 | 0.7183 | 0.6396 | 0.8190 | 1.5077 |
|
75 |
+
| 1.2599 | 0.5649 | 3600 | 0.6341 | 0.5071 | 0.8466 | 0.3620 | 0.8272 |
|
76 |
+
| 0.9866 | 0.5963 | 3800 | 0.6647 | 0.6939 | 0.6605 | 0.7308 | 1.4574 |
|
77 |
+
| 1.147 | 0.6276 | 4000 | 0.6824 | 0.64 | 0.7792 | 0.5430 | 1.2933 |
|
78 |
+
| 1.0307 | 0.6590 | 4200 | 0.6482 | 0.5795 | 0.7658 | 0.4661 | 1.1586 |
|
79 |
+
| 1.0616 | 0.6904 | 4400 | 0.6976 | 0.6877 | 0.7428 | 0.6403 | 1.2668 |
|
80 |
+
| 1.0724 | 0.7218 | 4600 | 0.6447 | 0.6225 | 0.6955 | 0.5633 | 1.1130 |
|
81 |
+
| 0.9499 | 0.7532 | 4800 | 0.7188 | 0.7046 | 0.7766 | 0.6448 | 1.1635 |
|
82 |
+
| 1.1302 | 0.7846 | 5000 | 0.7118 | 0.7357 | 0.7031 | 0.7715 | 1.2608 |
|
83 |
+
| 1.1921 | 0.8159 | 5200 | 0.7094 | 0.6754 | 0.8056 | 0.5814 | 1.1742 |
|
84 |
+
| 0.9532 | 0.8473 | 5400 | 0.7071 | 0.6860 | 0.7749 | 0.6154 | 1.1589 |
|
85 |
+
| 0.783 | 0.8787 | 5600 | 0.7 | 0.6577 | 0.8086 | 0.5543 | 1.3256 |
|
86 |
+
| 0.9835 | 0.9101 | 5800 | 0.7282 | 0.7166 | 0.7828 | 0.6606 | 1.1383 |
|
87 |
+
| 0.9898 | 0.9415 | 6000 | 0.7141 | 0.7158 | 0.7409 | 0.6923 | 1.0662 |
|
88 |
+
| 0.9768 | 0.9729 | 6200 | 0.7059 | 0.6711 | 0.8019 | 0.5769 | 1.1941 |
|
89 |
+
| 1.043 | 1.0042 | 6400 | 0.6729 | 0.5838 | 0.8628 | 0.4412 | 1.2302 |
|
90 |
+
| 0.9531 | 1.0356 | 6600 | 0.7106 | 0.7 | 0.7593 | 0.6493 | 1.1304 |
|
91 |
+
| 1.0585 | 1.0670 | 6800 | 0.7294 | 0.7132 | 0.7944 | 0.6471 | 1.0234 |
|
92 |
+
| 0.8862 | 1.0984 | 7000 | 0.6953 | 0.6230 | 0.8735 | 0.4842 | 1.1941 |
|
93 |
+
| 0.8721 | 1.1298 | 7200 | 0.7376 | 0.7271 | 0.792 | 0.6719 | 0.9352 |
|
94 |
+
| 0.8678 | 1.1611 | 7400 | 0.7388 | 0.7533 | 0.7402 | 0.7670 | 1.0473 |
|
95 |
+
| 0.7617 | 1.1925 | 7600 | 0.7294 | 0.7522 | 0.7181 | 0.7896 | 1.3020 |
|
96 |
+
| 1.0394 | 1.2239 | 7800 | 0.7212 | 0.7019 | 0.7904 | 0.6312 | 1.0322 |
|
97 |
+
| 0.822 | 1.2553 | 8000 | 0.7388 | 0.7266 | 0.7973 | 0.6674 | 1.0980 |
|
98 |
+
| 0.8406 | 1.2867 | 8200 | 0.7118 | 0.7357 | 0.7031 | 0.7715 | 1.4589 |
|
99 |
+
| 0.7059 | 1.3181 | 8400 | 0.7306 | 0.6999 | 0.8318 | 0.6041 | 1.0655 |
|
100 |
+
| 0.8649 | 1.3494 | 8600 | 0.7424 | 0.7266 | 0.8106 | 0.6584 | 0.9708 |
|
101 |
+
| 0.7142 | 1.3808 | 8800 | 0.7553 | 0.7419 | 0.8214 | 0.6765 | 1.1603 |
|
102 |
+
| 0.9057 | 1.4122 | 9000 | 0.76 | 0.7431 | 0.8381 | 0.6674 | 0.9389 |
|
103 |
+
| 0.9312 | 1.4436 | 9200 | 0.7553 | 0.7615 | 0.7721 | 0.7511 | 1.0568 |
|
104 |
+
| 0.8459 | 1.4750 | 9400 | 0.7459 | 0.7372 | 0.7974 | 0.6855 | 1.1646 |
|
105 |
+
| 0.8427 | 1.5064 | 9600 | 0.7459 | 0.7293 | 0.8174 | 0.6584 | 1.0133 |
|
106 |
+
| 0.7245 | 1.5377 | 9800 | 0.7341 | 0.6861 | 0.8885 | 0.5588 | 1.1397 |
|
107 |
+
| 0.6386 | 1.5691 | 10000 | 0.7294 | 0.6742 | 0.9015 | 0.5385 | 1.1112 |
|
108 |
+
| 0.7513 | 1.6005 | 10200 | 0.7671 | 0.7648 | 0.805 | 0.7285 | 0.9403 |
|
109 |
+
| 0.828 | 1.6319 | 10400 | 0.76 | 0.7639 | 0.7820 | 0.7466 | 0.9412 |
|
110 |
+
| 0.8393 | 1.6633 | 10600 | 0.7553 | 0.7219 | 0.8824 | 0.6109 | 0.9359 |
|
111 |
+
| 0.8679 | 1.6946 | 10800 | 0.7588 | 0.7402 | 0.8415 | 0.6606 | 0.8979 |
|
112 |
+
| 0.6735 | 1.7260 | 11000 | 0.7588 | 0.7285 | 0.8786 | 0.6222 | 1.0666 |
|
113 |
+
| 0.8702 | 1.7574 | 11200 | 0.7576 | 0.7553 | 0.795 | 0.7195 | 0.9554 |
|
114 |
+
| 0.7435 | 1.7888 | 11400 | 0.7588 | 0.7497 | 0.8143 | 0.6946 | 1.0937 |
|
115 |
+
| 0.8796 | 1.8202 | 11600 | 0.7824 | 0.7768 | 0.8320 | 0.7285 | 0.9257 |
|
116 |
+
| 0.6257 | 1.8516 | 11800 | 0.7659 | 0.7588 | 0.8172 | 0.7081 | 0.9606 |
|
117 |
+
| 0.8589 | 1.8829 | 12000 | 0.7659 | 0.7484 | 0.8481 | 0.6697 | 0.9013 |
|
118 |
+
| 0.865 | 1.9143 | 12200 | 0.7612 | 0.7717 | 0.7673 | 0.7760 | 1.0734 |
|
119 |
+
| 0.8068 | 1.9457 | 12400 | 0.76 | 0.7431 | 0.8381 | 0.6674 | 0.9214 |
|
120 |
+
| 0.6212 | 1.9771 | 12600 | 0.7706 | 0.7535 | 0.8539 | 0.6742 | 1.0116 |
|
121 |
+
| 0.7657 | 2.0085 | 12800 | 0.7718 | 0.7532 | 0.8605 | 0.6697 | 0.9830 |
|
122 |
+
| 0.6631 | 2.0399 | 13000 | 0.7776 | 0.7810 | 0.8005 | 0.7624 | 1.0075 |
|
123 |
+
| 0.3003 | 2.0712 | 13200 | 0.7812 | 0.7748 | 0.8333 | 0.7240 | 1.1456 |
|
124 |
+
| 0.5982 | 2.1026 | 13400 | 0.7753 | 0.7633 | 0.8438 | 0.6968 | 1.0728 |
|
125 |
+
| 0.4828 | 2.1340 | 13600 | 0.7753 | 0.7718 | 0.8177 | 0.7308 | 1.0474 |
|
126 |
+
| 0.5463 | 2.1654 | 13800 | 0.7776 | 0.7726 | 0.8252 | 0.7262 | 1.0521 |
|
127 |
+
| 0.5429 | 2.1968 | 14000 | 0.7706 | 0.7590 | 0.8365 | 0.6946 | 1.0990 |
|
128 |
+
| 0.7112 | 2.2282 | 14200 | 0.7729 | 0.7578 | 0.8507 | 0.6833 | 1.1072 |
|
129 |
+
| 0.4816 | 2.2595 | 14400 | 0.7753 | 0.7685 | 0.8277 | 0.7172 | 1.1528 |
|
130 |
+
| 0.7882 | 2.2909 | 14600 | 0.7765 | 0.7722 | 0.8214 | 0.7285 | 0.9670 |
|
131 |
+
| 0.5265 | 2.3223 | 14800 | 0.7765 | 0.7694 | 0.8298 | 0.7172 | 1.0724 |
|
132 |
+
| 0.6116 | 2.3537 | 15000 | 0.7776 | 0.7742 | 0.8203 | 0.7330 | 1.0316 |
|
133 |
+
| 0.575 | 2.3851 | 15200 | 0.7741 | 0.7624 | 0.8415 | 0.6968 | 1.1125 |
|
134 |
+
| 0.5599 | 2.4164 | 15400 | 0.7765 | 0.7754 | 0.8119 | 0.7421 | 1.0327 |
|
135 |
+
| 0.5821 | 2.4478 | 15600 | 0.7776 | 0.7784 | 0.8078 | 0.7511 | 1.0655 |
|
136 |
+
| 0.4777 | 2.4792 | 15800 | 0.7835 | 0.7880 | 0.8028 | 0.7738 | 1.1187 |
|
137 |
+
| 0.432 | 2.5106 | 16000 | 0.7788 | 0.7740 | 0.8256 | 0.7285 | 1.1973 |
|
138 |
+
| 0.4385 | 2.5420 | 16200 | 0.7729 | 0.7737 | 0.8029 | 0.7466 | 1.2155 |
|
139 |
+
| 0.6103 | 2.5734 | 16400 | 0.78 | 0.7771 | 0.8212 | 0.7376 | 1.0527 |
|
140 |
+
| 0.4618 | 2.6047 | 16600 | 0.78 | 0.7787 | 0.8164 | 0.7443 | 1.1377 |
|
141 |
+
| 0.471 | 2.6361 | 16800 | 0.7788 | 0.7814 | 0.8038 | 0.7602 | 1.1468 |
|
142 |
+
| 0.6206 | 2.6675 | 17000 | 0.7765 | 0.7791 | 0.8014 | 0.7579 | 1.1048 |
|
143 |
+
| 0.5869 | 2.6989 | 17200 | 0.7776 | 0.7850 | 0.7895 | 0.7805 | 1.1343 |
|
144 |
+
| 0.5647 | 2.7303 | 17400 | 0.7859 | 0.7849 | 0.8218 | 0.7511 | 1.0843 |
|
145 |
+
| 0.5527 | 2.7617 | 17600 | 0.7847 | 0.7875 | 0.8091 | 0.7670 | 1.0834 |
|
146 |
+
| 0.8013 | 2.7930 | 17800 | 0.7894 | 0.7926 | 0.8124 | 0.7738 | 0.9898 |
|
147 |
+
| 0.5232 | 2.8244 | 18000 | 0.7859 | 0.7884 | 0.8110 | 0.7670 | 1.0052 |
|
148 |
+
| 0.617 | 2.8558 | 18200 | 0.7824 | 0.7821 | 0.8157 | 0.7511 | 1.0083 |
|
149 |
+
| 0.5093 | 2.8872 | 18400 | 0.7835 | 0.7810 | 0.8241 | 0.7421 | 1.0510 |
|
150 |
+
| 0.5099 | 2.9186 | 18600 | 0.78 | 0.7797 | 0.8133 | 0.7489 | 1.0758 |
|
151 |
+
| 0.6239 | 2.9499 | 18800 | 0.7812 | 0.7801 | 0.8168 | 0.7466 | 1.0726 |
|
152 |
+
| 0.6592 | 2.9813 | 19000 | 0.7788 | 0.7773 | 0.8159 | 0.7421 | 1.0731 |
|
153 |
+
| 0.1354 | 3.0127 | 19200 | 0.7765 | 0.7791 | 0.8014 | 0.7579 | 1.3351 |
|
154 |
+
| 0.3893 | 3.0755 | 19600 | 1.1745 | 0.7835 | 0.7879 | 0.7986 | 0.7933 |
|
155 |
+
| 0.4753 | 3.1382 | 20000 | 1.2362 | 0.7882 | 0.8541 | 0.7149 | 0.7783 |
|
156 |
+
| 0.3696 | 3.2010 | 20400 | 1.2638 | 0.7847 | 0.8166 | 0.7557 | 0.7850 |
|
157 |
+
| 0.4476 | 3.2638 | 20800 | 1.3415 | 0.7706 | 0.8580 | 0.6697 | 0.7522 |
|
158 |
+
| 0.4819 | 3.3265 | 21200 | 1.2743 | 0.7776 | 0.8235 | 0.7285 | 0.7731 |
|
159 |
+
| 0.4361 | 3.3893 | 21600 | 1.1387 | 0.7765 | 0.8198 | 0.7308 | 0.7727 |
|
160 |
+
| 0.3522 | 3.4521 | 22000 | 1.4823 | 0.7694 | 0.7808 | 0.7738 | 0.7773 |
|
161 |
+
| 0.4655 | 3.5148 | 22400 | 1.4060 | 0.7529 | 0.7283 | 0.8371 | 0.7789 |
|
162 |
+
| 0.438 | 3.5776 | 22800 | 1.2310 | 0.7659 | 0.8 | 0.7330 | 0.7651 |
|
163 |
+
| 0.5766 | 3.6404 | 23200 | 1.2406 | 0.7776 | 0.7843 | 0.7896 | 0.7869 |
|
164 |
+
| 0.5009 | 3.7031 | 23600 | 1.1521 | 0.7882 | 0.8005 | 0.7896 | 0.7950 |
|
165 |
+
| 0.4282 | 3.7659 | 24000 | 1.2111 | 0.7765 | 0.7825 | 0.7896 | 0.7860 |
|
166 |
+
| 0.3733 | 3.8287 | 24400 | 1.3162 | 0.7765 | 0.8043 | 0.7534 | 0.7780 |
|
167 |
+
| 0.5248 | 3.8914 | 24800 | 1.2680 | 0.7788 | 0.8038 | 0.7602 | 0.7814 |
|
168 |
+
| 0.4236 | 3.9542 | 25200 | 1.3282 | 0.7706 | 0.7775 | 0.7828 | 0.7802 |
|
169 |
+
| 0.5013 | 4.0169 | 25600 | 1.2586 | 0.7753 | 0.7967 | 0.7624 | 0.7792 |
|
170 |
+
| 0.2414 | 4.0797 | 26000 | 1.2782 | 0.7741 | 0.7976 | 0.7579 | 0.7773 |
|
171 |
+
| 0.1564 | 4.1425 | 26400 | 1.4687 | 0.78 | 0.8133 | 0.7489 | 0.7797 |
|
172 |
+
| 0.2116 | 4.2052 | 26800 | 1.3950 | 0.7694 | 0.7821 | 0.7715 | 0.7768 |
|
173 |
+
| 0.288 | 4.2680 | 27200 | 1.3878 | 0.7753 | 0.8039 | 0.7511 | 0.7766 |
|
174 |
+
| 0.1134 | 4.3308 | 27600 | 1.6678 | 0.7729 | 0.8074 | 0.7398 | 0.7721 |
|
175 |
+
| 0.1839 | 4.3935 | 28000 | 1.6949 | 0.7729 | 0.8120 | 0.7330 | 0.7705 |
|
176 |
+
| 0.1831 | 4.4563 | 28400 | 1.5547 | 0.78 | 0.8102 | 0.7534 | 0.7808 |
|
177 |
+
| 0.2417 | 4.5191 | 28800 | 1.4637 | 0.7847 | 0.8246 | 0.7443 | 0.7824 |
|
178 |
+
| 0.1922 | 4.5818 | 29200 | 1.5513 | 0.7824 | 0.8157 | 0.7511 | 0.7821 |
|
179 |
+
| 0.252 | 4.6446 | 29600 | 1.5873 | 0.78 | 0.8102 | 0.7534 | 0.7808 |
|
180 |
+
| 0.2172 | 4.7074 | 30000 | 1.5172 | 0.7882 | 0.8259 | 0.7511 | 0.7867 |
|
181 |
+
| 0.1177 | 4.7701 | 30400 | 1.5945 | 0.7824 | 0.8024 | 0.7715 | 0.7866 |
|
182 |
+
| 0.1921 | 4.8329 | 30800 | 1.6031 | 0.7812 | 0.8048 | 0.7647 | 0.7842 |
|
183 |
+
| 0.1833 | 4.8957 | 31200 | 1.5854 | 0.7765 | 0.7903 | 0.7760 | 0.7831 |
|
184 |
+
| 0.243 | 4.9584 | 31600 | 1.5959 | 0.7812 | 0.8048 | 0.7647 | 0.7842 |
|
185 |
|
186 |
|
187 |
### Framework versions
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3005286112
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6732526a4bc2c6a335a8038fd67fdda30608cd2fd3ab6aa439f74d3405f3312f
|
3 |
size 3005286112
|
runs/Oct02_19-53-09_iit-p/events.out.tfevents.1727879006.iit-p
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cdc2c612dc379d6ac74cc601fb671863e32ebf2f5a10cb1dd5b9a1993cc17ba3
|
3 |
+
size 27283
|