TFG
Collection
Datasets and models leveraged and developed during my final degree work (TFG). Info and code can be found at https://github.com/enriquesaou/tfg-lm-qa
•
18 items
•
Updated
•
2
This model is a fine-tuned version of google-t5/t5-base on an MRQA sample. It achieves the following results on the evaluation set:
T5 base but trained at FP16 in the MRQA sample dataset. This model is the checkpoint at 3000 steps (3rd epoch), because there were instabilities during the late epochs.
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
0.7978 | 0.9996 | 833 | 0.6668 |
0.6516 | 1.9992 | 1666 | 0.6532 |
0.6275 | 3.0 | 2500 | 0.6532 |
(0.6443) | (3.9996) | (3333) | (0.6533) |
(2.0743) | (4.998) | (4165 | (nan) |
Note that this model is the checkpoint at 3000 steps (3rd epoch), because there were instabilities during the late epochs.
Base model
google-t5/t5-base