--- license: apache-2.0 base_model: google/flan-t5-large tags: - generated_from_trainer metrics: - rouge model-index: - name: flan-t5-large-finetuned-scope-summarization results: [] --- # flan-t5-large-finetuned-scope-summarization This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co./google/flan-t5-large) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.1195 - Rouge1: 24.038 - Rouge2: 21.4448 - Rougel: 23.6448 - Rougelsum: 23.7376 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5.6e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 20 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:| | 0.3649 | 1.0 | 158 | 0.2625 | 19.5356 | 12.4535 | 16.8939 | 17.0876 | | 0.2674 | 2.0 | 316 | 0.2422 | 19.7836 | 12.4864 | 16.9298 | 16.9928 | | 0.2516 | 3.0 | 474 | 0.2271 | 20.4584 | 13.593 | 17.9404 | 18.0498 | | 0.2407 | 4.0 | 632 | 0.2178 | 20.2729 | 13.6717 | 17.5 | 17.6375 | | 0.2304 | 5.0 | 790 | 0.2087 | 20.3933 | 14.4275 | 17.9315 | 18.0607 | | 0.2213 | 6.0 | 948 | 0.1969 | 21.4659 | 16.1078 | 19.4775 | 19.5604 | | 0.2134 | 7.0 | 1106 | 0.1863 | 23.3097 | 19.0603 | 21.9919 | 22.1651 | | 0.2069 | 8.0 | 1264 | 0.1803 | 22.5866 | 17.3665 | 20.4585 | 20.4009 | | 0.2 | 9.0 | 1422 | 0.1695 | 23.7295 | 19.7783 | 22.4861 | 22.5794 | | 0.1942 | 10.0 | 1580 | 0.1632 | 21.9543 | 16.572 | 19.539 | 19.5863 | | 0.1883 | 11.0 | 1738 | 0.1570 | 22.5164 | 18.8651 | 21.4345 | 21.6252 | | 0.1829 | 12.0 | 1896 | 0.1495 | 23.7871 | 20.6331 | 23.2495 | 23.4011 | | 0.178 | 13.0 | 2054 | 0.1425 | 23.789 | 21.1006 | 23.2292 | 23.4225 | | 0.1738 | 14.0 | 2212 | 0.1386 | 23.8972 | 21.2393 | 23.4578 | 23.5827 | | 0.1689 | 15.0 | 2370 | 0.1331 | 23.801 | 21.2013 | 23.3414 | 23.4499 | | 0.1654 | 16.0 | 2528 | 0.1286 | 24.1973 | 21.5666 | 23.7563 | 23.9153 | | 0.1629 | 17.0 | 2686 | 0.1257 | 23.8243 | 21.2713 | 23.4043 | 23.4941 | | 0.16 | 18.0 | 2844 | 0.1229 | 23.9496 | 21.3888 | 23.4687 | 23.6047 | | 0.1578 | 19.0 | 3002 | 0.1208 | 24.009 | 21.4585 | 23.5252 | 23.646 | | 0.156 | 20.0 | 3160 | 0.1195 | 24.038 | 21.4448 | 23.6448 | 23.7376 | ### Framework versions - Transformers 4.40.1 - Pytorch 2.3.0+cu121 - Datasets 2.19.0 - Tokenizers 0.19.1