gemma2b-summarize-gemini1.5flash
Collection
9 items
•
Updated
This model is a fine-tuned version of google/gemma-2b on the llama-duo/synth_summarize_dataset_dedup dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.0978 | 1.0 | 104 | 2.4831 |
0.9985 | 2.0 | 208 | 2.4666 |
0.9543 | 3.0 | 312 | 2.4561 |
0.92 | 4.0 | 416 | 2.4799 |
0.9016 | 5.0 | 520 | 2.4990 |
0.8871 | 6.0 | 624 | 2.5250 |
0.8635 | 7.0 | 728 | 2.5363 |
0.8535 | 8.0 | 832 | 2.5546 |
0.845 | 9.0 | 936 | 2.5566 |
0.853 | 10.0 | 1040 | 2.5573 |
Base model
google/gemma-2b