gemma2b-summarize-gemini1.5flash
Collection
9 items
•
Updated
This model is a fine-tuned version of google/gemma-2b on the llama-duo/synth_summarize_dataset_dedup dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.2288 | 0.9905 | 52 | 2.5091 |
1.0492 | 2.0 | 105 | 2.4525 |
1.0025 | 2.9905 | 157 | 2.4599 |
0.9597 | 4.0 | 210 | 2.4714 |
0.9394 | 4.9905 | 262 | 2.4863 |
0.912 | 6.0 | 315 | 2.5085 |
0.8901 | 6.9905 | 367 | 2.5172 |
0.8887 | 8.0 | 420 | 2.5294 |
0.8816 | 8.9905 | 472 | 2.5311 |
0.8797 | 9.9048 | 520 | 2.5322 |
Base model
google/gemma-2b