gemma2b-summarize-gemini1.5flash
Collection
9 items
•
Updated
This model is a fine-tuned version of google/gemma-2b on the llama-duo/synth_summarize_dataset_dedup dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.0246 | 0.9976 | 207 | 2.4550 |
0.9556 | 2.0 | 415 | 2.4530 |
0.9114 | 2.9976 | 622 | 2.4641 |
0.8927 | 4.0 | 830 | 2.4882 |
0.8752 | 4.9976 | 1037 | 2.5081 |
0.8602 | 6.0 | 1245 | 2.5277 |
0.8464 | 6.9976 | 1452 | 2.5513 |
0.8353 | 8.0 | 1660 | 2.5615 |
0.8267 | 8.9976 | 1867 | 2.5674 |
0.8289 | 9.9976 | 2070 | 2.5681 |
Base model
google/gemma-2b