gemma2b-summarize-gpt4o
Collection
9 items
•
Updated
This model is a fine-tuned version of google/gemma-2b on the llama-duo/synth_summarize_dataset_dedup dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.1249 | 1.0 | 293 | 2.4641 |
1.0415 | 2.0 | 586 | 2.4514 |
0.9915 | 3.0 | 879 | 2.4750 |
0.9551 | 4.0 | 1172 | 2.5292 |
0.9287 | 5.0 | 1465 | 2.5925 |
0.8733 | 6.0 | 1758 | 2.6555 |
0.8577 | 7.0 | 2051 | 2.7316 |
0.8364 | 8.0 | 2344 | 2.7742 |
0.8311 | 9.0 | 2637 | 2.7971 |
0.8243 | 10.0 | 2930 | 2.7978 |
Base model
google/gemma-2b