gemma2b-summarize-claude3sonnet
Collection
9 items
•
Updated
This model is a fine-tuned version of google/gemma-2b on the llama-duo/synth_summarize_dataset_dedup dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.187 | 0.9951 | 101 | 2.4971 |
1.0678 | 2.0 | 203 | 2.4806 |
1.0113 | 2.9951 | 304 | 2.4777 |
0.9692 | 4.0 | 406 | 2.5005 |
0.9448 | 4.9951 | 507 | 2.5205 |
0.906 | 6.0 | 609 | 2.5325 |
0.8867 | 6.9951 | 710 | 2.5459 |
0.8783 | 8.0 | 812 | 2.5595 |
0.8501 | 8.9951 | 913 | 2.5624 |
0.8569 | 9.9507 | 1010 | 2.5627 |
Base model
google/gemma-2b