gemma2b-summarize-claude3sonnet
Collection
9 items
•
Updated
This model is a fine-tuned version of google/gemma-2b on the llama-duo/synth_summarize_dataset_dedup dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.9711 | 0.9804 | 25 | 2.6222 |
1.3473 | 2.0 | 51 | 2.5571 |
1.1845 | 2.9804 | 76 | 2.5050 |
1.1242 | 4.0 | 102 | 2.5089 |
1.0937 | 4.9804 | 127 | 2.5239 |
1.0632 | 6.0 | 153 | 2.5275 |
1.0377 | 6.9804 | 178 | 2.5316 |
1.0278 | 8.0 | 204 | 2.5399 |
1.0176 | 8.9804 | 229 | 2.5422 |
1.007 | 9.8039 | 250 | 2.5421 |
Base model
google/gemma-2b