gemma2b-summarize-claude3sonnet
Collection
9 items
•
Updated
This model is a fine-tuned version of google/gemma-2b on the llama-duo/synth_summarize_dataset_dedup dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
0.9714 | 0.9994 | 808 | 2.4535 |
0.8916 | 2.0 | 1617 | 2.4785 |
0.8752 | 2.9994 | 2425 | 2.5144 |
0.8424 | 4.0 | 3234 | 2.5590 |
0.8173 | 4.9994 | 4042 | 2.6021 |
0.7949 | 6.0 | 4851 | 2.6446 |
0.7732 | 6.9994 | 5659 | 2.6786 |
0.7605 | 8.0 | 6468 | 2.6913 |
0.7532 | 8.9994 | 7276 | 2.6995 |
0.7647 | 9.9938 | 8080 | 2.6999 |
Base model
google/gemma-2b