Mistral-7B-0.3
Collection
36 items
•
Updated
This model is a fine-tuned version of mistralai/Mistral-7B-Instruct-v0.3 on the GaetanMichelet/chat-60_ft_task-3 dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
3.5059 | 0.8696 | 5 | 2.7689 |
1.5063 | 1.9130 | 11 | 0.8907 |
0.5013 | 2.9565 | 17 | 0.4844 |
0.2965 | 4.0 | 23 | 0.3944 |
0.189 | 4.8696 | 28 | 0.3692 |
0.0884 | 5.9130 | 34 | 0.4647 |
0.0487 | 6.9565 | 40 | 0.6312 |
0.0415 | 8.0 | 46 | 0.7006 |
0.0096 | 8.8696 | 51 | 0.6515 |
0.0226 | 9.9130 | 57 | 0.8801 |
0.019 | 10.9565 | 63 | 0.6534 |
0.0075 | 12.0 | 69 | 0.6282 |
Base model
mistralai/Mistral-7B-v0.3