metadata
base_model: mistralai/Mistral-Nemo-Base-2407
license: apache-2.0
datasets:
- BeaverAI/Nemo-Inst-Tune-ds
language:
- en
library_name: transformers
5.0bpw h6 exl2 quant of : (https://huggingface.co/BeaverAI/mistral-dory-12b)
Dory 12b
redone instruct finetune of mistral nemo 12b. not (E)RP-focused, leave that to drummer.
thanks to twisted for the compute :3
Prompting
alpaca-like:
### System:
[Optional system prompt]
### Instruction:
[Query]
### Response:
[Response]<EOT>
### Instruction:
[...]
Training details
Rank 64 QDoRA, trained on the following data mix:
- All of kalomaze/Opus_Instruct_3k
- All conversations with a reward model rating above 5 in Magpie-Align/Magpie-Gemma2-Pro-Preview-Filtered
- 50k of Gryphe/Sonnet3.5-SlimOrcaDedupCleaned
- All stories above 4.7 rating and published before 2020 in Fizzarolli/FallingThroughTheSkies-592k-Filtered-Filtered