license: apache-2.0 | |
datasets: | |
- HuggingFaceH4/ultrachat_200k | |
language: | |
- en | |
# Introduction | |
This model vistagi/Mixtral-8x7b-v0.1-sft is trained with Ultrachat-200K dataset through supervised finetuning using Mixtral-8x7b-v0.1 as the baseline model. | |
The training is done with bfloat16 precision using LoRA. | |
## Details | |
Used Librarys | |
- torch | |
- deepspeed | |
- pytorch lightning | |
- transformers | |
- peft | |