library_name: transformers | |
pipeline_tag: text-generation | |
license: apache-2.0 | |
datasets: | |
- euclaise/reddit-instruct | |
language: | |
- en | |
# Model Card for Model ID | |
This is the fine-tuned model of Mistral-7B-v0.1 (https://huggingface.co/mistralai/Mistral-7B-v0.1) with reddit instruct dataset | |
Model Architecture | |
Mistral-7B-v0.1 is a transformer model, with the following architecture choices: | |
Grouped-Query Attention | |
Sliding-Window Attention | |
Byte-fallback BPE tokenizer | |