File size: 473 Bytes
6311882
 
f3b764c
392acc9
 
 
 
 
6311882
 
 
392acc9
6311882
 
392acc9
 
6311882
392acc9
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
---
library_name: transformers
pipeline_tag: text-generation
license: apache-2.0
datasets:
- euclaise/reddit-instruct
language:
- en
---

# Model Card for Model ID
This is the fine-tuned model of Mistral-7B-v0.1  (https://huggingface.co/mistralai/Mistral-7B-v0.1) with reddit instruct dataset


Model Architecture
Mistral-7B-v0.1 is a transformer model, with the following architecture choices:

Grouped-Query Attention
Sliding-Window Attention
Byte-fallback BPE tokenizer