File size: 2,126 Bytes
72a337b
 
 
49c6ae6
72a337b
 
 
474ae3b
 
72a337b
 
474ae3b
72a337b
 
384cf0e
72a337b
 
 
384cf0e
72a337b
a6692eb
 
 
5c98e9e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
384cf0e
a6692eb
5c98e9e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
---
library_name: llamacpp
model_name: Vikhr-Gemma-2B-instruct
base_model: Vikhrmodels/Vikhr-Gemma-2B-instruct
language:
- ru
license: apache-2.0
tags:
- instruct
datasets:
- Vikhrmodels/GrandMaster-PRO-MAX
pipeline_tag: text-generation
---

# 💨 Vikhr-Gemma-2B-instruct

Мощная инструктивная модель на основе Gemma 2 2B, обученная на русскоязычном датасете GrandMaster-PRO-MAX.

- [HF model](https://huggingface.co/Vikhrmodels/Vikhr-Gemma-2B-instruct)

## Perplexity (ниже - лучше)

### Veles
| Model   | Perplexity         |
|:--------|:-------------------|
| Q4_K    | 4.7254 +/- 0.03867 |
| Q4_0    | 4.8067 +/- 0.03922 |
| Q8_0    | 4.6042 +/- 0.03751 |
| Q4_1    | 4.7798 +/- 0.03933 |
| F32     | 4.6013 +/- 0.03749 |
| Q6_K    | 4.6244 +/- 0.03760 |
| BF16    | 4.6015 +/- 0.03749 |
| Q2_K    | 5.6819 +/- 0.04737 |
| Q5_0    | 4.6876 +/- 0.03855 |
| Q5_K    | 4.6428 +/- 0.03789 |
| Q3_K_S  | 5.1485 +/- 0.04257 |
| Q2_K_S  | 6.3124 +/- 0.05359 |
| F16     | 4.6013 +/- 0.03749 |
| Q4_K_M  | 4.7254 +/- 0.03867 |
| Q5_K_M  | 4.6428 +/- 0.03789 |
| Q5_1    | 4.6518 +/- 0.03794 |
| Q4_K_S  | 4.7631 +/- 0.03916 |
| Q5_K_S  | 4.6509 +/- 0.03803 |
| Q3_K    | 4.8339 +/- 0.03965 |
| Q3_K_M  | 4.8339 +/- 0.03965 |
| Q3_K_L  | 4.7981 +/- 0.03934 |

### Wikitext-2
| Model   | Perplexity          |
|:--------|:--------------------|
| Q4_K    | 10.4374 +/- 0.07339 |
| Q4_0    | 10.6480 +/- 0.07452 |
| Q8_0    | 10.1209 +/- 0.07105 |
| Q4_1    | 10.5574 +/- 0.07476 |
| F32     | 10.1191 +/- 0.07099 |
| Q6_K    | 10.1503 +/- 0.07117 |
| BF16    | 10.1189 +/- 0.07098 |
| Q2_K    | 12.8851 +/- 0.09332 |
| Q5_0    | 10.2551 +/- 0.07251 |
| Q5_K    | 10.1975 +/- 0.07184 |
| Q3_K_S  | 11.6028 +/- 0.08333 |
| Q2_K_S  | 14.7951 +/- 0.10960 |
| F16     | 10.1191 +/- 0.07099 |
| Q4_K_M  | 10.4374 +/- 0.07339 |
| Q5_K_M  | 10.1975 +/- 0.07184 |
| Q5_1    | 10.2348 +/- 0.07208 |
| Q4_K_S  | 10.4924 +/- 0.07386 |
| Q5_K_S  | 10.2098 +/- 0.07198 |
| Q3_K    | 10.7416 +/- 0.07606 |
| Q3_K_M  | 10.7416 +/- 0.07606 |
| Q3_K_L  | 10.6242 +/- 0.07506 |