File size: 1,979 Bytes
4836fa5
 
 
 
 
 
 
 
 
 
 
ae92e93
4836fa5
 
e5ffce7
 
 
 
 
4836fa5
 
14ea297
4836fa5
 
06ff562
6a0396f
 
 
 
 
 
4836fa5
 
b9ab4f2
4836fa5
61e062c
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
---
license: apache-2.0
language:
- en
- es
- fr
tags:
- merge
---
![palmer-003 logo](https://huggingface.co/appvoid/palmer-002.5/resolve/main/003.png)

Creative writing has never been so accesible, palmer goes beyond what it was thought about small language models. This model is a "MErging of Experts" (MEoE) using an internal model `palmer-003-2401` as base, biased as an assistant, using dpo technique, without using any prompts—as a result of these efforts—palmer is better than most 1b language models on most benchmarks, despite being sometimes 40% smaller than its counterparts.

```
  Model      MMLU     ARC-C    OBQA   HellaSwag  PIQA  Winogrande Average  Params
tinyllama | 0.2577 | 0.3029 | 0.3600 | 0.5935 | 0.7329 | 0.5959 | 0.4738 | 1.1B  |
zyte 	  | 0.2397 | 0.3353 | 0.3700 | 0.6086 | 0.7541 | 0.5998 | 0.4845 | 1.1B  |
palmer    | 0.2523 | 0.3439 | 0.3740 | 0.6208 | 0.7524 | 0.6590 | 0.5004 | 1.1B  |
qwen      | 0.4536 | 0.3490 | 0.3320 | 0.5876 | 0.7307 | 0.5896 | 0.5070 | 1.8B  |
```

This work constitutes, given its compactness, an advancement towards SMLs, easily empowering edge devices such as mobile phones, raspberry pis and automated software/robots. Aditionally, palmer-003 follows the same philosophy as palmer-002.5 to become a more powerful model with more data instead of less.

```
prompt: Never give up.
output: Keep pushing forward. Remember, you are not alone.
You have the strength and the support of your family and friends.
As you continue on your journey, remember to take care of yourself.
Eat well, exercise regularly, and get enough sleep.
Stay positive and focused on your goals.
Remember, you are capable of achieving anything you set your mind to.
```

You can support me [through kofi](https://ko-fi.com/appvoid).

### Important
Keep in mind that if you quantize the fp16 gguf model to let's say q8 or q6 there is gonna be a extremely huge performance drop. So, if you want the most of this model, use it as it is, in fp16.