Update README.md
Browse files
README.md
CHANGED
@@ -13,6 +13,8 @@ tags:
|
|
13 |
|
14 |
palmer-004 a is merge of models targetting to get the performance of palmer-003 all they way to 32k context window. It was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [palmer-002-32k](https://huggingface.co/appvoid/palmer-002-32k) as a base.
|
15 |
|
|
|
|
|
16 |
The following models were included in the merge:
|
17 |
* [coven_tiny_1.1b_32k_orpo_alpha](https://huggingface.co/raidhon/coven_tiny_1.1b_32k_orpo_alpha)
|
18 |
* [palmer-003](https://huggingface.co/appvoid/palmer-003)
|
|
|
13 |
|
14 |
palmer-004 a is merge of models targetting to get the performance of palmer-003 all they way to 32k context window. It was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [palmer-002-32k](https://huggingface.co/appvoid/palmer-002-32k) as a base.
|
15 |
|
16 |
+
palmer-004 performs better than coven_tiny_1.1b_32k_orpo_alpha which is the current sota at open-llm-leaderboard, making this one the best overall 1b model on huggingface as of 06/01/2024.
|
17 |
+
|
18 |
The following models were included in the merge:
|
19 |
* [coven_tiny_1.1b_32k_orpo_alpha](https://huggingface.co/raidhon/coven_tiny_1.1b_32k_orpo_alpha)
|
20 |
* [palmer-003](https://huggingface.co/appvoid/palmer-003)
|