Update README.md
Browse files
README.md
CHANGED
@@ -2,8 +2,30 @@
|
|
2 |
license: gemma
|
3 |
language:
|
4 |
- en
|
|
|
5 |
---
|
6 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
# Abliterated Gemma 2 2B
|
8 |
|
9 |
[Abliterated](https://huggingface.co/blog/mlabonne/abliteration) version of [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it).
|
|
|
2 |
license: gemma
|
3 |
language:
|
4 |
- en
|
5 |
+
base_model: IlyaGusev/gemma-2-2b-it-abliterated
|
6 |
---
|
7 |
+
# gemma-2-2b-it-abliterated-exl2
|
8 |
+
Model: [gemma-2-2b-it-abliterated](https://huggingface.co/IlyaGusev/gemma-2-2b-it-abliterated)
|
9 |
+
Made by: [IlyaGusev](https://huggingface.co/IlyaGusev)
|
10 |
+
## Quants
|
11 |
+
[4bpw h6 (main)](https://huggingface.co/cgus/gemma-2-2b-it-abliterated-exl2/tree/main)
|
12 |
+
[4.5bpw h6](https://huggingface.co/cgus/gemma-2-2b-it-abliterated-exl2/tree/4.5bpw-h6)
|
13 |
+
[5bpw h6](https://huggingface.co/cgus/gemma-2-2b-it-abliterated-exl2/tree/5bpw-h6)
|
14 |
+
[6bpw h6](https://huggingface.co/cgus/gemma-2-2b-it-abliterated-exl2/tree/6bpw-h6)
|
15 |
+
[8bpw h8](https://huggingface.co/cgus/gemma-2-2b-it-abliterated-exl2/tree/8bpw-h8)
|
16 |
+
|
17 |
+
## Quantization notes
|
18 |
+
Made with Exllamav2 0.1.8 with the default dataset.
|
19 |
+
I hoped that this model could fit some 2GB GPUs like my old Quadro T400 that kinda what RTX2030 could be if it existed.
|
20 |
+
But even 4bpw version is bigger than 2GB, I didn't expect this since some 2-3B models had smaller model files.
|
21 |
+
It doesn't make much sense to make exl2 quants for such small models but I needed some small but good model for making prompts for Flux.
|
22 |
+
Since my system can't handle loading a big model and Flux at the same time.
|
23 |
+
I couldn't load this model with Exllamav2 0.1.7 so using 0.1.8 or newer might be necessary.
|
24 |
+
|
25 |
+
## How to use
|
26 |
+
This model version can be loaded with apps that have Exllamav2 loader: Text-Generation-WebUI, TabbyAPI, possibly KoboldAI, etc.
|
27 |
+
|
28 |
+
# Original model card
|
29 |
# Abliterated Gemma 2 2B
|
30 |
|
31 |
[Abliterated](https://huggingface.co/blog/mlabonne/abliteration) version of [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it).
|