SnakyMcSnekFace
commited on
Commit
•
8456d75
1
Parent(s):
ccca17f
New model version trained with 4096 context length
Browse files- Psyfighter2-13B-vore.Q4_K_M.gguf +2 -2
- README.md +33 -26
Psyfighter2-13B-vore.Q4_K_M.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:917bc2644119397263ae0c3ead67a549c5fff31729c82928c99f75ba42370c4e
|
3 |
+
size 7865956512
|
README.md
CHANGED
@@ -7,12 +7,8 @@ inference: false
|
|
7 |
tags:
|
8 |
- storywriting
|
9 |
- finetuned
|
10 |
-
- roleplay
|
11 |
-
- vore
|
12 |
- not-for-all-audiences
|
13 |
- gguf
|
14 |
-
- nsfw
|
15 |
-
- uncensored
|
16 |
base_model: SnakyMcSnekFace/Psyfighter2-13B-vore
|
17 |
model_type: llama
|
18 |
prompt_template: >
|
@@ -32,7 +28,7 @@ prompt_template: >
|
|
32 |
|
33 |
This is a quantized version of [SnakyMcSnekFace/Psyfighter2-13B-vore](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore) model.
|
34 |
|
35 |
-
This model is a version of [KoboldAI/LLaMA2-13B-Psyfighter2](https://huggingface.co/KoboldAI/LLaMA2-13B-Psyfighter2) finetuned to better understand vore context. The primary purpose of this model is to be a storywriting assistant,
|
36 |
|
37 |
The Adventure Mode is still work in progress, and will be added later.
|
38 |
|
@@ -60,17 +56,24 @@ The easiest way to try out the model is [Koboldcpp Colab Notebook](https://colab
|
|
60 |
- Paste the model URL into the field: `https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q4_K_M.gguf`
|
61 |
- Start the notebook, wait for the URL to CloudFlare tunnel to appear at the bottom and click it
|
62 |
- Use the model as a writing assistant
|
63 |
-
- You can try an adventure from [https://aetherroom.club/](https://aetherroom.club/), but keep in mind that the model will not let you take turn unless you stop it. Adventure mode is work-in-progress.
|
64 |
|
65 |
-
###
|
66 |
|
67 |
-
Another convenient way to use the model is [
|
68 |
|
69 |
-
Download
|
70 |
|
71 |
-
|
72 |
|
73 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
74 |
|
75 |
## Bias, Risks, and Limitations
|
76 |
|
@@ -78,27 +81,31 @@ By design, this model has a strong vorny bias. It's not intended for use by anyo
|
|
78 |
|
79 |
## Training Details
|
80 |
|
81 |
-
This model was fine-tuned on free-form text comprised of stories focused around the vore theme using
|
82 |
|
83 |
### Training Procedure
|
84 |
|
85 |
-
The model was fine-tuned
|
86 |
|
87 |
|
88 |
-
LoRa adapter configuration
|
89 |
|
90 |
-
- Rank:
|
91 |
-
- Alpha:
|
92 |
-
- Dropout rate: 0.
|
93 |
-
- Target weights:
|
|
|
94 |
|
95 |
-
Training parameters
|
96 |
|
97 |
-
-
|
98 |
-
- Samples per epoch:
|
99 |
- Number of epochs: 2
|
100 |
-
-
|
101 |
-
-
|
|
|
|
|
|
|
102 |
|
103 |
#### Preprocessing
|
104 |
|
@@ -106,7 +113,7 @@ The stories in dataset were pre-processed as follows:
|
|
106 |
|
107 |
- titles, foreword, tags, and anything not comprising the text of the story was removed
|
108 |
- non-ascii characters and character sequences serving as chapter separators were removed
|
109 |
-
- any story mentioning underage personas was
|
110 |
- names of private characters were replaced with randomized names across the dataset
|
111 |
|
112 |
## Environmental Impact
|
@@ -114,7 +121,7 @@ The stories in dataset were pre-processed as follows:
|
|
114 |
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
115 |
|
116 |
- **Hardware Type:** NVIDIA GeForce RTX 4060 Ti
|
117 |
-
- **Hours used:**
|
118 |
- **Cloud Provider:** N/A
|
119 |
- **Compute Region:** US-East
|
120 |
-
- **Carbon Emitted:**
|
|
|
7 |
tags:
|
8 |
- storywriting
|
9 |
- finetuned
|
|
|
|
|
10 |
- not-for-all-audiences
|
11 |
- gguf
|
|
|
|
|
12 |
base_model: SnakyMcSnekFace/Psyfighter2-13B-vore
|
13 |
model_type: llama
|
14 |
prompt_template: >
|
|
|
28 |
|
29 |
This is a quantized version of [SnakyMcSnekFace/Psyfighter2-13B-vore](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore) model.
|
30 |
|
31 |
+
This model is a version of [KoboldAI/LLaMA2-13B-Psyfighter2](https://huggingface.co/KoboldAI/LLaMA2-13B-Psyfighter2) finetuned to better understand vore context. The primary purpose of this model is to be a storywriting assistant, a conversational model in a chat, and an interactive choose-your-own-adventure text game.
|
32 |
|
33 |
The Adventure Mode is still work in progress, and will be added later.
|
34 |
|
|
|
56 |
- Paste the model URL into the field: `https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q4_K_M.gguf`
|
57 |
- Start the notebook, wait for the URL to CloudFlare tunnel to appear at the bottom and click it
|
58 |
- Use the model as a writing assistant
|
59 |
+
- You can try an adventure from [https://aetherroom.club/](https://aetherroom.club/), but keep in mind that the model will not let you take turn unless you stop it. Adventure mode is still work-in-progress, but it's getting there.
|
60 |
|
61 |
+
### Backyard AI
|
62 |
|
63 |
+
Another convenient way to use the model is [Backyard AI](https://backyard.ai/) application, which allows you to run the model locally on your computer. You'll need a graphics card with at least 8GB VRAM to use the model comfortably.
|
64 |
|
65 |
+
#### Download directly from HuggingFace (beta)
|
66 |
|
67 |
+
In the left panel, click `Manage Models`, then select `Hugging face models`. Paste `https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF` into the text field and press `Fetch Models`. Click `Download` button to the next to the model format. Once the model is downloaded, you can select it in your character card or set it as a default model.
|
68 |
|
69 |
+
#### Download manually
|
70 |
+
|
71 |
+
Download the [Psyfighter2-13B-vore.Q4_K_M.gguf](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q4_K_M.gguf) file into `%appdata%\faraday\models` folder on your computer. The model should appear in `Manage Models` menu under `Downloaded Models`. You can then select it in your character card or set it as a default model.
|
72 |
+
|
73 |
+
### Model updates
|
74 |
+
|
75 |
+
- 04/13/2024 - uploaded the first version of the model
|
76 |
+
- 05/25/2024 - updated training process, making the model more coherent and improving the writing quality
|
77 |
|
78 |
## Bias, Risks, and Limitations
|
79 |
|
|
|
81 |
|
82 |
## Training Details
|
83 |
|
84 |
+
This model was fine-tuned on free-form text comprised of stories focused around the vore theme using [rank-stabilized](https://arxiv.org/abs/2312.03732) [QLoRA adapter](https://arxiv.org/abs/2305.14314) [QLoRA method](https://arxiv.org/abs/2305.14314). The resulting adapter was merged into the FP16 precision base model. The quantized version of the model was prepared using [llama.cpp](https://github.com/ggerganov/llama.cpp).
|
85 |
|
86 |
### Training Procedure
|
87 |
|
88 |
+
The model was fine-tuned with a [rank-stabilized](https://arxiv.org/abs/2312.03732) [QLoRA adapter](https://arxiv.org/abs/2305.14314) on NVIDIA GeForce RTX 4060 Ti over the span of ~24 hours. Training was performed using [Unsloth AI](https://github.com/unslothai/unsloth) library on `Ubuntu 22.04.4 LTS` with `CUDA 12.1` and `Pytorch 2.3.0`.
|
89 |
|
90 |
|
91 |
+
#### LoRa adapter configuration
|
92 |
|
93 |
+
- Rank: 128
|
94 |
+
- Alpha: 16
|
95 |
+
- Dropout rate: 0.1
|
96 |
+
- Target weights: `["q_proj", "k_proj", "o_proj", "gate_proj", "up_proj"]`,
|
97 |
+
- `use_rslora=True`
|
98 |
|
99 |
+
#### Training parameters
|
100 |
|
101 |
+
- Max. sequence length: 4096 tokens
|
102 |
+
- Samples per epoch: 3783
|
103 |
- Number of epochs: 2
|
104 |
+
- Learning rate: 1e-4
|
105 |
+
- Warmup: 64 steps
|
106 |
+
- LR Schedule: linear
|
107 |
+
- Batch size: 1
|
108 |
+
- Gradient accumulation steps: 1
|
109 |
|
110 |
#### Preprocessing
|
111 |
|
|
|
113 |
|
114 |
- titles, foreword, tags, and anything not comprising the text of the story was removed
|
115 |
- non-ascii characters and character sequences serving as chapter separators were removed
|
116 |
+
- any story mentioning underage personas in any context was removed from the dataset
|
117 |
- names of private characters were replaced with randomized names across the dataset
|
118 |
|
119 |
## Environmental Impact
|
|
|
121 |
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
122 |
|
123 |
- **Hardware Type:** NVIDIA GeForce RTX 4060 Ti
|
124 |
+
- **Hours used:** 24
|
125 |
- **Cloud Provider:** N/A
|
126 |
- **Compute Region:** US-East
|
127 |
+
- **Carbon Emitted:** 0.83 kg CO2 eq.
|