SnakyMcSnekFace commited on
Commit
8456d75
1 Parent(s): ccca17f

New model version trained with 4096 context length

Browse files
Files changed (2) hide show
  1. Psyfighter2-13B-vore.Q4_K_M.gguf +2 -2
  2. README.md +33 -26
Psyfighter2-13B-vore.Q4_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a1d9536c7245271758e7e846ba4e0d6b3061d295e0ca2777c04318b719ab8de
3
- size 7865956352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:917bc2644119397263ae0c3ead67a549c5fff31729c82928c99f75ba42370c4e
3
+ size 7865956512
README.md CHANGED
@@ -7,12 +7,8 @@ inference: false
7
  tags:
8
  - storywriting
9
  - finetuned
10
- - roleplay
11
- - vore
12
  - not-for-all-audiences
13
  - gguf
14
- - nsfw
15
- - uncensored
16
  base_model: SnakyMcSnekFace/Psyfighter2-13B-vore
17
  model_type: llama
18
  prompt_template: >
@@ -32,7 +28,7 @@ prompt_template: >
32
 
33
  This is a quantized version of [SnakyMcSnekFace/Psyfighter2-13B-vore](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore) model.
34
 
35
- This model is a version of [KoboldAI/LLaMA2-13B-Psyfighter2](https://huggingface.co/KoboldAI/LLaMA2-13B-Psyfighter2) finetuned to better understand vore context. The primary purpose of this model is to be a storywriting assistant, as well as a conversational model in a chat.
36
 
37
  The Adventure Mode is still work in progress, and will be added later.
38
 
@@ -60,17 +56,24 @@ The easiest way to try out the model is [Koboldcpp Colab Notebook](https://colab
60
  - Paste the model URL into the field: `https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q4_K_M.gguf`
61
  - Start the notebook, wait for the URL to CloudFlare tunnel to appear at the bottom and click it
62
  - Use the model as a writing assistant
63
- - You can try an adventure from [https://aetherroom.club/](https://aetherroom.club/), but keep in mind that the model will not let you take turn unless you stop it. Adventure mode is work-in-progress.
64
 
65
- ### Faraday
66
 
67
- Another convenient way to use the model is [Faraday.dev](https://faraday.dev/) application, which allows you to run the model locally on your computer. You'll need a graphics card with at least 8GB VRAM to use `Q4_K_M` version comfortably, and 16GB VRAM for `Q8_0`. (`Q4_K_M` version is smaller and faster, `Q8_0` is slower but more coherent.)
68
 
69
- Download the [Psyfighter2-13B-vore.Q4_K_M.gguf](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q4_K_M.gguf) or [Psyfighter2-13B-vore.Q8_0.gguf](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q8_0.gguf) file into `%appdata%\faraday\models` folder on your computer. The model should appear in `Manage Models` menu under `Downloaded Models`. You can then select it in your character card or set it as a default model.
70
 
71
- ### Others
72
 
73
- TBD
 
 
 
 
 
 
 
74
 
75
  ## Bias, Risks, and Limitations
76
 
@@ -78,27 +81,31 @@ By design, this model has a strong vorny bias. It's not intended for use by anyo
78
 
79
  ## Training Details
80
 
81
- This model was fine-tuned on free-form text comprised of stories focused around the vore theme using the [QLoRA method](https://arxiv.org/abs/2305.14314). The resulting adapter was merged into the base model. The quantized version of the model was prepared using [llama.cpp](https://github.com/ggerganov/llama.cpp).
82
 
83
  ### Training Procedure
84
 
85
- The model was fine-tuned using the [QLoRA method](https://arxiv.org/abs/2305.14314) on NVIDIA GeForce RTX 4060 Ti over the span of ~7 days. Training was performed using [text-generation-webui by oobabooga](https://github.com/oobabooga/text-generation-webui) with [Training PRO plug-in by FartyPants](https://github.com/FartyPants/Training_PRO).
86
 
87
 
88
- LoRa adapter configuration:
89
 
90
- - Rank: 512
91
- - Alpha: 1024
92
- - Dropout rate: 0.05
93
- - Target weights: v_prog, q_proj
 
94
 
95
- Training parameters:
96
 
97
- - Sample size: 768 tokens
98
- - Samples per epoch: 47420
99
  - Number of epochs: 2
100
- - First epoch: Learning rate = 3e-4, 1000 steps warmup, cosine schedule
101
- - Second epoch: Learning rate = 1e-4, 256 steps warmup, inverse sqrt schedule
 
 
 
102
 
103
  #### Preprocessing
104
 
@@ -106,7 +113,7 @@ The stories in dataset were pre-processed as follows:
106
 
107
  - titles, foreword, tags, and anything not comprising the text of the story was removed
108
  - non-ascii characters and character sequences serving as chapter separators were removed
109
- - any story mentioning underage personas was taken out of the dataset
110
  - names of private characters were replaced with randomized names across the dataset
111
 
112
  ## Environmental Impact
@@ -114,7 +121,7 @@ The stories in dataset were pre-processed as follows:
114
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
115
 
116
  - **Hardware Type:** NVIDIA GeForce RTX 4060 Ti
117
- - **Hours used:** 168
118
  - **Cloud Provider:** N/A
119
  - **Compute Region:** US-East
120
- - **Carbon Emitted:** 5.8 kg CO2 eq.
 
7
  tags:
8
  - storywriting
9
  - finetuned
 
 
10
  - not-for-all-audiences
11
  - gguf
 
 
12
  base_model: SnakyMcSnekFace/Psyfighter2-13B-vore
13
  model_type: llama
14
  prompt_template: >
 
28
 
29
  This is a quantized version of [SnakyMcSnekFace/Psyfighter2-13B-vore](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore) model.
30
 
31
+ This model is a version of [KoboldAI/LLaMA2-13B-Psyfighter2](https://huggingface.co/KoboldAI/LLaMA2-13B-Psyfighter2) finetuned to better understand vore context. The primary purpose of this model is to be a storywriting assistant, a conversational model in a chat, and an interactive choose-your-own-adventure text game.
32
 
33
  The Adventure Mode is still work in progress, and will be added later.
34
 
 
56
  - Paste the model URL into the field: `https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q4_K_M.gguf`
57
  - Start the notebook, wait for the URL to CloudFlare tunnel to appear at the bottom and click it
58
  - Use the model as a writing assistant
59
+ - You can try an adventure from [https://aetherroom.club/](https://aetherroom.club/), but keep in mind that the model will not let you take turn unless you stop it. Adventure mode is still work-in-progress, but it's getting there.
60
 
61
+ ### Backyard AI
62
 
63
+ Another convenient way to use the model is [Backyard AI](https://backyard.ai/) application, which allows you to run the model locally on your computer. You'll need a graphics card with at least 8GB VRAM to use the model comfortably.
64
 
65
+ #### Download directly from HuggingFace (beta)
66
 
67
+ In the left panel, click `Manage Models`, then select `Hugging face models`. Paste `https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF` into the text field and press `Fetch Models`. Click `Download` button to the next to the model format. Once the model is downloaded, you can select it in your character card or set it as a default model.
68
 
69
+ #### Download manually
70
+
71
+ Download the [Psyfighter2-13B-vore.Q4_K_M.gguf](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q4_K_M.gguf) file into `%appdata%\faraday\models` folder on your computer. The model should appear in `Manage Models` menu under `Downloaded Models`. You can then select it in your character card or set it as a default model.
72
+
73
+ ### Model updates
74
+
75
+ - 04/13/2024 - uploaded the first version of the model
76
+ - 05/25/2024 - updated training process, making the model more coherent and improving the writing quality
77
 
78
  ## Bias, Risks, and Limitations
79
 
 
81
 
82
  ## Training Details
83
 
84
+ This model was fine-tuned on free-form text comprised of stories focused around the vore theme using [rank-stabilized](https://arxiv.org/abs/2312.03732) [QLoRA adapter](https://arxiv.org/abs/2305.14314) [QLoRA method](https://arxiv.org/abs/2305.14314). The resulting adapter was merged into the FP16 precision base model. The quantized version of the model was prepared using [llama.cpp](https://github.com/ggerganov/llama.cpp).
85
 
86
  ### Training Procedure
87
 
88
+ The model was fine-tuned with a [rank-stabilized](https://arxiv.org/abs/2312.03732) [QLoRA adapter](https://arxiv.org/abs/2305.14314) on NVIDIA GeForce RTX 4060 Ti over the span of ~24 hours. Training was performed using [Unsloth AI](https://github.com/unslothai/unsloth) library on `Ubuntu 22.04.4 LTS` with `CUDA 12.1` and `Pytorch 2.3.0`.
89
 
90
 
91
+ #### LoRa adapter configuration
92
 
93
+ - Rank: 128
94
+ - Alpha: 16
95
+ - Dropout rate: 0.1
96
+ - Target weights: `["q_proj", "k_proj", "o_proj", "gate_proj", "up_proj"]`,
97
+ - `use_rslora=True`
98
 
99
+ #### Training parameters
100
 
101
+ - Max. sequence length: 4096 tokens
102
+ - Samples per epoch: 3783
103
  - Number of epochs: 2
104
+ - Learning rate: 1e-4
105
+ - Warmup: 64 steps
106
+ - LR Schedule: linear
107
+ - Batch size: 1
108
+ - Gradient accumulation steps: 1
109
 
110
  #### Preprocessing
111
 
 
113
 
114
  - titles, foreword, tags, and anything not comprising the text of the story was removed
115
  - non-ascii characters and character sequences serving as chapter separators were removed
116
+ - any story mentioning underage personas in any context was removed from the dataset
117
  - names of private characters were replaced with randomized names across the dataset
118
 
119
  ## Environmental Impact
 
121
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
122
 
123
  - **Hardware Type:** NVIDIA GeForce RTX 4060 Ti
124
+ - **Hours used:** 24
125
  - **Cloud Provider:** N/A
126
  - **Compute Region:** US-East
127
+ - **Carbon Emitted:** 0.83 kg CO2 eq.