SnakyMcSnekFace commited on
Commit
adfd91d
1 Parent(s): 4c373df

Upload 4 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Psyfighter2-13B-vore.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Psyfighter2-13B-vore.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ Psyfighter2-13B-vore.F16.gguf
Psyfighter2-13B-vore.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a1d9536c7245271758e7e846ba4e0d6b3061d295e0ca2777c04318b719ab8de
3
+ size 7865956352
Psyfighter2-13B-vore.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:342cd5a50f47804c345e0b4b9be970d8e49726ca60ace39bd735b1ff57540f08
3
+ size 13831319552
README.md CHANGED
@@ -1,3 +1,117 @@
1
  ---
2
  license: llama2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: llama2
3
+ language:
4
+ - en
5
+ pipeline_tag: text-generation
6
+ inference: false
7
+ tags:
8
+ - pytorch
9
+ - storywriting
10
+ - finetuned
11
+ - not-for-all-audiences
12
+ - gguf
13
+ base_model: SnakyMcSnekFace/Psyfighter2-13B-vore
14
+ model_type: llama
15
+ prompt_template: >
16
+ Below is an instruction that describes a task. Write a response that
17
+ appropriately completes the request.
18
+
19
+
20
+ ### Instruction:
21
+
22
+ {prompt}
23
+
24
+
25
+ ### Response:
26
  ---
27
+
28
+ # Model Card for Psyfighter2-13B-vore-GGUF
29
+
30
+ This is a quantized version of [SnakyMcSnekFace/Psyfighter2-13B-vore](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore) model.
31
+
32
+ This model is a version of [KoboldAI/LLaMA2-13B-Psyfighter2](https://huggingface.co/KoboldAI/LLaMA2-13B-Psyfighter2) finetuned to better understand vore context. The primary purpose of this model is to be a storywriting assistant, as well as a conversational model in a chat.
33
+
34
+ The Adventure Mode is still work in progress, and will be added later.
35
+
36
+ ## Model Details
37
+
38
+ ### Model Description
39
+
40
+ The model behaves similarly to `LLaMA2-13B-Psyfighter2`, which it was derived from. Please [see the README.md here](https://huggingface.co/KoboldAI/LLaMA2-13B-Psyfighter2/blob/main/README.md) to learn more.
41
+
42
+ This model was fine-tuned on ~55 MiB of free-form text, containing stories focused around the vore theme. As a result, it has a strong vorny bias.
43
+
44
+ ## How to Get Started with the Model
45
+
46
+ The model can be used with any AI chatbots and front-ends designed to work with `.gguf` models. The model fits fully into 8GB VRAM, but can also run with degraded performance on smaller graphics cards.
47
+
48
+ Similarly to the base model, the less prompt the model receives, the more creative is the output. For example, the writing assistant will generate an entire story when prompted with only 2-3 words.
49
+
50
+ In the chat mode, if the conversation is not going where you would like it to go, edit the model's output and let it continue generation. The model will also match the style of the conversation.
51
+
52
+ ### Koboldcpp Colab Notebook
53
+
54
+ The easiest way to try out the model is [Koboldcpp Colab Notebook](https://colab.research.google.com/github/lostruins/koboldcpp/blob/concedo/colab.ipynb). This method doesn't require you to have a powerful graphics card.
55
+
56
+ - Open the notebook
57
+ - Paste the model URL into the field: `https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q4_K_M.gguf`
58
+ - Start the notebook, wait for the URL to CloudFlare tunnel to appear at the bottom and click it
59
+ - Use the model as a writing assistant
60
+ - You can try an adventure from [https://aetherroom.club/](https://aetherroom.club/), but keep in mind that the model will not let you take turn unless you stop it. Adventure mode is work-in-progress.
61
+
62
+ ### Faraday
63
+
64
+ Another convenient way to use the model is [Faraday.dev](https://faraday.dev/) application, which allows you to run the model locally on your computer. You'll need a graphics card with at least 8GB VRAM to use this method comfortably.
65
+
66
+ Download the [Psyfighter2-13B-vore_q4_k_m.gguf](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q4_K_M.gguf) file into `%appdata%\faraday\models` folder on your computer. The model should appear in `Manage Models` menu under `Downloaded Models`. You can then select it in your character card or set it as a default model.
67
+
68
+ ### Others
69
+
70
+ TBD
71
+
72
+ ## Bias, Risks, and Limitations
73
+
74
+ By design, this model has a strong vorny bias. It's not intended for use by anyone below 18 years old.
75
+
76
+ ## Training Details
77
+
78
+ This model was fine-tuned on free-form text comprised of stories focused around the vore theme using the [QLoRA method](https://arxiv.org/abs/2305.14314). The resulting adapter was merged into the base model. The quantized version of the model was prepared using [llama.cpp](https://github.com/ggerganov/llama.cpp).
79
+
80
+ ### Training Procedure
81
+
82
+ The model was fine-tuned using the [QLoRA method](https://arxiv.org/abs/2305.14314) on NVIDIA GeForce RTX 4060 Ti over the span of ~7 days. Training was performed using [text-generation-webui by oobabooga](https://github.com/oobabooga/text-generation-webui) with [Training PRO plug-in by FartyPants](https://github.com/FartyPants/Training_PRO).
83
+
84
+
85
+ LoRa adapter configuration:
86
+
87
+ - Rank: 512
88
+ - Alpha: 1024
89
+ - Dropout rate: 0.05
90
+ - Target weights: v_prog, q_proj
91
+
92
+ Training parameters:
93
+
94
+ - Sample size: 768 tokens
95
+ - Samples per epoch: 47420
96
+ - Number of epochs: 2
97
+ - First epoch: Learning rate = 3e-4, 1000 steps warmup, cosine schedule
98
+ - Second epoch: Learning rate = 1e-4, 256 steps warmup, inverse sqrt schedule
99
+
100
+ #### Preprocessing
101
+
102
+ The stories in dataset were pre-processed as follows:
103
+
104
+ - titles, foreword, tags, and anything not comprising the text of the story was removed
105
+ - non-ascii characters and character sequences serving as chapter separators were removed
106
+ - any story mentioning underage personas was taken out of the dataset
107
+ - names of private characters were replaced with randomized names across the dataset
108
+
109
+ ## Environmental Impact
110
+
111
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
112
+
113
+ - **Hardware Type:** NVIDIA GeForce RTX 4060 Ti
114
+ - **Hours used:** 168
115
+ - **Cloud Provider:** N/A
116
+ - **Compute Region:** US-East
117
+ - **Carbon Emitted:** 5.8 kg CO2 eq.