File size: 5,680 Bytes
4c373df adfd91d be6f506 adfd91d ccca17f adfd91d 4c373df adfd91d 1aaaa24 adfd91d 1aaaa24 adfd91d 1aaaa24 adfd91d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 |
---
license: llama2
language:
- en
pipeline_tag: text-generation
inference: false
tags:
- storywriting
- finetuned
- roleplay
- vore
- not-for-all-audiences
- gguf
- nsfw
- uncensored
base_model: SnakyMcSnekFace/Psyfighter2-13B-vore
model_type: llama
prompt_template: >
Below is an instruction that describes a task. Write a response that
appropriately completes the request.
### Instruction:
{prompt}
### Response:
---
# Model Card for Psyfighter2-13B-vore-GGUF
This is a quantized version of [SnakyMcSnekFace/Psyfighter2-13B-vore](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore) model.
This model is a version of [KoboldAI/LLaMA2-13B-Psyfighter2](https://huggingface.co/KoboldAI/LLaMA2-13B-Psyfighter2) finetuned to better understand vore context. The primary purpose of this model is to be a storywriting assistant, as well as a conversational model in a chat.
The Adventure Mode is still work in progress, and will be added later.
## Model Details
### Model Description
The model behaves similarly to `KoboldAI/LLaMA2-13B-Psyfighter2`, which it was derived from. Please [see the README.md here](https://huggingface.co/KoboldAI/LLaMA2-13B-Psyfighter2/blob/main/README.md) to learn more.
This model was fine-tuned on ~55 MiB of free-form text, containing stories focused around the vore theme. As a result, it has a strong vorny bias.
## How to Get Started with the Model
The model can be used with any AI chatbots and front-ends designed to work with `.gguf` models. The model fits fully into 8GB VRAM, but can also run with degraded performance on smaller graphics cards.
Similarly to the base model, the less prompt the model receives, the more creative is the output. For example, the writing assistant will generate an entire story when prompted with only 2-3 words.
In the chat mode, if the conversation is not going where you would like it to go, edit the model's output and let it continue generation. The model will also match the style of the conversation.
### Koboldcpp Colab Notebook
The easiest way to try out the model is [Koboldcpp Colab Notebook](https://colab.research.google.com/github/lostruins/koboldcpp/blob/concedo/colab.ipynb). This method doesn't require you to have a powerful graphics card.
- Open the notebook
- Paste the model URL into the field: `https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q4_K_M.gguf`
- Start the notebook, wait for the URL to CloudFlare tunnel to appear at the bottom and click it
- Use the model as a writing assistant
- You can try an adventure from [https://aetherroom.club/](https://aetherroom.club/), but keep in mind that the model will not let you take turn unless you stop it. Adventure mode is work-in-progress.
### Faraday
Another convenient way to use the model is [Faraday.dev](https://faraday.dev/) application, which allows you to run the model locally on your computer. You'll need a graphics card with at least 8GB VRAM to use `Q4_K_M` version comfortably, and 16GB VRAM for `Q8_0`. (`Q4_K_M` version is smaller and faster, `Q8_0` is slower but more coherent.)
Download the [Psyfighter2-13B-vore.Q4_K_M.gguf](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q4_K_M.gguf) or [Psyfighter2-13B-vore.Q8_0.gguf](https://huggingface.co/SnakyMcSnekFace/Psyfighter2-13B-vore-GGUF/resolve/main/Psyfighter2-13B-vore.Q8_0.gguf) file into `%appdata%\faraday\models` folder on your computer. The model should appear in `Manage Models` menu under `Downloaded Models`. You can then select it in your character card or set it as a default model.
### Others
TBD
## Bias, Risks, and Limitations
By design, this model has a strong vorny bias. It's not intended for use by anyone below 18 years old.
## Training Details
This model was fine-tuned on free-form text comprised of stories focused around the vore theme using the [QLoRA method](https://arxiv.org/abs/2305.14314). The resulting adapter was merged into the base model. The quantized version of the model was prepared using [llama.cpp](https://github.com/ggerganov/llama.cpp).
### Training Procedure
The model was fine-tuned using the [QLoRA method](https://arxiv.org/abs/2305.14314) on NVIDIA GeForce RTX 4060 Ti over the span of ~7 days. Training was performed using [text-generation-webui by oobabooga](https://github.com/oobabooga/text-generation-webui) with [Training PRO plug-in by FartyPants](https://github.com/FartyPants/Training_PRO).
LoRa adapter configuration:
- Rank: 512
- Alpha: 1024
- Dropout rate: 0.05
- Target weights: v_prog, q_proj
Training parameters:
- Sample size: 768 tokens
- Samples per epoch: 47420
- Number of epochs: 2
- First epoch: Learning rate = 3e-4, 1000 steps warmup, cosine schedule
- Second epoch: Learning rate = 1e-4, 256 steps warmup, inverse sqrt schedule
#### Preprocessing
The stories in dataset were pre-processed as follows:
- titles, foreword, tags, and anything not comprising the text of the story was removed
- non-ascii characters and character sequences serving as chapter separators were removed
- any story mentioning underage personas was taken out of the dataset
- names of private characters were replaced with randomized names across the dataset
## Environmental Impact
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** NVIDIA GeForce RTX 4060 Ti
- **Hours used:** 168
- **Cloud Provider:** N/A
- **Compute Region:** US-East
- **Carbon Emitted:** 5.8 kg CO2 eq.
|