Casual-Autopsy's picture
Update README.md
afbf701 verified
metadata
language:
  - en
license: llama3
library_name: transformers
tags:
  - merge
  - mergekit
  - lazymergekit
  - not-for-all-audiences
  - nsfw
  - rp
  - roleplay
  - role-play
base_model:
  - Casual-Autopsy/L3-Umbral-Mind-RP-v1.0-8B
  - bluuwhale/L3-SthenoMaidBlackroot-8B-V1
  - Cas-Warehouse/Llama-3-SOVL-MopeyMule-Blackroot-8B
  - Cas-Warehouse/Llama-3-Mopeyfied-Psychology-v2
  - tannedbum/L3-Nymeria-8B
  - migtissera/Llama-3-8B-Synthia-v3.5
  - Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
  - tannedbum/L3-Nymeria-Maid-8B
  - Casual-Autopsy/L3-Uncen-Merger-Omelette-RP-v0.2-8B
  - aifeifei798/llama3-8B-DarkIdol-2.1-Uncensored-32K
  - Nitral-AI/Hathor_Stable-v0.2-L3-8B
  - Sao10K/L3-8B-Stheno-v3.1
pipeline_tag: text-generation
model-index:
  - name: L3-Umbral-Mind-RP-v2.0-8B
    results:
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: IFEval (0-Shot)
          type: HuggingFaceH4/ifeval
          args:
            num_few_shot: 0
        metrics:
          - type: inst_level_strict_acc and prompt_level_strict_acc
            value: 71.23
            name: strict accuracy
        source:
          url: >-
            https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: BBH (3-Shot)
          type: BBH
          args:
            num_few_shot: 3
        metrics:
          - type: acc_norm
            value: 32.49
            name: normalized accuracy
        source:
          url: >-
            https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: MATH Lvl 5 (4-Shot)
          type: hendrycks/competition_math
          args:
            num_few_shot: 4
        metrics:
          - type: exact_match
            value: 10.12
            name: exact match
        source:
          url: >-
            https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: GPQA (0-shot)
          type: Idavidrein/gpqa
          args:
            num_few_shot: 0
        metrics:
          - type: acc_norm
            value: 4.92
            name: acc_norm
        source:
          url: >-
            https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: MuSR (0-shot)
          type: TAUR-Lab/MuSR
          args:
            num_few_shot: 0
        metrics:
          - type: acc_norm
            value: 5.55
            name: acc_norm
        source:
          url: >-
            https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: MMLU-PRO (5-shot)
          type: TIGER-Lab/MMLU-Pro
          config: main
          split: test
          args:
            num_few_shot: 5
        metrics:
          - type: acc
            value: 30.26
            name: accuracy
        source:
          url: >-
            https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
          name: Open LLM Leaderboard
Image by ろ47

Merge

This is a merge of pre-trained language models created using mergekit.


Merge Details

The goal of this merge was to make an RP model better suited for role-plays with heavy themes such as but not limited to:

  • Mental illness
  • Self-harm
  • Trauma
  • Suicide

I hated how RP models tended to be overly positive and hopeful with role-plays involving such themes, but thanks to failspy/Llama-3-8B-Instruct-MopeyMule this problem has been lessened considerably.

If you're an enjoyer of savior/reverse savior type role-plays like myself, then this model is for you.


Usage Info

This model is meant to be used with asterisks/quotes RPing formats, any other format that isn't asterisks/quotes is likely to cause issues


Quants


Merge Method

This model was merged using several Task Arithmetic merges and then tied together with a Model Stock merge, followed by another Task Arithmetic merge with a model containing psychology data.

Models Merged

The following models were included in the merge:


Evaluation Results

Open LLM Leaderboard

Detailed results can be found here

Explaination for AI RP newbies: IFEval is the most important evaluation for RP AIs as it determines how well it can follow OOC, Lorebooks, and most importantly character cards. The rest don't matter. At least not nearly as much as IFEval.

Metric Value
Avg. 25.76
IFEval (0-Shot) 71.23
BBH (3-Shot) 32.49
MATH Lvl 5 (4-Shot) 10.12
GPQA (0-shot) 4.92
MuSR (0-shot) 5.55
MMLU-PRO (5-shot) 30.26

UGI Leaderboard

Information about the metrics can be found at the bottom of the UGI Leaderboard in the respective tabs.

Metric(UGI-Leaderboard) Value Value Metric(Writing Style)
UGI(Avg.) 31.82 0.107 RegV1
W/10 5.83 0.096 RegV2
Unruly 43.3 0.05 MyScore
Internet 20 9.12 ASSS
Stats 23.6 0 SMOG
Writing 33.8 1.47 Yule
PolContro 38.3

Secret Sauce

The following YAML configurations were used to produce this model:

Umbral-1

models:
  - model: Casual-Autopsy/L3-Umbral-Mind-RP-v1.0-8B
  - model: bluuwhale/L3-SthenoMaidBlackroot-8B-V1
    parameters:
      density: 0.45
      weight: 0.4
  - model: Cas-Warehouse/Llama-3-SOVL-MopeyMule-Blackroot-8B
    parameters:
      density: 0.65
      weight: 0.1
merge_method: dare_ties
base_model: Casual-Autopsy/L3-Umbral-Mind-RP-v1.0-8B
parameters:
  int8_mask: true
dtype: bfloat16

Umbral-2

models:
  - model: Cas-Warehouse/Llama-3-Mopeyfied-Psychology-v2
  - model: tannedbum/L3-Nymeria-8B
    parameters:
      density: 0.45
      weight: 0.25
  - model: migtissera/Llama-3-8B-Synthia-v3.5
    parameters:
      density: 0.65
      weight: 0.25
merge_method: dare_ties
base_model: Cas-Warehouse/Llama-3-Mopeyfied-Psychology-v2
parameters:
  int8_mask: true
dtype: bfloat16

Umbral-3

models:
  - model: Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
  - model: tannedbum/L3-Nymeria-Maid-8B
    parameters:
      density: 0.4
      weight: 0.3
  - model: Cas-Warehouse/Llama-3-SOVL-MopeyMule-Blackroot-8B
    parameters:
      density: 0.6
      weight: 0.2
merge_method: dare_ties
base_model: Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
parameters:
  int8_mask: true
dtype: bfloat16

Mopey-Omelette

models:
  - model: Casual-Autopsy/L3-Uncen-Merger-Omelette-RP-v0.2-8B
  - model: Cas-Warehouse/Llama-3-SOVL-MopeyMule-Blackroot-8B
    parameters:
      weight: 0.15
merge_method: task_arithmetic
base_model: Casual-Autopsy/L3-Uncen-Merger-Omelette-RP-v0.2-8B
dtype: bfloat16

Umbral-Mind-1

models:
  - model: Casual-Autopsy/Umbral-1
  - model: Casual-Autopsy/Umbral-3
merge_method: slerp
base_model: Casual-Autopsy/Umbral-1
parameters:
  t:
    - value: [0.7, 0.5, 0.3, 0.25, 0.2, 0.25, 0.3, 0.5, 0.7]
  embed_slerp: true
dtype: bfloat16

Umbral-Mind-2

models:
  - model: Casual-Autopsy/Umbral-Mind-1
  - model: Casual-Autopsy/Umbral-2
merge_method: slerp
base_model: Casual-Autopsy/Umbral-Mind-1
parameters:
  t:
    - value: [0.1, 0.15, 0.2, 0.4, 0.6, 0.4, 0.2, 0.15, 0.1]
  embed_slerp: true
dtype: bfloat16

Umbral-Mind-3

models:
  - model: Casual-Autopsy/Umbral-Mind-2
  - model: Casual-Autopsy/Mopey-Omelette
merge_method: slerp
base_model: Casual-Autopsy/Umbral-Mind-2
parameters:
  t:
    - value: [0.2, 0.25, 0.3, 0.4, 0.3, 0.25, 0.2, 0.25, 0.3, 0.4, 0.3, 0.25, 0.2]
  embed_slerp: true
dtype: bfloat16

L3-Umbral-Mind-RP-v2.0-8B

models:
  - model: Casual-Autopsy/Umbral-Mind-3
  - model: Casual-Autopsy/L3-Umbral-Mind-RP-v1.0-8B
    parameters:
      weight: 0.04
  - model: aifeifei798/llama3-8B-DarkIdol-2.1-Uncensored-32K
    parameters:
      weight: 0.02
  - model: Nitral-AI/Hathor_Stable-v0.2-L3-8B
    parameters:
      weight: 0.02
  - model: Sao10K/L3-8B-Stheno-v3.1
    parameters:
      weight: 0.01
merge_method: task_arithmetic
base_model: Casual-Autopsy/Umbral-Mind-3
dtype: bfloat16