Triangle104 commited on
Commit
57d0446
1 Parent(s): 1c7c4b2

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +174 -0
README.md ADDED
@@ -0,0 +1,174 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - ja
5
+ license: cc-by-nc-4.0
6
+ library_name: transformers
7
+ tags:
8
+ - nsfw
9
+ - Visual novel
10
+ - roleplay
11
+ - mergekit
12
+ - merge
13
+ - llama-cpp
14
+ - gguf-my-repo
15
+ base_model: spow12/ChatWaifu_v2.0_22B
16
+ datasets:
17
+ - roleplay4fun/aesir-v1.1
18
+ - kalomaze/Opus_Instruct_3k
19
+ - Gryphe/Sonnet3.5-SlimOrcaDedupCleaned
20
+ - Aratako/Synthetic-JP-EN-Coding-Dataset-567k
21
+ - Aratako/Synthetic-Japanese-Roleplay-gpt-4o-mini-39.6k-formatted
22
+ - Aratako/Synthetic-Japanese-Roleplay-NSFW-Claude-3.5s-15.3k-formatted
23
+ - Aratako_Rosebleu_1on1_Dialogues_RP
24
+ - SkunkworksAI/reasoning-0.01
25
+ - jondurbin_gutenberg_dpo
26
+ - nbeerbower_gutenberg2_dpo
27
+ - jondurbi_py_dpo
28
+ - jondurbin_truthy_dpo
29
+ - flammenai_character_roleplay_DPO
30
+ - kyujinpy_orca_math_dpo
31
+ - argilla_Capybara_Preferences
32
+ - antiven0m_physical_reasoning_dpo
33
+ - aixsatoshi_Swallow_MX_chatbot_DPO
34
+ pipeline_tag: text-generation
35
+ model-index:
36
+ - name: ChatWaifu_v2.0_22B
37
+ results:
38
+ - task:
39
+ type: text-generation
40
+ name: Text Generation
41
+ dataset:
42
+ name: IFEval (0-Shot)
43
+ type: HuggingFaceH4/ifeval
44
+ args:
45
+ num_few_shot: 0
46
+ metrics:
47
+ - type: inst_level_strict_acc and prompt_level_strict_acc
48
+ value: 65.11
49
+ name: strict accuracy
50
+ source:
51
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=spow12/ChatWaifu_v2.0_22B
52
+ name: Open LLM Leaderboard
53
+ - task:
54
+ type: text-generation
55
+ name: Text Generation
56
+ dataset:
57
+ name: BBH (3-Shot)
58
+ type: BBH
59
+ args:
60
+ num_few_shot: 3
61
+ metrics:
62
+ - type: acc_norm
63
+ value: 42.29
64
+ name: normalized accuracy
65
+ source:
66
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=spow12/ChatWaifu_v2.0_22B
67
+ name: Open LLM Leaderboard
68
+ - task:
69
+ type: text-generation
70
+ name: Text Generation
71
+ dataset:
72
+ name: MATH Lvl 5 (4-Shot)
73
+ type: hendrycks/competition_math
74
+ args:
75
+ num_few_shot: 4
76
+ metrics:
77
+ - type: exact_match
78
+ value: 18.58
79
+ name: exact match
80
+ source:
81
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=spow12/ChatWaifu_v2.0_22B
82
+ name: Open LLM Leaderboard
83
+ - task:
84
+ type: text-generation
85
+ name: Text Generation
86
+ dataset:
87
+ name: GPQA (0-shot)
88
+ type: Idavidrein/gpqa
89
+ args:
90
+ num_few_shot: 0
91
+ metrics:
92
+ - type: acc_norm
93
+ value: 9.96
94
+ name: acc_norm
95
+ source:
96
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=spow12/ChatWaifu_v2.0_22B
97
+ name: Open LLM Leaderboard
98
+ - task:
99
+ type: text-generation
100
+ name: Text Generation
101
+ dataset:
102
+ name: MuSR (0-shot)
103
+ type: TAUR-Lab/MuSR
104
+ args:
105
+ num_few_shot: 0
106
+ metrics:
107
+ - type: acc_norm
108
+ value: 5.59
109
+ name: acc_norm
110
+ source:
111
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=spow12/ChatWaifu_v2.0_22B
112
+ name: Open LLM Leaderboard
113
+ - task:
114
+ type: text-generation
115
+ name: Text Generation
116
+ dataset:
117
+ name: MMLU-PRO (5-shot)
118
+ type: TIGER-Lab/MMLU-Pro
119
+ config: main
120
+ split: test
121
+ args:
122
+ num_few_shot: 5
123
+ metrics:
124
+ - type: acc
125
+ value: 31.51
126
+ name: accuracy
127
+ source:
128
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=spow12/ChatWaifu_v2.0_22B
129
+ name: Open LLM Leaderboard
130
+ ---
131
+
132
+ # Triangle104/ChatWaifu_v2.0_22B-Q8_0-GGUF
133
+ This model was converted to GGUF format from [`spow12/ChatWaifu_v2.0_22B`](https://huggingface.co/spow12/ChatWaifu_v2.0_22B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
134
+ Refer to the [original model card](https://huggingface.co/spow12/ChatWaifu_v2.0_22B) for more details on the model.
135
+
136
+ ## Use with llama.cpp
137
+ Install llama.cpp through brew (works on Mac and Linux)
138
+
139
+ ```bash
140
+ brew install llama.cpp
141
+
142
+ ```
143
+ Invoke the llama.cpp server or the CLI.
144
+
145
+ ### CLI:
146
+ ```bash
147
+ llama-cli --hf-repo Triangle104/ChatWaifu_v2.0_22B-Q8_0-GGUF --hf-file chatwaifu_v2.0_22b-q8_0.gguf -p "The meaning to life and the universe is"
148
+ ```
149
+
150
+ ### Server:
151
+ ```bash
152
+ llama-server --hf-repo Triangle104/ChatWaifu_v2.0_22B-Q8_0-GGUF --hf-file chatwaifu_v2.0_22b-q8_0.gguf -c 2048
153
+ ```
154
+
155
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
156
+
157
+ Step 1: Clone llama.cpp from GitHub.
158
+ ```
159
+ git clone https://github.com/ggerganov/llama.cpp
160
+ ```
161
+
162
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
163
+ ```
164
+ cd llama.cpp && LLAMA_CURL=1 make
165
+ ```
166
+
167
+ Step 3: Run inference through the main binary.
168
+ ```
169
+ ./llama-cli --hf-repo Triangle104/ChatWaifu_v2.0_22B-Q8_0-GGUF --hf-file chatwaifu_v2.0_22b-q8_0.gguf -p "The meaning to life and the universe is"
170
+ ```
171
+ or
172
+ ```
173
+ ./llama-server --hf-repo Triangle104/ChatWaifu_v2.0_22B-Q8_0-GGUF --hf-file chatwaifu_v2.0_22b-q8_0.gguf -c 2048
174
+ ```