EmanuelOverride commited on
Commit
bbd3be0
1 Parent(s): fe059fd

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +162 -0
README.md ADDED
@@ -0,0 +1,162 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: ResplendentAI/Paradigm_7B
3
+ datasets:
4
+ - ResplendentAI/Alpaca_NSFW_Shuffled
5
+ - unalignment/toxic-dpo-v0.2
6
+ language:
7
+ - en
8
+ library_name: transformers
9
+ license: cc-by-sa-4.0
10
+ tags:
11
+ - mergekit
12
+ - merge
13
+ - llama-cpp
14
+ - gguf-my-repo
15
+ model-index:
16
+ - name: Paradigm_7B
17
+ results:
18
+ - task:
19
+ type: text-generation
20
+ name: Text Generation
21
+ dataset:
22
+ name: AI2 Reasoning Challenge (25-Shot)
23
+ type: ai2_arc
24
+ config: ARC-Challenge
25
+ split: test
26
+ args:
27
+ num_few_shot: 25
28
+ metrics:
29
+ - type: acc_norm
30
+ value: 73.63
31
+ name: normalized accuracy
32
+ source:
33
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ResplendentAI/Paradigm_7B
34
+ name: Open LLM Leaderboard
35
+ - task:
36
+ type: text-generation
37
+ name: Text Generation
38
+ dataset:
39
+ name: HellaSwag (10-Shot)
40
+ type: hellaswag
41
+ split: validation
42
+ args:
43
+ num_few_shot: 10
44
+ metrics:
45
+ - type: acc_norm
46
+ value: 88.66
47
+ name: normalized accuracy
48
+ source:
49
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ResplendentAI/Paradigm_7B
50
+ name: Open LLM Leaderboard
51
+ - task:
52
+ type: text-generation
53
+ name: Text Generation
54
+ dataset:
55
+ name: MMLU (5-Shot)
56
+ type: cais/mmlu
57
+ config: all
58
+ split: test
59
+ args:
60
+ num_few_shot: 5
61
+ metrics:
62
+ - type: acc
63
+ value: 64.02
64
+ name: accuracy
65
+ source:
66
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ResplendentAI/Paradigm_7B
67
+ name: Open LLM Leaderboard
68
+ - task:
69
+ type: text-generation
70
+ name: Text Generation
71
+ dataset:
72
+ name: TruthfulQA (0-shot)
73
+ type: truthful_qa
74
+ config: multiple_choice
75
+ split: validation
76
+ args:
77
+ num_few_shot: 0
78
+ metrics:
79
+ - type: mc2
80
+ value: 75.19
81
+ source:
82
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ResplendentAI/Paradigm_7B
83
+ name: Open LLM Leaderboard
84
+ - task:
85
+ type: text-generation
86
+ name: Text Generation
87
+ dataset:
88
+ name: Winogrande (5-shot)
89
+ type: winogrande
90
+ config: winogrande_xl
91
+ split: validation
92
+ args:
93
+ num_few_shot: 5
94
+ metrics:
95
+ - type: acc
96
+ value: 84.53
97
+ name: accuracy
98
+ source:
99
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ResplendentAI/Paradigm_7B
100
+ name: Open LLM Leaderboard
101
+ - task:
102
+ type: text-generation
103
+ name: Text Generation
104
+ dataset:
105
+ name: GSM8k (5-shot)
106
+ type: gsm8k
107
+ config: main
108
+ split: test
109
+ args:
110
+ num_few_shot: 5
111
+ metrics:
112
+ - type: acc
113
+ value: 66.79
114
+ name: accuracy
115
+ source:
116
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ResplendentAI/Paradigm_7B
117
+ name: Open LLM Leaderboard
118
+ ---
119
+
120
+ # EmanuelOverride/Paradigm_7B-Q4_K_M-GGUF
121
+ This model was converted to GGUF format from [`ResplendentAI/Paradigm_7B`](https://huggingface.co/ResplendentAI/Paradigm_7B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
122
+ Refer to the [original model card](https://huggingface.co/ResplendentAI/Paradigm_7B) for more details on the model.
123
+
124
+ ## Use with llama.cpp
125
+ Install llama.cpp through brew (works on Mac and Linux)
126
+
127
+ ```bash
128
+ brew install llama.cpp
129
+
130
+ ```
131
+ Invoke the llama.cpp server or the CLI.
132
+
133
+ ### CLI:
134
+ ```bash
135
+ llama-cli --hf-repo EmanuelOverride/Paradigm_7B-Q4_K_M-GGUF --hf-file paradigm_7b-q4_k_m.gguf -p "The meaning to life and the universe is"
136
+ ```
137
+
138
+ ### Server:
139
+ ```bash
140
+ llama-server --hf-repo EmanuelOverride/Paradigm_7B-Q4_K_M-GGUF --hf-file paradigm_7b-q4_k_m.gguf -c 2048
141
+ ```
142
+
143
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
144
+
145
+ Step 1: Clone llama.cpp from GitHub.
146
+ ```
147
+ git clone https://github.com/ggerganov/llama.cpp
148
+ ```
149
+
150
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
151
+ ```
152
+ cd llama.cpp && LLAMA_CURL=1 make
153
+ ```
154
+
155
+ Step 3: Run inference through the main binary.
156
+ ```
157
+ ./llama-cli --hf-repo EmanuelOverride/Paradigm_7B-Q4_K_M-GGUF --hf-file paradigm_7b-q4_k_m.gguf -p "The meaning to life and the universe is"
158
+ ```
159
+ or
160
+ ```
161
+ ./llama-server --hf-repo EmanuelOverride/Paradigm_7B-Q4_K_M-GGUF --hf-file paradigm_7b-q4_k_m.gguf -c 2048
162
+ ```