v000000 commited on
Commit
44b5b71
1 Parent(s): eef4015

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +112 -42
README.md CHANGED
@@ -6,49 +6,119 @@ tags:
6
  - merge
7
  - mistral
8
  - llama-cpp
9
- - gguf-my-repo
10
  ---
11
 
12
- # v000000/MN-12B-Estrella-v1-Q6_K-GGUF
13
- This model was converted to GGUF format from [`v000000/MN-12B-Estrella-v1`](https://huggingface.co/v000000/MN-12B-Estrella-v1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
 
14
  Refer to the [original model card](https://huggingface.co/v000000/MN-12B-Estrella-v1) for more details on the model.
15
 
16
- ## Use with llama.cpp
17
- Install llama.cpp through brew (works on Mac and Linux)
18
-
19
- ```bash
20
- brew install llama.cpp
21
-
22
- ```
23
- Invoke the llama.cpp server or the CLI.
24
-
25
- ### CLI:
26
- ```bash
27
- llama-cli --hf-repo v000000/MN-12B-Estrella-v1-Q6_K-GGUF --hf-file mn-12b-estrella-v1-q6_k.gguf -p "The meaning to life and the universe is"
28
- ```
29
-
30
- ### Server:
31
- ```bash
32
- llama-server --hf-repo v000000/MN-12B-Estrella-v1-Q6_K-GGUF --hf-file mn-12b-estrella-v1-q6_k.gguf -c 2048
33
- ```
34
-
35
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
36
-
37
- Step 1: Clone llama.cpp from GitHub.
38
- ```
39
- git clone https://github.com/ggerganov/llama.cpp
40
- ```
41
-
42
- Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
43
- ```
44
- cd llama.cpp && LLAMA_CURL=1 make
45
- ```
46
-
47
- Step 3: Run inference through the main binary.
48
- ```
49
- ./llama-cli --hf-repo v000000/MN-12B-Estrella-v1-Q6_K-GGUF --hf-file mn-12b-estrella-v1-q6_k.gguf -p "The meaning to life and the universe is"
50
- ```
51
- or
52
- ```
53
- ./llama-server --hf-repo v000000/MN-12B-Estrella-v1-Q6_K-GGUF --hf-file mn-12b-estrella-v1-q6_k.gguf -c 2048
54
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  - merge
7
  - mistral
8
  - llama-cpp
 
9
  ---
10
 
11
+ GGUF version.
12
+
13
+ This model was converted to GGUF format from [`v000000/MN-12B-Estrella-v1`](https://huggingface.co/v000000/MN-12B-Estrella-v1) using llama.cpp'
14
  Refer to the [original model card](https://huggingface.co/v000000/MN-12B-Estrella-v1) for more details on the model.
15
 
16
+ Mistral-Nemo-12B-Estrella-v1-Q6_K-GGUF
17
+ ---------------------------------------------------------------------
18
+
19
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64f74b6e6389380c77562762/MyveknmJhuj43YrukIDAU.png)
20
+
21
+ Untested! Untested! Untested!
22
+
23
+ <b>Mistral Instruct / ChatML format.</b>
24
+
25
+ ----------------------------------------------------------------------
26
+ ## merge
27
+
28
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
29
+
30
+ ## Merge Details
31
+ ### Merge Method
32
+
33
+ This model was merged with a multi-step method using the <b>DELLA</b>, <b>DELLA_LINEAR</b> and <b>SLERP</b> merge algorithms.
34
+
35
+ ### Models Merged
36
+
37
+ The following models were included in the merge:
38
+ * [nothingiisreal/MN-12B-Celeste-V1.9](https://huggingface.co/nothingiisreal/MN-12B-Celeste-V1.9)
39
+ * [shuttleai/shuttle-2.5-mini](https://huggingface.co/shuttleai/shuttle-2.5-mini)
40
+ * [anthracite-org/magnum-12b-v2](https://huggingface.co/anthracite-org/magnum-12b-v2)
41
+ * [Sao10K/MN-12B-Lyra-v1](https://huggingface.co/Sao10K/MN-12B-Lyra-v1)
42
+ * [unsloth/Mistral-Nemo-Instruct-2407](https://huggingface.co/unsloth/Mistral-Nemo-Instruct-2407)
43
+ * [NeverSleep/Lumimaid-v0.2-12B](https://huggingface.co/NeverSleep/Lumimaid-v0.2-12B)
44
+ * [UsernameJustAnother/Nemo-12B-Marlin-v5](https://huggingface.co/UsernameJustAnother/Nemo-12B-Marlin-v5)
45
+ * [BeaverAI/mistral-doryV2-12b](https://huggingface.co/BeaverAI/mistral-doryV2-12b)
46
+ * [invisietch/Atlantis-v0.1-12B](https://huggingface.co/invisietch/Atlantis-v0.1-12B)
47
+
48
+ ### Configuration
49
+
50
+ The following YAML configuration was used to produce this model:
51
+
52
+ ```yaml
53
+ #Step 1 (Part1)
54
+ models:
55
+ - model: Sao10K/MN-12B-Lyra-v1
56
+ parameters:
57
+ weight: 0.15
58
+ density: 0.77
59
+ - model: shuttleai/shuttle-2.5-mini
60
+ parameters:
61
+ weight: 0.20
62
+ density: 0.78
63
+ - model: anthracite-org/magnum-12b-v2
64
+ parameters:
65
+ weight: 0.35
66
+ density: 0.85
67
+ - model: nothingiisreal/MN-12B-Celeste-V1.9
68
+ parameters:
69
+ weight: 0.55
70
+ density: 0.90
71
+ merge_method: della
72
+ base_model: Sao10K/MN-12B-Lyra-v1
73
+ parameters:
74
+ int8_mask: true
75
+ epsilon: 0.05
76
+ lambda: 1
77
+ dtype: bfloat16
78
+ #Step 2 (Part2)
79
+ models:
80
+ - model: BeaverAI/mistral-doryV2-12b
81
+ parameters:
82
+ weight: 0.10
83
+ density: 0.4
84
+ - model: unsloth/Mistral-Nemo-Instruct-2407
85
+ parameters:
86
+ weight: 0.20
87
+ density: 0.4
88
+ - model: UsernameJustAnother/Nemo-12B-Marlin-v5
89
+ parameters:
90
+ weight: 0.25
91
+ density: 0.5
92
+ - model: invisietch/Atlantis-v0.1-12B
93
+ parameters:
94
+ weight: 0.3
95
+ density: 0.5
96
+ - model: NeverSleep/Lumimaid-v0.2-12B
97
+ parameters:
98
+ weight: 0.4
99
+ density: 0.8
100
+ merge_method: della_linear
101
+ base_model: anthracite-org/magnum-12b-v2
102
+ parameters:
103
+ int8_mask: true
104
+ epsilon: 0.05
105
+ lambda: 1
106
+ dtype: bfloat16
107
+ #Step 3 (Estrella)
108
+ slices:
109
+ - sources:
110
+ - model: v000000/MN-12B-Part2
111
+ layer_range: [0, 40]
112
+ - model: v000000/MN-12B-Part1
113
+ layer_range: [0, 40]
114
+ merge_method: slerp
115
+ base_model: v000000/MN-12B-Part1
116
+ parameters: #smooth gradient prio part1
117
+ t:
118
+ - filter: self_attn
119
+ value: [0, 0.5, 0.3, 0.7, 0.6, 0.1, 0.6, 0.3, 0.8, 0.5]
120
+ - filter: mlp
121
+ value: [0, 0.5, 0.4, 0.3, 0, 0.3, 0.4, 0.7, 0.2, 0.5]
122
+ - value: 0.5
123
+ dtype: bfloat16
124
+ ```