Updating model files
Browse files
README.md
CHANGED
@@ -2,6 +2,17 @@
|
|
2 |
inference: false
|
3 |
license: other
|
4 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
# LLaMa 7B GGML
|
6 |
|
7 |
This repo contains GGML format model files for the original LLaMa.
|
@@ -10,7 +21,7 @@ These files are for CPU (+ CUDA) inference using [llama.cpp](https://github.com/
|
|
10 |
|
11 |
I've uploaded them mostly for my own convenience, allowing me to easily grab them if and when I need them for future testing and comparisons.
|
12 |
|
13 |
-
## Provided files
|
14 |
|
15 |
The following formats are included:
|
16 |
* float16
|
@@ -27,3 +38,14 @@ llama.cpp recently made a breaking change to its quantisation methods.
|
|
27 |
I have quantised the GGML files in this repo with the latest version. Therefore you will require llama.cpp compiled on May 12th or later (commit `b9fd7ee` or later) to use them.
|
28 |
|
29 |
I will not be providing GGML formats for the older llama.cpp code. They're already uploaded all over HF if you really need them!
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
inference: false
|
3 |
license: other
|
4 |
---
|
5 |
+
<div style="width: 100%;">
|
6 |
+
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
7 |
+
</div>
|
8 |
+
<div style="display: flex; justify-content: space-between; width: 100%;">
|
9 |
+
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
10 |
+
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
|
11 |
+
</div>
|
12 |
+
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
13 |
+
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
|
14 |
+
</div>
|
15 |
+
</div>
|
16 |
# LLaMa 7B GGML
|
17 |
|
18 |
This repo contains GGML format model files for the original LLaMa.
|
|
|
21 |
|
22 |
I've uploaded them mostly for my own convenience, allowing me to easily grab them if and when I need them for future testing and comparisons.
|
23 |
|
24 |
+
## Provided files
|
25 |
|
26 |
The following formats are included:
|
27 |
* float16
|
|
|
38 |
I have quantised the GGML files in this repo with the latest version. Therefore you will require llama.cpp compiled on May 12th or later (commit `b9fd7ee` or later) to use them.
|
39 |
|
40 |
I will not be providing GGML formats for the older llama.cpp code. They're already uploaded all over HF if you really need them!
|
41 |
+
## Want to support my work?
|
42 |
+
|
43 |
+
I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
|
44 |
+
|
45 |
+
So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
|
46 |
+
|
47 |
+
Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
|
48 |
+
|
49 |
+
* Patreon: coming soon! (just awaiting approval)
|
50 |
+
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
51 |
+
* Discord: https://discord.gg/UBgz4VXf
|