Update README.md
Browse files
README.md
CHANGED
@@ -1,11 +1,17 @@
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
-
tags:
|
|
|
|
|
|
|
|
|
4 |
---
|
5 |
|
6 |
-
# Model Card for
|
7 |
|
8 |
-
|
|
|
|
|
9 |
|
10 |
|
11 |
|
@@ -196,6 +202,4 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
|
|
196 |
|
197 |
## Model Card Contact
|
198 |
|
199 |
-
[More Information Needed]
|
200 |
-
|
201 |
-
|
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
+
tags:
|
4 |
+
- moe
|
5 |
+
license: llama2
|
6 |
+
language:
|
7 |
+
- en
|
8 |
---
|
9 |
|
10 |
+
# Model Card for Giant Hydra 240B
|
11 |
|
12 |
+
Yes, you read that correctly, this is a 4x70b MOE model with ~240B parameters. I doubt there is any way that I will have the benchmarks run here anytime soon to be on the leaderboard but I am looking into renting time on runpod to get the scores myself and put them here.
|
13 |
+
|
14 |
+
This model should cover multiple different disciplines and behaviors well as I tried to use and gate correctly a wide set of models including one I fine tuned myself.
|
15 |
|
16 |
|
17 |
|
|
|
202 |
|
203 |
## Model Card Contact
|
204 |
|
205 |
+
[More Information Needed]
|
|
|
|