ArkaAbacus
commited on
Commit
•
8f558d6
1
Parent(s):
5af6f9d
Update README.md
Browse files
README.md
CHANGED
@@ -133,10 +133,10 @@ Meta-Llama-3-70B-Instruct 9.006250
|
|
133 |
|
134 |
### OpenLLM Leaderboard Manual Evaluation
|
135 |
|
136 |
-
| Model | ARC | Hellaswag | MMLU | TruthfulQA | Winogrande | GSM8K* |
|
137 |
-
| :---- | ---: | ------: | ---: | ---: | ---: | ---: |
|
138 |
-
| Smaug-Llama-3-70B-Instruct | 70.6 | 86.1 | 79.2 | 62.5 | 83.5 | 90.5 |
|
139 |
-
| Llama-3-70B-Instruct | 71.4 | 85.7 | 80.0 | 61.8 | 82.9 | 91.1 |
|
140 |
|
141 |
**GSM8K** The GSM8K numbers quoted here are computed using a recent release
|
142 |
of the [LM Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness/).
|
|
|
133 |
|
134 |
### OpenLLM Leaderboard Manual Evaluation
|
135 |
|
136 |
+
| Model | ARC | Hellaswag | MMLU | TruthfulQA | Winogrande | GSM8K* | Average |
|
137 |
+
| :---- | ---: | ------: | ---: | ---: | ---: | ---: | ---: |
|
138 |
+
| Smaug-Llama-3-70B-Instruct | 70.6 | 86.1 | 79.2 | 62.5 | 83.5 | 90.5 | 78.7 |
|
139 |
+
| Llama-3-70B-Instruct | 71.4 | 85.7 | 80.0 | 61.8 | 82.9 | 91.1 | 78.8 |
|
140 |
|
141 |
**GSM8K** The GSM8K numbers quoted here are computed using a recent release
|
142 |
of the [LM Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness/).
|