Update README.md
Browse files
README.md
CHANGED
@@ -46,7 +46,7 @@ Here is the performance of this model across benchmarks explored in our paper [H
|
|
46 |
|
47 |
| MMLU 0-shot | MMLU 5-shot | GSM Direct | GSM CoT | BBH Direct | BBH CoT | TydiQA Gold-Passage | TydiQA Closed-book | Codex-Eval Pass@1 | Codex-Eval Pass@10 | AlpacaFarm vs Davinci-003 | Average |
|
48 |
|:-----------:|:-----------:|:----------:|:-------:|:----------:|:-------:|:-------------------:|:------------------:|:-----------------:|:------------------:|:-------------------------:|---------|
|
49 |
-
|
|
50 |
|
51 |
If you use this model, please cite our work, the llama paper, and the original dataset:
|
52 |
|
|
|
46 |
|
47 |
| MMLU 0-shot | MMLU 5-shot | GSM Direct | GSM CoT | BBH Direct | BBH CoT | TydiQA Gold-Passage | TydiQA Closed-book | Codex-Eval Pass@1 | Codex-Eval Pass@10 | AlpacaFarm vs Davinci-003 | Average |
|
48 |
|:-----------:|:-----------:|:----------:|:-------:|:----------:|:-------:|:-------------------:|:------------------:|:-----------------:|:------------------:|:-------------------------:|---------|
|
49 |
+
| 34.7 | 34.5 | 6.5 | 7.5 | 29.6 | 30.5 | 36.7 | 10.5 | 16.5 | 29.2 | 17.5 | 22.6 |
|
50 |
|
51 |
If you use this model, please cite our work, the llama paper, and the original dataset:
|
52 |
|