Update README.md
Browse files
README.md
CHANGED
@@ -124,6 +124,12 @@ We place #1 for all 13B models at release time!
|
|
124 |
|
125 |
Since our release, a new model which merges an Orca-style model with a Platypus (trained on STEM and logic) model places narrowly above ours, but we were #1 at release time.
|
126 |
|
|
|
|
|
|
|
|
|
|
|
|
|
127 |
## GPT4ALL Leaderboard Performance
|
128 |
|
129 |
We have tested using parameters matching the GPT4ALL Benchmark Suite and report our results and placement vs their official reporting below.
|
|
|
124 |
|
125 |
Since our release, a new model which merges an Orca-style model with a Platypus (trained on STEM and logic) model places narrowly above ours, but we were #1 at release time.
|
126 |
|
127 |
+
Below we also highlight how our model fits relative to models of all sizes on the current (as of Aug 10th, 2023) leaderboard.
|
128 |
+
|
129 |
+
![OpenOrca Preview2 HuggingFace Leaderboard Performance](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B/resolve/main/Images/OpenOrcaP2HFLeaderboardFull.png "HuggingFace Full Leaderboard")
|
130 |
+
|
131 |
+
Notably, performance is beyond falcon-40b-instruct, and close to LLaMA1-65B base.
|
132 |
+
|
133 |
## GPT4ALL Leaderboard Performance
|
134 |
|
135 |
We have tested using parameters matching the GPT4ALL Benchmark Suite and report our results and placement vs their official reporting below.
|