Update README.md
Browse files
README.md
CHANGED
@@ -29,7 +29,7 @@ We evaluate the model on [RewardBench](https://github.com/allenai/reward-bench):
|
|
29 |
| Model | Score | Chat | Chat Hard | Safety | Reasoning | Prior Sets (0.5 weight) |
|
30 |
|------------------|-------|-------|-----------|--------|-----------|-------------------------|
|
31 |
| [Llama 3 Tulu 2 8b UF RM](https://huggingface.co/allenai/llama-3-tulu-2-8b-uf-mean-rm) | 66.3 | 96.6 | 59.4 | 61.4 | 80.7 | |
|
32 |
-
| **[Llama 3 Tulu 2 70b UF RM](https://huggingface.co/allenai/llama-3-tulu-2-70b-uf-mean-rm) (this model)** |
|
33 |
|
34 |
|
35 |
|
|
|
29 |
| Model | Score | Chat | Chat Hard | Safety | Reasoning | Prior Sets (0.5 weight) |
|
30 |
|------------------|-------|-------|-----------|--------|-----------|-------------------------|
|
31 |
| [Llama 3 Tulu 2 8b UF RM](https://huggingface.co/allenai/llama-3-tulu-2-8b-uf-mean-rm) | 66.3 | 96.6 | 59.4 | 61.4 | 80.7 | |
|
32 |
+
| **[Llama 3 Tulu 2 70b UF RM](https://huggingface.co/allenai/llama-3-tulu-2-70b-uf-mean-rm) (this model)** | 65.3 | 89.1 | 52.6 | 64.0 | 88.3 | |
|
33 |
|
34 |
|
35 |
|