migtissera leaderboard-pr-bot commited on
Commit
9452a82
1 Parent(s): 63513c3

Adding Evaluation Results (#1)

Browse files

- Adding Evaluation Results (d23e943dcb8330a894d60efe557d160d47adc591)


Co-authored-by: Open LLM Leaderboard PR Bot <[email protected]>

Files changed (1) hide show
  1. README.md +15 -2
README.md CHANGED
@@ -1,8 +1,8 @@
1
  ---
2
  license: other
3
- base_model: mistralai/Codestral-22B-v0.1
4
  tags:
5
  - generated_from_trainer
 
6
  model-index:
7
  - name: home/ubuntu/trinity-codestral-1
8
  results: []
@@ -79,4 +79,17 @@ while True:
79
  answer = generate_text(llm_prompt)
80
  print(answer)
81
  conversation = f"{llm_prompt}{answer}<|im_end|>\n<|im_start|>user\n"
82
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: other
 
3
  tags:
4
  - generated_from_trainer
5
+ base_model: mistralai/Codestral-22B-v0.1
6
  model-index:
7
  - name: home/ubuntu/trinity-codestral-1
8
  results: []
 
79
  answer = generate_text(llm_prompt)
80
  print(answer)
81
  conversation = f"{llm_prompt}{answer}<|im_end|>\n<|im_start|>user\n"
82
+ ```
83
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
84
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_migtissera__Trinity-2-Codestral-22B-v0.2)
85
+
86
+ | Metric |Value|
87
+ |-------------------|----:|
88
+ |Avg. |21.87|
89
+ |IFEval (0-Shot) |43.45|
90
+ |BBH (3-Shot) |37.61|
91
+ |MATH Lvl 5 (4-Shot)| 8.38|
92
+ |GPQA (0-shot) | 6.71|
93
+ |MuSR (0-shot) | 9.06|
94
+ |MMLU-PRO (5-shot) |26.00|
95
+