Adding the Open Portuguese LLM Leaderboard Evaluation Results

#1
Files changed (1) hide show
  1. README.md +22 -2
README.md CHANGED
@@ -1,6 +1,8 @@
1
  ---
 
 
2
  license: mit
3
- library_name: "trl"
4
  tags:
5
  - DPO
6
  - WeniGPT
@@ -8,7 +10,6 @@ base_model: Weni/WeniGPT-2.2.3-Zephyr-7B-merged-LLM_Base_2.0.3_SFT
8
  model-index:
9
  - name: Weni/WeniGPT-2.8.1-Zephyr-7B-zephyr-prompt-binarized
10
  results: []
11
- language: ['en']
12
  ---
13
 
14
  # Weni/WeniGPT-2.8.1-Zephyr-7B-zephyr-prompt-binarized
@@ -79,3 +80,22 @@ The following hyperparameters were used during training:
79
 
80
  ### Hardware
81
  - Cloud provided: runpod.io
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
  license: mit
5
+ library_name: trl
6
  tags:
7
  - DPO
8
  - WeniGPT
 
10
  model-index:
11
  - name: Weni/WeniGPT-2.8.1-Zephyr-7B-zephyr-prompt-binarized
12
  results: []
 
13
  ---
14
 
15
  # Weni/WeniGPT-2.8.1-Zephyr-7B-zephyr-prompt-binarized
 
80
 
81
  ### Hardware
82
  - Cloud provided: runpod.io
83
+
84
+
85
+ # Open Portuguese LLM Leaderboard Evaluation Results
86
+
87
+ Detailed results can be found [here](https://huggingface.co/datasets/eduagarcia-temp/llm_pt_leaderboard_raw_results/tree/main/Weni/WeniGPT-2.8.1-Zephyr-7B-zephyr-prompt-binarized) and on the [🚀 Open Portuguese LLM Leaderboard](https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard)
88
+
89
+ | Metric | Value |
90
+ |--------------------------|---------|
91
+ |Average |**50.64**|
92
+ |ENEM Challenge (No Images)| 37.44|
93
+ |BLUEX (No Images) | 36.58|
94
+ |OAB Exams | 35.03|
95
+ |Assin2 RTE | 88.48|
96
+ |Assin2 STS | 33.02|
97
+ |FaQuAD NLI | 45.45|
98
+ |HateBR Binary | 68.62|
99
+ |PT Hate Speech Binary | 56.79|
100
+ |tweetSentBR | 54.32|
101
+