Spaces:
Running
Running
orionweller
commited on
Commit
•
f632e87
1
Parent(s):
c49e5d1
Automated Leaderboard Update
Browse files- all_data_tasks/0/default.jsonl +0 -0
- all_data_tasks/1/default.jsonl +0 -0
- all_data_tasks/10/default.jsonl +0 -0
- all_data_tasks/11/default.jsonl +0 -0
- all_data_tasks/12/default.jsonl +0 -0
- all_data_tasks/13/default.jsonl +0 -0
- all_data_tasks/2/default.jsonl +0 -0
- all_data_tasks/3/default.jsonl +0 -0
- all_data_tasks/31/default.jsonl +2 -2
- all_data_tasks/33/default.jsonl +2 -2
- all_data_tasks/35/default.jsonl +16 -16
- all_data_tasks/39/default.jsonl +0 -0
- all_data_tasks/4/default.jsonl +0 -0
- all_data_tasks/40/default.jsonl +0 -0
- all_data_tasks/5/default.jsonl +0 -0
- all_data_tasks/6/default.jsonl +0 -0
- all_data_tasks/8/default.jsonl +0 -0
- all_data_tasks/9/default.jsonl +0 -0
- boards_data/en/data_overall/default.jsonl +0 -0
- boards_data/en/data_tasks/Classification/default.jsonl +0 -0
- boards_data/en/data_tasks/Clustering/default.jsonl +0 -0
- boards_data/en/data_tasks/PairClassification/default.jsonl +0 -0
- boards_data/en/data_tasks/Reranking/default.jsonl +0 -0
- boards_data/en/data_tasks/Retrieval/default.jsonl +0 -0
- boards_data/en/data_tasks/STS/default.jsonl +0 -0
- boards_data/en/data_tasks/Summarization/default.jsonl +0 -0
- boards_data/other-cls/data_tasks/Classification/default.jsonl +0 -0
- boards_data/other-sts/data_tasks/STS/default.jsonl +0 -0
- boards_data/ru/data_overall/default.jsonl +20 -20
- boards_data/ru/data_tasks/Classification/default.jsonl +2 -2
- boards_data/ru/data_tasks/PairClassification/default.jsonl +2 -2
- boards_data/ru/data_tasks/Retrieval/default.jsonl +16 -16
- boards_data/zh/data_overall/default.jsonl +0 -0
- boards_data/zh/data_tasks/Classification/default.jsonl +0 -0
- boards_data/zh/data_tasks/Clustering/default.jsonl +0 -0
- boards_data/zh/data_tasks/PairClassification/default.jsonl +0 -0
- boards_data/zh/data_tasks/Reranking/default.jsonl +0 -0
- boards_data/zh/data_tasks/Retrieval/default.jsonl +0 -0
- boards_data/zh/data_tasks/STS/default.jsonl +0 -0
all_data_tasks/0/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/1/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/10/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/11/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/12/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/13/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/2/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/3/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/31/default.jsonl
CHANGED
@@ -2,8 +2,8 @@
|
|
2 |
{"index":10,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":67.48,"GeoreviewClassification":53.47,"HeadlineClassification":85.66,"InappropriatenessClassification":65.29,"KinopoiskClassification":64.25,"RuReviewsClassification":68.58,"RuSciBenchGRNTIClassification":64.56,"RuSciBenchOECDClassification":51.2,"MassiveIntentClassification (ru)":76.01,"MassiveScenarioClassification (ru)":78.28}
|
3 |
{"index":35,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.31,"GeoreviewClassification":55.9,"HeadlineClassification":86.18,"InappropriatenessClassification":65.53,"KinopoiskClassification":66.12,"RuReviewsClassification":68.56,"RuSciBenchGRNTIClassification":65.07,"RuSciBenchOECDClassification":50.21,"MassiveIntentClassification (ru)":67.6,"MassiveScenarioClassification (ru)":71.59}
|
4 |
{"index":11,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":65.21,"GeoreviewClassification":52.04,"HeadlineClassification":83.4,"InappropriatenessClassification":64.14,"KinopoiskClassification":63.15,"RuReviewsClassification":68.24,"RuSciBenchGRNTIClassification":61.24,"RuSciBenchOECDClassification":48.1,"MassiveIntentClassification (ru)":68.61,"MassiveScenarioClassification (ru)":77.98}
|
5 |
-
{"index":
|
6 |
-
{"index":
|
7 |
{"index":18,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":62.74,"GeoreviewClassification":49.7,"HeadlineClassification":78.0,"InappropriatenessClassification":61.32,"KinopoiskClassification":63.27,"RuReviewsClassification":67.96,"RuSciBenchGRNTIClassification":59.33,"RuSciBenchOECDClassification":46.33,"MassiveIntentClassification (ru)":66.97,"MassiveScenarioClassification (ru)":71.8}
|
8 |
{"index":29,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average":61.92,"GeoreviewClassification":50.98,"HeadlineClassification":70.09,"InappropriatenessClassification":60.76,"KinopoiskClassification":63.33,"RuReviewsClassification":68.52,"RuSciBenchGRNTIClassification":57.67,"RuSciBenchOECDClassification":44.2,"MassiveIntentClassification (ru)":68.85,"MassiveScenarioClassification (ru)":72.9}
|
9 |
{"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":61.58,"GeoreviewClassification":47.69,"HeadlineClassification":83.46,"InappropriatenessClassification":61.32,"KinopoiskClassification":59.04,"RuReviewsClassification":66.09,"RuSciBenchGRNTIClassification":61.41,"RuSciBenchOECDClassification":48.67,"MassiveIntentClassification (ru)":60.08,"MassiveScenarioClassification (ru)":66.44}
|
|
|
2 |
{"index":10,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":67.48,"GeoreviewClassification":53.47,"HeadlineClassification":85.66,"InappropriatenessClassification":65.29,"KinopoiskClassification":64.25,"RuReviewsClassification":68.58,"RuSciBenchGRNTIClassification":64.56,"RuSciBenchOECDClassification":51.2,"MassiveIntentClassification (ru)":76.01,"MassiveScenarioClassification (ru)":78.28}
|
3 |
{"index":35,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.31,"GeoreviewClassification":55.9,"HeadlineClassification":86.18,"InappropriatenessClassification":65.53,"KinopoiskClassification":66.12,"RuReviewsClassification":68.56,"RuSciBenchGRNTIClassification":65.07,"RuSciBenchOECDClassification":50.21,"MassiveIntentClassification (ru)":67.6,"MassiveScenarioClassification (ru)":71.59}
|
4 |
{"index":11,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":65.21,"GeoreviewClassification":52.04,"HeadlineClassification":83.4,"InappropriatenessClassification":64.14,"KinopoiskClassification":63.15,"RuReviewsClassification":68.24,"RuSciBenchGRNTIClassification":61.24,"RuSciBenchOECDClassification":48.1,"MassiveIntentClassification (ru)":68.61,"MassiveScenarioClassification (ru)":77.98}
|
5 |
+
{"index":22,"Rank":5,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/arkohut\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
|
6 |
+
{"index":39,"Rank":6,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/jinaai\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
|
7 |
{"index":18,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":62.74,"GeoreviewClassification":49.7,"HeadlineClassification":78.0,"InappropriatenessClassification":61.32,"KinopoiskClassification":63.27,"RuReviewsClassification":67.96,"RuSciBenchGRNTIClassification":59.33,"RuSciBenchOECDClassification":46.33,"MassiveIntentClassification (ru)":66.97,"MassiveScenarioClassification (ru)":71.8}
|
8 |
{"index":29,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average":61.92,"GeoreviewClassification":50.98,"HeadlineClassification":70.09,"InappropriatenessClassification":60.76,"KinopoiskClassification":63.33,"RuReviewsClassification":68.52,"RuSciBenchGRNTIClassification":57.67,"RuSciBenchOECDClassification":44.2,"MassiveIntentClassification (ru)":68.85,"MassiveScenarioClassification (ru)":72.9}
|
9 |
{"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":61.58,"GeoreviewClassification":47.69,"HeadlineClassification":83.46,"InappropriatenessClassification":61.32,"KinopoiskClassification":59.04,"RuReviewsClassification":66.09,"RuSciBenchGRNTIClassification":61.41,"RuSciBenchOECDClassification":48.67,"MassiveIntentClassification (ru)":60.08,"MassiveScenarioClassification (ru)":66.44}
|
all_data_tasks/33/default.jsonl
CHANGED
@@ -18,8 +18,8 @@
|
|
18 |
{"index":45,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/LaBSE\">LaBSE<\/a>","Model Size (Million Parameters)":471,"Memory Usage (GB, fp32)":1.75,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.71}
|
19 |
{"index":25,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.61}
|
20 |
{"index":11,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","TERRa":55.57}
|
21 |
-
{"index":
|
22 |
-
{"index":
|
23 |
{"index":33,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"TERRa":54.98}
|
24 |
{"index":8,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.17}
|
25 |
{"index":7,"Rank":25,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.02}
|
|
|
18 |
{"index":45,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/LaBSE\">LaBSE<\/a>","Model Size (Million Parameters)":471,"Memory Usage (GB, fp32)":1.75,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.71}
|
19 |
{"index":25,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.61}
|
20 |
{"index":11,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","TERRa":55.57}
|
21 |
+
{"index":15,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
|
22 |
+
{"index":36,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
|
23 |
{"index":33,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"TERRa":54.98}
|
24 |
{"index":8,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.17}
|
25 |
{"index":7,"Rank":25,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.02}
|
all_data_tasks/35/default.jsonl
CHANGED
@@ -7,22 +7,22 @@
|
|
7 |
{"index":33,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"Average":67.14,"RiaNewsRetrieval":70.24,"RuBQRetrieval":69.58,"MIRACLRetrieval (ru)":61.6}
|
8 |
{"index":18,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.51,"RiaNewsRetrieval":78.86,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":53.91}
|
9 |
{"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":66.17,"RiaNewsRetrieval":77.42,"RuBQRetrieval":67.75,"MIRACLRetrieval (ru)":53.33}
|
10 |
-
{"index":
|
11 |
-
{"index":
|
12 |
-
{"index":
|
13 |
-
{"index":
|
14 |
-
{"index":
|
15 |
-
{"index":
|
16 |
-
{"index":
|
17 |
-
{"index":
|
18 |
-
{"index":
|
19 |
-
{"index":
|
20 |
-
{"index":
|
21 |
-
{"index":
|
22 |
-
{"index":
|
23 |
-
{"index":
|
24 |
-
{"index":
|
25 |
-
{"index":
|
26 |
{"index":13,"Rank":29,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
27 |
{"index":15,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":"","RiaNewsRetrieval":70.01,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":""}
|
28 |
{"index":16,"Rank":32,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
|
|
7 |
{"index":33,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"Average":67.14,"RiaNewsRetrieval":70.24,"RuBQRetrieval":69.58,"MIRACLRetrieval (ru)":61.6}
|
8 |
{"index":18,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.51,"RiaNewsRetrieval":78.86,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":53.91}
|
9 |
{"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":66.17,"RiaNewsRetrieval":77.42,"RuBQRetrieval":67.75,"MIRACLRetrieval (ru)":53.33}
|
10 |
+
{"index":11,"Rank":10,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":66.13,"RiaNewsRetrieval":79.42,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":52.21}
|
11 |
+
{"index":36,"Rank":11,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":65.85,"RiaNewsRetrieval":70.0,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":59.01}
|
12 |
+
{"index":59,"Rank":12,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/LaBSE-ru-turbo\">LaBSE-ru-turbo<\/a>","Model Size (Million Parameters)":128,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":63.68,"RiaNewsRetrieval":69.36,"RuBQRetrieval":65.71,"MIRACLRetrieval (ru)":55.97}
|
13 |
+
{"index":28,"Rank":13,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-base\">USER-base<\/a>","Model Size (Million Parameters)":124,"Memory Usage (GB, fp32)":0.46,"Embedding Dimensions":768,"Max Tokens":512,"Average":56.64,"RiaNewsRetrieval":77.83,"RuBQRetrieval":56.86,"MIRACLRetrieval (ru)":35.22}
|
14 |
+
{"index":60,"Rank":14,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/rubert-tiny-turbo\">rubert-tiny-turbo<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":312,"Max Tokens":2048,"Average":46.69,"RiaNewsRetrieval":51.27,"RuBQRetrieval":51.73,"MIRACLRetrieval (ru)":37.07}
|
15 |
+
{"index":25,"Rank":15,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":24.78,"RiaNewsRetrieval":34.73,"RuBQRetrieval":29.03,"MIRACLRetrieval (ru)":10.58}
|
16 |
+
{"index":19,"Rank":16,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_mt_nlu_ru\">sbert_large_mt_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":514,"Average":19.13,"RiaNewsRetrieval":21.4,"RuBQRetrieval":29.8,"MIRACLRetrieval (ru)":6.2}
|
17 |
+
{"index":27,"Rank":17,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/rubert-tiny2\">rubert-tiny2<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":2048,"Max Tokens":514,"Average":8.89,"RiaNewsRetrieval":13.92,"RuBQRetrieval":10.87,"MIRACLRetrieval (ru)":1.89}
|
18 |
+
{"index":20,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_nlu_ru\">sbert_large_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":512,"Average":8.51,"RiaNewsRetrieval":11.11,"RuBQRetrieval":12.45,"MIRACLRetrieval (ru)":1.98}
|
19 |
+
{"index":9,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased-sentence\">rubert-base-cased-sentence<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":7.09,"RiaNewsRetrieval":6.72,"RuBQRetrieval":12.63,"MIRACLRetrieval (ru)":1.92}
|
20 |
+
{"index":7,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.71,"RiaNewsRetrieval":4.14,"RuBQRetrieval":10.6,"MIRACLRetrieval (ru)":2.39}
|
21 |
+
{"index":8,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.33,"RiaNewsRetrieval":5.58,"RuBQRetrieval":9.52,"MIRACLRetrieval (ru)":0.88}
|
22 |
+
{"index":47,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/all-MiniLM-L6-v2\">all-MiniLM-L6-v2<\/a>","Model Size (Million Parameters)":23,"Memory Usage (GB, fp32)":0.09,"Embedding Dimensions":384,"Max Tokens":512,"Average":1.23,"RiaNewsRetrieval":0.67,"RuBQRetrieval":2.64,"MIRACLRetrieval (ru)":0.39}
|
23 |
+
{"index":1,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/github.com\/facebookresearch\/LASER\">LASER2<\/a>","Model Size (Million Parameters)":43,"Memory Usage (GB, fp32)":0.16,"Embedding Dimensions":1024,"Max Tokens":"N\/A","Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
24 |
+
{"index":2,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Alibaba-NLP\/gte-multilingual-base\">gte-multilingual-base<\/a>","Model Size (Million Parameters)":305,"Memory Usage (GB, fp32)":1.14,"Embedding Dimensions":768,"Max Tokens":8192,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
25 |
+
{"index":10,"Rank":28,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":70.94,"MIRACLRetrieval (ru)":""}
|
26 |
{"index":13,"Rank":29,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
27 |
{"index":15,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":"","RiaNewsRetrieval":70.01,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":""}
|
28 |
{"index":16,"Rank":32,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
all_data_tasks/39/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/4/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/40/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/5/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/6/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/8/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
all_data_tasks/9/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/en/data_overall/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/en/data_tasks/Classification/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/en/data_tasks/Clustering/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/en/data_tasks/PairClassification/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/en/data_tasks/Reranking/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/en/data_tasks/Retrieval/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/en/data_tasks/STS/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/en/data_tasks/Summarization/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/other-cls/data_tasks/Classification/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/other-sts/data_tasks/STS/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/ru/data_overall/default.jsonl
CHANGED
@@ -2,26 +2,26 @@
|
|
2 |
{"index":35,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":66.03,"Classification Average (9 datasets)":66.31,"Clustering Average (3 datasets)":63.21,"PairClassification Average (1 datasets)":63.89,"Reranking Average (2 datasets)":69.17,"Retrieval Average (3 datasets)":74.41,"STS Average (3 datasets)":74.85,"MultilabelClassification Average (2 datasets)":41.15}
|
3 |
{"index":29,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average (23 datasets)":62.58,"Classification Average (9 datasets)":61.92,"Clustering Average (3 datasets)":53.61,"PairClassification Average (1 datasets)":65.07,"Reranking Average (2 datasets)":68.72,"Retrieval Average (3 datasets)":73.63,"STS Average (3 datasets)":75.38,"MultilabelClassification Average (2 datasets)":35.88}
|
4 |
{"index":18,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":61.77,"Classification Average (9 datasets)":62.74,"Clustering Average (3 datasets)":56.06,"PairClassification Average (1 datasets)":60.79,"Reranking Average (2 datasets)":63.89,"Retrieval Average (3 datasets)":66.51,"STS Average (3 datasets)":73.97,"MultilabelClassification Average (2 datasets)":38.88}
|
5 |
-
{"index":
|
6 |
-
{"index":
|
7 |
-
{"index":
|
8 |
-
{"index":
|
9 |
-
{"index":
|
10 |
-
{"index":
|
11 |
-
{"index":
|
12 |
-
{"index":
|
13 |
-
{"index":
|
14 |
-
{"index":
|
15 |
-
{"index":
|
16 |
-
{"index":
|
17 |
-
{"index":
|
18 |
-
{"index":
|
19 |
-
{"index":
|
20 |
-
{"index":
|
21 |
-
{"index":
|
22 |
-
{"index":
|
23 |
-
{"index":
|
24 |
-
{"index":
|
25 |
{"index":13,"Rank":28,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
|
26 |
{"index":15,"Rank":30,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average (23 datasets)":"","Classification Average (9 datasets)":56.45,"Clustering Average (3 datasets)":51.35,"PairClassification Average (1 datasets)":55.14,"Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":69.48,"MultilabelClassification Average (2 datasets)":31.44}
|
27 |
{"index":16,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
|
|
|
2 |
{"index":35,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":66.03,"Classification Average (9 datasets)":66.31,"Clustering Average (3 datasets)":63.21,"PairClassification Average (1 datasets)":63.89,"Reranking Average (2 datasets)":69.17,"Retrieval Average (3 datasets)":74.41,"STS Average (3 datasets)":74.85,"MultilabelClassification Average (2 datasets)":41.15}
|
3 |
{"index":29,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average (23 datasets)":62.58,"Classification Average (9 datasets)":61.92,"Clustering Average (3 datasets)":53.61,"PairClassification Average (1 datasets)":65.07,"Reranking Average (2 datasets)":68.72,"Retrieval Average (3 datasets)":73.63,"STS Average (3 datasets)":75.38,"MultilabelClassification Average (2 datasets)":35.88}
|
4 |
{"index":18,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":61.77,"Classification Average (9 datasets)":62.74,"Clustering Average (3 datasets)":56.06,"PairClassification Average (1 datasets)":60.79,"Reranking Average (2 datasets)":63.89,"Retrieval Average (3 datasets)":66.51,"STS Average (3 datasets)":73.97,"MultilabelClassification Average (2 datasets)":38.88}
|
5 |
+
{"index":11,"Rank":5,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average (23 datasets)":61.64,"Classification Average (9 datasets)":65.21,"Clustering Average (3 datasets)":55.28,"PairClassification Average (1 datasets)":55.57,"Reranking Average (2 datasets)":62.3,"Retrieval Average (3 datasets)":66.13,"STS Average (3 datasets)":71.18,"MultilabelClassification Average (2 datasets)":36.46}
|
6 |
+
{"index":3,"Rank":6,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/BAAI\/bge-m3\">bge-m3<\/a>","Model Size (Million Parameters)":567,"Memory Usage (GB, fp32)":2.11,"Embedding Dimensions":1024,"Max Tokens":8192,"Average (23 datasets)":61.58,"Classification Average (9 datasets)":60.46,"Clustering Average (3 datasets)":52.38,"PairClassification Average (1 datasets)":60.6,"Reranking Average (2 datasets)":69.7,"Retrieval Average (3 datasets)":74.79,"STS Average (3 datasets)":73.68,"MultilabelClassification Average (2 datasets)":34.86}
|
7 |
+
{"index":34,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large\">multilingual-e5-large<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":61.41,"Classification Average (9 datasets)":61.01,"Clustering Average (3 datasets)":52.23,"PairClassification Average (1 datasets)":58.42,"Reranking Average (2 datasets)":69.66,"Retrieval Average (3 datasets)":74.04,"STS Average (3 datasets)":71.62,"MultilabelClassification Average (2 datasets)":36.01}
|
8 |
+
{"index":12,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average (23 datasets)":60.48,"Classification Average (9 datasets)":61.58,"Clustering Average (3 datasets)":55.67,"PairClassification Average (1 datasets)":57.71,"Reranking Average (2 datasets)":63.16,"Retrieval Average (3 datasets)":66.17,"STS Average (3 datasets)":71.7,"MultilabelClassification Average (2 datasets)":36.1}
|
9 |
+
{"index":59,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/LaBSE-ru-turbo\">LaBSE-ru-turbo<\/a>","Model Size (Million Parameters)":128,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":59.04,"Classification Average (9 datasets)":59.23,"Clustering Average (3 datasets)":53.22,"PairClassification Average (1 datasets)":57.81,"Reranking Average (2 datasets)":63.04,"Retrieval Average (3 datasets)":63.68,"STS Average (3 datasets)":72.54,"MultilabelClassification Average (2 datasets)":36.32}
|
10 |
+
{"index":33,"Rank":10,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"Average (23 datasets)":58.34,"Classification Average (9 datasets)":58.26,"Clustering Average (3 datasets)":50.27,"PairClassification Average (1 datasets)":54.98,"Reranking Average (2 datasets)":66.24,"Retrieval Average (3 datasets)":67.14,"STS Average (3 datasets)":70.16,"MultilabelClassification Average (2 datasets)":33.65}
|
11 |
+
{"index":28,"Rank":11,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-base\">USER-base<\/a>","Model Size (Million Parameters)":124,"Memory Usage (GB, fp32)":0.46,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":57.98,"Classification Average (9 datasets)":59.88,"Clustering Average (3 datasets)":53.42,"PairClassification Average (1 datasets)":60.11,"Reranking Average (2 datasets)":55.58,"Retrieval Average (3 datasets)":56.64,"STS Average (3 datasets)":73.07,"MultilabelClassification Average (2 datasets)":36.98}
|
12 |
+
{"index":36,"Rank":12,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average (23 datasets)":57.29,"Classification Average (9 datasets)":56.44,"Clustering Average (3 datasets)":51.35,"PairClassification Average (1 datasets)":55.14,"Reranking Average (2 datasets)":65.28,"Retrieval Average (3 datasets)":65.85,"STS Average (3 datasets)":69.48,"MultilabelClassification Average (2 datasets)":31.99}
|
13 |
+
{"index":60,"Rank":13,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/rubert-tiny-turbo\">rubert-tiny-turbo<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":312,"Max Tokens":2048,"Average (23 datasets)":53.22,"Classification Average (9 datasets)":55.01,"Clustering Average (3 datasets)":49.57,"PairClassification Average (1 datasets)":56.27,"Reranking Average (2 datasets)":54.94,"Retrieval Average (3 datasets)":46.69,"STS Average (3 datasets)":70.23,"MultilabelClassification Average (2 datasets)":31.7}
|
14 |
+
{"index":19,"Rank":14,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_mt_nlu_ru\">sbert_large_mt_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":48.72,"Classification Average (9 datasets)":57.52,"Clustering Average (3 datasets)":51.29,"PairClassification Average (1 datasets)":51.97,"Reranking Average (2 datasets)":40.56,"Retrieval Average (3 datasets)":19.13,"STS Average (3 datasets)":64.4,"MultilabelClassification Average (2 datasets)":32.67}
|
15 |
+
{"index":25,"Rank":15,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":48.23,"Classification Average (9 datasets)":54.98,"Clustering Average (3 datasets)":46.84,"PairClassification Average (1 datasets)":55.61,"Reranking Average (2 datasets)":41.84,"Retrieval Average (3 datasets)":24.78,"STS Average (3 datasets)":65.91,"MultilabelClassification Average (2 datasets)":31.27}
|
16 |
+
{"index":20,"Rank":16,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_nlu_ru\">sbert_large_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":512,"Average (23 datasets)":45.35,"Classification Average (9 datasets)":57.24,"Clustering Average (3 datasets)":50.43,"PairClassification Average (1 datasets)":50.17,"Reranking Average (2 datasets)":32.8,"Retrieval Average (3 datasets)":8.51,"STS Average (3 datasets)":57.21,"MultilabelClassification Average (2 datasets)":31.87}
|
17 |
+
{"index":27,"Rank":17,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/rubert-tiny2\">rubert-tiny2<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":2048,"Max Tokens":514,"Average (23 datasets)":42.22,"Classification Average (9 datasets)":52.17,"Clustering Average (3 datasets)":39.11,"PairClassification Average (1 datasets)":51.87,"Reranking Average (2 datasets)":30.95,"Retrieval Average (3 datasets)":8.89,"STS Average (3 datasets)":61.6,"MultilabelClassification Average (2 datasets)":29.45}
|
18 |
+
{"index":7,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":41.94,"Classification Average (9 datasets)":55.15,"Clustering Average (3 datasets)":38.41,"PairClassification Average (1 datasets)":53.02,"Reranking Average (2 datasets)":27.84,"Retrieval Average (3 datasets)":5.71,"STS Average (3 datasets)":56.2,"MultilabelClassification Average (2 datasets)":29.32}
|
19 |
+
{"index":9,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased-sentence\">rubert-base-cased-sentence<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":41.91,"Classification Average (9 datasets)":51.27,"Clustering Average (3 datasets)":43.13,"PairClassification Average (1 datasets)":59.12,"Reranking Average (2 datasets)":26.83,"Retrieval Average (3 datasets)":7.09,"STS Average (3 datasets)":61.18,"MultilabelClassification Average (2 datasets)":27.8}
|
20 |
+
{"index":8,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":37.53,"Classification Average (9 datasets)":51.6,"Clustering Average (3 datasets)":27.91,"PairClassification Average (1 datasets)":53.17,"Reranking Average (2 datasets)":27.46,"Retrieval Average (3 datasets)":5.33,"STS Average (3 datasets)":46.22,"MultilabelClassification Average (2 datasets)":26.2}
|
21 |
+
{"index":0,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/docs.voyageai.com\/embeddings\/\">voyage-3<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":1024,"Max Tokens":32000,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":76.0,"STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
|
22 |
+
{"index":1,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/github.com\/facebookresearch\/LASER\">LASER2<\/a>","Model Size (Million Parameters)":43,"Memory Usage (GB, fp32)":0.16,"Embedding Dimensions":1024,"Max Tokens":"N\/A","Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
|
23 |
+
{"index":2,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Alibaba-NLP\/gte-multilingual-base\">gte-multilingual-base<\/a>","Model Size (Million Parameters)":305,"Memory Usage (GB, fp32)":1.14,"Embedding Dimensions":768,"Max Tokens":8192,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
|
24 |
+
{"index":10,"Rank":27,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average (23 datasets)":"","Classification Average (9 datasets)":67.48,"Clustering Average (3 datasets)":61.91,"PairClassification Average (1 datasets)":58.93,"Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":35.6}
|
25 |
{"index":13,"Rank":28,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
|
26 |
{"index":15,"Rank":30,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average (23 datasets)":"","Classification Average (9 datasets)":56.45,"Clustering Average (3 datasets)":51.35,"PairClassification Average (1 datasets)":55.14,"Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":69.48,"MultilabelClassification Average (2 datasets)":31.44}
|
27 |
{"index":16,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
|
boards_data/ru/data_tasks/Classification/default.jsonl
CHANGED
@@ -2,8 +2,8 @@
|
|
2 |
{"index":10,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":67.48,"GeoreviewClassification":53.47,"HeadlineClassification":85.66,"InappropriatenessClassification":65.29,"KinopoiskClassification":64.25,"RuReviewsClassification":68.58,"RuSciBenchGRNTIClassification":64.56,"RuSciBenchOECDClassification":51.2,"MassiveIntentClassification (ru)":76.01,"MassiveScenarioClassification (ru)":78.28}
|
3 |
{"index":35,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.31,"GeoreviewClassification":55.9,"HeadlineClassification":86.18,"InappropriatenessClassification":65.53,"KinopoiskClassification":66.12,"RuReviewsClassification":68.56,"RuSciBenchGRNTIClassification":65.07,"RuSciBenchOECDClassification":50.21,"MassiveIntentClassification (ru)":67.6,"MassiveScenarioClassification (ru)":71.59}
|
4 |
{"index":11,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":65.21,"GeoreviewClassification":52.04,"HeadlineClassification":83.4,"InappropriatenessClassification":64.14,"KinopoiskClassification":63.15,"RuReviewsClassification":68.24,"RuSciBenchGRNTIClassification":61.24,"RuSciBenchOECDClassification":48.1,"MassiveIntentClassification (ru)":68.61,"MassiveScenarioClassification (ru)":77.98}
|
5 |
-
{"index":
|
6 |
-
{"index":
|
7 |
{"index":18,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":62.74,"GeoreviewClassification":49.7,"HeadlineClassification":78.0,"InappropriatenessClassification":61.32,"KinopoiskClassification":63.27,"RuReviewsClassification":67.96,"RuSciBenchGRNTIClassification":59.33,"RuSciBenchOECDClassification":46.33,"MassiveIntentClassification (ru)":66.97,"MassiveScenarioClassification (ru)":71.8}
|
8 |
{"index":29,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average":61.92,"GeoreviewClassification":50.98,"HeadlineClassification":70.09,"InappropriatenessClassification":60.76,"KinopoiskClassification":63.33,"RuReviewsClassification":68.52,"RuSciBenchGRNTIClassification":57.67,"RuSciBenchOECDClassification":44.2,"MassiveIntentClassification (ru)":68.85,"MassiveScenarioClassification (ru)":72.9}
|
9 |
{"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":61.58,"GeoreviewClassification":47.69,"HeadlineClassification":83.46,"InappropriatenessClassification":61.32,"KinopoiskClassification":59.04,"RuReviewsClassification":66.09,"RuSciBenchGRNTIClassification":61.41,"RuSciBenchOECDClassification":48.67,"MassiveIntentClassification (ru)":60.08,"MassiveScenarioClassification (ru)":66.44}
|
|
|
2 |
{"index":10,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":67.48,"GeoreviewClassification":53.47,"HeadlineClassification":85.66,"InappropriatenessClassification":65.29,"KinopoiskClassification":64.25,"RuReviewsClassification":68.58,"RuSciBenchGRNTIClassification":64.56,"RuSciBenchOECDClassification":51.2,"MassiveIntentClassification (ru)":76.01,"MassiveScenarioClassification (ru)":78.28}
|
3 |
{"index":35,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.31,"GeoreviewClassification":55.9,"HeadlineClassification":86.18,"InappropriatenessClassification":65.53,"KinopoiskClassification":66.12,"RuReviewsClassification":68.56,"RuSciBenchGRNTIClassification":65.07,"RuSciBenchOECDClassification":50.21,"MassiveIntentClassification (ru)":67.6,"MassiveScenarioClassification (ru)":71.59}
|
4 |
{"index":11,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":65.21,"GeoreviewClassification":52.04,"HeadlineClassification":83.4,"InappropriatenessClassification":64.14,"KinopoiskClassification":63.15,"RuReviewsClassification":68.24,"RuSciBenchGRNTIClassification":61.24,"RuSciBenchOECDClassification":48.1,"MassiveIntentClassification (ru)":68.61,"MassiveScenarioClassification (ru)":77.98}
|
5 |
+
{"index":22,"Rank":5,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/arkohut\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
|
6 |
+
{"index":39,"Rank":6,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/jinaai\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
|
7 |
{"index":18,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":62.74,"GeoreviewClassification":49.7,"HeadlineClassification":78.0,"InappropriatenessClassification":61.32,"KinopoiskClassification":63.27,"RuReviewsClassification":67.96,"RuSciBenchGRNTIClassification":59.33,"RuSciBenchOECDClassification":46.33,"MassiveIntentClassification (ru)":66.97,"MassiveScenarioClassification (ru)":71.8}
|
8 |
{"index":29,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average":61.92,"GeoreviewClassification":50.98,"HeadlineClassification":70.09,"InappropriatenessClassification":60.76,"KinopoiskClassification":63.33,"RuReviewsClassification":68.52,"RuSciBenchGRNTIClassification":57.67,"RuSciBenchOECDClassification":44.2,"MassiveIntentClassification (ru)":68.85,"MassiveScenarioClassification (ru)":72.9}
|
9 |
{"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":61.58,"GeoreviewClassification":47.69,"HeadlineClassification":83.46,"InappropriatenessClassification":61.32,"KinopoiskClassification":59.04,"RuReviewsClassification":66.09,"RuSciBenchGRNTIClassification":61.41,"RuSciBenchOECDClassification":48.67,"MassiveIntentClassification (ru)":60.08,"MassiveScenarioClassification (ru)":66.44}
|
boards_data/ru/data_tasks/PairClassification/default.jsonl
CHANGED
@@ -18,8 +18,8 @@
|
|
18 |
{"index":45,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/LaBSE\">LaBSE<\/a>","Model Size (Million Parameters)":471,"Memory Usage (GB, fp32)":1.75,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.71}
|
19 |
{"index":25,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.61}
|
20 |
{"index":11,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","TERRa":55.57}
|
21 |
-
{"index":
|
22 |
-
{"index":
|
23 |
{"index":33,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"TERRa":54.98}
|
24 |
{"index":8,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.17}
|
25 |
{"index":7,"Rank":25,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.02}
|
|
|
18 |
{"index":45,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/LaBSE\">LaBSE<\/a>","Model Size (Million Parameters)":471,"Memory Usage (GB, fp32)":1.75,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.71}
|
19 |
{"index":25,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.61}
|
20 |
{"index":11,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","TERRa":55.57}
|
21 |
+
{"index":15,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
|
22 |
+
{"index":36,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
|
23 |
{"index":33,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"TERRa":54.98}
|
24 |
{"index":8,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.17}
|
25 |
{"index":7,"Rank":25,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.02}
|
boards_data/ru/data_tasks/Retrieval/default.jsonl
CHANGED
@@ -7,22 +7,22 @@
|
|
7 |
{"index":33,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"Average":67.14,"RiaNewsRetrieval":70.24,"RuBQRetrieval":69.58,"MIRACLRetrieval (ru)":61.6}
|
8 |
{"index":18,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.51,"RiaNewsRetrieval":78.86,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":53.91}
|
9 |
{"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":66.17,"RiaNewsRetrieval":77.42,"RuBQRetrieval":67.75,"MIRACLRetrieval (ru)":53.33}
|
10 |
-
{"index":
|
11 |
-
{"index":
|
12 |
-
{"index":
|
13 |
-
{"index":
|
14 |
-
{"index":
|
15 |
-
{"index":
|
16 |
-
{"index":
|
17 |
-
{"index":
|
18 |
-
{"index":
|
19 |
-
{"index":
|
20 |
-
{"index":
|
21 |
-
{"index":
|
22 |
-
{"index":
|
23 |
-
{"index":
|
24 |
-
{"index":
|
25 |
-
{"index":
|
26 |
{"index":13,"Rank":29,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
27 |
{"index":15,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":"","RiaNewsRetrieval":70.01,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":""}
|
28 |
{"index":16,"Rank":32,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
|
|
7 |
{"index":33,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"Average":67.14,"RiaNewsRetrieval":70.24,"RuBQRetrieval":69.58,"MIRACLRetrieval (ru)":61.6}
|
8 |
{"index":18,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.51,"RiaNewsRetrieval":78.86,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":53.91}
|
9 |
{"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":66.17,"RiaNewsRetrieval":77.42,"RuBQRetrieval":67.75,"MIRACLRetrieval (ru)":53.33}
|
10 |
+
{"index":11,"Rank":10,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":66.13,"RiaNewsRetrieval":79.42,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":52.21}
|
11 |
+
{"index":36,"Rank":11,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":65.85,"RiaNewsRetrieval":70.0,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":59.01}
|
12 |
+
{"index":59,"Rank":12,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/LaBSE-ru-turbo\">LaBSE-ru-turbo<\/a>","Model Size (Million Parameters)":128,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":63.68,"RiaNewsRetrieval":69.36,"RuBQRetrieval":65.71,"MIRACLRetrieval (ru)":55.97}
|
13 |
+
{"index":28,"Rank":13,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-base\">USER-base<\/a>","Model Size (Million Parameters)":124,"Memory Usage (GB, fp32)":0.46,"Embedding Dimensions":768,"Max Tokens":512,"Average":56.64,"RiaNewsRetrieval":77.83,"RuBQRetrieval":56.86,"MIRACLRetrieval (ru)":35.22}
|
14 |
+
{"index":60,"Rank":14,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/rubert-tiny-turbo\">rubert-tiny-turbo<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":312,"Max Tokens":2048,"Average":46.69,"RiaNewsRetrieval":51.27,"RuBQRetrieval":51.73,"MIRACLRetrieval (ru)":37.07}
|
15 |
+
{"index":25,"Rank":15,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":24.78,"RiaNewsRetrieval":34.73,"RuBQRetrieval":29.03,"MIRACLRetrieval (ru)":10.58}
|
16 |
+
{"index":19,"Rank":16,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_mt_nlu_ru\">sbert_large_mt_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":514,"Average":19.13,"RiaNewsRetrieval":21.4,"RuBQRetrieval":29.8,"MIRACLRetrieval (ru)":6.2}
|
17 |
+
{"index":27,"Rank":17,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/rubert-tiny2\">rubert-tiny2<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":2048,"Max Tokens":514,"Average":8.89,"RiaNewsRetrieval":13.92,"RuBQRetrieval":10.87,"MIRACLRetrieval (ru)":1.89}
|
18 |
+
{"index":20,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_nlu_ru\">sbert_large_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":512,"Average":8.51,"RiaNewsRetrieval":11.11,"RuBQRetrieval":12.45,"MIRACLRetrieval (ru)":1.98}
|
19 |
+
{"index":9,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased-sentence\">rubert-base-cased-sentence<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":7.09,"RiaNewsRetrieval":6.72,"RuBQRetrieval":12.63,"MIRACLRetrieval (ru)":1.92}
|
20 |
+
{"index":7,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.71,"RiaNewsRetrieval":4.14,"RuBQRetrieval":10.6,"MIRACLRetrieval (ru)":2.39}
|
21 |
+
{"index":8,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.33,"RiaNewsRetrieval":5.58,"RuBQRetrieval":9.52,"MIRACLRetrieval (ru)":0.88}
|
22 |
+
{"index":47,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/all-MiniLM-L6-v2\">all-MiniLM-L6-v2<\/a>","Model Size (Million Parameters)":23,"Memory Usage (GB, fp32)":0.09,"Embedding Dimensions":384,"Max Tokens":512,"Average":1.23,"RiaNewsRetrieval":0.67,"RuBQRetrieval":2.64,"MIRACLRetrieval (ru)":0.39}
|
23 |
+
{"index":1,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/github.com\/facebookresearch\/LASER\">LASER2<\/a>","Model Size (Million Parameters)":43,"Memory Usage (GB, fp32)":0.16,"Embedding Dimensions":1024,"Max Tokens":"N\/A","Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
24 |
+
{"index":2,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Alibaba-NLP\/gte-multilingual-base\">gte-multilingual-base<\/a>","Model Size (Million Parameters)":305,"Memory Usage (GB, fp32)":1.14,"Embedding Dimensions":768,"Max Tokens":8192,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
25 |
+
{"index":10,"Rank":28,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":70.94,"MIRACLRetrieval (ru)":""}
|
26 |
{"index":13,"Rank":29,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
27 |
{"index":15,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":"","RiaNewsRetrieval":70.01,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":""}
|
28 |
{"index":16,"Rank":32,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
|
boards_data/zh/data_overall/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/zh/data_tasks/Classification/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/zh/data_tasks/Clustering/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/zh/data_tasks/PairClassification/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/zh/data_tasks/Reranking/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/zh/data_tasks/Retrieval/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
boards_data/zh/data_tasks/STS/default.jsonl
CHANGED
The diff for this file is too large to render.
See raw diff
|
|