orionweller commited on
Commit
f632e87
1 Parent(s): c49e5d1

Automated Leaderboard Update

Browse files
Files changed (39) hide show
  1. all_data_tasks/0/default.jsonl +0 -0
  2. all_data_tasks/1/default.jsonl +0 -0
  3. all_data_tasks/10/default.jsonl +0 -0
  4. all_data_tasks/11/default.jsonl +0 -0
  5. all_data_tasks/12/default.jsonl +0 -0
  6. all_data_tasks/13/default.jsonl +0 -0
  7. all_data_tasks/2/default.jsonl +0 -0
  8. all_data_tasks/3/default.jsonl +0 -0
  9. all_data_tasks/31/default.jsonl +2 -2
  10. all_data_tasks/33/default.jsonl +2 -2
  11. all_data_tasks/35/default.jsonl +16 -16
  12. all_data_tasks/39/default.jsonl +0 -0
  13. all_data_tasks/4/default.jsonl +0 -0
  14. all_data_tasks/40/default.jsonl +0 -0
  15. all_data_tasks/5/default.jsonl +0 -0
  16. all_data_tasks/6/default.jsonl +0 -0
  17. all_data_tasks/8/default.jsonl +0 -0
  18. all_data_tasks/9/default.jsonl +0 -0
  19. boards_data/en/data_overall/default.jsonl +0 -0
  20. boards_data/en/data_tasks/Classification/default.jsonl +0 -0
  21. boards_data/en/data_tasks/Clustering/default.jsonl +0 -0
  22. boards_data/en/data_tasks/PairClassification/default.jsonl +0 -0
  23. boards_data/en/data_tasks/Reranking/default.jsonl +0 -0
  24. boards_data/en/data_tasks/Retrieval/default.jsonl +0 -0
  25. boards_data/en/data_tasks/STS/default.jsonl +0 -0
  26. boards_data/en/data_tasks/Summarization/default.jsonl +0 -0
  27. boards_data/other-cls/data_tasks/Classification/default.jsonl +0 -0
  28. boards_data/other-sts/data_tasks/STS/default.jsonl +0 -0
  29. boards_data/ru/data_overall/default.jsonl +20 -20
  30. boards_data/ru/data_tasks/Classification/default.jsonl +2 -2
  31. boards_data/ru/data_tasks/PairClassification/default.jsonl +2 -2
  32. boards_data/ru/data_tasks/Retrieval/default.jsonl +16 -16
  33. boards_data/zh/data_overall/default.jsonl +0 -0
  34. boards_data/zh/data_tasks/Classification/default.jsonl +0 -0
  35. boards_data/zh/data_tasks/Clustering/default.jsonl +0 -0
  36. boards_data/zh/data_tasks/PairClassification/default.jsonl +0 -0
  37. boards_data/zh/data_tasks/Reranking/default.jsonl +0 -0
  38. boards_data/zh/data_tasks/Retrieval/default.jsonl +0 -0
  39. boards_data/zh/data_tasks/STS/default.jsonl +0 -0
all_data_tasks/0/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/1/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/10/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/11/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/12/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/13/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/2/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/3/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/31/default.jsonl CHANGED
@@ -2,8 +2,8 @@
2
  {"index":10,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":67.48,"GeoreviewClassification":53.47,"HeadlineClassification":85.66,"InappropriatenessClassification":65.29,"KinopoiskClassification":64.25,"RuReviewsClassification":68.58,"RuSciBenchGRNTIClassification":64.56,"RuSciBenchOECDClassification":51.2,"MassiveIntentClassification (ru)":76.01,"MassiveScenarioClassification (ru)":78.28}
3
  {"index":35,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.31,"GeoreviewClassification":55.9,"HeadlineClassification":86.18,"InappropriatenessClassification":65.53,"KinopoiskClassification":66.12,"RuReviewsClassification":68.56,"RuSciBenchGRNTIClassification":65.07,"RuSciBenchOECDClassification":50.21,"MassiveIntentClassification (ru)":67.6,"MassiveScenarioClassification (ru)":71.59}
4
  {"index":11,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":65.21,"GeoreviewClassification":52.04,"HeadlineClassification":83.4,"InappropriatenessClassification":64.14,"KinopoiskClassification":63.15,"RuReviewsClassification":68.24,"RuSciBenchGRNTIClassification":61.24,"RuSciBenchOECDClassification":48.1,"MassiveIntentClassification (ru)":68.61,"MassiveScenarioClassification (ru)":77.98}
5
- {"index":39,"Rank":5,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/jinaai\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
6
- {"index":22,"Rank":6,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/arkohut\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
7
  {"index":18,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":62.74,"GeoreviewClassification":49.7,"HeadlineClassification":78.0,"InappropriatenessClassification":61.32,"KinopoiskClassification":63.27,"RuReviewsClassification":67.96,"RuSciBenchGRNTIClassification":59.33,"RuSciBenchOECDClassification":46.33,"MassiveIntentClassification (ru)":66.97,"MassiveScenarioClassification (ru)":71.8}
8
  {"index":29,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average":61.92,"GeoreviewClassification":50.98,"HeadlineClassification":70.09,"InappropriatenessClassification":60.76,"KinopoiskClassification":63.33,"RuReviewsClassification":68.52,"RuSciBenchGRNTIClassification":57.67,"RuSciBenchOECDClassification":44.2,"MassiveIntentClassification (ru)":68.85,"MassiveScenarioClassification (ru)":72.9}
9
  {"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":61.58,"GeoreviewClassification":47.69,"HeadlineClassification":83.46,"InappropriatenessClassification":61.32,"KinopoiskClassification":59.04,"RuReviewsClassification":66.09,"RuSciBenchGRNTIClassification":61.41,"RuSciBenchOECDClassification":48.67,"MassiveIntentClassification (ru)":60.08,"MassiveScenarioClassification (ru)":66.44}
 
2
  {"index":10,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":67.48,"GeoreviewClassification":53.47,"HeadlineClassification":85.66,"InappropriatenessClassification":65.29,"KinopoiskClassification":64.25,"RuReviewsClassification":68.58,"RuSciBenchGRNTIClassification":64.56,"RuSciBenchOECDClassification":51.2,"MassiveIntentClassification (ru)":76.01,"MassiveScenarioClassification (ru)":78.28}
3
  {"index":35,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.31,"GeoreviewClassification":55.9,"HeadlineClassification":86.18,"InappropriatenessClassification":65.53,"KinopoiskClassification":66.12,"RuReviewsClassification":68.56,"RuSciBenchGRNTIClassification":65.07,"RuSciBenchOECDClassification":50.21,"MassiveIntentClassification (ru)":67.6,"MassiveScenarioClassification (ru)":71.59}
4
  {"index":11,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":65.21,"GeoreviewClassification":52.04,"HeadlineClassification":83.4,"InappropriatenessClassification":64.14,"KinopoiskClassification":63.15,"RuReviewsClassification":68.24,"RuSciBenchGRNTIClassification":61.24,"RuSciBenchOECDClassification":48.1,"MassiveIntentClassification (ru)":68.61,"MassiveScenarioClassification (ru)":77.98}
5
+ {"index":22,"Rank":5,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/arkohut\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
6
+ {"index":39,"Rank":6,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/jinaai\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
7
  {"index":18,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":62.74,"GeoreviewClassification":49.7,"HeadlineClassification":78.0,"InappropriatenessClassification":61.32,"KinopoiskClassification":63.27,"RuReviewsClassification":67.96,"RuSciBenchGRNTIClassification":59.33,"RuSciBenchOECDClassification":46.33,"MassiveIntentClassification (ru)":66.97,"MassiveScenarioClassification (ru)":71.8}
8
  {"index":29,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average":61.92,"GeoreviewClassification":50.98,"HeadlineClassification":70.09,"InappropriatenessClassification":60.76,"KinopoiskClassification":63.33,"RuReviewsClassification":68.52,"RuSciBenchGRNTIClassification":57.67,"RuSciBenchOECDClassification":44.2,"MassiveIntentClassification (ru)":68.85,"MassiveScenarioClassification (ru)":72.9}
9
  {"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":61.58,"GeoreviewClassification":47.69,"HeadlineClassification":83.46,"InappropriatenessClassification":61.32,"KinopoiskClassification":59.04,"RuReviewsClassification":66.09,"RuSciBenchGRNTIClassification":61.41,"RuSciBenchOECDClassification":48.67,"MassiveIntentClassification (ru)":60.08,"MassiveScenarioClassification (ru)":66.44}
all_data_tasks/33/default.jsonl CHANGED
@@ -18,8 +18,8 @@
18
  {"index":45,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/LaBSE\">LaBSE<\/a>","Model Size (Million Parameters)":471,"Memory Usage (GB, fp32)":1.75,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.71}
19
  {"index":25,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.61}
20
  {"index":11,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","TERRa":55.57}
21
- {"index":36,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
22
- {"index":15,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
23
  {"index":33,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"TERRa":54.98}
24
  {"index":8,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.17}
25
  {"index":7,"Rank":25,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.02}
 
18
  {"index":45,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/LaBSE\">LaBSE<\/a>","Model Size (Million Parameters)":471,"Memory Usage (GB, fp32)":1.75,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.71}
19
  {"index":25,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.61}
20
  {"index":11,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","TERRa":55.57}
21
+ {"index":15,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
22
+ {"index":36,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
23
  {"index":33,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"TERRa":54.98}
24
  {"index":8,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.17}
25
  {"index":7,"Rank":25,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.02}
all_data_tasks/35/default.jsonl CHANGED
@@ -7,22 +7,22 @@
7
  {"index":33,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"Average":67.14,"RiaNewsRetrieval":70.24,"RuBQRetrieval":69.58,"MIRACLRetrieval (ru)":61.6}
8
  {"index":18,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.51,"RiaNewsRetrieval":78.86,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":53.91}
9
  {"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":66.17,"RiaNewsRetrieval":77.42,"RuBQRetrieval":67.75,"MIRACLRetrieval (ru)":53.33}
10
- {"index":36,"Rank":10,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":65.85,"RiaNewsRetrieval":70.0,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":59.01}
11
- {"index":59,"Rank":11,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/LaBSE-ru-turbo\">LaBSE-ru-turbo<\/a>","Model Size (Million Parameters)":128,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":63.68,"RiaNewsRetrieval":69.36,"RuBQRetrieval":65.71,"MIRACLRetrieval (ru)":55.97}
12
- {"index":28,"Rank":12,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-base\">USER-base<\/a>","Model Size (Million Parameters)":124,"Memory Usage (GB, fp32)":0.46,"Embedding Dimensions":768,"Max Tokens":512,"Average":56.64,"RiaNewsRetrieval":77.83,"RuBQRetrieval":56.86,"MIRACLRetrieval (ru)":35.22}
13
- {"index":60,"Rank":13,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/rubert-tiny-turbo\">rubert-tiny-turbo<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":312,"Max Tokens":2048,"Average":46.69,"RiaNewsRetrieval":51.27,"RuBQRetrieval":51.73,"MIRACLRetrieval (ru)":37.07}
14
- {"index":25,"Rank":14,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":24.78,"RiaNewsRetrieval":34.73,"RuBQRetrieval":29.03,"MIRACLRetrieval (ru)":10.58}
15
- {"index":19,"Rank":15,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_mt_nlu_ru\">sbert_large_mt_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":514,"Average":19.13,"RiaNewsRetrieval":21.4,"RuBQRetrieval":29.8,"MIRACLRetrieval (ru)":6.2}
16
- {"index":27,"Rank":16,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/rubert-tiny2\">rubert-tiny2<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":2048,"Max Tokens":514,"Average":8.89,"RiaNewsRetrieval":13.92,"RuBQRetrieval":10.87,"MIRACLRetrieval (ru)":1.89}
17
- {"index":20,"Rank":17,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_nlu_ru\">sbert_large_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":512,"Average":8.51,"RiaNewsRetrieval":11.11,"RuBQRetrieval":12.45,"MIRACLRetrieval (ru)":1.98}
18
- {"index":9,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased-sentence\">rubert-base-cased-sentence<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":7.09,"RiaNewsRetrieval":6.72,"RuBQRetrieval":12.63,"MIRACLRetrieval (ru)":1.92}
19
- {"index":7,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.71,"RiaNewsRetrieval":4.14,"RuBQRetrieval":10.6,"MIRACLRetrieval (ru)":2.39}
20
- {"index":8,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.33,"RiaNewsRetrieval":5.58,"RuBQRetrieval":9.52,"MIRACLRetrieval (ru)":0.88}
21
- {"index":47,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/all-MiniLM-L6-v2\">all-MiniLM-L6-v2<\/a>","Model Size (Million Parameters)":23,"Memory Usage (GB, fp32)":0.09,"Embedding Dimensions":384,"Max Tokens":512,"Average":1.23,"RiaNewsRetrieval":0.67,"RuBQRetrieval":2.64,"MIRACLRetrieval (ru)":0.39}
22
- {"index":1,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/github.com\/facebookresearch\/LASER\">LASER2<\/a>","Model Size (Million Parameters)":43,"Memory Usage (GB, fp32)":0.16,"Embedding Dimensions":1024,"Max Tokens":"N\/A","Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
23
- {"index":2,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Alibaba-NLP\/gte-multilingual-base\">gte-multilingual-base<\/a>","Model Size (Million Parameters)":305,"Memory Usage (GB, fp32)":1.14,"Embedding Dimensions":768,"Max Tokens":8192,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
24
- {"index":10,"Rank":27,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":70.94,"MIRACLRetrieval (ru)":""}
25
- {"index":11,"Rank":28,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":"","RiaNewsRetrieval":79.42,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":""}
26
  {"index":13,"Rank":29,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
27
  {"index":15,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":"","RiaNewsRetrieval":70.01,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":""}
28
  {"index":16,"Rank":32,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
 
7
  {"index":33,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"Average":67.14,"RiaNewsRetrieval":70.24,"RuBQRetrieval":69.58,"MIRACLRetrieval (ru)":61.6}
8
  {"index":18,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.51,"RiaNewsRetrieval":78.86,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":53.91}
9
  {"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":66.17,"RiaNewsRetrieval":77.42,"RuBQRetrieval":67.75,"MIRACLRetrieval (ru)":53.33}
10
+ {"index":11,"Rank":10,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":66.13,"RiaNewsRetrieval":79.42,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":52.21}
11
+ {"index":36,"Rank":11,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":65.85,"RiaNewsRetrieval":70.0,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":59.01}
12
+ {"index":59,"Rank":12,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/LaBSE-ru-turbo\">LaBSE-ru-turbo<\/a>","Model Size (Million Parameters)":128,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":63.68,"RiaNewsRetrieval":69.36,"RuBQRetrieval":65.71,"MIRACLRetrieval (ru)":55.97}
13
+ {"index":28,"Rank":13,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-base\">USER-base<\/a>","Model Size (Million Parameters)":124,"Memory Usage (GB, fp32)":0.46,"Embedding Dimensions":768,"Max Tokens":512,"Average":56.64,"RiaNewsRetrieval":77.83,"RuBQRetrieval":56.86,"MIRACLRetrieval (ru)":35.22}
14
+ {"index":60,"Rank":14,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/rubert-tiny-turbo\">rubert-tiny-turbo<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":312,"Max Tokens":2048,"Average":46.69,"RiaNewsRetrieval":51.27,"RuBQRetrieval":51.73,"MIRACLRetrieval (ru)":37.07}
15
+ {"index":25,"Rank":15,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":24.78,"RiaNewsRetrieval":34.73,"RuBQRetrieval":29.03,"MIRACLRetrieval (ru)":10.58}
16
+ {"index":19,"Rank":16,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_mt_nlu_ru\">sbert_large_mt_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":514,"Average":19.13,"RiaNewsRetrieval":21.4,"RuBQRetrieval":29.8,"MIRACLRetrieval (ru)":6.2}
17
+ {"index":27,"Rank":17,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/rubert-tiny2\">rubert-tiny2<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":2048,"Max Tokens":514,"Average":8.89,"RiaNewsRetrieval":13.92,"RuBQRetrieval":10.87,"MIRACLRetrieval (ru)":1.89}
18
+ {"index":20,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_nlu_ru\">sbert_large_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":512,"Average":8.51,"RiaNewsRetrieval":11.11,"RuBQRetrieval":12.45,"MIRACLRetrieval (ru)":1.98}
19
+ {"index":9,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased-sentence\">rubert-base-cased-sentence<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":7.09,"RiaNewsRetrieval":6.72,"RuBQRetrieval":12.63,"MIRACLRetrieval (ru)":1.92}
20
+ {"index":7,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.71,"RiaNewsRetrieval":4.14,"RuBQRetrieval":10.6,"MIRACLRetrieval (ru)":2.39}
21
+ {"index":8,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.33,"RiaNewsRetrieval":5.58,"RuBQRetrieval":9.52,"MIRACLRetrieval (ru)":0.88}
22
+ {"index":47,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/all-MiniLM-L6-v2\">all-MiniLM-L6-v2<\/a>","Model Size (Million Parameters)":23,"Memory Usage (GB, fp32)":0.09,"Embedding Dimensions":384,"Max Tokens":512,"Average":1.23,"RiaNewsRetrieval":0.67,"RuBQRetrieval":2.64,"MIRACLRetrieval (ru)":0.39}
23
+ {"index":1,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/github.com\/facebookresearch\/LASER\">LASER2<\/a>","Model Size (Million Parameters)":43,"Memory Usage (GB, fp32)":0.16,"Embedding Dimensions":1024,"Max Tokens":"N\/A","Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
24
+ {"index":2,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Alibaba-NLP\/gte-multilingual-base\">gte-multilingual-base<\/a>","Model Size (Million Parameters)":305,"Memory Usage (GB, fp32)":1.14,"Embedding Dimensions":768,"Max Tokens":8192,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
25
+ {"index":10,"Rank":28,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":70.94,"MIRACLRetrieval (ru)":""}
26
  {"index":13,"Rank":29,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
27
  {"index":15,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":"","RiaNewsRetrieval":70.01,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":""}
28
  {"index":16,"Rank":32,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
all_data_tasks/39/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/4/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/40/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/5/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/6/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/8/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
all_data_tasks/9/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/en/data_overall/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/en/data_tasks/Classification/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/en/data_tasks/Clustering/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/en/data_tasks/PairClassification/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/en/data_tasks/Reranking/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/en/data_tasks/Retrieval/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/en/data_tasks/STS/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/en/data_tasks/Summarization/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/other-cls/data_tasks/Classification/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/other-sts/data_tasks/STS/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/ru/data_overall/default.jsonl CHANGED
@@ -2,26 +2,26 @@
2
  {"index":35,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":66.03,"Classification Average (9 datasets)":66.31,"Clustering Average (3 datasets)":63.21,"PairClassification Average (1 datasets)":63.89,"Reranking Average (2 datasets)":69.17,"Retrieval Average (3 datasets)":74.41,"STS Average (3 datasets)":74.85,"MultilabelClassification Average (2 datasets)":41.15}
3
  {"index":29,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average (23 datasets)":62.58,"Classification Average (9 datasets)":61.92,"Clustering Average (3 datasets)":53.61,"PairClassification Average (1 datasets)":65.07,"Reranking Average (2 datasets)":68.72,"Retrieval Average (3 datasets)":73.63,"STS Average (3 datasets)":75.38,"MultilabelClassification Average (2 datasets)":35.88}
4
  {"index":18,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":61.77,"Classification Average (9 datasets)":62.74,"Clustering Average (3 datasets)":56.06,"PairClassification Average (1 datasets)":60.79,"Reranking Average (2 datasets)":63.89,"Retrieval Average (3 datasets)":66.51,"STS Average (3 datasets)":73.97,"MultilabelClassification Average (2 datasets)":38.88}
5
- {"index":3,"Rank":5,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/BAAI\/bge-m3\">bge-m3<\/a>","Model Size (Million Parameters)":567,"Memory Usage (GB, fp32)":2.11,"Embedding Dimensions":1024,"Max Tokens":8192,"Average (23 datasets)":61.58,"Classification Average (9 datasets)":60.46,"Clustering Average (3 datasets)":52.38,"PairClassification Average (1 datasets)":60.6,"Reranking Average (2 datasets)":69.7,"Retrieval Average (3 datasets)":74.79,"STS Average (3 datasets)":73.68,"MultilabelClassification Average (2 datasets)":34.86}
6
- {"index":34,"Rank":6,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large\">multilingual-e5-large<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":61.41,"Classification Average (9 datasets)":61.01,"Clustering Average (3 datasets)":52.23,"PairClassification Average (1 datasets)":58.42,"Reranking Average (2 datasets)":69.66,"Retrieval Average (3 datasets)":74.04,"STS Average (3 datasets)":71.62,"MultilabelClassification Average (2 datasets)":36.01}
7
- {"index":12,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average (23 datasets)":60.48,"Classification Average (9 datasets)":61.58,"Clustering Average (3 datasets)":55.67,"PairClassification Average (1 datasets)":57.71,"Reranking Average (2 datasets)":63.16,"Retrieval Average (3 datasets)":66.17,"STS Average (3 datasets)":71.7,"MultilabelClassification Average (2 datasets)":36.1}
8
- {"index":59,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/LaBSE-ru-turbo\">LaBSE-ru-turbo<\/a>","Model Size (Million Parameters)":128,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":59.04,"Classification Average (9 datasets)":59.23,"Clustering Average (3 datasets)":53.22,"PairClassification Average (1 datasets)":57.81,"Reranking Average (2 datasets)":63.04,"Retrieval Average (3 datasets)":63.68,"STS Average (3 datasets)":72.54,"MultilabelClassification Average (2 datasets)":36.32}
9
- {"index":33,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"Average (23 datasets)":58.34,"Classification Average (9 datasets)":58.26,"Clustering Average (3 datasets)":50.27,"PairClassification Average (1 datasets)":54.98,"Reranking Average (2 datasets)":66.24,"Retrieval Average (3 datasets)":67.14,"STS Average (3 datasets)":70.16,"MultilabelClassification Average (2 datasets)":33.65}
10
- {"index":28,"Rank":10,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-base\">USER-base<\/a>","Model Size (Million Parameters)":124,"Memory Usage (GB, fp32)":0.46,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":57.98,"Classification Average (9 datasets)":59.88,"Clustering Average (3 datasets)":53.42,"PairClassification Average (1 datasets)":60.11,"Reranking Average (2 datasets)":55.58,"Retrieval Average (3 datasets)":56.64,"STS Average (3 datasets)":73.07,"MultilabelClassification Average (2 datasets)":36.98}
11
- {"index":36,"Rank":11,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average (23 datasets)":57.29,"Classification Average (9 datasets)":56.44,"Clustering Average (3 datasets)":51.35,"PairClassification Average (1 datasets)":55.14,"Reranking Average (2 datasets)":65.28,"Retrieval Average (3 datasets)":65.85,"STS Average (3 datasets)":69.48,"MultilabelClassification Average (2 datasets)":31.99}
12
- {"index":60,"Rank":12,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/rubert-tiny-turbo\">rubert-tiny-turbo<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":312,"Max Tokens":2048,"Average (23 datasets)":53.22,"Classification Average (9 datasets)":55.01,"Clustering Average (3 datasets)":49.57,"PairClassification Average (1 datasets)":56.27,"Reranking Average (2 datasets)":54.94,"Retrieval Average (3 datasets)":46.69,"STS Average (3 datasets)":70.23,"MultilabelClassification Average (2 datasets)":31.7}
13
- {"index":19,"Rank":13,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_mt_nlu_ru\">sbert_large_mt_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":48.72,"Classification Average (9 datasets)":57.52,"Clustering Average (3 datasets)":51.29,"PairClassification Average (1 datasets)":51.97,"Reranking Average (2 datasets)":40.56,"Retrieval Average (3 datasets)":19.13,"STS Average (3 datasets)":64.4,"MultilabelClassification Average (2 datasets)":32.67}
14
- {"index":25,"Rank":14,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":48.23,"Classification Average (9 datasets)":54.98,"Clustering Average (3 datasets)":46.84,"PairClassification Average (1 datasets)":55.61,"Reranking Average (2 datasets)":41.84,"Retrieval Average (3 datasets)":24.78,"STS Average (3 datasets)":65.91,"MultilabelClassification Average (2 datasets)":31.27}
15
- {"index":20,"Rank":15,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_nlu_ru\">sbert_large_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":512,"Average (23 datasets)":45.35,"Classification Average (9 datasets)":57.24,"Clustering Average (3 datasets)":50.43,"PairClassification Average (1 datasets)":50.17,"Reranking Average (2 datasets)":32.8,"Retrieval Average (3 datasets)":8.51,"STS Average (3 datasets)":57.21,"MultilabelClassification Average (2 datasets)":31.87}
16
- {"index":27,"Rank":16,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/rubert-tiny2\">rubert-tiny2<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":2048,"Max Tokens":514,"Average (23 datasets)":42.22,"Classification Average (9 datasets)":52.17,"Clustering Average (3 datasets)":39.11,"PairClassification Average (1 datasets)":51.87,"Reranking Average (2 datasets)":30.95,"Retrieval Average (3 datasets)":8.89,"STS Average (3 datasets)":61.6,"MultilabelClassification Average (2 datasets)":29.45}
17
- {"index":7,"Rank":17,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":41.94,"Classification Average (9 datasets)":55.15,"Clustering Average (3 datasets)":38.41,"PairClassification Average (1 datasets)":53.02,"Reranking Average (2 datasets)":27.84,"Retrieval Average (3 datasets)":5.71,"STS Average (3 datasets)":56.2,"MultilabelClassification Average (2 datasets)":29.32}
18
- {"index":9,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased-sentence\">rubert-base-cased-sentence<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":41.91,"Classification Average (9 datasets)":51.27,"Clustering Average (3 datasets)":43.13,"PairClassification Average (1 datasets)":59.12,"Reranking Average (2 datasets)":26.83,"Retrieval Average (3 datasets)":7.09,"STS Average (3 datasets)":61.18,"MultilabelClassification Average (2 datasets)":27.8}
19
- {"index":8,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":37.53,"Classification Average (9 datasets)":51.6,"Clustering Average (3 datasets)":27.91,"PairClassification Average (1 datasets)":53.17,"Reranking Average (2 datasets)":27.46,"Retrieval Average (3 datasets)":5.33,"STS Average (3 datasets)":46.22,"MultilabelClassification Average (2 datasets)":26.2}
20
- {"index":0,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/docs.voyageai.com\/embeddings\/\">voyage-3<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":1024,"Max Tokens":32000,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":76.0,"STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
21
- {"index":1,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/github.com\/facebookresearch\/LASER\">LASER2<\/a>","Model Size (Million Parameters)":43,"Memory Usage (GB, fp32)":0.16,"Embedding Dimensions":1024,"Max Tokens":"N\/A","Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
22
- {"index":2,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Alibaba-NLP\/gte-multilingual-base\">gte-multilingual-base<\/a>","Model Size (Million Parameters)":305,"Memory Usage (GB, fp32)":1.14,"Embedding Dimensions":768,"Max Tokens":8192,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
23
- {"index":10,"Rank":26,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average (23 datasets)":"","Classification Average (9 datasets)":67.48,"Clustering Average (3 datasets)":61.91,"PairClassification Average (1 datasets)":58.93,"Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":35.6}
24
- {"index":11,"Rank":27,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average (23 datasets)":"","Classification Average (9 datasets)":65.21,"Clustering Average (3 datasets)":55.28,"PairClassification Average (1 datasets)":55.57,"Reranking Average (2 datasets)":62.3,"Retrieval Average (3 datasets)":"","STS Average (3 datasets)":71.18,"MultilabelClassification Average (2 datasets)":36.46}
25
  {"index":13,"Rank":28,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
26
  {"index":15,"Rank":30,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average (23 datasets)":"","Classification Average (9 datasets)":56.45,"Clustering Average (3 datasets)":51.35,"PairClassification Average (1 datasets)":55.14,"Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":69.48,"MultilabelClassification Average (2 datasets)":31.44}
27
  {"index":16,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
 
2
  {"index":35,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":66.03,"Classification Average (9 datasets)":66.31,"Clustering Average (3 datasets)":63.21,"PairClassification Average (1 datasets)":63.89,"Reranking Average (2 datasets)":69.17,"Retrieval Average (3 datasets)":74.41,"STS Average (3 datasets)":74.85,"MultilabelClassification Average (2 datasets)":41.15}
3
  {"index":29,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average (23 datasets)":62.58,"Classification Average (9 datasets)":61.92,"Clustering Average (3 datasets)":53.61,"PairClassification Average (1 datasets)":65.07,"Reranking Average (2 datasets)":68.72,"Retrieval Average (3 datasets)":73.63,"STS Average (3 datasets)":75.38,"MultilabelClassification Average (2 datasets)":35.88}
4
  {"index":18,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":61.77,"Classification Average (9 datasets)":62.74,"Clustering Average (3 datasets)":56.06,"PairClassification Average (1 datasets)":60.79,"Reranking Average (2 datasets)":63.89,"Retrieval Average (3 datasets)":66.51,"STS Average (3 datasets)":73.97,"MultilabelClassification Average (2 datasets)":38.88}
5
+ {"index":11,"Rank":5,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average (23 datasets)":61.64,"Classification Average (9 datasets)":65.21,"Clustering Average (3 datasets)":55.28,"PairClassification Average (1 datasets)":55.57,"Reranking Average (2 datasets)":62.3,"Retrieval Average (3 datasets)":66.13,"STS Average (3 datasets)":71.18,"MultilabelClassification Average (2 datasets)":36.46}
6
+ {"index":3,"Rank":6,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/BAAI\/bge-m3\">bge-m3<\/a>","Model Size (Million Parameters)":567,"Memory Usage (GB, fp32)":2.11,"Embedding Dimensions":1024,"Max Tokens":8192,"Average (23 datasets)":61.58,"Classification Average (9 datasets)":60.46,"Clustering Average (3 datasets)":52.38,"PairClassification Average (1 datasets)":60.6,"Reranking Average (2 datasets)":69.7,"Retrieval Average (3 datasets)":74.79,"STS Average (3 datasets)":73.68,"MultilabelClassification Average (2 datasets)":34.86}
7
+ {"index":34,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large\">multilingual-e5-large<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":61.41,"Classification Average (9 datasets)":61.01,"Clustering Average (3 datasets)":52.23,"PairClassification Average (1 datasets)":58.42,"Reranking Average (2 datasets)":69.66,"Retrieval Average (3 datasets)":74.04,"STS Average (3 datasets)":71.62,"MultilabelClassification Average (2 datasets)":36.01}
8
+ {"index":12,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average (23 datasets)":60.48,"Classification Average (9 datasets)":61.58,"Clustering Average (3 datasets)":55.67,"PairClassification Average (1 datasets)":57.71,"Reranking Average (2 datasets)":63.16,"Retrieval Average (3 datasets)":66.17,"STS Average (3 datasets)":71.7,"MultilabelClassification Average (2 datasets)":36.1}
9
+ {"index":59,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/LaBSE-ru-turbo\">LaBSE-ru-turbo<\/a>","Model Size (Million Parameters)":128,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":59.04,"Classification Average (9 datasets)":59.23,"Clustering Average (3 datasets)":53.22,"PairClassification Average (1 datasets)":57.81,"Reranking Average (2 datasets)":63.04,"Retrieval Average (3 datasets)":63.68,"STS Average (3 datasets)":72.54,"MultilabelClassification Average (2 datasets)":36.32}
10
+ {"index":33,"Rank":10,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"Average (23 datasets)":58.34,"Classification Average (9 datasets)":58.26,"Clustering Average (3 datasets)":50.27,"PairClassification Average (1 datasets)":54.98,"Reranking Average (2 datasets)":66.24,"Retrieval Average (3 datasets)":67.14,"STS Average (3 datasets)":70.16,"MultilabelClassification Average (2 datasets)":33.65}
11
+ {"index":28,"Rank":11,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-base\">USER-base<\/a>","Model Size (Million Parameters)":124,"Memory Usage (GB, fp32)":0.46,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":57.98,"Classification Average (9 datasets)":59.88,"Clustering Average (3 datasets)":53.42,"PairClassification Average (1 datasets)":60.11,"Reranking Average (2 datasets)":55.58,"Retrieval Average (3 datasets)":56.64,"STS Average (3 datasets)":73.07,"MultilabelClassification Average (2 datasets)":36.98}
12
+ {"index":36,"Rank":12,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average (23 datasets)":57.29,"Classification Average (9 datasets)":56.44,"Clustering Average (3 datasets)":51.35,"PairClassification Average (1 datasets)":55.14,"Reranking Average (2 datasets)":65.28,"Retrieval Average (3 datasets)":65.85,"STS Average (3 datasets)":69.48,"MultilabelClassification Average (2 datasets)":31.99}
13
+ {"index":60,"Rank":13,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/rubert-tiny-turbo\">rubert-tiny-turbo<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":312,"Max Tokens":2048,"Average (23 datasets)":53.22,"Classification Average (9 datasets)":55.01,"Clustering Average (3 datasets)":49.57,"PairClassification Average (1 datasets)":56.27,"Reranking Average (2 datasets)":54.94,"Retrieval Average (3 datasets)":46.69,"STS Average (3 datasets)":70.23,"MultilabelClassification Average (2 datasets)":31.7}
14
+ {"index":19,"Rank":14,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_mt_nlu_ru\">sbert_large_mt_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":514,"Average (23 datasets)":48.72,"Classification Average (9 datasets)":57.52,"Clustering Average (3 datasets)":51.29,"PairClassification Average (1 datasets)":51.97,"Reranking Average (2 datasets)":40.56,"Retrieval Average (3 datasets)":19.13,"STS Average (3 datasets)":64.4,"MultilabelClassification Average (2 datasets)":32.67}
15
+ {"index":25,"Rank":15,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":48.23,"Classification Average (9 datasets)":54.98,"Clustering Average (3 datasets)":46.84,"PairClassification Average (1 datasets)":55.61,"Reranking Average (2 datasets)":41.84,"Retrieval Average (3 datasets)":24.78,"STS Average (3 datasets)":65.91,"MultilabelClassification Average (2 datasets)":31.27}
16
+ {"index":20,"Rank":16,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_nlu_ru\">sbert_large_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":512,"Average (23 datasets)":45.35,"Classification Average (9 datasets)":57.24,"Clustering Average (3 datasets)":50.43,"PairClassification Average (1 datasets)":50.17,"Reranking Average (2 datasets)":32.8,"Retrieval Average (3 datasets)":8.51,"STS Average (3 datasets)":57.21,"MultilabelClassification Average (2 datasets)":31.87}
17
+ {"index":27,"Rank":17,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/rubert-tiny2\">rubert-tiny2<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":2048,"Max Tokens":514,"Average (23 datasets)":42.22,"Classification Average (9 datasets)":52.17,"Clustering Average (3 datasets)":39.11,"PairClassification Average (1 datasets)":51.87,"Reranking Average (2 datasets)":30.95,"Retrieval Average (3 datasets)":8.89,"STS Average (3 datasets)":61.6,"MultilabelClassification Average (2 datasets)":29.45}
18
+ {"index":7,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":41.94,"Classification Average (9 datasets)":55.15,"Clustering Average (3 datasets)":38.41,"PairClassification Average (1 datasets)":53.02,"Reranking Average (2 datasets)":27.84,"Retrieval Average (3 datasets)":5.71,"STS Average (3 datasets)":56.2,"MultilabelClassification Average (2 datasets)":29.32}
19
+ {"index":9,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased-sentence\">rubert-base-cased-sentence<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":41.91,"Classification Average (9 datasets)":51.27,"Clustering Average (3 datasets)":43.13,"PairClassification Average (1 datasets)":59.12,"Reranking Average (2 datasets)":26.83,"Retrieval Average (3 datasets)":7.09,"STS Average (3 datasets)":61.18,"MultilabelClassification Average (2 datasets)":27.8}
20
+ {"index":8,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average (23 datasets)":37.53,"Classification Average (9 datasets)":51.6,"Clustering Average (3 datasets)":27.91,"PairClassification Average (1 datasets)":53.17,"Reranking Average (2 datasets)":27.46,"Retrieval Average (3 datasets)":5.33,"STS Average (3 datasets)":46.22,"MultilabelClassification Average (2 datasets)":26.2}
21
+ {"index":0,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/docs.voyageai.com\/embeddings\/\">voyage-3<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":1024,"Max Tokens":32000,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":76.0,"STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
22
+ {"index":1,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/github.com\/facebookresearch\/LASER\">LASER2<\/a>","Model Size (Million Parameters)":43,"Memory Usage (GB, fp32)":0.16,"Embedding Dimensions":1024,"Max Tokens":"N\/A","Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
23
+ {"index":2,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Alibaba-NLP\/gte-multilingual-base\">gte-multilingual-base<\/a>","Model Size (Million Parameters)":305,"Memory Usage (GB, fp32)":1.14,"Embedding Dimensions":768,"Max Tokens":8192,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
24
+ {"index":10,"Rank":27,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average (23 datasets)":"","Classification Average (9 datasets)":67.48,"Clustering Average (3 datasets)":61.91,"PairClassification Average (1 datasets)":58.93,"Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":35.6}
25
  {"index":13,"Rank":28,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
26
  {"index":15,"Rank":30,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average (23 datasets)":"","Classification Average (9 datasets)":56.45,"Clustering Average (3 datasets)":51.35,"PairClassification Average (1 datasets)":55.14,"Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":69.48,"MultilabelClassification Average (2 datasets)":31.44}
27
  {"index":16,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average (23 datasets)":"","Classification Average (9 datasets)":"","Clustering Average (3 datasets)":"","PairClassification Average (1 datasets)":"","Reranking Average (2 datasets)":"","Retrieval Average (3 datasets)":"","STS Average (3 datasets)":"","MultilabelClassification Average (2 datasets)":""}
boards_data/ru/data_tasks/Classification/default.jsonl CHANGED
@@ -2,8 +2,8 @@
2
  {"index":10,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":67.48,"GeoreviewClassification":53.47,"HeadlineClassification":85.66,"InappropriatenessClassification":65.29,"KinopoiskClassification":64.25,"RuReviewsClassification":68.58,"RuSciBenchGRNTIClassification":64.56,"RuSciBenchOECDClassification":51.2,"MassiveIntentClassification (ru)":76.01,"MassiveScenarioClassification (ru)":78.28}
3
  {"index":35,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.31,"GeoreviewClassification":55.9,"HeadlineClassification":86.18,"InappropriatenessClassification":65.53,"KinopoiskClassification":66.12,"RuReviewsClassification":68.56,"RuSciBenchGRNTIClassification":65.07,"RuSciBenchOECDClassification":50.21,"MassiveIntentClassification (ru)":67.6,"MassiveScenarioClassification (ru)":71.59}
4
  {"index":11,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":65.21,"GeoreviewClassification":52.04,"HeadlineClassification":83.4,"InappropriatenessClassification":64.14,"KinopoiskClassification":63.15,"RuReviewsClassification":68.24,"RuSciBenchGRNTIClassification":61.24,"RuSciBenchOECDClassification":48.1,"MassiveIntentClassification (ru)":68.61,"MassiveScenarioClassification (ru)":77.98}
5
- {"index":39,"Rank":5,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/jinaai\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
6
- {"index":22,"Rank":6,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/arkohut\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
7
  {"index":18,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":62.74,"GeoreviewClassification":49.7,"HeadlineClassification":78.0,"InappropriatenessClassification":61.32,"KinopoiskClassification":63.27,"RuReviewsClassification":67.96,"RuSciBenchGRNTIClassification":59.33,"RuSciBenchOECDClassification":46.33,"MassiveIntentClassification (ru)":66.97,"MassiveScenarioClassification (ru)":71.8}
8
  {"index":29,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average":61.92,"GeoreviewClassification":50.98,"HeadlineClassification":70.09,"InappropriatenessClassification":60.76,"KinopoiskClassification":63.33,"RuReviewsClassification":68.52,"RuSciBenchGRNTIClassification":57.67,"RuSciBenchOECDClassification":44.2,"MassiveIntentClassification (ru)":68.85,"MassiveScenarioClassification (ru)":72.9}
9
  {"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":61.58,"GeoreviewClassification":47.69,"HeadlineClassification":83.46,"InappropriatenessClassification":61.32,"KinopoiskClassification":59.04,"RuReviewsClassification":66.09,"RuSciBenchGRNTIClassification":61.41,"RuSciBenchOECDClassification":48.67,"MassiveIntentClassification (ru)":60.08,"MassiveScenarioClassification (ru)":66.44}
 
2
  {"index":10,"Rank":2,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":67.48,"GeoreviewClassification":53.47,"HeadlineClassification":85.66,"InappropriatenessClassification":65.29,"KinopoiskClassification":64.25,"RuReviewsClassification":68.58,"RuSciBenchGRNTIClassification":64.56,"RuSciBenchOECDClassification":51.2,"MassiveIntentClassification (ru)":76.01,"MassiveScenarioClassification (ru)":78.28}
3
  {"index":35,"Rank":3,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-large-instruct\">multilingual-e5-large-instruct<\/a>","Model Size (Million Parameters)":560,"Memory Usage (GB, fp32)":2.09,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.31,"GeoreviewClassification":55.9,"HeadlineClassification":86.18,"InappropriatenessClassification":65.53,"KinopoiskClassification":66.12,"RuReviewsClassification":68.56,"RuSciBenchGRNTIClassification":65.07,"RuSciBenchOECDClassification":50.21,"MassiveIntentClassification (ru)":67.6,"MassiveScenarioClassification (ru)":71.59}
4
  {"index":11,"Rank":4,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":65.21,"GeoreviewClassification":52.04,"HeadlineClassification":83.4,"InappropriatenessClassification":64.14,"KinopoiskClassification":63.15,"RuReviewsClassification":68.24,"RuSciBenchGRNTIClassification":61.24,"RuSciBenchOECDClassification":48.1,"MassiveIntentClassification (ru)":68.61,"MassiveScenarioClassification (ru)":77.98}
5
+ {"index":22,"Rank":5,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/arkohut\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
6
+ {"index":39,"Rank":6,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/jinaai\/jina-embeddings-v3\">jina-embeddings-v3<\/a>","Model Size (Million Parameters)":572,"Memory Usage (GB, fp32)":2.13,"Embedding Dimensions":1024,"Max Tokens":8194,"Average":64.41,"GeoreviewClassification":48.01,"HeadlineClassification":75.08,"InappropriatenessClassification":61.05,"KinopoiskClassification":62.39,"RuReviewsClassification":67.58,"RuSciBenchGRNTIClassification":59.19,"RuSciBenchOECDClassification":45.56,"MassiveIntentClassification (ru)":76.8,"MassiveScenarioClassification (ru)":84.06}
7
  {"index":18,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":62.74,"GeoreviewClassification":49.7,"HeadlineClassification":78.0,"InappropriatenessClassification":61.32,"KinopoiskClassification":63.27,"RuReviewsClassification":67.96,"RuSciBenchGRNTIClassification":59.33,"RuSciBenchOECDClassification":46.33,"MassiveIntentClassification (ru)":66.97,"MassiveScenarioClassification (ru)":71.8}
8
  {"index":29,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-bge-m3\">USER-bge-m3<\/a>","Model Size (Million Parameters)":359,"Memory Usage (GB, fp32)":1.34,"Embedding Dimensions":1024,"Max Tokens":8192,"Average":61.92,"GeoreviewClassification":50.98,"HeadlineClassification":70.09,"InappropriatenessClassification":60.76,"KinopoiskClassification":63.33,"RuReviewsClassification":68.52,"RuSciBenchGRNTIClassification":57.67,"RuSciBenchOECDClassification":44.2,"MassiveIntentClassification (ru)":68.85,"MassiveScenarioClassification (ru)":72.9}
9
  {"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":61.58,"GeoreviewClassification":47.69,"HeadlineClassification":83.46,"InappropriatenessClassification":61.32,"KinopoiskClassification":59.04,"RuReviewsClassification":66.09,"RuSciBenchGRNTIClassification":61.41,"RuSciBenchOECDClassification":48.67,"MassiveIntentClassification (ru)":60.08,"MassiveScenarioClassification (ru)":66.44}
boards_data/ru/data_tasks/PairClassification/default.jsonl CHANGED
@@ -18,8 +18,8 @@
18
  {"index":45,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/LaBSE\">LaBSE<\/a>","Model Size (Million Parameters)":471,"Memory Usage (GB, fp32)":1.75,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.71}
19
  {"index":25,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.61}
20
  {"index":11,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","TERRa":55.57}
21
- {"index":36,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
22
- {"index":15,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
23
  {"index":33,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"TERRa":54.98}
24
  {"index":8,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.17}
25
  {"index":7,"Rank":25,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.02}
 
18
  {"index":45,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/LaBSE\">LaBSE<\/a>","Model Size (Million Parameters)":471,"Memory Usage (GB, fp32)":1.75,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.71}
19
  {"index":25,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":55.61}
20
  {"index":11,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","TERRa":55.57}
21
+ {"index":15,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
22
+ {"index":36,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"TERRa":55.14}
23
  {"index":33,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"TERRa":54.98}
24
  {"index":8,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.17}
25
  {"index":7,"Rank":25,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"TERRa":53.02}
boards_data/ru/data_tasks/Retrieval/default.jsonl CHANGED
@@ -7,22 +7,22 @@
7
  {"index":33,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"Average":67.14,"RiaNewsRetrieval":70.24,"RuBQRetrieval":69.58,"MIRACLRetrieval (ru)":61.6}
8
  {"index":18,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.51,"RiaNewsRetrieval":78.86,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":53.91}
9
  {"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":66.17,"RiaNewsRetrieval":77.42,"RuBQRetrieval":67.75,"MIRACLRetrieval (ru)":53.33}
10
- {"index":36,"Rank":10,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":65.85,"RiaNewsRetrieval":70.0,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":59.01}
11
- {"index":59,"Rank":11,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/LaBSE-ru-turbo\">LaBSE-ru-turbo<\/a>","Model Size (Million Parameters)":128,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":63.68,"RiaNewsRetrieval":69.36,"RuBQRetrieval":65.71,"MIRACLRetrieval (ru)":55.97}
12
- {"index":28,"Rank":12,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-base\">USER-base<\/a>","Model Size (Million Parameters)":124,"Memory Usage (GB, fp32)":0.46,"Embedding Dimensions":768,"Max Tokens":512,"Average":56.64,"RiaNewsRetrieval":77.83,"RuBQRetrieval":56.86,"MIRACLRetrieval (ru)":35.22}
13
- {"index":60,"Rank":13,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/rubert-tiny-turbo\">rubert-tiny-turbo<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":312,"Max Tokens":2048,"Average":46.69,"RiaNewsRetrieval":51.27,"RuBQRetrieval":51.73,"MIRACLRetrieval (ru)":37.07}
14
- {"index":25,"Rank":14,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":24.78,"RiaNewsRetrieval":34.73,"RuBQRetrieval":29.03,"MIRACLRetrieval (ru)":10.58}
15
- {"index":19,"Rank":15,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_mt_nlu_ru\">sbert_large_mt_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":514,"Average":19.13,"RiaNewsRetrieval":21.4,"RuBQRetrieval":29.8,"MIRACLRetrieval (ru)":6.2}
16
- {"index":27,"Rank":16,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/rubert-tiny2\">rubert-tiny2<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":2048,"Max Tokens":514,"Average":8.89,"RiaNewsRetrieval":13.92,"RuBQRetrieval":10.87,"MIRACLRetrieval (ru)":1.89}
17
- {"index":20,"Rank":17,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_nlu_ru\">sbert_large_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":512,"Average":8.51,"RiaNewsRetrieval":11.11,"RuBQRetrieval":12.45,"MIRACLRetrieval (ru)":1.98}
18
- {"index":9,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased-sentence\">rubert-base-cased-sentence<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":7.09,"RiaNewsRetrieval":6.72,"RuBQRetrieval":12.63,"MIRACLRetrieval (ru)":1.92}
19
- {"index":7,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.71,"RiaNewsRetrieval":4.14,"RuBQRetrieval":10.6,"MIRACLRetrieval (ru)":2.39}
20
- {"index":8,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.33,"RiaNewsRetrieval":5.58,"RuBQRetrieval":9.52,"MIRACLRetrieval (ru)":0.88}
21
- {"index":47,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/all-MiniLM-L6-v2\">all-MiniLM-L6-v2<\/a>","Model Size (Million Parameters)":23,"Memory Usage (GB, fp32)":0.09,"Embedding Dimensions":384,"Max Tokens":512,"Average":1.23,"RiaNewsRetrieval":0.67,"RuBQRetrieval":2.64,"MIRACLRetrieval (ru)":0.39}
22
- {"index":1,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/github.com\/facebookresearch\/LASER\">LASER2<\/a>","Model Size (Million Parameters)":43,"Memory Usage (GB, fp32)":0.16,"Embedding Dimensions":1024,"Max Tokens":"N\/A","Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
23
- {"index":2,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Alibaba-NLP\/gte-multilingual-base\">gte-multilingual-base<\/a>","Model Size (Million Parameters)":305,"Memory Usage (GB, fp32)":1.14,"Embedding Dimensions":768,"Max Tokens":8192,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
24
- {"index":10,"Rank":27,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":70.94,"MIRACLRetrieval (ru)":""}
25
- {"index":11,"Rank":28,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":"","RiaNewsRetrieval":79.42,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":""}
26
  {"index":13,"Rank":29,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
27
  {"index":15,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":"","RiaNewsRetrieval":70.01,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":""}
28
  {"index":16,"Rank":32,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
 
7
  {"index":33,"Rank":7,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-base\">multilingual-e5-base<\/a>","Model Size (Million Parameters)":278,"Memory Usage (GB, fp32)":1.04,"Embedding Dimensions":768,"Max Tokens":514,"Average":67.14,"RiaNewsRetrieval":70.24,"RuBQRetrieval":69.58,"MIRACLRetrieval (ru)":61.6}
8
  {"index":18,"Rank":8,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/ru-en-RoSBERTa\">ru-en-RoSBERTa<\/a>","Model Size (Million Parameters)":404,"Memory Usage (GB, fp32)":1.5,"Embedding Dimensions":1024,"Max Tokens":514,"Average":66.51,"RiaNewsRetrieval":78.86,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":53.91}
9
  {"index":12,"Rank":9,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-v1\">KaLM-embedding-multilingual-mini-v1<\/a>","Model Size (Million Parameters)":494,"Memory Usage (GB, fp32)":1.84,"Embedding Dimensions":896,"Max Tokens":131072,"Average":66.17,"RiaNewsRetrieval":77.42,"RuBQRetrieval":67.75,"MIRACLRetrieval (ru)":53.33}
10
+ {"index":11,"Rank":10,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/HIT-TMG\/KaLM-embedding-multilingual-mini-instruct-v1\">KaLM-embedding-multilingual-mini-instruct-v1<\/a>","Model Size (Million Parameters)":"","Memory Usage (GB, fp32)":"","Embedding Dimensions":"","Max Tokens":"","Average":66.13,"RiaNewsRetrieval":79.42,"RuBQRetrieval":66.77,"MIRACLRetrieval (ru)":52.21}
11
+ {"index":36,"Rank":11,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/intfloat\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":65.85,"RiaNewsRetrieval":70.0,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":59.01}
12
+ {"index":59,"Rank":12,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/LaBSE-ru-turbo\">LaBSE-ru-turbo<\/a>","Model Size (Million Parameters)":128,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":63.68,"RiaNewsRetrieval":69.36,"RuBQRetrieval":65.71,"MIRACLRetrieval (ru)":55.97}
13
+ {"index":28,"Rank":13,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/deepvk\/USER-base\">USER-base<\/a>","Model Size (Million Parameters)":124,"Memory Usage (GB, fp32)":0.46,"Embedding Dimensions":768,"Max Tokens":512,"Average":56.64,"RiaNewsRetrieval":77.83,"RuBQRetrieval":56.86,"MIRACLRetrieval (ru)":35.22}
14
+ {"index":60,"Rank":14,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sergeyzh\/rubert-tiny-turbo\">rubert-tiny-turbo<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":312,"Max Tokens":2048,"Average":46.69,"RiaNewsRetrieval":51.27,"RuBQRetrieval":51.73,"MIRACLRetrieval (ru)":37.07}
15
+ {"index":25,"Rank":15,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/LaBSE-en-ru\">LaBSE-en-ru<\/a>","Model Size (Million Parameters)":129,"Memory Usage (GB, fp32)":0.48,"Embedding Dimensions":768,"Max Tokens":512,"Average":24.78,"RiaNewsRetrieval":34.73,"RuBQRetrieval":29.03,"MIRACLRetrieval (ru)":10.58}
16
+ {"index":19,"Rank":16,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_mt_nlu_ru\">sbert_large_mt_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":514,"Average":19.13,"RiaNewsRetrieval":21.4,"RuBQRetrieval":29.8,"MIRACLRetrieval (ru)":6.2}
17
+ {"index":27,"Rank":17,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/cointegrated\/rubert-tiny2\">rubert-tiny2<\/a>","Model Size (Million Parameters)":29,"Memory Usage (GB, fp32)":0.11,"Embedding Dimensions":2048,"Max Tokens":514,"Average":8.89,"RiaNewsRetrieval":13.92,"RuBQRetrieval":10.87,"MIRACLRetrieval (ru)":1.89}
18
+ {"index":20,"Rank":18,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/ai-forever\/sbert_large_nlu_ru\">sbert_large_nlu_ru<\/a>","Model Size (Million Parameters)":427,"Memory Usage (GB, fp32)":1.59,"Embedding Dimensions":1024,"Max Tokens":512,"Average":8.51,"RiaNewsRetrieval":11.11,"RuBQRetrieval":12.45,"MIRACLRetrieval (ru)":1.98}
19
+ {"index":9,"Rank":19,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased-sentence\">rubert-base-cased-sentence<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":7.09,"RiaNewsRetrieval":6.72,"RuBQRetrieval":12.63,"MIRACLRetrieval (ru)":1.92}
20
+ {"index":7,"Rank":20,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/distilrubert-small-cased-conversational\">distilrubert-small-cased-conversational<\/a>","Model Size (Million Parameters)":106,"Memory Usage (GB, fp32)":0.39,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.71,"RiaNewsRetrieval":4.14,"RuBQRetrieval":10.6,"MIRACLRetrieval (ru)":2.39}
21
+ {"index":8,"Rank":21,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/DeepPavlov\/rubert-base-cased\">rubert-base-cased<\/a>","Model Size (Million Parameters)":180,"Memory Usage (GB, fp32)":0.67,"Embedding Dimensions":768,"Max Tokens":512,"Average":5.33,"RiaNewsRetrieval":5.58,"RuBQRetrieval":9.52,"MIRACLRetrieval (ru)":0.88}
22
+ {"index":47,"Rank":22,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/sentence-transformers\/all-MiniLM-L6-v2\">all-MiniLM-L6-v2<\/a>","Model Size (Million Parameters)":23,"Memory Usage (GB, fp32)":0.09,"Embedding Dimensions":384,"Max Tokens":512,"Average":1.23,"RiaNewsRetrieval":0.67,"RuBQRetrieval":2.64,"MIRACLRetrieval (ru)":0.39}
23
+ {"index":1,"Rank":23,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/github.com\/facebookresearch\/LASER\">LASER2<\/a>","Model Size (Million Parameters)":43,"Memory Usage (GB, fp32)":0.16,"Embedding Dimensions":1024,"Max Tokens":"N\/A","Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
24
+ {"index":2,"Rank":24,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Alibaba-NLP\/gte-multilingual-base\">gte-multilingual-base<\/a>","Model Size (Million Parameters)":305,"Memory Usage (GB, fp32)":1.14,"Embedding Dimensions":768,"Max Tokens":8192,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
25
+ {"index":10,"Rank":28,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/GritLM\/GritLM-7B\">GritLM-7B<\/a>","Model Size (Million Parameters)":7240,"Memory Usage (GB, fp32)":26.97,"Embedding Dimensions":"","Max Tokens":4096,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":70.94,"MIRACLRetrieval (ru)":""}
26
  {"index":13,"Rank":29,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Labib11\/MUG-B-1.6\">MUG-B-1.6<\/a>","Model Size (Million Parameters)":335,"Memory Usage (GB, fp32)":1.25,"Embedding Dimensions":1024,"Max Tokens":512,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
27
  {"index":15,"Rank":31,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Marqo\/multilingual-e5-small\">multilingual-e5-small<\/a>","Model Size (Million Parameters)":118,"Memory Usage (GB, fp32)":0.44,"Embedding Dimensions":384,"Max Tokens":512,"Average":"","RiaNewsRetrieval":70.01,"RuBQRetrieval":68.53,"MIRACLRetrieval (ru)":""}
28
  {"index":16,"Rank":32,"Model":"<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https:\/\/huggingface.co\/Muennighoff\/SGPT-125M-weightedmean-msmarco-specb-bitfit\">SGPT-125M-weightedmean-msmarco-specb-bitfit<\/a>","Model Size (Million Parameters)":138,"Memory Usage (GB, fp32)":0.51,"Embedding Dimensions":768,"Max Tokens":2048,"Average":"","RiaNewsRetrieval":"","RuBQRetrieval":"","MIRACLRetrieval (ru)":""}
boards_data/zh/data_overall/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/zh/data_tasks/Classification/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/zh/data_tasks/Clustering/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/zh/data_tasks/PairClassification/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/zh/data_tasks/Reranking/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/zh/data_tasks/Retrieval/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
boards_data/zh/data_tasks/STS/default.jsonl CHANGED
The diff for this file is too large to render. See raw diff