|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""The SuperGLUE benchmark.""" |
|
|
|
import json |
|
import os |
|
import datasets |
|
import pandas as pd |
|
|
|
_CITATION = """\ |
|
""" |
|
|
|
|
|
_DESCRIPTION = """\ |
|
""" |
|
|
|
_HOMEPAGE = "" |
|
|
|
_LICENSE = "" |
|
|
|
_SUPERLIM_CITATION = """\ |
|
Yvonne Adesam, Aleksandrs Berdicevskis, Felix Morger (2020): SwedishGLUE – Towards a Swedish Test Set for Evaluating Natural Language Understanding Models BibTeX |
|
[1] Original Absabank: |
|
Jacobo Rouces, Lars Borin, Nina Tahmasebi (2020): Creating an Annotated Corpus for Aspect-Based Sentiment Analysis in Swedish, in Proceedings of the 5th conference in Digital Humanities in the Nordic Countries, Riga, Latvia, October 21-23, 2020. BibTeX |
|
[2] DaLAJ: |
|
Volodina, Elena, Yousuf Ali Mohammed, and Julia Klezl (2021). DaLAJ - a dataset for linguistic acceptability judgments for Swedish. In Proceedings of the 10th Workshop on Natural Language Processing for Computer Assisted Language Learning (NLP4CALL 2021). Linköping Electronic Conference Proceedings 177:3, s. 28-37. https://ep.liu.se/ecp/177/003/ecp2021177003.pdf |
|
[3] Analogy: |
|
Tosin Adewumi, Foteini Liwicki, Markus Liwicki. (2020). Corpora compared: The case of the Swedish Gigaword & Wikipedia corpora. In: Proceedings of the 8th SLTC, Gothenburg. arXiv preprint arXiv:2011.03281 |
|
[4] Swedish Test Set for SemEval 2020 Task 1: |
|
Unsupervised Lexical Semantic Change Detection: Dominik Schlechtweg, Barbara McGillivray, Simon Hengchen, Haim Dubossarsky, Nina Tahmasebi (2020): SemEval-2020 Task 1: Unsupervised Lexical Semantic Change Detection, in Proceedings of the Fourteenth Workshop on Semantic Evaluation (SemEval2020), Barcelona, Spain (Online), December 12, 2020. BibTeX |
|
[5] Winogender: |
|
Saga Hansson, Konstantinos Mavromatakis, Yvonne Adesam, Gerlof Bouma and Dana Dannélls (2021). The Swedish Winogender Dataset. In The 23rd Nordic Conference on Computational Linguistics (NoDaLiDa 2021), Reykjavik. |
|
[6] SuperSim: |
|
Hengchen, Simon and Tahmasebi, Nina (2021). SuperSim: a test set for word similarity and relatedness in Swedish. In The 23rd Nordic Conference on Computational Linguistics (NoDaLiDa 2021), Reykjavik. arXiv preprint arXiv:2014.05228 |
|
|
|
""" |
|
|
|
_SUPERLIM_DESCRIPTION = """\ |
|
SuperLim, A standardized suite for evaluation and analysis of Swedish natural language understanding systems. |
|
|
|
""" |
|
_DaLAJ_DESCRIPTION = """\ |
|
Determine whether a sentence is correct Swedish or not. |
|
""" |
|
_DaLAJ_CITATION = """\ |
|
[1] Original Absabank: |
|
Jacobo Rouces, Lars Borin, Nina Tahmasebi (2020): Creating an Annotated Corpus for Aspect-Based Sentiment Analysis in Swedish, in Proceedings of the 5th conference in Digital Humanities in the Nordic Countries, Riga, Latvia, October 21-23, 2020. BibTeX |
|
[2] DaLAJ: |
|
Volodina, Elena, Yousuf Ali Mohammed, and Julia Klezl (2021). DaLAJ - a dataset for linguistic acceptability judgments for Swedish. In Proceedings of the 10th Workshop on Natural Language Processing for Computer Assisted Language Learning (NLP4CALL 2021). Linköping Electronic Conference Proceedings 177:3, s. 28-37. https://ep.liu.se/ecp/177/003/ecp2021177003.pdf |
|
""" |
|
|
|
_SweAna_DESCRIPTION = """\ |
|
The Swedish analogy test set follows the format of the original Google version. However, it is bigger and balanced across the 2 major categories, |
|
having a total of 20,638 samples, made up of 10,381 semantic and 10,257 syntactic samples. It is also roughly balanced across the syntactic subsections. |
|
There are 5 semantic subsections and 6 syntactic subsections. The dataset was constructed, partly using the samples in the English version, |
|
with the help of tools dedicated to Swedish translation and it was proof-read for corrections by two native speakers (with a percentage agreement of 98.93\%).""" |
|
_SweAna_CITATION = """\ |
|
[1] Original Absabank: |
|
Jacobo Rouces, Lars Borin, Nina Tahmasebi (2020): Creating an Annotated Corpus for Aspect-Based Sentiment Analysis in Swedish, in Proceedings of the 5th conference in Digital Humanities in the Nordic Countries, Riga, Latvia, October 21-23, 2020. BibTeX |
|
""" |
|
|
|
_SweDiag_DESCRIPTION = """\ |
|
Färdig preliminär översättning av SuperGLUE diagnostik. Datan innehåller alla ursprungliga annoterade satspar från SuperGLUE tillsammans |
|
med deras svenska översättningar.""" |
|
_SweDiag_CITATION = """\ |
|
""" |
|
_SweFaq_DESCRIPTION = """\ |
|
Vanliga frågor från svenska myndigheters webbsidor med svar i randomiserad ordning""" |
|
_SweFaq_CITATION = """\ |
|
""" |
|
_SweFracas_DESCRIPTION = """\ |
|
A textual inference/entailment problem set, derived from FraCas. The original English Fracas [1] was converted to html and edited by Bill MacCartney [2], |
|
and then automatically translated to Swedish by Peter Ljunglöf and Magdalena Siverbo [3]. The current tabular form of the set was created by Aleksandrs Berdicevskis |
|
by merging the Swedish and English versions and removing some of the problems. Finally, Lars Borin went through all the translations, correcting and Swedifying them manually. |
|
As a result, many translations are rather liberal and diverge noticeably from the English original.""" |
|
_SweFracas_CITATION = """\ |
|
""" |
|
_SwePar_DESCRIPTION = """\ |
|
SweParaphrase is a subset of the automatically translated Swedish Semantic Textual Similarity dataset (Isbister and Sahlgren, 2020). |
|
It consists of 165 manually corrected Swedish sentence pairs paired with the original English sentences and their similarity scores |
|
ranging between 0 (no meaning overlap) and 5 (meaning equivalence). These scores were taken from the English data, they were assigned |
|
by Crowdsourcing through Mechanical Turk. Each sentence pair belongs to one genre (e.g. news, forums or captions). |
|
The task is to determine how similar two sentences are.""" |
|
_SwePar_CITATION = """\ |
|
""" |
|
_SweSat_DESCRIPTION = """\ |
|
The dataset provides a gold standard for Swedish word synonymy/definition. The test items are collected from the Swedish Scholastic |
|
Aptitude Test (högskoleprovet), currently spanning the years 2006--2021 and 822 vocabulary test items. The task for the tested system |
|
is to determine which synonym or definition of five alternatives is correct for each test item. |
|
""" |
|
_SweSat_CITATION = """\ |
|
""" |
|
|
|
_SweSim_DESCRIPTION = """\ |
|
SuperSim is a large-scale similarity and relatedness test set for Swedish built with expert human judgments. The test set is composed of 1360 word-pairs independently judged for both relatedness and similarity by five annotators.""" |
|
|
|
_SweWgr_DESCRIPTION = """\ |
|
The SweWinogender test set is diagnostic dataset to measure gender bias in coreference resolution. It is modelled after the English Winogender benchmark, |
|
and is released with reference statistics on the distribution of men and women between occupations and the association between gender and occupation in modern corpus material.""" |
|
|
|
_SweWsc_DESCRIPTION = """\ |
|
SweWinograd is a pronoun resolution test set, containing constructed items in the style of Winograd schema’s. The interpretation of the target pronouns is determined by (common sense) |
|
reasoning and knowledge, and not by syntactic constraints, lexical distributional information or discourse structuring patterns. |
|
The dataset contains 90 multiple choice with multiple correct answers test items.""" |
|
|
|
_SweWic_DESCRIPTION = """\ |
|
The Swedish Word-in-Context dataset provides a benchmark for evaluating distributional models of word meaning, in particular context-sensitive/dynamic models. Constructed following the principles of the (English) |
|
Word-in-Context dataset, SweWiC consists of 1000 sentence pairs, where each sentence in a pair contains an occurence of a potentially ambiguous focus word specific to that pair. The question posed to the tested |
|
system is whether these two occurrences represent instances of the same word sense. There are 500 same-sense pairs and 500 different-sense pairs.""" |
|
|
|
|
|
|
|
|
|
_URL = "https://huggingface.co/datasets/AI-Sweden/SuperLim/resolve/main/data/" |
|
_TASKS = { |
|
"dalaj": "DaLAJ", |
|
"sweana": "SweAna", |
|
"swediag": "SweDiag", |
|
"swefaq": "SweFaq", |
|
"swefracas": "SweFracas", |
|
"swepar": "SwePar", |
|
"swesat": "SweSat", |
|
"swesim_relatedness": "SweSim_relatedness", |
|
"swesim_similarity": "SweSim_similarity", |
|
"swewgr": "SweWgr", |
|
"swewic": "SweWic", |
|
"swewsc": "SweWsc" |
|
} |
|
|
|
|
|
|
|
class SuperLim(datasets.GeneratorBasedBuilder): |
|
"""TODO: Short description of my dataset.""" |
|
|
|
VERSION = datasets.Version("1.1.0") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
BUILDER_CONFIGS = [ |
|
datasets.BuilderConfig(name="dalaj", version=VERSION, description=_DaLAJ_DESCRIPTION), |
|
datasets.BuilderConfig(name="sweana", version=VERSION, description=_SweAna_DESCRIPTION), |
|
datasets.BuilderConfig(name="swediag", version=VERSION, description=_SweDiag_DESCRIPTION), |
|
datasets.BuilderConfig(name="swefaq", version=VERSION, description=_SweFaq_DESCRIPTION), |
|
datasets.BuilderConfig(name="swefracas", version=VERSION, description=_SweFracas_DESCRIPTION), |
|
datasets.BuilderConfig(name="swepar", version=VERSION, description=_SwePar_DESCRIPTION), |
|
datasets.BuilderConfig(name="swesat", version=VERSION, description=_SweSat_DESCRIPTION), |
|
datasets.BuilderConfig(name="swesim_relatedness", version=VERSION, description=_SweSim_DESCRIPTION), |
|
datasets.BuilderConfig(name="swesim_similarity", version=VERSION, description=_SweSim_DESCRIPTION), |
|
datasets.BuilderConfig(name="swewgr", version=VERSION, description=_SweWgr_DESCRIPTION), |
|
datasets.BuilderConfig(name="swewic", version=VERSION, description=_SweWic_DESCRIPTION), |
|
datasets.BuilderConfig(name="swewsc", version=VERSION, description=_SweWsc_DESCRIPTION), |
|
] |
|
|
|
|
|
def _info(self): |
|
|
|
if self.config.name == "dalaj": |
|
features = datasets.Features( |
|
{ |
|
"original_sentence": datasets.Value("string"), |
|
"corrected_sentence": datasets.Value("string"), |
|
"error_indices": datasets.Value("string"), |
|
"corrected_indices": datasets.Value("string"), |
|
"error_corr_pair": datasets.Value("string"), |
|
"error_label": datasets.Value("string"), |
|
"l1": datasets.Value("string"), |
|
"approximate_level": datasets.Value("string"), |
|
|
|
} |
|
) |
|
elif self.config.name == "sweana": |
|
features = datasets.Features( |
|
{ |
|
"a": datasets.Value("string"), |
|
"b": datasets.Value("string"), |
|
"c": datasets.Value("string"), |
|
"d": datasets.Value("string"), |
|
"relation": datasets.Value("string"), |
|
} |
|
) |
|
elif self.config.name == "swediag": |
|
features = datasets.Features( |
|
{ |
|
"premise": datasets.Value("string"), |
|
"hypothesis": datasets.Value("string"), |
|
"label": datasets.Value("string"), |
|
} |
|
) |
|
elif self.config.name == "swefaq": |
|
features = datasets.Features( |
|
{ |
|
"question": datasets.Value("string"), |
|
"candidate_answer": datasets.Value("string"), |
|
"correct_answer": datasets.Value("string"), |
|
} |
|
) |
|
elif self.config.name == "swefracas": |
|
features = datasets.Features( |
|
{ |
|
"answer": datasets.Value("string"), |
|
"question": datasets.Value("string"), |
|
"premiss_1": datasets.Value("string"), |
|
"premiss_2": datasets.Value("string"), |
|
"premiss_3": datasets.Value("string"), |
|
"premiss_4": datasets.Value("string"), |
|
"premiss_5": datasets.Value("string"), |
|
} |
|
) |
|
elif self.config.name == "swepar": |
|
features = datasets.Features( |
|
{ |
|
"sentence_1": datasets.Value("string"), |
|
"sentence_2": datasets.Value("string"), |
|
"similarity_score": datasets.Value("string"), |
|
} |
|
) |
|
elif self.config.name == "swesat": |
|
features = datasets.Features( |
|
{ |
|
"target_item": datasets.Value("string"), |
|
"answer_1": datasets.Value("string"), |
|
"answer_2": datasets.Value("string"), |
|
"answer_3": datasets.Value("string"), |
|
"answer_4": datasets.Value("string"), |
|
"answer_5": datasets.Value("string"), |
|
} |
|
) |
|
elif self.config.name == "swesim_relatedness": |
|
features = datasets.Features( |
|
{ |
|
"word_1": datasets.Value("string"), |
|
"word_2": datasets.Value("string"), |
|
"relatedness": datasets.Value("string"), |
|
} |
|
) |
|
elif self.config.name == "swesim_similarity": |
|
features = datasets.Features( |
|
{ |
|
"word_1": datasets.Value("string"), |
|
"word_2": datasets.Value("string"), |
|
"similarity": datasets.Value("string"), |
|
} |
|
) |
|
elif self.config.name == "swewgr": |
|
features = datasets.Features( |
|
{ |
|
"text": datasets.Value("string"), |
|
"challenge": datasets.Value("string"), |
|
"responses": datasets.Value("string"), |
|
} |
|
) |
|
elif self.config.name == "swewic": |
|
features = datasets.Features( |
|
{ |
|
"sentence_1": datasets.Value("string"), |
|
"word_1": datasets.Value("string"), |
|
"sentence_2": datasets.Value("string"), |
|
"word_2": datasets.Value("string"), |
|
"same_sense": datasets.Value("string"), |
|
"start_1": datasets.Value("string"), |
|
"start_2": datasets.Value("string"), |
|
"end_1": datasets.Value("string"), |
|
"end_2": datasets.Value("string"), |
|
} |
|
) |
|
elif self.config.name == "swewsc": |
|
features = datasets.Features( |
|
{ |
|
"passage": datasets.Value("string"), |
|
"challenge_text": datasets.Value("string"), |
|
"response_text": datasets.Value("string"), |
|
"challenge_begin": datasets.Value("string"), |
|
"challenge_end": datasets.Value("string"), |
|
"response_begin": datasets.Value("string"), |
|
"response_end": datasets.Value("string"), |
|
"label": datasets.Value("string") |
|
} |
|
) |
|
else: |
|
features = datasets.Features( |
|
{ |
|
"sentence": datasets.Value("string"), |
|
"option2": datasets.Value("string"), |
|
"second_domain_answer": datasets.Value("string") |
|
|
|
} |
|
) |
|
return datasets.DatasetInfo( |
|
|
|
description=_DESCRIPTION, |
|
|
|
features=features, |
|
|
|
|
|
|
|
|
|
homepage=_HOMEPAGE, |
|
|
|
license=_LICENSE, |
|
|
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if self.config.name == "dalaj": |
|
data_dir_test = dl_manager.download_and_extract(os.path.join(_URL,_TASKS[self.config.name],"test.csv")) |
|
data_dir_train = dl_manager.download_and_extract(os.path.join(_URL,_TASKS[self.config.name],"train.csv")) |
|
data_dir_dev = dl_manager.download_and_extract(os.path.join(_URL,_TASKS[self.config.name],"dev.csv")) |
|
|
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
|
|
gen_kwargs={ |
|
"filepath": data_dir_train, |
|
"split": "train", |
|
}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
|
|
gen_kwargs={ |
|
"filepath": data_dir_test, |
|
"split": "test" |
|
}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.VALIDATION, |
|
|
|
gen_kwargs={ |
|
"filepath": data_dir_dev, |
|
"split": "dev", |
|
}, |
|
), |
|
] |
|
else: |
|
data_dir_test = dl_manager.download_and_extract(os.path.join(_URL, _TASKS[self.config.name], "test.csv")) |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
|
|
gen_kwargs={ |
|
"filepath": data_dir_test, |
|
"split": "test" |
|
}, |
|
), |
|
] |
|
|
|
|
|
def _generate_examples(self, filepath, split): |
|
|
|
|
|
df = pd.read_csv(filepath) |
|
for key, row in df.iterrows(): |
|
if self.config.name == "dalaj": |
|
|
|
yield key, { |
|
"original_sentence": row["original sentence"], |
|
"corrected_sentence": row["corrected sentence"], |
|
"error_indices": row["error indices"], |
|
"corrected_indices": row["corrected indices"], |
|
"error_corr_pair": row["error-corr pair"], |
|
"error_label": row["error label"], |
|
"l1": row["l1"], |
|
"approximate_level": row["approximate level"], |
|
} |
|
elif self.config.name == "sweana": |
|
yield key, { |
|
"a": row["A"], |
|
"b": row["B"], |
|
"c": row["C"], |
|
"d": row["D"], |
|
"relation": row["relation"], |
|
} |
|
elif self.config.name == "swediag": |
|
yield key, { |
|
"premise": row["Premise_SE"], |
|
"hypothesis": row["Hypothesis_SE"], |
|
"label": row["Label"], |
|
} |
|
elif self.config.name == "swefaq": |
|
yield key, { |
|
"question": row["question"], |
|
"candidate_answer": row["candidate_answer"], |
|
"correct_answer": row["correct_answer"], |
|
} |
|
elif self.config.name == "swefracas": |
|
yield key, { |
|
"answer": row["svar"], |
|
"question": row["fråga"], |
|
"premiss_1": row["premiss_1"], |
|
"premiss_2": row["premiss_2"], |
|
"premiss_3": row["premiss_3"], |
|
"premiss_4": row["premiss_4"], |
|
"premiss_5": row["premiss_5"], |
|
} |
|
elif self.config.name == "swepar": |
|
yield key, { |
|
"sentence_1": row["sentence_1"], |
|
"sentence_2": row["sentence_2"], |
|
"similarity_score": row["similarity_score"], |
|
} |
|
elif self.config.name == "swesat": |
|
yield key, { |
|
"target_item": row["target_item"], |
|
"answer_1": row["answer_1"], |
|
"answer_2": row["answer_2"], |
|
"answer_3": row["answer_3"], |
|
"answer_4": row["answer_4"], |
|
"answer_5": row["answer_5"], |
|
} |
|
elif self.config.name == "swesim_relatedness": |
|
yield key, { |
|
"word_1": row["Word 1 "], |
|
"word_2": row[" Word 2 "], |
|
"relatedness": row[" Average "], |
|
} |
|
elif self.config.name == "swesim_similarity": |
|
yield key, { |
|
"word_1": row["Word 1 "], |
|
"word_2": row[" Word 2 "], |
|
"similarity": row[" Average "], |
|
} |
|
elif self.config.name == "swewgr": |
|
yield key, { |
|
"text": row["text"], |
|
"challenge": row["challenge"], |
|
"responses": row["responses"], |
|
} |
|
elif self.config.name == "swewic": |
|
yield key, { |
|
"sentence_1": row["sentence1"], |
|
"word_1": row["word1"], |
|
"sentence_2": row["sentence2"], |
|
"word_2": row["word2"], |
|
"same_sense": row["same_sense"], |
|
"start_1": row["start1"], |
|
"end_1": row["end1"], |
|
"start_2": row["start2"], |
|
"end_2": row["end2"], |
|
} |
|
elif self.config.name == "swewsc": |
|
yield key, { |
|
"passage": row["passage"], |
|
"challenge_text": row["challenge_text"], |
|
"response_text": row["response_text"], |
|
"challenge_begin":row["challenge_begin"], |
|
"challenge_end":row["challenge_end"], |
|
"response_begin":row["response_begin"], |
|
"response_end":row["response_end"], |
|
"label":row["label"] |
|
} |
|
else: |
|
yield key, { |
|
"sentence": data["sentence"], |
|
"option2": data["option2"], |
|
"second_domain_answer": "" if split == "test" else data["second_domain_answer"], |
|
} |
|
|
|
|