Datasets:

ArXiv:
License:
the-vault-function / the-vault-function.py
NamCyan's picture
update
9ff26ff
raw
history blame
8.13 kB
import os
import pyarrow as pa
import pyarrow.parquet as pq
import datasets
_REPO_NAME = 'Fsoft-AIC/the-vault'
_LANG_TO_TEXT = {
"python": "python",
"c": "c",
# "c#": "c_sharp",
# "c++": "cpp",
# "go": "go",
# "Java": "java",
# "javascript": "javascript",
# "php": "php",
# "ruby": "ruby",
# "rust": "rust",
}
_DESCRIPTION = """The Vault"""
_HOMEPAGE = "https://huggingface.co/Fsoft-AIC"
_TEXT_TO_LANG = {}
for lang in _LANG_TO_TEXT:
_TEXT_TO_LANG[_LANG_TO_TEXT[lang]] = lang
_LANG_CONFIGS = ["all"] + list(_TEXT_TO_LANG.keys())
num_shard_split = {
'train/small/python': 1,
'train/medium/python': 1,
'train/small/c': 1,
'train/medium/c': 1
}
_SPLIT_CONFIGS = ["all", "train", "train/small", "train/medium"]
class TheVaultFunctionConfig(datasets.BuilderConfig):
"""BuilderConfig for The Vault dataset."""
def __init__(self, *args, languages=["all"], split_set= ["all"], **kwargs):
"""BuilderConfig for the GitHub Code dataset.
Args:
split_set (:obj:`List[str]`): List of split set to load.
languages (:obj:`List[str]`): List of languages to load.
**kwargs: keyword arguments forwarded to super.
"""
super().__init__(
*args,
name= "+".join([split.replace("/", "_") for split in split_set]) + "-" + "+".join(languages),
**kwargs,
)
languages = set([lang.lower() for lang in languages])
split_set = set([split.lower() for split in split_set])
assert all([language in _LANG_CONFIGS for language in languages]), f"languages {languages} contains language not in {_LANG_CONFIGS}."
assert all([split in _SPLIT_CONFIGS for split in split_set]), f"split_set {split_set} contains element not in {_SPLIT_CONFIGS}."
if "all" in split_set:
assert len(split_set)==1, f"Passed 'all' together with other split sets. {split_set}"
elif "train" in split_set:
for split in split_set:
if "train" in split and split != "train":
raise ValueError(f"Split set 'train' already contains '{split}'. Please only include one.")
if "all" in languages:
assert len(languages)==1, f"Passed 'all' together with other languages. {languages}"
# self.filter_languages = False
# else:
# self.filter_languages = True
self.languages = set(languages)
self.split_set= split_set
class TheVaultFunction(datasets.GeneratorBasedBuilder):
"""The Vault dataset."""
VERSION = datasets.Version("1.0.0")
BUILDER_CONFIG_CLASS = TheVaultFunctionConfig
BUILDER_CONFIGS = [TheVaultFunctionConfig(languages=[lang], split_set=[spl]) for lang in _LANG_CONFIGS for spl in _SPLIT_CONFIGS]
DEFAULT_CONFIG_NAME = "all-all"
def _info(self):
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=datasets.Features({
"repo": datasets.Value("string"),
"path": datasets.Value("string"),
"license": datasets.Value("string"),
"language": datasets.Value("string"),
"identifier": datasets.Value("string"),
"return_type": datasets.Value("string"),
# "original_string": datasets.Value("string"),
"original_docstring": datasets.Value("string"),
"docstring": datasets.Value("string"),
"code": datasets.Value("string"),
"code_tokens": datasets.Value("string"),
"docstring_tokens": datasets.Value("string"),
"short_docstring": datasets.Value("string"),
"short_docstring_tokens": datasets.Value("string"),
"comment": datasets.Value("string"),
}),
supervised_keys=None,
homepage=_HOMEPAGE,
)
def _split_generators(self, dl_manager):
generators = []
split_set = list(self.config.split_set)
languages = list(self.config.languages)
if "all" in split_set:
split_set = _SPLIT_CONFIGS[1:]
load_full_train = False
if "train" in split_set:
split_set.remove('train')
split_set.extend(["train/small", "train/medium"])
load_full_train = True
if "all" in languages:
languages = _LANG_CONFIGS[1:]
train_split_files = []
for split in split_set:
split_files = []
for language in languages:
num_shards = num_shard_split[f"{split}/{language}"]
data_files = [
f"data/{split}/{language}-{_index:05d}-of-{num_shards:05d}.parquet"
for _index in range(num_shards)
]
files = dl_manager.download(data_files)
split_files.extend(files)
if load_full_train and "train" in split:
train_split_files.extend(split_files)
else:
generators.append(
datasets.SplitGenerator(
name=split.replace("/", "_"),
gen_kwargs={
"files": split_files,
},
),
)
if load_full_train and train_split_files:
generators = [datasets.SplitGenerator(name="train", gen_kwargs={"files": train_split_files})] + generators
return generators
def _generate_examples(self, files):
key = 0
for file_idx, file in enumerate(files):
with open(file, "rb") as f:
parquet_file = pq.ParquetFile(f)
for batch_idx, record_batch in enumerate(parquet_file.iter_batches(batch_size=10_000)):
pa_table = pa.Table.from_batches([record_batch])
for row_index in range(pa_table.num_rows):
row = pa_table.slice(row_index, 1).to_pydict()
# lang = row['language'][0]
# if self.config.filter_languages and not lang in self.config.languages:
# continue
yield key, {
"repo": row['repo'][0],
"path": row['path'][0],
"license": row['license'][0],
"language": row['language'][0],
"identifier": row['identifier'][0],
"return_type": row['return_type'][0],
# "original_string": row['original_string'][0],
"original_docstring": row['original_docstring'][0],
"docstring": row['docstring'][0],
"docstring_tokens": row['docstring_tokens'][0],
"code": row['code'][0],
"code_tokens": row['code_tokens'][0],
"short_docstring": row['short_docstring'][0],
"short_docstring_tokens": row['short_docstring_tokens'][0],
"comment": row['comment'][0]
}
key += 1