| import datasets |
| from glob import glob |
| import json |
| import zipfile |
| from random import shuffle |
|
|
| _DESCRIPTION = """\ |
| Masader is the largest public catalogue for Arabic NLP datasets, which consists of more than 200 datasets annotated with 25 attributes. |
| """ |
|
|
| _CITATION = """\ |
| @misc{alyafeai2021masader, |
| title={Masader: Metadata Sourcing for Arabic Text and Speech Data Resources}, |
| author={Zaid Alyafeai and Maraim Masoud and Mustafa Ghaleb and Maged S. Al-shaibani}, |
| year={2021}, |
| eprint={2110.06744}, |
| archivePrefix={arXiv}, |
| primaryClass={cs.CL} |
| } |
| """ |
|
|
|
|
| class MasaderConfig(datasets.BuilderConfig): |
| """BuilderConfig for Masader.""" |
|
|
| def __init__(self, **kwargs): |
| """BuilderConfig for MetRec. |
| Args: |
| **kwargs: keyword arguments forwarded to super. |
| """ |
| super(MasaderConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs) |
|
|
|
|
| class Masader(datasets.GeneratorBasedBuilder): |
| """Masaderdataset.""" |
|
|
| BUILDER_CONFIGS = [ |
| MasaderConfig( |
| name="plain_text", |
| description="Plain text", |
| ) |
| ] |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features( |
| { |
| 'Name': datasets.Value("string"), |
| 'Subsets': [{'Name':datasets.Value("string"), |
| 'Dialect':datasets.Value("string") , |
| 'Volume':datasets.Value("string") , |
| 'Unit':datasets.Value("string")}], |
| 'HF Link': datasets.Value("string"), |
| 'Link': datasets.Value("string"), |
| 'License': datasets.Value("string"), |
| 'Year': datasets.Value("int32"), |
| 'Language': datasets.Value("string"), |
| 'Dialect': datasets.Value("string"), |
| 'Domain': datasets.Value("string"), |
| 'Form': datasets.Value("string"), |
| 'Collection Style': datasets.Value("string"), |
| 'Description': datasets.Value("string"), |
| 'Volume': datasets.Value("string"), |
| 'Unit': datasets.Value("string"), |
| 'Ethical Risks': datasets.Value("string"), |
| 'Provider': datasets.Value("string"), |
| 'Derived From': datasets.Value("string"), |
| 'Paper Title': datasets.Value("string"), |
| 'Paper Link': datasets.Value("string"), |
| 'Script': datasets.Value("string"), |
| 'Tokenized': datasets.Value("string"), |
| 'Host': datasets.Value("string"), |
| 'Access': datasets.Value("string"), |
| 'Cost': datasets.Value("string"), |
| 'Test Split': datasets.Value("string"), |
| 'Tasks': datasets.Value("string"), |
| 'Venue Title': datasets.Value("string"), |
| 'Citations': datasets.Value("string"), |
| 'Venue Type': datasets.Value("string"), |
| 'Venue Name': datasets.Value("string"), |
| 'Authors': datasets.Value("string"), |
| 'Affiliations': datasets.Value("string"), |
| 'Abstract': datasets.Value("string"), |
| 'Added By': datasets.Value("string"), |
| } |
| ), |
| supervised_keys=None, |
| homepage="https://github.com/arbml/Masader", |
| citation=_CITATION,) |
| |
| def extract_all(self, dir): |
| zip_files = glob(dir+'/**/**.zip', recursive=True) |
| for file in zip_files: |
| with zipfile.ZipFile(file) as item: |
| item.extractall('/'.join(file.split('/')[:-1])) |
|
|
|
|
|
|
| def _split_generators(self, dl_manager): |
| url = ['https://github.com/ARBML/masader/archive/main.zip'] |
| downloaded_files = dl_manager.download_and_extract(url) |
| self.extract_all(downloaded_files[0]) |
| all_files = sorted(glob(downloaded_files[0]+'/masader-main/datasets/**.json')) |
| shuffle(all_files) |
| return [datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={'filepaths':{'inputs':all_files} })] |
| |
| def _generate_examples(self, filepaths): |
| for idx,filepath in enumerate(filepaths['inputs']): |
| with open(filepath, 'r') as f: |
| data = json.load(f) |
| yield idx, data |