| import json |
| import os |
| from pathlib import Path |
| import datasets |
|
|
|
|
| _CITATION = """ |
| @misc{gurari2018vizwiz, |
| title={VizWiz Grand Challenge: Answering Visual Questions from Blind People}, |
| author={Danna Gurari and Qing Li and Abigale J. Stangl and Anhong Guo and Chi Lin and Kristen Grauman and Jiebo Luo and Jeffrey P. Bigham}, |
| year={2018}, |
| eprint={1802.08218}, |
| archivePrefix={arXiv}, |
| primaryClass={cs.CV} |
| } |
| """ |
|
|
| _HOMEPAGE = "https://vizwiz.org/tasks-and-datasets/vqa/" |
|
|
| _DESCRIPTION = """ |
| The VizWiz-VQA dataset originates from a natural visual question answering setting where blind people |
| each took an image and recorded a spoken question about it, together with 10 crowdsourced answers per |
| visual question. The proposed challenge addresses the following two tasks for this dataset: predict the |
| answer to a visual question and (2) predict whether a visual question cannot be answered. |
| """ |
|
|
| _LICENSE = " Creative Commons Attribution 4.0 International License." |
|
|
| _DATA_URL = {"train" : "https://vizwiz.cs.colorado.edu/VizWiz_final/images/train.zip", |
| "test" : "https://vizwiz.cs.colorado.edu/VizWiz_final/images/test.zip", |
| "val" : "https://vizwiz.cs.colorado.edu/VizWiz_final/images/val.zip" } |
|
|
| _ANNOTATION_URL = "https://vizwiz.cs.colorado.edu/VizWiz_final/vqa_data/Annotations.zip" |
|
|
| _FEATURES = datasets.Features( |
| { |
| "id" : datasets.Value("int32"), |
| "image": datasets.Image(), |
| "filename": datasets.Value("string"), |
| "question": datasets.Value("string"), |
| "answers": datasets.Sequence(datasets.Value("string")), |
| "answers_original": [ |
| { |
| "answer": datasets.Value("string"), |
| "answer_confidence": datasets.Value("string"), |
| } |
| ], |
| "answer_type": datasets.Value("string"), |
| "answerable": datasets.Value("int32") |
| } |
| ) |
|
|
|
|
| class VizWiz(datasets.GeneratorBasedBuilder): |
| VERSION = datasets.Version("1.0.0") |
| def _info(self): |
| return datasets.DatasetInfo( |
| description = _DESCRIPTION, |
| features = _FEATURES, |
| homepage = _HOMEPAGE, |
| license = _LICENSE, |
| citation = _CITATION, |
| ) |
| |
| def _split_generators(self, dl_manager): |
| ann_file_train = os.path.join(dl_manager.download_and_extract(_ANNOTATION_URL), "train.json") |
| ann_file_val = os.path.join(dl_manager.download_and_extract(_ANNOTATION_URL), "val.json") |
| ann_file_test = os.path.join(dl_manager.download_and_extract(_ANNOTATION_URL), "test.json") |
| image_folders = {k: Path(v) for k, v in dl_manager.download_and_extract(_DATA_URL).items()} |
|
|
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={ |
| "annotation_file": ann_file_train, |
| "image_folders": image_folders, |
| "split_key": 'train' |
| }, |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.VALIDATION, |
| gen_kwargs={ |
| "annotation_file": ann_file_val, |
| "image_folders": image_folders, |
| "split_key": "val" |
| }, |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.TEST, |
| gen_kwargs={ |
| "annotation_file": ann_file_test, |
| "image_folders": image_folders, |
| "split_key": "test" |
| }, |
| ), |
| ] |
| |
| def _generate_examples(self, annotation_file,image_folders,split_key): |
| counter = 0 |
| annotations = json.load(open(annotation_file)) |
| for ann in annotations: |
| if split_key in ['train','val']: |
| answers = [answer["answer"] for answer in ann["answers"]] |
| answers_original = ann['answers'] |
| answer_type = ann["answer_type"] |
| answerable = ann["answerable"] |
| |
| else: |
| |
| answers = None |
| answers_original = None |
| answer_type = None |
| answerable = None |
| |
| yield counter, { |
| "id" : counter, |
| "image": str(image_folders[split_key]/split_key/ann['image']), |
| "filename" : ann['image'], |
| "question" : ann["question"], |
| "answers" : answers, |
| "answers_original" : answers_original, |
| "answer_type" : answer_type, |
| "answerable" : answerable |
| } |
| counter += 1 |
|
|