├── images ├── vq2a_examples.png └── gif_vq2a_approach.gif ├── LICENSE └── README.md /images/vq2a_examples.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/google-research-datasets/maverics/HEAD/images/vq2a_examples.png -------------------------------------------------------------------------------- /images/gif_vq2a_approach.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/google-research-datasets/maverics/HEAD/images/gif_vq2a_approach.gif -------------------------------------------------------------------------------- /LICENSE: -------------------------------------------------------------------------------- 1 | The dataset may be freely used for any purpose, although acknowledgement of 2 | Google LLC ("Google") as the data source would be appreciated. The dataset is 3 | provided "AS IS" without any warranty, express or implied. Google disclaims all 4 | liability for any damages, direct or indirect, resulting from the use of the 5 | dataset. 6 | -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- 1 | # MAVERICS 2 | 3 |
4 |
5 |
12 |
13 |
23 | dataset str: dataset name
24 | split str: dataset split
25 | annotations List of image-question-answers triplets, each of which is
26 | -- image_id str: image ID
27 | -- caption str: image caption
28 | -- qa_pairs List of question-answer pairs, each of which is
29 | ---- question_id str: question ID
30 | ---- raw_question str: raw question
31 | ---- question str: processed question
32 | ---- answers List of str: 10 ground-truth answers
33 |
44 | @inproceedings{changpinyo2022vq2a,
45 | title = {All You May Need for VQA are Image Captions},
46 | author = {Changpinyo, Soravit and Kukliansky, Doron and Szpektor, Idan and Chen, Xi and Ding, Nan and Soricut, Radu},
47 | booktitle = {NAACL},
48 | year = {2022},
49 | }
50 |