Info hash | f075ad12eccbbd665aec68db5d208dc68e7a384f |
Last mirror activity | 21:01 ago |
Size | 7.98GB (7,984,418,554 bytes) |
Added | 2015-10-08 16:06:59 |
Views | 1850 |
Hits | 2570 |
ID | 2982 |
Type | multi |
Downloaded | 175 time(s) |
Uploaded by | joecohen |
Folder | data |
Num files | 23 files [See full list] |
Mirrors | 8 complete, 0 downloading = 8 mirror(s) total [Log in to see full list] |
data (23 files)
abstract_v002/captions_abstract_v002_train-val2015.zip | 3.56MB |
abstract_v002/scene_img/scene_img_abstract_v002_test2015.zip | 2.89GB |
abstract_v002/scene_img/scene_img_abstract_v002_train2015.zip | 2.91GB |
abstract_v002/scene_img/scene_img_abstract_v002_val2015.zip | 1.45GB |
abstract_v002/scene_json/scene_json_abstract_v002_test2015.zip | 233.66MB |
abstract_v002/scene_json/scene_json_abstract_v002_train2015.zip | 235.00MB |
abstract_v002/scene_json/scene_json_abstract_v002_val2015.zip | 117.53MB |
abstract_v002/vqa/Annotations_Train_abstract_v002.zip | 2.65MB |
abstract_v002/vqa/Annotations_Val_abstract_v002.zip | 1.32MB |
abstract_v002/vqa/Questions_Test_abstract_v002.zip | 4.39MB |
abstract_v002/vqa/Questions_Train_abstract_v002.zip | 4.61MB |
abstract_v002/vqa/Questions_Val_abstract_v002.zip | 2.24MB |
mscoco/prev_rel/Beta_v0.1/Annotations_Train_mscoco.zip | 1.52MB |
mscoco/prev_rel/Beta_v0.1/Questions_Train_mscoco.zip | 2.81MB |
mscoco/prev_rel/Beta_v0.9/Annotations_Train_mscoco.zip | 11.82MB |
mscoco/prev_rel/Beta_v0.9/Annotations_Val_mscoco.zip | 5.87MB |
mscoco/prev_rel/Beta_v0.9/Questions_Train_mscoco.zip | 22.07MB |
mscoco/prev_rel/Beta_v0.9/Questions_Val_mscoco.zip | 10.67MB |
mscoco/vqa/Annotations_Train_mscoco.zip | 12.17MB |
mscoco/vqa/Annotations_Val_mscoco.zip | 6.03MB |
mscoco/vqa/Questions_Test_mscoco.zip | 26.51MB |
mscoco/vqa/Questions_Train_mscoco.zip | 21.99MB |
mscoco/vqa/Questions_Val_mscoco.zip | 10.59MB |
Type: Dataset
Tags: deeplearning
Bibtex:
Tags: deeplearning
Bibtex:
@article{, title= {VQA: Visual Question Answering Dataset}, journal= {}, author= {Stanislaw Antol and Aishwarya Agrawal and Jiasen Lu and Margaret Mitchell and Dhruv Batra and C. Lawrence Zitnick and Devi Parikh}, booktitle= {International Conference on Computer Vision (ICCV)}, year= {2015}, url= {http://visualqa.org/}, abstract= {254,721 images, 764,163 questions, 9,934,119 answers! ===What is VQA? VQA is a new dataset containing open-ended questions about images. These questions require an understanding of vision, language and commonsense knowledge to answer. Over 250K images (MSCOCO and abstract scenes) 3 questions per image 10 ground truth answers per question 3 plausible (but likely incorrect) answers per question Open-ended and multiple-choice answering tasks Automatic evaluation metric ===Overview For every image, we collected 3 free-form natural-language questions with 10 concise open-ended answers each. We provide two formats of the VQA task: open-ended and multiple-choice. For additional details, please see the VQA paper. The annotations we release are the result of the following post-processing steps on the raw crowdsourced data: Spelling correction (using Bing Speller) of question and answer strings Question normalization (first char uppercase, last char ???) Answer normalization (all chars lowercase, no period except as decimal point, number words ?> digits, strip articles (a, an the)) Adding apostrophe if a contraction is missing it (e.g., convert "dont" to "don't") Please follow the instructions in the README to download and setup the VQA data (annotations and images). ==October 2015: Full release (v1.0) Real Images 204,721 MSCOCO images (all of current train/val/test) 614,163 questions 6,141,630 ground truth answers 1,842,489 plausible answers Abstract Scenes 50,000 abstract scenes 150,000 questions 1,500,000 ground truth answers 450,000 plausible answers 250,000 captions ==July 2015: Beta v0.9 release 123,287 MSCOCO images (all of train/val) 369,861 questions 3,698,610 ground truth answers 1,109,583 plausible answers ==June 2015: Beta v0.1 release 10,000 MSCOCO images (from train) 30,000 questions 300,000 ground truth answers 90,000 plausible answers }, keywords= {deeplearning}, terms= {View the terms here: http://visualqa.org/terms.html} }