Dynamic Memory Networks for Visual and Textual Question Answering

4 Mar 2016  ·  Caiming Xiong, Stephen Merity, Richard Socher ·

Neural network architectures with memory and attention mechanisms exhibit certain reasoning capabilities required for question answering. One such architecture, the dynamic memory network (DMN), obtained high accuracy on a variety of language tasks. However, it was not shown whether the architecture achieves strong results for question answering when supporting facts are not marked during training or whether it could be applied to other modalities such as images. Based on an analysis of the DMN, we propose several improvements to its memory and input modules. Together with these changes we introduce a novel input module for images in order to be able to answer visual questions. Our new DMN+ model improves the state of the art on both the Visual Question Answering dataset and the \babi-10k text question-answering dataset without supporting fact supervision.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Visual Question Answering (VQA) COCO Visual Question Answering (VQA) real images 1.0 open ended DMN+ [xiong2016dynamic] Percentage correct 60.4 # 8
Visual Question Answering (VQA) VQA v1 test-dev DMN+ Accuracy 60.3 # 6
Visual Question Answering (VQA) VQA v1 test-std DMN+ Accuracy 60.4 # 4

Methods