Vietnamese Visual Question Answering
4 papers with code • 0 benchmarks • 0 datasets
This task has no description! Would you like to contribute one?
Benchmarks
These leaderboards are used to track progress in Vietnamese Visual Question Answering
No evaluation results yet. Help compare methods by
submitting
evaluation metrics.
Most implemented papers
OpenViVQA: Task, Dataset, and Multimodal Fusion Models for Visual Question Answering in Vietnamese
The VQA task requires methods that have the ability to fuse the information from questions and images to produce appropriate answers.
ViCLEVR: A Visual Reasoning Dataset and Hybrid Multimodal Fusion Model for Visual Question Answering in Vietnamese
Neural models for VQA have made remarkable progress on large-scale datasets, with a primary focus on resource-rich languages like English.
ViTextVQA: A Large-Scale Visual Question Answering Dataset for Evaluating Vietnamese Text Comprehension in Images
Visual Question Answering (VQA) is a complicated task that requires the capability of simultaneously processing natural language and images.