Vietnamese Visual Question Answering

4 papers with code • 0 benchmarks • 0 datasets

This task has no description! Would you like to contribute one?

Most implemented papers

OpenViVQA: Task, Dataset, and Multimodal Fusion Models for Visual Question Answering in Vietnamese

hieunghia-pat/openvivqa-dataset 7 May 2023

The VQA task requires methods that have the ability to fuse the information from questions and images to produce appropriate answers.

ViCLEVR: A Visual Reasoning Dataset and Hybrid Multimodal Fusion Model for Visual Question Answering in Vietnamese

kvt0012/viclevr 27 Oct 2023

Neural models for VQA have made remarkable progress on large-scale datasets, with a primary focus on resource-rich languages like English.

ViTextVQA: A Large-Scale Visual Question Answering Dataset for Evaluating Vietnamese Text Comprehension in Images

minhquan6203/vitextvqa-dataset 16 Apr 2024

Visual Question Answering (VQA) is a complicated task that requires the capability of simultaneously processing natural language and images.