Question Answering
2883 papers with code • 130 benchmarks • 360 datasets
Question Answering is the task of answering questions (typically reading comprehension questions), but abstaining when presented with a question that cannot be answered based on the provided context.
Question answering can be segmented into domain-specific tasks like community question answering and knowledge-base question answering. Popular benchmark datasets for evaluation question answering systems include SQuAD, HotPotQA, bAbI, TriviaQA, WikiQA, and many others. Models for question answering are typically evaluated on metrics like EM and F1. Some recent top performing models are T5 and XLNet.
( Image credit: SQuAD )
Libraries
Use these libraries to find Question Answering models and implementationsDatasets
Subtasks
- Open-Ended Question Answering
- Open-Domain Question Answering
- Conversational Question Answering
- Answer Selection
- Answer Selection
- Knowledge Base Question Answering
- Community Question Answering
- Zero-Shot Video Question Answer
- Multiple Choice Question Answering (MCQA)
- Long Form Question Answering
- Science Question Answering
- Generative Question Answering
- Cross-Lingual Question Answering
- Mathematical Question Answering
- Temporal/Casual QA
- Logical Reasoning Question Answering
- Multilingual Machine Comprehension in English Hindi
- True or False Question Answering
- Question Quality Assessment
Latest papers with no code
Retrieval Augmented Generation for Domain-specific Question Answering
Question answering (QA) has become an important application in the advanced development of large language models.
Grounded Knowledge-Enhanced Medical VLP for Chest X-Ray
Medical vision-language pre-training has emerged as a promising approach for learning domain-general representations of medical image and text.
Pegasus-v1 Technical Report
This technical report introduces Pegasus-1, a multimodal language model specialized in video content understanding and interaction through natural language.
MedDr: Diagnosis-Guided Bootstrapping for Large-Scale Medical Vision-Language Learning
The rapid advancement of large-scale vision-language models has showcased remarkable capabilities across various tasks.
Multi-view Content-aware Indexing for Long Document Retrieval
As they do not consider content structures, the resultant chunks can exclude vital information or include irrelevant content.
Med42 -- Evaluating Fine-Tuning Strategies for Medical LLMs: Full-Parameter vs. Parameter-Efficient Approaches
This study presents a comprehensive analysis and comparison of two predominant fine-tuning methodologies - full-parameter fine-tuning and parameter-efficient tuning - within the context of medical Large Language Models (LLMs).
Insights into Alignment: Evaluating DPO and its Variants Across Multiple Tasks
Key observations reveal that alignment methods achieve optimal performance with smaller training data subsets, exhibit limited effectiveness in reasoning tasks yet significantly impact mathematical problem-solving, and employing an instruction-tuned model notably influences truthfulness.
XC-Cache: Cross-Attending to Cached Context for Efficient LLM Inference
Just-in-time processing of a context is inefficient due to the quadratic cost of self-attention operations, and caching is desirable.
Wiki-LLaVA: Hierarchical Retrieval-Augmented Generation for Multimodal LLMs
Multimodal LLMs are the natural evolution of LLMs, and enlarge their capabilities so as to work beyond the pure textual modality.
Boter: Bootstrapping Knowledge Selection and Question Answering for Knowledge-based VQA
Knowledge-based Visual Question Answering (VQA) requires models to incorporate external knowledge to respond to questions about visual content.