Question Answering

2883 papers with code • 130 benchmarks • 360 datasets

Question Answering is the task of answering questions (typically reading comprehension questions), but abstaining when presented with a question that cannot be answered based on the provided context.

Question answering can be segmented into domain-specific tasks like community question answering and knowledge-base question answering. Popular benchmark datasets for evaluation question answering systems include SQuAD, HotPotQA, bAbI, TriviaQA, WikiQA, and many others. Models for question answering are typically evaluated on metrics like EM and F1. Some recent top performing models are T5 and XLNet.

( Image credit: SQuAD )

Libraries

Use these libraries to find Question Answering models and implementations
27 papers
125,059
5 papers
2,548
See all 11 libraries.

Latest papers with no code

Retrieval Augmented Generation for Domain-specific Question Answering

no code yet • 23 Apr 2024

Question answering (QA) has become an important application in the advanced development of large language models.

Grounded Knowledge-Enhanced Medical VLP for Chest X-Ray

no code yet • 23 Apr 2024

Medical vision-language pre-training has emerged as a promising approach for learning domain-general representations of medical image and text.

Pegasus-v1 Technical Report

no code yet • 23 Apr 2024

This technical report introduces Pegasus-1, a multimodal language model specialized in video content understanding and interaction through natural language.

MedDr: Diagnosis-Guided Bootstrapping for Large-Scale Medical Vision-Language Learning

no code yet • 23 Apr 2024

The rapid advancement of large-scale vision-language models has showcased remarkable capabilities across various tasks.

Multi-view Content-aware Indexing for Long Document Retrieval

no code yet • 23 Apr 2024

As they do not consider content structures, the resultant chunks can exclude vital information or include irrelevant content.

Med42 -- Evaluating Fine-Tuning Strategies for Medical LLMs: Full-Parameter vs. Parameter-Efficient Approaches

no code yet • 23 Apr 2024

This study presents a comprehensive analysis and comparison of two predominant fine-tuning methodologies - full-parameter fine-tuning and parameter-efficient tuning - within the context of medical Large Language Models (LLMs).

Insights into Alignment: Evaluating DPO and its Variants Across Multiple Tasks

no code yet • 23 Apr 2024

Key observations reveal that alignment methods achieve optimal performance with smaller training data subsets, exhibit limited effectiveness in reasoning tasks yet significantly impact mathematical problem-solving, and employing an instruction-tuned model notably influences truthfulness.

XC-Cache: Cross-Attending to Cached Context for Efficient LLM Inference

no code yet • 23 Apr 2024

Just-in-time processing of a context is inefficient due to the quadratic cost of self-attention operations, and caching is desirable.

Wiki-LLaVA: Hierarchical Retrieval-Augmented Generation for Multimodal LLMs

no code yet • 23 Apr 2024

Multimodal LLMs are the natural evolution of LLMs, and enlarge their capabilities so as to work beyond the pure textual modality.

Boter: Bootstrapping Knowledge Selection and Question Answering for Knowledge-based VQA

no code yet • 22 Apr 2024

Knowledge-based Visual Question Answering (VQA) requires models to incorporate external knowledge to respond to questions about visual content.