Open-Domain Question Answering

197 papers with code • 15 benchmarks • 26 datasets

Open-domain question answering is the task of question answering on open-domain datasets such as Wikipedia.

Libraries

Use these libraries to find Open-Domain Question Answering models and implementations

Beyond Memorization: The Challenge of Random Memory Access in Language Models

sail-sg/lm-random-memory-access 12 Mar 2024

Through carefully-designed synthetic tasks, covering the scenarios of full recitation, selective recitation and grounded question answering, we reveal that LMs manage to sequentially access their memory while encountering challenges in randomly accessing memorized content.

6
12 Mar 2024

REAR: A Relevance-Aware Retrieval-Augmented Framework for Open-Domain Question Answering

rucaibox/rear 27 Feb 2024

By combining the improvements in both architecture and training, our proposed REAR can better utilize external knowledge by effectively perceiving the relevance of retrieved documents.

18
27 Feb 2024

RetrievalQA: Assessing Adaptive Retrieval-Augmented Generation for Short-form Open-Domain Question Answering

hyintell/retrievalqa 26 Feb 2024

Based on our findings, we propose Time-Aware Adaptive Retrieval (TA-ARE), a simple yet effective method that helps LLMs assess the necessity of retrieval without calibration or additional training.

42
26 Feb 2024

Pre-training Cross-lingual Open Domain Question Answering with Large-scale Synthetic Supervision

fantabulous-j/class 26 Feb 2024

Cross-lingual question answering (CLQA) is a complex problem, comprising cross-lingual retrieval from a multilingual knowledge base, followed by answer generation either in English or the query language.

0
26 Feb 2024

VerAs: Verify then Assess STEM Lab Reports

psunlpgroup/veras 7 Feb 2024

With an increasing focus in STEM education on critical thinking skills, science writing plays an ever more important role in curricula that stress inquiry skills.

0
07 Feb 2024

Can AI Assistants Know What They Don't Know?

openmoss/say-i-dont-know 24 Jan 2024

To answer this question, we construct a model-specific "I don't know" (Idk) dataset for an assistant, which contains its known and unknown questions, based on existing open-domain question answering datasets.

41
24 Jan 2024

Mitigating the Impact of False Negatives in Dense Retrieval with Contrastive Confidence Regularization

wangskygit/passage-sieve 30 Dec 2023

Hard negative sampling, which is commonly used to improve contrastive learning, can introduce more noise in training.

12
30 Dec 2023

Learning to Filter Context for Retrieval-Augmented Generation

zorazrw/filco 14 Nov 2023

To alleviate these problems, we propose FILCO, a method that improves the quality of the context provided to the generator by (1) identifying useful context based on lexical and information-theoretic approaches, and (2) training context filtering models that can filter retrieved contexts at test time.

148
14 Nov 2023

Detrimental Contexts in Open-Domain Question Answering

xfactlab/emnlp2023-damaging-retrieval 27 Oct 2023

However, counter-intuitively, too much context can have a negative impact on the model when evaluated on common question answering (QA) datasets.

5
27 Oct 2023

Knowledge Corpus Error in Question Answering

xfactlab/emnlp2023-knowledge-corpus-error 27 Oct 2023

This error arises when the knowledge corpus used for retrieval is only a subset of the entire string space, potentially excluding more helpful passages that exist outside the corpus.

0
27 Oct 2023