no code implementations • 16 Nov 2023 • Michael J. Q. Zhang, Eunsol Choi
In this work, we study such behavior in LMs by proposing a task-agnostic framework for resolving ambiguity by asking users clarifying questions.
1 code implementation • NeurIPS 2023 • Shankar Padmanabhan, Yasumasa Onoe, Michael J. Q. Zhang, Greg Durrett, Eunsol Choi
Then, we update the model parameters so that the distribution of the LM (the student) matches the distribution of the LM conditioned on the definition (the teacher) on the transfer set.
no code implementations • 24 May 2023 • Jeremy R. Cole, Michael J. Q. Zhang, Daniel Gillick, Julian Martin Eisenschlos, Bhuwan Dhingra, Jacob Eisenstein
We investigate question answering from this perspective, focusing on answering a subset of questions with a high degree of accuracy, from a set of questions in which many are inherently ambiguous.
1 code implementation • 24 May 2023 • Michael J. Q. Zhang, Eunsol Choi
While large language models are able to retain vast amounts of world knowledge seen during pretraining, such knowledge is prone to going out of date and is nontrivial to update.
1 code implementation • 2 May 2023 • Yasumasa Onoe, Michael J. Q. Zhang, Shankar Padmanabhan, Greg Durrett, Eunsol Choi
Pre-trained language models (LMs) are used for knowledge intensive tasks like question answering, but their knowledge gets continuously outdated as the world changes.
no code implementations • 1 Mar 2023 • Jeremy R. Cole, Palak Jain, Julian Martin Eisenschlos, Michael J. Q. Zhang, Eunsol Choi, Bhuwan Dhingra
We propose representing factual changes between paired documents as question-answer pairs, where the answer to the same question differs between two versions.
no code implementations • 25 Oct 2022 • Hung-Ting Chen, Michael J. Q. Zhang, Eunsol Choi
Question answering models can use rich knowledge sources -- up to one hundred retrieved passages and parametric knowledge in the large-scale language model (LM).
no code implementations • Findings (NAACL) 2022 • Yasumasa Onoe, Michael J. Q. Zhang, Eunsol Choi, Greg Durrett
Given its wide coverage on entity knowledge and temporal indexing, our dataset can be used to evaluate LMs and techniques designed to modify or extend their knowledge.
1 code implementation • EMNLP 2021 • Michael J. Q. Zhang, Eunsol Choi
To construct SituatedQA, we first identify such questions in existing QA datasets.
2 code implementations • 3 Sep 2021 • Yasumasa Onoe, Michael J. Q. Zhang, Eunsol Choi, Greg Durrett
We introduce CREAK, a testbed for commonsense reasoning about entity knowledge, bridging fact-checking about entities (Harry Potter is a wizard and is skilled at riding a broomstick) with commonsense inferences (if you're good at a skill you can teach others how to do it).