Search Results for author: Shankar Padmanabhan

Found 4 papers, 2 papers with code

Propagating Knowledge Updates to LMs Through Distillation

1 code implementation NeurIPS 2023 Shankar Padmanabhan, Yasumasa Onoe, Michael J. Q. Zhang, Greg Durrett, Eunsol Choi

Then, we update the model parameters so that the distribution of the LM (the student) matches the distribution of the LM conditioned on the definition (the teacher) on the transfer set.

knowledge editing Language Modelling

Can LMs Learn New Entities from Descriptions? Challenges in Propagating Injected Knowledge

1 code implementation2 May 2023 Yasumasa Onoe, Michael J. Q. Zhang, Shankar Padmanabhan, Greg Durrett, Eunsol Choi

Pre-trained language models (LMs) are used for knowledge intensive tasks like question answering, but their knowledge gets continuously outdated as the world changes.

Question Answering

Self-Programming Artificial Intelligence Using Code-Generating Language Models

no code implementations30 Apr 2022 Alex Sheng, Shankar Padmanabhan

Prior work in meta-learning and neural architecture search has led to substantial successes across various task domains, spawning myriad approaches for algorithmically optimizing the design and learning dynamics of deep learning models.

Code Generation Language Modelling +2

Cannot find the paper you are looking for? You can Submit a new open access paper.