no code implementations • 29 Apr 2023 • Rajshekhar Das, Jonathan Francis, Sanket Vaibhav Mehta, Jean Oh, Emma Strubell, Jose Moura
Self-training based on pseudo-labels has emerged as a dominant approach for addressing conditional distribution shifts in unsupervised domain adaptation (UDA) for semantic segmentation problems.
no code implementations • 19 Dec 2022 • Sanket Vaibhav Mehta, Jai Gupta, Yi Tay, Mostafa Dehghani, Vinh Q. Tran, Jinfeng Rao, Marc Najork, Emma Strubell, Donald Metzler
In this work, we introduce DSI++, a continual learning challenge for DSI to incrementally index new documents while being able to answer queries related to both previously and newly indexed documents.
no code implementations • 10 Jul 2022 • Shagun Sodhani, Mojtaba Faramarzi, Sanket Vaibhav Mehta, Pranshu Malviya, Mohamed Abdelsalam, Janarthanan Janarthanan, Sarath Chandar
Following these different classes of learning algorithms, we discuss the commonly used evaluation benchmarks and metrics for lifelong learning (Chapter 6) and wrap up with a discussion of future challenges and important research directions in Chapter 7.
no code implementations • 25 May 2022 • Clara Na, Sanket Vaibhav Mehta, Emma Strubell
Model compression by way of parameter pruning, quantization, or distillation has recently gained popularity as an approach for reducing the computational requirements of modern deep neural network models for NLP.
1 code implementation • NeurIPS 2023 • Sanket Vaibhav Mehta, Darshan Patil, Sarath Chandar, Emma Strubell
The lifelong learning paradigm in machine learning is an attractive alternative to the more prominent isolated learning scheme not only due to its resemblance to biological learning but also its potential to reduce energy waste by obviating excessive model re-training.
3 code implementations • ICLR 2022 • Vamsi Aribandi, Yi Tay, Tal Schuster, Jinfeng Rao, Huaixiu Steven Zheng, Sanket Vaibhav Mehta, Honglei Zhuang, Vinh Q. Tran, Dara Bahri, Jianmo Ni, Jai Gupta, Kai Hui, Sebastian Ruder, Donald Metzler
Despite the recent success of multi-task learning and transfer learning for natural language processing (NLP), few works have systematically studied the effect of scaling up the number of tasks during pre-training.
1 code implementation • ACL 2022 • Sanket Vaibhav Mehta, Jinfeng Rao, Yi Tay, Mihir Kale, Ankur P. Parikh, Emma Strubell
Data-to-text generation focuses on generating fluent natural language responses from structured meaning representations (MRs).
no code implementations • 29 Sep 2021 • Rajshekhar Das, Jonathan Francis, Sanket Vaibhav Mehta, Jean Oh, Emma Strubell, Jose Moura
Crucially, the objectness constraint is agnostic to the ground-truth semantic segmentation labels and, therefore, remains appropriate for unsupervised adaptation settings.
no code implementations • EMNLP 2020 • ZiRui Wang, Sanket Vaibhav Mehta, Barnabás Póczos, Jaime Carbonell
State-of-the-art lifelong language learning methods store past examples in episodic memory and replay them at both training and inference time.
1 code implementation • IJCNLP 2019 • Harsh Jhamtani, Sanket Vaibhav Mehta, Jaime Carbonell, Taylor Berg-Kirkpatrick
Existing recurrent neural language models often fail to capture higher-level structure present in text: for example, rhyming patterns present in poetry.
1 code implementation • 21 Oct 2018 • Sanket Vaibhav Mehta, Shagun Sodhani, Dhaval Patel
Spatial co-location pattern mining refers to the task of discovering the group of objects or events that co-occur at many places.
Databases Distributed, Parallel, and Cluster Computing
no code implementations • EMNLP 2018 • Sanket Vaibhav Mehta, Jay Yoon Lee, Jaime Carbonell
The paper proposes a semi-supervised semantic role labeling method that outperforms the state-of-the-art in limited SRL training corpora.
no code implementations • 26 Jul 2017 • Jay Yoon Lee, Sanket Vaibhav Mehta, Michael Wick, Jean-Baptiste Tristan, Jaime Carbonell
Practitioners apply neural networks to increasingly complex problems in natural language processing, such as syntactic parsing and semantic role labeling that have rich output structures.