1 code implementation • 17 Nov 2022 • Lukas Hedegaard, Aman Alok, Juby Jose, Alexandros Iosifidis
To improve on this, we propose Structured Pruning Adapters (SPAs), a family of compressing, task-switching network adapters, that accelerate and specialize networks using tiny parameter sets and structured pruning.
no code implementations • 31 Oct 2022 • Aman Alok, Rahul Gupta, Shankar Ananthakrishnan
Hypothesis rejection modules in both schemes reject/accept a hypothesis based on features drawn from the utterance directed to the SLU system, the associated SLU hypothesis and SLU confidence score.
no code implementations • 28 Jan 2021 • Manoj Kumar, Varun Kumar, Hadrien Glaude, Cyprien delichy, Aman Alok, Rahul Gupta
We make use of a conditional generator for data augmentation that is trained directly using the meta-learning objective and simultaneously with prototypical networks, hence ensuring that data augmentation is customized to the task.
no code implementations • 22 May 2020 • Nikhita Vedula, Rahul Gupta, Aman Alok, Mukund Sridhar
We propose a novel framework, ADVIN, to automatically discover novel domains and intents from large volumes of unlabeled data.
no code implementations • 16 May 2020 • Aarsh Patel, Rahul Gupta, Mukund Harakere, Satyapriya Krishna, Aman Alok, Peng Liu
In this research work, we aim to achieve classification parity across explicit as well as implicit sensitive features.
no code implementations • 20 Jun 2019 • Rahul Gupta, Aman Alok, Shankar Ananthakrishnan
An OVA system consists of as many OVA models as the number of classes, providing the advantage of asynchrony, where each OVA model can be re-trained independent of other models.