no code implementations • 16 Nov 2023 • Ramya Ramakrishnan, Ethan R. Elenberg, Hashan Narangodage, Ryan Mcdonald
In task-oriented dialogue, a system often needs to follow a sequence of actions, called a workflow, that complies with a set of guidelines in order to complete a task.
1 code implementation • 16 Nov 2023 • Shivanshu Gupta, Clemens Rosenbaum, Ethan R. Elenberg
Further, we experiment with two variations: (1) fine-tuning gist models for each dataset and (2) multi-task training a single model on a large collection of datasets.
1 code implementation • 23 Jul 2023 • Paloma Sodhi, Felix Wu, Ethan R. Elenberg, Kilian Q. Weinberger, Ryan Mcdonald
A common training technique for language models is teacher forcing (TF).
no code implementations • 26 May 2023 • Loay Mualem, Ethan R. Elenberg, Moran Feldman, Amin Karbasi
Despite the rich existing literature about minimax optimization in continuous settings, only very partial results of this kind have been obtained for combinatorial settings.
1 code implementation • 23 May 2023 • Anmol Kabra, Ethan R. Elenberg
Large, general purpose language models have demonstrated impressive performance across many different conversational domains.
no code implementations • 30 Sep 2022 • Nihal V. Nayak, Ethan R. Elenberg, Clemens Rosenbaum
We adapt existing approaches from the few-sample model evaluation literature to actively sub-sample, with a learned surrogate model, the most informative data points for annotation to estimate the evaluation metric.
2 code implementations • NeurIPS 2020 • Geoff Pleiss, Tianyi Zhang, Ethan R. Elenberg, Kilian Q. Weinberger
Not all data in a typical training set help with generalization; some samples can be overly ambiguous or outrightly mislabeled.
1 code implementation • WS 2019 • Jeremy Wohlwend, Ethan R. Elenberg, Samuel Altschul, Shawn Henry, Tao Lei
However, in many real-world applications the label set is frequently changing.
no code implementations • 25 Sep 2019 • Geoff Pleiss, Tianyi Zhang, Ethan R. Elenberg, Kilian Q. Weinberger
This paper introduces a new method to discover mislabeled training samples and to mitigate their impact on the training process of deep networks.
no code implementations • 7 Jun 2018 • Maurice Diesendruck, Ethan R. Elenberg, Rajat Sen, Guy W. Cole, Sanjay Shakkottai, Sinead A. Williamson
Deep generative networks can simulate from a complex target distribution, by minimizing a loss with respect to samples from that distribution.
1 code implementation • NeurIPS 2017 • Ethan R. Elenberg, Alexandros G. Dimakis, Moran Feldman, Amin Karbasi
In many machine learning applications, it is important to explain the predictions of a black-box classifier.
no code implementations • 2 Dec 2016 • Ethan R. Elenberg, Rajiv Khanna, Alexandros G. Dimakis, Sahand Negahban
Our results extend the work of Das and Kempe (2011) from the setting of linear regression to arbitrary objective functions.