Search Results for author: Tejaswini Pedapati

Found 17 papers, 1 papers with code

NeuroPrune: A Neuro-inspired Topological Sparse Training Algorithm for Large Language Models

no code implementations28 Feb 2024 Amit Dhurandhar, Tejaswini Pedapati, Ronny Luss, Soham Dan, Aurelie Lozano, Payel Das, Georgios Kollias

Transformer-based Language Models have become ubiquitous in Natural Language Processing (NLP) due to their impressive performance on various tasks.

Machine Translation Natural Language Inference

From PEFT to DEFT: Parameter Efficient Finetuning for Reducing Activation Density in Transformers

1 code implementation2 Feb 2024 Bharat Runwal, Tejaswini Pedapati, Pin-Yu Chen

Building upon this insight, in this work, we propose a novel density loss that encourages higher activation sparsity (equivalently, lower activation density) in the pre-trained models.

Neural Capacitance: A New Perspective of Neural Network Selection via Edge Dynamics

no code implementations11 Jan 2022 Chunheng Jiang, Tejaswini Pedapati, Pin-Yu Chen, Yizhou Sun, Jianxi Gao

To this end, we construct a network mapping $\phi$, converting a neural network $G_A$ to a directed line graph $G_B$ that is defined on those edges in $G_A$.

Model Selection

Contrastive Explanations for Comparing Preferences of Reinforcement Learning Agents

no code implementations17 Dec 2021 Jasmina Gajcin, Rahul Nair, Tejaswini Pedapati, Radu Marinescu, Elizabeth Daly, Ivana Dusparic

In complex tasks where the reward function is not straightforward and consists of a set of objectives, multiple reinforcement learning (RL) policies that perform task adequately, but employ different strategies can be trained by adjusting the impact of individual objectives on reward function.

Autonomous Driving reinforcement-learning +1

CoFrNets: Interpretable Neural Architecture Inspired by Continued Fractions

no code implementations NeurIPS 2021 Isha Puri, Amit Dhurandhar, Tejaswini Pedapati, Karthikeyan Shanmugam, Dennis Wei, Kush R. Varshney

We experiment on nonlinear synthetic functions and are able to accurately model as well as estimate feature attributions and even higher order terms in some cases, which is a testament to the representational power as well as interpretability of such architectures.

Multihop: Leveraging Complex Models to Learn Accurate Simple Models

no code implementations14 Sep 2021 Amit Dhurandhar, Tejaswini Pedapati

In this paper, we propose a meta-approach where we transfer information from the complex model to the simple model by dynamically selecting and/or constructing a sequence of intermediate models of decreasing complexity that are less intricate than the original complex model.

Explainable artificial intelligence Knowledge Distillation +2

Learning to Rank Learning Curves

no code implementations ICML 2020 Martin Wistuba, Tejaswini Pedapati

Many automated machine learning methods, such as those for hyperparameter and neural architecture optimization, are computationally expensive because they involve training many different model configurations.

Learning-To-Rank Neural Architecture Search +1

Learning Global Transparent Models Consistent with Local Contrastive Explanations

no code implementations NeurIPS 2020 Tejaswini Pedapati, Avinash Balakrishnan, Karthikeyan Shanmugam, Amit Dhurandhar

Based on a key insight we propose a novel method where we create custom boolean features from sparse local contrastive explanations of the black-box model and then train a globally transparent model on just these, and showcase empirically that such models have higher local consistency compared with other known strategies, while still being close in performance to models that are trained with access to the original data.

counterfactual

Model Agnostic Contrastive Explanations for Structured Data

no code implementations31 May 2019 Amit Dhurandhar, Tejaswini Pedapati, Avinash Balakrishnan, Pin-Yu Chen, Karthikeyan Shanmugam, Ruchir Puri

Recently, a method [7] was proposed to generate contrastive explanations for differentiable models such as deep neural networks, where one has complete access to the model.

A Survey on Neural Architecture Search

no code implementations4 May 2019 Martin Wistuba, Ambrish Rawat, Tejaswini Pedapati

The growing interest in both the automation of machine learning and deep learning has inevitably led to the development of a wide variety of automated methods for neural architecture search.

Data Augmentation Evolutionary Algorithms +1

Inductive Transfer for Neural Architecture Optimization

no code implementations8 Mar 2019 Martin Wistuba, Tejaswini Pedapati

First, we propose a novel neural architecture selection method which employs this knowledge to identify strong and weak characteristics of neural architectures across datasets.

Image Classification Neural Architecture Search

Understanding Unequal Gender Classification Accuracy from Face Images

no code implementations30 Nov 2018 Vidya Muthukumar, Tejaswini Pedapati, Nalini Ratha, Prasanna Sattigeri, Chai-Wah Wu, Brian Kingsbury, Abhishek Kumar, Samuel Thomas, Aleksandra Mojsilovic, Kush R. Varshney

Recent work shows unequal performance of commercial face classification services in the gender classification task across intersectional groups defined by skin type and gender.

Classification Gender Classification +1

Neurology-as-a-Service for the Developing World

no code implementations16 Nov 2017 Tejas Dharamsi, Payel Das, Tejaswini Pedapati, Gregory Bramble, Vinod Muthusamy, Horst Samulowitz, Kush R. Varshney, Yuvaraj Rajamanickam, John Thomas, Justin Dauwels

In this work, we present a cloud-based deep neural network approach to provide decision support for non-specialist physicians in EEG analysis and interpretation.

EEG Feature Engineering

Cannot find the paper you are looking for? You can Submit a new open access paper.