Search Results for author: Khushbu Pahwa

Found 5 papers, 0 papers with code

GNNX-BENCH: Unravelling the Utility of Perturbation-based GNN Explainers through In-depth Benchmarking

no code implementations3 Oct 2023 Mert Kosan, Samidha Verma, Burouj Armgaan, Khushbu Pahwa, Ambuj Singh, Sourav Medya, Sayan Ranu

Motivated by this need, we present a benchmarking study on perturbation-based explainability methods for GNNs, aiming to systematically evaluate and compare a wide range of explainability techniques.

Benchmarking counterfactual

Scaling Distributed Multi-task Reinforcement Learning with Experience Sharing

no code implementations11 Jul 2023 Sanae Amani, Khushbu Pahwa, Vladimir Braverman, Lin F. Yang

Our research demonstrates that to achieve $\epsilon$-optimal policies for all $M$ tasks, a single agent using DistMT-LSVI needs to run a total number of episodes that is at most $\tilde{\mathcal{O}}({d^3H^6(\epsilon^{-2}+c_{\rm sep}^{-2})}\cdot M/N)$, where $c_{\rm sep}>0$ is a constant representing task separability, $H$ is the horizon of each episode, and $d$ is the feature dimension of the dynamics and rewards.

OpenAI Gym reinforcement-learning +1

FACTIFY3M: A Benchmark for Multimodal Fact Verification with Explainability through 5W Question-Answering

no code implementations22 May 2023 Megha Chakraborty, Khushbu Pahwa, Anku Rani, Shreyas Chatterjee, Dwip Dalal, Harshit Dave, Ritvik G, Preethi Gurumurthy, Adarsh Mahor, Samahriti Mukherjee, Aditya Pakala, Ishan Paul, Janvita Reddy, Arghya Sarkar, Kinjal Sensharma, Aman Chadha, Amit P. Sheth, Amitava Das

To address this gap, we introduce FACTIFY 3M, a dataset of 3 million samples that pushes the boundaries of the domain of fact verification via a multimodal fake news dataset, in addition to offering explainability through the concept of 5W question-answering.

Fact Verification Question Answering

Variance of Twitter Embeddings and Temporal Trends of COVID-19 cases

no code implementations30 Sep 2021 Mayank Sethi, Ambika Sadhu, Khushbu Pahwa, Sargun Nagpal, Tavpritesh Sethi

Using word embeddings to capture the semantic meaning of tweets, we identify Significant Dimensions (SDs). Our methodology predicts the rise in cases with a lead time of 15 days and 30 days with R2 scores of 0. 80 and 0. 62 respectively.

Word Embeddings

WiseR: An end-to-end structure learning and deployment framework for causal graphical models

no code implementations16 Aug 2021 Shubham Maheshwari, Khushbu Pahwa, Tavpritesh Sethi

Structure learning offers an expressive, versatile and explainable approach to causal and mechanistic modeling of complex biological data.

Cannot find the paper you are looking for? You can Submit a new open access paper.