no code implementations • 3 Jan 2024 • Wei Qian, Chenxu Zhao, Yangyi Li, Fenglong Ma, Chao Zhang, Mengdi Huai
To tackle the aforementioned challenges, in this paper, we design a novel uncertainty modeling framework for self-explaining networks, which not only demonstrates strong distribution-free uncertainty modeling performance for the generated explanations in the interpretation layer but also excels in producing efficient and effective prediction sets for the final predictions based on the informative high-level basis explanations.
1 code implementation • 20 Dec 2023 • Guangtao Zheng, Mengdi Huai, Aidong Zhang
Then, we propose Adversarial learning with Semantics Transformations (AdvST) that augments the source domain data with semantics transformations and learns a robust model with the augmented data.
no code implementations • 29 Nov 2023 • Lijie Hu, Yixin Liu, Ninghao Liu, Mengdi Huai, Lichao Sun, Di Wang
However, ViTs suffer from issues with explanation faithfulness, as their focal points are fragile to adversarial attacks and can be easily changed with even slight perturbations on the input image.
no code implementations • 16 Oct 2023 • Chenxu Zhao, Wei Qian, Yucheng Shi, Mengdi Huai, Ninghao Liu
Deep neural networks have exhibited remarkable performance across a wide range of real-world tasks.
no code implementations • 4 Oct 2023 • Yuan Zhong, Suhan Cui, Jiaqi Wang, Xiaochen Wang, Ziyi Yin, Yaqing Wang, Houping Xiao, Mengdi Huai, Ting Wang, Fenglong Ma
Health risk prediction is one of the fundamental tasks under predictive modeling in the medical domain, which aims to forecast the potential health risks that patients may face in the future using their historical Electronic Health Records (EHR).
1 code implementation • 6 Apr 2023 • Cheng-Long Wang, Mengdi Huai, Di Wang
To extend machine unlearning to graph data, \textit{GraphEraser} has been proposed.
no code implementations • 29 Nov 2022 • Sanchit Sinha, Mengdi Huai, Jianhui Sun, Aidong Zhang
Subsequently, we propose a potential general adversarial training-based defense mechanism to increase robustness of these systems to the proposed malicious attacks.
no code implementations • 23 Nov 2022 • Lijie Hu, Yixin Liu, Ninghao Liu, Mengdi Huai, Lichao Sun, Di Wang
Results show that SEAT is more stable against different perturbations and randomness while also keeps the explainability of attention, which indicates it is a more faithful explanation.
no code implementations • 29 Sep 2021 • Liuyi Yao, Yaliang Li, Bolin Ding, Jingren Zhou, Jinduo Liu, Mengdi Huai, Jing Gao
To tackle these challenges, we propose a novel casual graph based fair prediction framework, which integrates graph structure learning into fair prediction to ensure that unfair pathways are excluded in the causal graph.
1 code implementation • NeurIPS 2018 • Liuyi Yao, Sheng Li, Yaliang Li, Mengdi Huai, Jing Gao, Aidong Zhang
Estimating individual treatment effect (ITE) is a challenging problem in causal inference, due to the missing counterfactuals and the selection bias.