1 code implementation • 2 Nov 2023 • Yinghua Yao, Yuangang Pan, Jing Li, Ivor W. Tsang, Xin Yao
Therein, the interested clustering factor and the confounding factor are coarsely considered in the raw feature space, where the correlation between the data and the confounding factor is ideally assumed to be linear for convenient solutions.
1 code implementation • 28 Apr 2023 • Jing Li, Yuangang Pan, Yueming Lyu, Yinghua Yao, Yulei Sui, Ivor W. Tsang
Unlike existing model tuning methods where the target data is always ready for calculating model gradients, the model providers in EXPECTED only see some feedbacks which could be as simple as scalars, such as inference accuracy or usage rate.
no code implementations • 8 Apr 2023 • Dashan Gao, Yunce Zhao, Yinghua Yao, Zeqi Zhang, Bifei Mao, Xin Yao
In this paper, we study the robustness of deep learning models against joint perturbations by proposing a novel attack mechanism named Semantic-Preserving Adversarial (SPA) attack, which can then be used to enhance adversarial training.
no code implementations • 26 Nov 2021 • Yinghua Yao, Yuangang Pan, Ivor W. Tsang, Xin Yao
In particular, we simultaneously train two modules: a generator that translates an input image to the desired image with smooth subtle changes with respect to the interested attributes; and a ranker that ranks rival preferences consisting of the input image and the desired image.
no code implementations • 29 Sep 2021 • Jing Li, Yuangang Pan, Yueming Lyu, Yinghua Yao, Ivor Tsang
Instead of learning from scratch, fine-tuning a pre-trained model to fit a related target dataset of interest or downstream tasks has been a handy trick to achieve the desired performance.
1 code implementation • 14 Jul 2021 • Yinghua Yao, Yuangang Pan, Ivor W. Tsang, Xin Yao
This paper proposes Differential-Critic Generative Adversarial Network (DiCGAN) to learn the distribution of user-desired data when only partial instead of the entire dataset possesses the desired property.