1 code implementation • 7 Mar 2024 • Amber Yijia Zheng, Tong He, Yixuan Qiu, Minjie Wang, David Wipf
These optimal features typically depend on tunable parameters of the lower-level energy in such a way that the entire bilevel pipeline can be trained end-to-end.
1 code implementation • 8 Apr 2023 • Yixuan Qiu, Xiao Wang
Sampling from high-dimensional distributions is a fundamental problem in statistical research and practice.
1 code implementation • 23 Feb 2023 • Yijia Zheng, Tong He, Yixuan Qiu, David Wipf
Although the variational autoencoder (VAE) and its conditional extension (CVAE) are capable of state-of-the-art results across multiple domains, their precise behavior is still not fully understood, particularly in the context of data (like images) that lie on or near a low-dimensional manifold.
no code implementations • 11 May 2022 • Shuai Liu, Yixuan Qiu, Baojuan Li, Huaning Wang, Xiangyu Chang
We consider the problem of identifying alterations of brain functional connectivity for a single MDD patient.
no code implementations • 29 Sep 2021 • Yijia Zheng, Yixuan Qiu
Deep generative modeling has long been viewed as a challenging unsupervised learning problem, partly due to the lack of labels and the high dimension of the data.
1 code implementation • ICLR 2020 • Yixuan Qiu, Lingsong Zhang, Xiao Wang
The contrastive divergence algorithm is a popular approach to training energy-based latent variable models, which has been widely used in many machine learning models such as the restricted Boltzmann machines and deep belief nets.
no code implementations • 25 Apr 2020 • Xiao Guo, Yixuan Qiu, Hai Zhang, Xiangyu Chang
Directed networks are broadly used to represent asymmetric relationships among units.
no code implementations • 13 Feb 2020 • Yixuan Qiu, Xiao Wang
We introduce a novel and efficient algorithm called the stochastic approximate gradient descent (SAGD), as an alternative to the stochastic gradient descent for cases where unbiased stochastic gradients cannot be trivially obtained.
2 code implementations • 19 Nov 2019 • Yixuan Qiu, Jing Lei, Kathryn Roeder
In this work we study sparse PCA based on the convex FPS formulation, and propose a new algorithm that is computationally efficient and applicable to large and high-dimensional data sets.