no code implementations • 1 Apr 2024 • Qi Zhang, Yi Zhou, Ashley Prater-Bennette, Lixin Shen, Shaofeng Zou
We prove that our algorithm finds an $\epsilon$-stationary point with a computational complexity of $\mathcal O(\epsilon^{-3k_*-5})$, where $k_*$ is the parameter of the Cressie-Read divergence.
no code implementations • 4 Jan 2024 • Feng Yu, Lixin Shen, Guohui Song
Sparse Bayesian Learning (SBL) models are extensively used in signal processing and machine learning for promoting sparsity through hierarchical priors.
no code implementations • 29 Sep 2021 • Yuzhen Liu, Lixin Shen
The coefficients of this linear combination are served as the weights between the hidden layer and the output layer of the neural network while the mean square error between the exact solution and the approximation solution at the training set as the cost function.
no code implementations • 3 Mar 2021 • Ashley Prater-Bennette, Lixin Shen, Erin E. Tripp
The log-sum penalty is often adopted as a replacement for the $\ell_0$ pseudo-norm in compressive sensing and low-rank optimization.
Compressive Sensing Optimization and Control 49J53, 49J52, 90C26
no code implementations • 19 Feb 2015 • Ashley Prater, Lixin Shen, Bruce W. Suter
In this paper, we study a simple iterative method for finding the Dantzig selector, which was designed for linear regression problems.
no code implementations • 20 Jan 2015 • Ashley Prater, Lixin Shen
In this paper, we propose using the Dantzig selector model incorporating an overcomplete dictionary to separate a noisy undersampled collection of composite signals, and present an algorithm to efficiently solve the model.