Search Results for author: Linyun He

Found 1 papers, 1 papers with code

Nonparametric Learning of Two-Layer ReLU Residual Units

1 code implementation17 Aug 2020 Zhunxuan Wang, Linyun He, Chunchuan Lyu, Shay B. Cohen

We describe an algorithm that learns two-layer residual units using rectified linear unit (ReLU) activation: suppose the input $\mathbf{x}$ is from a distribution with support space $\mathbb{R}^d$ and the ground-truth generative model is a residual unit of this type, given by $\mathbf{y} = \boldsymbol{B}^\ast\left[\left(\boldsymbol{A}^\ast\mathbf{x}\right)^+ + \mathbf{x}\right]$, where ground-truth network parameters $\boldsymbol{A}^\ast \in \mathbb{R}^{d\times d}$ represent a full-rank matrix with nonnegative entries and $\boldsymbol{B}^\ast \in \mathbb{R}^{m\times d}$ is full-rank with $m \geq d$ and for $\boldsymbol{c} \in \mathbb{R}^d$, $[\boldsymbol{c}^{+}]_i = \max\{0, c_i\}$.

Vocal Bursts Valence Prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.