Search Results for author: Junze Yin

Found 20 papers, 0 papers with code

How to Inverting the Leverage Score Distribution?

no code implementations21 Apr 2024 Zhihang Li, Zhao Song, Weixin Wang, Junze Yin, Zheng Yu

Leverage score is a fundamental problem in machine learning and theoretical computer science.

Local Convergence of Approximate Newton Method for Two Layer Nonlinear Regression

no code implementations26 Nov 2023 Zhihang Li, Zhao Song, Zifan Wang, Junze Yin

Our main results involve analyzing the convergence properties of an approximate Newton method used to minimize the regularized training loss.

Question Answering regression +2

Revisiting Quantum Algorithms for Linear Regressions: Quadratic Speedups without Data-Dependent Parameters

no code implementations24 Nov 2023 Zhao Song, Junze Yin, Ruizhe Zhang

However, the running times of these algorithms depend on some quantum linear algebra-related parameters, such as $\kappa(A)$, the condition number of $A$.

regression

The Expressibility of Polynomial based Attention Scheme

no code implementations30 Oct 2023 Zhao Song, Guangyi Xu, Junze Yin

In this paper, we offer a theoretical analysis of the expressive capabilities of polynomial attention.

Decision Making

A Unified Scheme of ResNet and Softmax

no code implementations23 Sep 2023 Zhao Song, Weixin Wang, Junze Yin

The Hessian is shown to be positive semidefinite, and its structure is characterized as the sum of a low-rank matrix and a diagonal matrix.

Image Classification object-detection +3

A Fast Optimization View: Reformulating Single Layer Attention in LLM Based on Tensor and SVM Trick, and Solving It in Matrix Multiplication Time

no code implementations14 Sep 2023 Yeqi Gao, Zhao Song, Weixin Wang, Junze Yin

$A_3$ is a matrix in $\mathbb{R}^{n \times d}$, $\mathsf{A}_{j_0} \in \mathbb{R}^{n \times d^2}$ is the $j_0$-th block of $\mathsf{A}$.

Solving Attention Kernel Regression Problem via Pre-conditioner

no code implementations28 Aug 2023 Zhao Song, Junze Yin, Lichen Zhang

Given an input matrix $A\in \mathbb{R}^{n\times d}$ with $n\gg d$ and a response vector $b$, we first consider the matrix exponential of the matrix $A^\top A$ as a proxy, and we in turn design algorithms for two types of regression problems: $\min_{x\in \mathbb{R}^d}\|(A^\top A)^jx-b\|_2$ and $\min_{x\in \mathbb{R}^d}\|A(A^\top A)^jx-b\|_2$ for any positive integer $j$.

regression

GradientCoin: A Peer-to-Peer Decentralized Large Language Models

no code implementations21 Aug 2023 Yeqi Gao, Zhao Song, Junze Yin

It is likely that only two types of people would be interested in setting up a practical system for it: $\bullet$ Those who prefer to use a decentralized ChatGPT-like software.

Efficient Alternating Minimization with Applications to Weighted Low Rank Approximation

no code implementations7 Jun 2023 Zhao Song, Mingquan Ye, Junze Yin, Lichen Zhang

For weighted low rank approximation, this improves the runtime of [LLR16] from $n^2 k^2$ to $n^2k$.

2k

Faster Robust Tensor Power Method for Arbitrary Order

no code implementations1 Jun 2023 Yichuan Deng, Zhao Song, Junze Yin

Tensor decomposition is a fundamental method used in various areas to deal with high-dimensional data.

Tensor Decomposition

Federated Empirical Risk Minimization via Second-Order Method

no code implementations27 May 2023 Song Bian, Zhao Song, Junze Yin

Many convex optimization problems with important applications in machine learning are formulated as empirical risk minimization (ERM).

Federated Learning regression +1

Fast and Efficient Matching Algorithm with Deadline Instances

no code implementations15 May 2023 Zhao Song, Weixin Wang, Chenbo Yin, Junze Yin

But in \textsc{FastPostponedGreedy} algorithm, the status of each node is unknown at first.

An Iterative Algorithm for Rescaled Hyperbolic Functions Regression

no code implementations1 May 2023 Yeqi Gao, Zhao Song, Junze Yin

LLMs have shown great promise in improving the accuracy and efficiency of these tasks, and have the potential to revolutionize the field of natural language processing (NLP) in the years to come.

In-Context Learning Language Modelling +4

Low Rank Matrix Completion via Robust Alternating Minimization in Nearly Linear Time

no code implementations21 Feb 2023 Yuzhou Gu, Zhao Song, Junze Yin, Lichen Zhang

Moreover, our algorithm runs in time $\widetilde O(|\Omega| k)$, which is nearly linear in the time to verify the solution while preserving the sample complexity.

Low-Rank Matrix Completion regression

A Nearly-Optimal Bound for Fast Regression with $\ell_\infty$ Guarantee

no code implementations1 Feb 2023 Zhao Song, Mingquan Ye, Junze Yin, Lichen Zhang

One popular approach for solving such $\ell_2$ regression problem is via sketching: picking a structured random matrix $S\in \mathbb{R}^{m\times n}$ with $m\ll n$ and $SA$ can be quickly computed, solve the ``sketched'' regression problem $\arg\min_{x\in \mathbb{R}^d} \|SAx-Sb\|_2$.

regression

A Faster $k$-means++ Algorithm

no code implementations28 Nov 2022 Jiehao Liang, Somdeb Sarkhel, Zhao Song, Chenbo Yin, Junze Yin, Danyang Zhuo

We propose a new algorithm \textsc{FastKmeans++} that only takes in $\widetilde{O}(nd + nk^2)$ time, in total.

Clustering

Dynamic Maintenance of Kernel Density Estimation Data Structure: From Practice to Theory

no code implementations8 Aug 2022 Jiehao Liang, Zhao Song, Zhaozhuo Xu, Junze Yin, Danyang Zhuo

In this work, we focus on the dynamic maintenance of KDE data structures with robustness to adversarial queries.

Density Estimation

InstaHide's Sample Complexity When Mixing Two Private Images

no code implementations24 Nov 2020 Baihe Huang, Zhao Song, Runzhou Tao, Junze Yin, Ruizhe Zhang, Danyang Zhuo

On the current InstaHide challenge setup, where each InstaHide image is a mixture of two private images, we present a new algorithm to recover all the private images with a provable guarantee and optimal sample complexity.

Vocal Bursts Valence Prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.