no code implementations • 20 Mar 2024 • Ruizhe Zhang, Qingyao Ai, Ziyi Ye, Yueyue Wu, Xiaohui Xie, Yiqun Liu
Traditional feedback signal such as clicks is too coarse to use as they do not reflect any fine-grained relevance information.
no code implementations • 17 Mar 2024 • Ruizhe Zhang, Haitao Li, Yueyue Wu, Qingyao Ai, Yiqun Liu, Min Zhang, Shaoping Ma
In recent years, the utilization of large language models for natural language dialogue has gained momentum, leading to their widespread adoption across various domains.
no code implementations • 25 Feb 2024 • Ruizhe Zhang, Qingyao Ai, Yiqun Liu, Yueyue Wu, Beining Wang
Gender of the defendants in both the task and relevant cases was edited to statistically measure the effect of gender bias in the legal case search results on participants' perceptions.
no code implementations • 10 Feb 2024 • Yeqi Gao, Zhao Song, Ruizhe Zhang
Given its widespread application in machine learning and optimization, the Kronecker product emerges as a pivotal linear algebra operator.
1 code implementation • 18 Jan 2024 • Ruizhe Zhang, Xinke Jiang, Yuchen Fang, Jiayuan Luo, Yongxin Xu, Yichen Zhu, Xu Chu, Junfeng Zhao, Yasha Wang
Graph Neural Networks (GNNs) have shown considerable effectiveness in a variety of graph learning tasks, particularly those based on the message-passing approach in recent years.
no code implementations • 26 Dec 2023 • Xinke Jiang, Ruizhe Zhang, Yongxin Xu, Rihong Qiu, Yue Fang, Zhiyuan Wang, Jinyi Tang, Hongxin Ding, Xu Chu, Junfeng Zhao, Yasha Wang
In this paper, we investigate the retrieval-augmented generation (RAG) based on Knowledge Graphs (KGs) to improve the accuracy and reliability of Large Language Models (LLMs).
no code implementations • 24 Nov 2023 • Zhao Song, Junze Yin, Ruizhe Zhang
However, the running times of these algorithms depend on some quantum linear algebra-related parameters, such as $\kappa(A)$, the condition number of $A$.
no code implementations • 7 Oct 2023 • Beining Wang, Ruizhe Zhang, Yueyue Wu, Qingyao Ai, Min Zhang, Yiqun Liu
Given a specific query case, legal case retrieval systems aim to retrieve a set of case documents relevant to the case at hand.
no code implementations • 16 Jul 2023 • Yeqi Gao, Zhao Song, Xin Yang, Ruizhe Zhang
It is well-known that quantum machine has certain computational advantages compared to the classical machine.
no code implementations • 22 Mar 2023 • Lianke Qin, Zhao Song, Ruizhe Zhang
In this paper, we relax that rank-$k$ assumption and solve a much more general matrix sensing problem.
no code implementations • 29 Jan 2023 • Ruizhe Zhang, Qingyao Ai, Yueyue Wu, Yixiao Ma, Yiqun Liu
In the process of searching, legal practitioners often need the search results under several different causes of cases as reference.
no code implementations • 1 Jan 2023 • Hongru Yang, Ziyu Jiang, Ruizhe Zhang, Zhangyang Wang, Yingbin Liang
This work studies training one-hidden-layer overparameterized ReLU networks via gradient descent in the neural tangent kernel (NTK) regime, where the networks' biases are initialized to some constant rather than zero.
no code implementations • 12 Oct 2022 • Andrew M. Childs, Tongyang Li, Jin-Peng Liu, Chunhao Wang, Ruizhe Zhang
We also prove a $1/\epsilon^{1-o(1)}$ quantum lower bound for estimating normalizing constants, implying near-optimality of our quantum algorithms in $\epsilon$.
no code implementations • 26 Sep 2022 • Tongyang Li, Ruizhe Zhang
As an application, we give a quantum algorithm for zeroth-order stochastic convex bandits with $\tilde{O}(n^{5}\log^{2} T)$ regret, an exponential speedup in $T$ compared to the classical $\Omega(\sqrt{T})$ lower bound.
no code implementations • 14 Dec 2021 • Zhao Song, Lichen Zhang, Ruizhe Zhang
We consider the problem of training a multi-layer over-parametrized neural network to minimize the empirical risk induced by a loss function.
no code implementations • NeurIPS 2021 • Zhao Song, Shuo Yang, Ruizhe Zhang
The classical training method requires paying $\Omega(mnd)$ cost for both forward computation and backward computation, where $m$ is the width of the neural network, and we are given $n$ training points in $d$-dimensional space.
no code implementations • 29 Sep 2021 • Baihe Huang, Zhao Song, Runzhou Tao, Ruizhe Zhang, Danyang Zhuo
Inspired by InstaHide challenge [Huang, Song, Li and Arora'20], [Chen, Song and Zhuo'20] recently provides one mathematical formulation of InstaHide attack problem under Gaussian images distribution.
no code implementations • 2 Feb 2021 • Sitan Chen, Zhao Song, Runzhou Tao, Ruizhe Zhang
As this problem is hard in the worst-case, we study a natural average-case variant that arises in the context of these reconstruction attacks: $\mathbf{M} = \mathbf{W}\mathbf{W}^{\top}$ for $\mathbf{W}$ a random Boolean matrix with $k$-sparse rows, and the goal is to recover $\mathbf{W}$ up to column permutation.
no code implementations • 24 Nov 2020 • Baihe Huang, Zhao Song, Runzhou Tao, Junze Yin, Ruizhe Zhang, Danyang Zhuo
On the current InstaHide challenge setup, where each InstaHide image is a mixture of two private images, we present a new algorithm to recover all the private images with a provable guarantee and optimal sample complexity.