Search Results for author: Yun Yue

Found 10 papers, 6 papers with code

Understanding Hyperbolic Metric Learning through Hard Negative Sampling

1 code implementation23 Apr 2024 Yun Yue, Fangzhou Lin, Guanyi Mou, Ziming Zhang

In recent years, there has been a growing trend of incorporating hyperbolic geometry methods into computer vision.

Metric Learning

AGD: an Auto-switchable Optimizer using Stepwise Gradient Difference for Preconditioning Matrix

1 code implementation NeurIPS 2023 Yun Yue, Zhiling Ye, Jiadi Jiang, Yongchao Liu, Ke Zhang

Additionally, we introduce an auto-switching function that enables the preconditioning matrix to switch dynamically between Stochastic Gradient Descent (SGD) and the adaptive optimizer.

Recommendation Systems

Sharpness-Aware Minimization Revisited: Weighted Sharpness as a Regularization Term

1 code implementation25 May 2023 Yun Yue, Jiadi Jiang, Zhiling Ye, Ning Gao, Yongchao Liu, Ke Zhang

Deep Neural Networks (DNNs) generalization is known to be closely related to the flatness of minima, leading to the development of Sharpness-Aware Minimization (SAM) for seeking flatter minima and better generalization.

Hyperbolic Contrastive Learning

no code implementations2 Feb 2023 Yun Yue, Fangzhou Lin, Kazunori D Yamada, Ziming Zhang

Learning good image representations that are beneficial to downstream tasks is a challenging task in computer vision.

Adversarial Robustness Contrastive Learning +2

Hyperbolic Chamfer Distance for Point Cloud Completion

1 code implementation ICCV 2023 Fangzhou Lin, Yun Yue, Songlin Hou, Xuechu Yu, Yajun Xu, Kazunori D Yamada, Ziming Zhang

Chamfer distance (CD) is a standard metric to measure the shape dissimilarity between point clouds in point cloud completion, as well as a loss function for (deep) learning.

Point Cloud Completion

SBO-RNN: Reformulating Recurrent Neural Networks via Stochastic Bilevel Optimization

no code implementations NeurIPS 2021 Ziming Zhang, Yun Yue, Guojun Wu, Yanhua Li, Haichong Zhang

In this paper we consider the training stability of recurrent neural networks (RNNs) and propose a family of RNNs, namely SBO-RNN, that can be formulated using stochastic bilevel optimization (SBO).

Bilevel Optimization

Learning Lightweight Neural Networks via Channel-Split Recurrent Convolution

no code implementations29 Sep 2021 Guojun Wu, Yun Yue, Yanhua Li, Ziming Zhang

Lightweight neural networks refer to deep networks with small numbers of parameters, which are allowed to be implemented in resource-limited hardware such as embedded systems.

Adaptive Optimizers with Sparse Group Lasso for Neural Networks in CTR Prediction

1 code implementation30 Jul 2021 Yun Yue, Yongchao Liu, Suo Tong, Minghao Li, Zhen Zhang, Chunyang Wen, Huanjun Bao, Lihong Gu, Jinjie Gu, Yixiang Mu

We develop a novel framework that adds the regularizers of the sparse group lasso to a family of adaptive optimizers in deep learning, such as Momentum, Adagrad, Adam, AMSGrad, AdaHessian, and create a new class of optimizers, which are named Group Momentum, Group Adagrad, Group Adam, Group AMSGrad and Group AdaHessian, etc., accordingly.

Click-Through Rate Prediction

Adaptive Optimizers with Sparse Group Lasso

no code implementations1 Jan 2021 Yun Yue, Suo Tong, Zhen Zhang, Yongchao Liu, Chunyang Wen, Huanjun Bao, Jinjie Gu, Yixiang Mu

We develop a novel framework that adds the regularizers to a family of adaptive optimizers in deep learning, such as MOMENTUM, ADAGRAD, ADAM, AMSGRAD, ADAHESSIAN, and create a new class of optimizers, which are named GROUP MOMENTUM, GROUP ADAGRAD, GROUP ADAM, GROUP AMSGRAD and GROUP ADAHESSIAN, etc., accordingly.

Cannot find the paper you are looking for? You can Submit a new open access paper.