Search Results for author: Zhenmei Shi

Found 11 papers, 7 papers with code

Towards Few-Shot Adaptation of Foundation Models via Multitask Finetuning

1 code implementation22 Feb 2024 Zhuoyan Xu, Zhenmei Shi, Junyi Wei, Fangzhou Mu, Yin Li, YIngyu Liang

An emerging solution with recent success in vision and NLP involves finetuning a foundation model on a selection of relevant tasks, before its adaptation to a target task with limited labeled samples.

Fourier Circuits in Neural Networks: Unlocking the Potential of Large Language Models in Mathematical Reasoning and Modular Arithmetic

no code implementations12 Feb 2024 Jiuxiang Gu, Chenyang Li, YIngyu Liang, Zhenmei Shi, Zhao Song, Tianyi Zhou

Our research presents a thorough analytical characterization of the features learned by stylized one-hidden layer neural networks and one-layer Transformers in addressing this task.

2k Mathematical Reasoning

A Graph-Theoretic Framework for Understanding Open-World Semi-Supervised Learning

1 code implementation NeurIPS 2023 Yiyou Sun, Zhenmei Shi, Yixuan Li

Open-world semi-supervised learning aims at inferring both known and novel classes in unlabeled data, by harnessing prior knowledge from a labeled set with known classes.

Clustering Open-World Semi-Supervised Learning +1

When and How Does Known Class Help Discover Unknown Ones? Provable Understanding Through Spectral Analysis

1 code implementation9 Aug 2023 Yiyou Sun, Zhenmei Shi, YIngyu Liang, Yixuan Li

This paper bridges the gap by providing an analytical framework to formalize and investigate when and how known classes can help discover novel classes.

Novel Class Discovery

Domain Generalization via Nuclear Norm Regularization

1 code implementation13 Mar 2023 Zhenmei Shi, Yifei Ming, Ying Fan, Frederic Sala, YIngyu Liang

In this paper, we propose a simple and effective regularization method based on the nuclear norm of the learned features for domain generalization.

Domain Generalization

The Trade-off between Universality and Label Efficiency of Representations from Contrastive Learning

1 code implementation28 Feb 2023 Zhenmei Shi, Jiefeng Chen, Kunyang Li, Jayaram Raghuram, Xi Wu, YIngyu Liang, Somesh Jha

foundation models) has recently become a prevalent learning paradigm, where one first pre-trains a representation using large-scale unlabeled data, and then learns simple predictors on top of the representation using small labeled data from the downstream tasks.

Contrastive Learning

A Theoretical Analysis on Feature Learning in Neural Networks: Emergence from Inputs and Advantage over Fixed Features

no code implementations ICLR 2022 Zhenmei Shi, Junyi Wei, YIngyu Liang

These results provide theoretical evidence showing that feature learning in neural networks depends strongly on the input structure and leads to the superior performance.

Attentive Walk-Aggregating Graph Neural Networks

1 code implementation6 Oct 2021 Mehmet F. Demirel, Shengchao Liu, Siddhant Garg, Zhenmei Shi, YIngyu Liang

Our experiments demonstrate the strong performance of AWARE in graph-level prediction tasks in the standard setting in the domains of molecular property prediction and social networks.

Molecular Property Prediction Property Prediction

Deep Online Fused Video Stabilization

1 code implementation2 Feb 2021 Zhenmei Shi, Fuhao Shi, Wei-Sheng Lai, Chia-Kai Liang, YIngyu Liang

We present a deep neural network (DNN) that uses both sensor data (gyroscope) and image content (optical flow) to stabilize videos through unsupervised learning.

Video Stabilization

SF-Net: Structured Feature Network for Continuous Sign Language Recognition

no code implementations4 Aug 2019 Zhaoyang Yang, Zhenmei Shi, Xiaoyong Shen, Yu-Wing Tai

The proposed SF-Net extracts features in a structured manner and gradually encodes information at the frame level, the gloss level and the sentence level into the feature representation.

Sentence Sign Language Recognition

DAWN: Dual Augmented Memory Network for Unsupervised Video Object Tracking

no code implementations2 Aug 2019 Zhenmei Shi, Haoyang Fang, Yu-Wing Tai, Chi-Keung Tang

Our Dual Augmented Memory Network (DAWN) is unique in remembering both target and background, and using an improved attention LSTM memory to guide the focus on memorized features.

Video Object Tracking Visual Tracking

Cannot find the paper you are looking for? You can Submit a new open access paper.