Search Results for author: Shitong Shao

Found 13 papers, 7 papers with code

Elucidating the Design Space of Dataset Condensation

no code implementations21 Apr 2024 Shitong Shao, Zikai Zhou, Huanran Chen, Zhiqiang Shen

Dataset condensation, a concept within data-centric learning, efficiently transfers critical attributes from an original dataset to a synthetic version, maintaining both diversity and realism.

Dataset Condensation

Self-supervised Dataset Distillation: A Good Compression Is All You Need

2 code implementations11 Apr 2024 Muxin Zhou, Zeyuan Yin, Shitong Shao, Zhiqiang Shen

In this work, we consider addressing this task through the new lens of model informativeness in the compression stage on the original dataset pretraining.

Informativeness

Precise Knowledge Transfer via Flow Matching

no code implementations3 Feb 2024 Shitong Shao, Zhiqiang Shen, Linrui Gong, Huanran Chen, Xu Dai

We name this framework Knowledge Transfer with Flow Matching (FM-KT), which can be integrated with a metric-based distillation method with any form (\textit{e. g.} vanilla KD, DKD, PKD and DIST) and a meta-encoder with any available architecture (\textit{e. g.} CNN, MLP and Transformer).

Transfer Learning

Rethinking Centered Kernel Alignment in Knowledge Distillation

1 code implementation22 Jan 2024 Zikai Zhou, Yunhang Shen, Shitong Shao, Linrui Gong, Shaohui Lin

This paper first provides a theoretical perspective to illustrate the effectiveness of CKA, which decouples CKA to the upper bound of Maximum Mean Discrepancy~(MMD) and a constant term.

Image Classification Knowledge Distillation +2

Generalized Large-Scale Data Condensation via Various Backbone and Statistical Matching

1 code implementation29 Nov 2023 Shitong Shao, Zeyuan Yin, Muxin Zhou, Xindong Zhang, Zhiqiang Shen

We call this perspective "generalized matching" and propose Generalized Various Backbone and Statistical Matching (G-VBSM) in this work, which aims to create a synthetic dataset with densities, ensuring consistency with the complete dataset across various backbones, layers, and statistics.

Dataset Condensation

Catch-Up Distillation: You Only Need to Train Once for Accelerating Sampling

1 code implementation18 May 2023 Shitong Shao, Xu Dai, Shouyi Yin, Lujun Li, Huanran Chen, Yang Hu

On CIFAR-10, we obtain a FID of 2. 80 by sampling in 15 steps under one-session training and the new state-of-the-art FID of 3. 37 by sampling in one step with additional training.

Knowledge Distillation

Black-box Source-free Domain Adaptation via Two-stage Knowledge Distillation

no code implementations13 May 2023 Shuai Wang, Daoan Zhang, Zipei Yan, Shitong Shao, Rui Li

In Stage \uppercase\expandafter{\romannumeral1}, we train the target model from scratch with soft pseudo-labels generated by the source model in a knowledge distillation manner.

Knowledge Distillation Source-Free Domain Adaptation +1

DiffuseExpand: Expanding dataset for 2D medical image segmentation using diffusion models

1 code implementation26 Apr 2023 Shitong Shao, Xiaohan Yuan, Zhen Huang, Ziming Qiu, Shuai Wang, Kevin Zhou

Based on this insight, we propose an approach called DiffuseExpand for expanding datasets for 2D medical image segmentation using DPM, which first samples a variety of masks from Gaussian noise to ensure the diversity, and then synthesizes images to ensure the alignment of images and masks.

Image Generation Image Segmentation +3

AIIR-MIX: Multi-Agent Reinforcement Learning Meets Attention Individual Intrinsic Reward Mixing Network

no code implementations19 Feb 2023 Wei Li, Weiyan Liu, Shitong Shao, Shiyi Huang

The results show that AIIR-MIX can dynamically assign each agent a real-time intrinsic reward in accordance with their actual contribution.

Multi-agent Reinforcement Learning Starcraft +1

Teaching What You Should Teach: A Data-Based Distillation Method

no code implementations11 Dec 2022 Shitong Shao, Huanran Chen, Zhen Huang, Linrui Gong, Shuai Wang, Xinxiao wu

To be specific, we design a neural network-based data augmentation module with priori bias, which assists in finding what meets the teacher's strengths but the student's weaknesses, by learning magnitudes and probabilities to generate suitable data samples.

Data Augmentation Knowledge Distillation +1

Bootstrap Generalization Ability from Loss Landscape Perspective

1 code implementation18 Sep 2022 Huanran Chen, Shitong Shao, Ziyi Wang, Zirui Shang, Jin Chen, Xiaofeng Ji, Xinxiao wu

Domain generalization aims to learn a model that can generalize well on the unseen test dataset, i. e., out-of-distribution data, which has different distribution from the training dataset.

Domain Generalization

Cannot find the paper you are looking for? You can Submit a new open access paper.