Search Results for author: Zexi Li

Found 14 papers, 4 papers with code

Improving Group Connectivity for Generalization of Federated Deep Learning

no code implementations29 Feb 2024 Zexi Li, Jie Lin, Zhiqi Li, Didi Zhu, Chao Wu

Bridging the gap between LMC and FL, in this paper, we leverage fixed anchor models to empirically and theoretically study the transitivity property of connectivity from two models (LMC) to a group of models (model fusion in FL).

Federated Learning Linear Mode Connectivity

Model Tailor: Mitigating Catastrophic Forgetting in Multi-modal Large Language Models

no code implementations19 Feb 2024 Didi Zhu, Zhongyi Sun, Zexi Li, Tao Shen, Ke Yan, Shouhong Ding, Kun Kuang, Chao Wu

Catastrophic forgetting emerges as a critical challenge when fine-tuning multi-modal large language models (MLLMs), where improving performance on unseen tasks often leads to a significant performance drop on the original tasks.

Image Captioning Question Answering +1

Training-time Neuron Alignment through Permutation Subspace for Improving Linear Mode Connectivity and Model Fusion

no code implementations2 Feb 2024 Zexi Li, Zhiqi Li, Jie Lin, Tao Shen, Tao Lin, Chao Wu

In deep learning, stochastic gradient descent often yields functionally similar yet widely scattered solutions in the weight space even under the same initialization, causing barriers in the Linear Mode Connectivity (LMC) landscape.

Federated Learning Linear Mode Connectivity

Scalable Geometric Fracture Assembly via Co-creation Space among Assemblers

1 code implementation19 Dec 2023 Ruiyuan Zhang, Jiaxiang Liu, Zexi Li, Hao Dong, Jie Fu, Chao Wu

Therefore, there is a need to develop a scalable framework for geometric fracture assembly without relying on semantic information.

3D Assembly

FediOS: Decoupling Orthogonal Subspaces for Personalization in Feature-skew Federated Learning

no code implementations30 Nov 2023 Lingzhi Gao, Zexi Li, Yang Lu, Chao Wu

A typical way of pFL focuses on label distribution skew, and they adopt a decoupling scheme where the model is split into a common feature extractor and two prediction heads (generic and personalized).

Personalized Federated Learning

Understanding Prompt Tuning for V-L Models Through the Lens of Neural Collapse

no code implementations28 Jun 2023 Didi Zhu, Zexi Li, Min Zhang, Junkun Yuan, Yunfeng Shao, Jiashuo Liu, Kun Kuang, Yinchuan Li, Chao Wu

It is found that NC optimality of text-to-image representations shows a positive correlation with downstream generalizability, which is more severe under class imbalance settings.

Universal Domain Adaptation via Compressive Attention Matching

no code implementations ICCV 2023 Didi Zhu, Yincuan Li, Junkun Yuan, Zexi Li, Kun Kuang, Chao Wu

To address this issue, we propose a Universal Attention Matching (UniAM) framework by exploiting the self-attention mechanism in vision transformer to capture the crucial object information.

Universal Domain Adaptation

No Fear of Classifier Biases: Neural Collapse Inspired Federated Learning with Synthetic and Fixed Classifier

1 code implementation ICCV 2023 Zexi Li, Xinyi Shang, Rui He, Tao Lin, Chao Wu

Recent advances in neural collapse have shown that the classifiers and feature prototypes under perfect training scenarios collapse into an optimal structure called simplex equiangular tight frame (ETF).

Classifier calibration Federated Learning

Revisiting Weighted Aggregation in Federated Learning with Neural Networks

1 code implementation14 Feb 2023 Zexi Li, Tao Lin, Xinyi Shang, Chao Wu

In federated learning (FL), weighted aggregation of local models is conducted to generate a global model, and the aggregation weights are normalized (the sum of weights is 1) and proportional to the local data sizes.

Federated Learning

Towards Effective Clustered Federated Learning: A Peer-to-peer Framework with Adaptive Neighbor Matching

no code implementations23 Mar 2022 Zexi Li, Jiaxun Lu, Shuang Luo, Didi Zhu, Yunfeng Shao, Yinchuan Li, Zhimeng Zhang, Yongheng Wang, Chao Wu

In the literature, centralized clustered FL algorithms require the assumption of the number of clusters and hence are not effective enough to explore the latent relationships among clients.

Federated Learning

Ensemble Federated Adversarial Training with Non-IID data

no code implementations26 Oct 2021 Shuang Luo, Didi Zhu, Zexi Li, Chao Wu

Despite federated learning endows distributed clients with a cooperative training mode under the premise of protecting data privacy and security, the clients are still vulnerable when encountering adversarial samples due to the lack of robustness.

Federated Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.