Search Results for author: ZhaoXin Huan

Found 7 papers, 0 papers with code

AntDT: A Self-Adaptive Distributed Training Framework for Leader and Straggler Nodes

no code implementations15 Apr 2024 Youshao Xiao, Lin Ju, Zhenglei Zhou, Siyuan Li, ZhaoXin Huan, Dalong Zhang, Rujie Jiang, Lin Wang, Xiaolu Zhang, Lei Liang, Jun Zhou

Previous works only address part of the stragglers and could not adaptively solve various stragglers in practice.

Breaking the Length Barrier: LLM-Enhanced CTR Prediction in Long Textual User Behaviors

no code implementations28 Mar 2024 Binzong Geng, ZhaoXin Huan, Xiaolu Zhang, Yong He, Liang Zhang, Fajie Yuan, Jun Zhou, Linjian Mo

However, we argue that a critical obstacle remains in deploying LLMs for practical use: the efficiency of LLMs when processing long textual user behaviors.

Click-Through Rate Prediction

G-Meta: Distributed Meta Learning in GPU Clusters for Large-Scale Recommender Systems

no code implementations9 Jan 2024 Youshao Xiao, Shangchun Zhao, Zhenglei Zhou, ZhaoXin Huan, Lin Ju, Xiaolu Zhang, Lin Wang, Jun Zhou

However, the existing systems are not tailored for meta learning based DLRM models and have critical problems regarding efficiency in distributed training in the GPU cluster.

Meta-Learning Recommendation Systems

One Model for All: Large Language Models are Domain-Agnostic Recommendation Systems

no code implementations22 Oct 2023 Zuoli Tang, ZhaoXin Huan, Zihao Li, Xiaolu Zhang, Jun Hu, Chilin Fu, Jun Zhou, Chenliang Li

We expect that by mixing the user's behaviors across different domains, we can exploit the common knowledge encoded in the pre-trained language model to alleviate the problems of data sparsity and cold start problems.

Language Modelling Question Answering +3

Rethinking Memory and Communication Cost for Efficient Large Language Model Training

no code implementations9 Oct 2023 Chan Wu, Hanxiao Zhang, Lin Ju, Jinjing Huang, Youshao Xiao, ZhaoXin Huan, Siyuan Li, Fanzhuang Meng, Lei Liang, Xiaolu Zhang, Jun Zhou

In this paper, we rethink the impact of memory consumption and communication costs on the training speed of large language models, and propose a memory-communication balanced strategy set Partial Redundancy Optimizer (PaRO).

Language Modelling Large Language Model

Data-Free Adversarial Perturbations for Practical Black-Box Attack

no code implementations3 Mar 2020 ZhaoXin Huan, Yulong Wang, Xiaolu Zhang, Lin Shang, Chilin Fu, Jun Zhou

Adversarial examples often exhibit black-box attacking transferability, which allows that adversarial examples crafted for one model can fool another model.

Cannot find the paper you are looking for? You can Submit a new open access paper.