Search Results for author: Xiucheng Wang

Found 13 papers, 3 papers with code

Imperfect Digital Twin Assisted Low Cost Reinforcement Training for Multi-UAV Networks

no code implementations25 Oct 2023 Xiucheng Wang, Nan Cheng, Longfei Ma, Zhisheng Yin, Tom. Luan, Ning Lu

Two cascade neural networks (NN) are used to optimize the joint number of virtually generated UAVs, the DT construction cost, and the performance of multi-UAV networks.

reinforcement-learning

Label-free Deep Learning Driven Secure Access Selection in Space-Air-Ground Integrated Networks

no code implementations28 Aug 2023 Zhaowei Wang, Zhisheng Yin, Xiucheng Wang, Nan Cheng, Yuan Zhang, Tom H. Luan

Considering the inherent co-channel interference due to spectrum sharing among multi-tier access networks in SAGIN, it can be leveraged to assist the physical layer security among heterogeneous transmissions.

Effectively Heterogeneous Federated Learning: A Pairing and Split Learning Based Approach

no code implementations26 Aug 2023 Jinglong Shen, Xiucheng Wang, Nan Cheng, Longfei Ma, Conghao Zhou, Yuan Zhang

As a promising paradigm federated Learning (FL) is widely used in privacy-preserving machine learning, which allows distributed devices to collaboratively train a model while avoiding data transmission among clients.

Federated Learning Privacy Preserving

Mobility-Aware Computation Offloading for Swarm Robotics using Deep Reinforcement Learning

no code implementations22 Aug 2023 Xiucheng Wang, Hongzhi Guo

Therefore, current swarm robotics have a small number of robots, which can only provide limited spatio-temporal information.

Edge-computing reinforcement-learning +1

Distilling Knowledge from Resource Management Algorithms to Neural Networks: A Unified Training Assistance Approach

no code implementations15 Aug 2023 Longfei Ma, Nan Cheng, Xiucheng Wang, Zhisheng Yin, Haibo Zhou, Wei Quan

To fully leverage the high performance of traditional model-based methods and the low complexity of the NN-based method, a knowledge distillation (KD) based algorithm distillation (AD) method is proposed in this paper to improve the performance and convergence speed of the NN-based method, where traditional SINR optimization methods are employed as ``teachers" to assist the training of NNs, which are ``students", thus enhancing the performance of unsupervised and reinforcement learning techniques.

Knowledge Distillation Management +1

Knowledge-Driven Multi-Agent Reinforcement Learning for Computation Offloading in Cybertwin-Enabled Internet of Vehicles

no code implementations4 Aug 2023 Ruijin Sun, Xiao Yang, Nan Cheng, Xiucheng Wang, Changle Li

By offloading computation-intensive tasks of vehicles to roadside units (RSUs), mobile edge computing (MEC) in the Internet of Vehicles (IoV) can relieve the onboard computation burden.

Edge-computing Multi-agent Reinforcement Learning

Interpretable and Secure Trajectory Optimization for UAV-Assisted Communication

no code implementations5 Jul 2023 Yunhao Quan, Nan Cheng, Xiucheng Wang, Jinglong Shen, Longfei Ma, Zhisheng Yin

Unmanned aerial vehicles (UAVs) have gained popularity due to their flexible mobility, on-demand deployment, and the ability to establish high probability line-of-sight wireless communication.

Collision Avoidance Explainable Artificial Intelligence (XAI)

Scalable Resource Management for Dynamic MEC: An Unsupervised Link-Output Graph Neural Network Approach

1 code implementation15 Jun 2023 Xiucheng Wang, Nan Cheng, Lianhao Fu, Wei Quan, Ruijin Sun, Yilong Hui, Tom Luan, Xuemin Shen

However, the dynamics of edge networks raise two challenges in neural network (NN)-based optimization methods: low scalability and high training costs.

Edge-computing Management

Digital Twin-Assisted Knowledge Distillation Framework for Heterogeneous Federated Learning

no code implementations10 Mar 2023 Xiucheng Wang, Nan Cheng, Longfei Ma, Ruijin Sun, Rong Chai, Ning Lu

In this paper, to deal with the heterogeneity in federated learning (FL) systems, a knowledge distillation (KD) driven training framework for FL is proposed, where each user can select its neural network model on demand and distill knowledge from a big teacher model using its own private dataset.

Federated Learning Knowledge Distillation +2

SigT: An Efficient End-to-End MIMO-OFDM Receiver Framework Based on Transformer

1 code implementation2 Nov 2022 Ziyou Ren, Nan Cheng, Ruijin Sun, Xiucheng Wang, Ning Lu, Wenchao Xu

Multiple-input multiple-output and orthogonal frequency-division multiplexing (MIMO-OFDM) are the key technologies in 4G and subsequent wireless communication systems.

On-Demand Resource Management for 6G Wireless Networks Using Knowledge-Assisted Dynamic Neural Networks

no code implementations2 Aug 2022 Longfei Ma, Nan Cheng, Xiucheng Wang, Ruijin Sun, Ning Lu

On-demand service provisioning is a critical yet challenging issue in 6G wireless communication networks, since emerging services have significantly diverse requirements and the network resources become increasingly heterogeneous and dynamic.

Decision Making Management

Digital Twin-Assisted Efficient Reinforcement Learning for Edge Task Scheduling

no code implementations2 Aug 2022 Xiucheng Wang, Longfei Ma, Haocheng Li, Zhisheng Yin, Tom. Luan, Nan Cheng

We use DT to simulate the results of different decisions made by the agent, so that one agent can try multiple actions at a time, or, similarly, multiple agents can interact with environment in parallel in DT.

Q-Learning reinforcement-learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.