no code implementations • 28 Jul 2023 • Xueming Liu, Kunda Liu, Tianjiang Hu, Qingrui Zhang
Based on the generation strategy of desired formation pattern and relative localization estimates, a cooperative formation tracking control scheme is proposed, which enables the formation geometric center to asymptotically converge to the moving target.
1 code implementation • 7 Dec 2022 • Zheng Zhang, Qingrui Zhang, Bo Zhu, Xiaohan Wang, Tianjiang Hu
In this paper, a novel algorithm named EASpace (Enhanced Action Space) is proposed, which formulates macro actions in an alternative form to accelerate the learning process using multiple available sub-optimal expert policies.
no code implementations • 28 Nov 2022 • Qingrui Zhang
This is a brief tutorial on the least square estimation technique that is straightforward yet effective for parameter estimation.
no code implementations • 9 Mar 2022 • Zheng Zhang, Xiaohan Wang, Qingrui Zhang, Tianjiang Hu
It is shown by numerical simulations that the proposed hybrid design outperforms the pursuit policies either learned from vanilla reinforcement learning or designed by the potential field method.
no code implementations • 20 Sep 2020 • Qingrui Zhang, Hao Dong, Wei Pan
More importantly, the existing multi-agent reinforcement learning (MARL) algorithms cannot ensure the closed-loop stability of a multi-agent system from a control-theoretic perspective, so the learned control polices are highly possible to generate abnormal or dangerous behaviors in real applications.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 17 Aug 2020 • Qingrui Zhang, Wei Pan, Vasso Reppa
This paper presents a novel model-reference reinforcement learning algorithm for the intelligent tracking control of uncertain autonomous surface vehicles with collision avoidance.
no code implementations • 30 Mar 2020 • Qingrui Zhang, Wei Pan, Vasso Reppa
With the conventional control, we can ensure the learning-based control law provides closed-loop stability for the overall system, and potentially increase the sample efficiency of the deep reinforcement learning.