no code implementations • 28 Nov 2023 • Hao Pei, Si Lin, Chuanfu Li, Che Wang, Haoming Chen, Sizhe Li
To enhance the intelligence degree in operation and maintenance, a novel method for fault detection in power grids is proposed.
no code implementations • 15 Nov 2023 • Xiang Li, Che Wang, Bing Li, Hao Chen, Sizhe Li
In this paper, we propose a method for knowledge graph construction in power distribution networks.
no code implementations • 1 Oct 2023 • Zecheng Wang, Che Wang, Zixuan Dong, Keith Ross
Recently, it has been shown that for offline deep reinforcement learning (DRL), pre-training Decision Transformer with a large language corpus can improve downstream performance (Reid et al., 2022).
no code implementations • 12 Oct 2022 • Tairan He, Yuge Zhang, Kan Ren, Minghuan Liu, Che Wang, Weinan Zhang, Yuqing Yang, Dongsheng Li
A good state representation is crucial to solving complicated reinforcement learning (RL) challenges.
no code implementations • 7 Sep 2022 • Zixuan Dong, Che Wang, Keith Ross
We nevertheless show that for a large class of MDPs, which includes stochastic MDPs such as blackjack and deterministic MDPs such as Go, the Q-function in MC-UCB converges almost surely to the optimal Q function.
1 code implementation • 17 Feb 2022 • Che Wang, Xufang Luo, Keith Ross, Dongsheng Li
We propose VRL3, a powerful data-driven framework with a simple design for solving challenging visual deep reinforcement learning (DRL) tasks.
no code implementations • 17 Nov 2021 • Yanqiu Wu, Xinyue Chen, Che Wang, Yiming Zhang, Keith W. Ross
In particular, Truncated Quantile Critics (TQC) achieves state-of-the-art asymptotic training performance on the MuJoCo benchmark with a distributional representation of critics; and Randomized Ensemble Double Q-Learning (REDQ) achieves high sample efficiency that is competitive with state-of-the-art model-based methods using a high update-to-data ratio and target randomization.
no code implementations • 29 Sep 2021 • Tairan He, Yuge Zhang, Kan Ren, Che Wang, Weinan Zhang, Dongsheng Li, Yuqing Yang
A good state representation is crucial to reinforcement learning (RL) while an ideal representation is hard to learn only with signals from the RL objective.
6 code implementations • ICLR 2021 • Xinyue Chen, Che Wang, Zijian Zhou, Keith Ross
Using a high Update-To-Data (UTD) ratio, model-based methods have recently achieved much higher sample efficiency than previous model-free methods for continuous-action DRL benchmarks.
no code implementations • ICLR 2022 • Che Wang, Shuhan Yuan, Kai Shao, Keith Ross
A simple and natural algorithm for reinforcement learning (RL) is Monte Carlo Exploring Starts (MCES), where the Q-function is estimated by averaging the Monte Carlo returns, and the policy is improved by choosing actions that maximize the current estimate of the Q-function.
1 code implementation • NeurIPS 2020 • Xinyue Chen, Zijian Zhou, Zheng Wang, Che Wang, Yanqiu Wu, Keith Ross
There has recently been a surge in research in batch Deep Reinforcement Learning (DRL), which aims for learning a high-performing policy from a given dataset without additional interactions with the environment.
3 code implementations • ICML 2020 • Che Wang, Yanqiu Wu, Quan Vuong, Keith Ross
We aim to develop off-policy DRL algorithms that not only exceed state-of-the-art performance but are also simple and minimalistic.
no code implementations • 25 Sep 2019 • Che Wang, Yanqiu Wu, Quan Vuong, Keith Ross
The field of Deep Reinforcement Learning (DRL) has recently seen a surge in the popularity of maximum entropy reinforcement learning algorithms.
3 code implementations • 10 Jun 2019 • Che Wang, Keith Ross
The ERE algorithm samples more aggressively from recent experience, and also orders the updates to ensure that updates from old data do not overwrite updates from new data.