Pretraining Deep Actor-Critic Reinforcement Learning Algorithms With Expert Demonstrations

31 Jan 2018 Xiaoqin Zhang Huimin Ma

Pretraining with expert demonstrations have been found useful in speeding up the training process of deep reinforcement learning algorithms since less online simulation data is required. Some people use supervised learning to speed up the process of feature learning, others pretrain the policies by imitating expert demonstrations... (read more)

PDF Abstract
No code implementations yet. Submit your code now

Tasks


Datasets


Results from the Paper


  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods used in the Paper


METHOD TYPE
Entropy Regularization
Regularization
Softmax
Output Functions
TRPO
Policy Gradient Methods
Retrace
Value Function Estimation
Stochastic Dueling Network
Value Function Estimation
ACER
Policy Gradient Methods
Experience Replay
Replay Memory
Dense Connections
Feedforward Networks
Weight Decay
Regularization
ReLU
Activation Functions
Adam
Stochastic Optimization
Convolution
Convolutions
Batch Normalization
Normalization
DDPG
Policy Gradient Methods