Search Results for author: Zhenghao Peng

Found 17 papers, 9 papers with code

SwinTextSpotter v2: Towards Better Synergy for Scene Text Spotting

no code implementations15 Jan 2024 Mingxin Huang, Dezhi Peng, Hongliang Li, Zhenghao Peng, Chongyu Liu, Dahua Lin, Yuliang Liu, Xiang Bai, Lianwen Jin

In this paper, we propose a new end-to-end scene text spotting framework termed SwinTextSpotter v2, which seeks to find a better synergy between text detection and recognition.

Text Detection Text Spotting

Guarded Policy Optimization with Imperfect Online Demonstrations

no code implementations3 Mar 2023 Zhenghai Xue, Zhenghao Peng, Quanyi Li, Zhihan Liu, Bolei Zhou

Assuming optimal, the teacher policy has the perfect timing and capability to intervene in the learning process of the student agent, providing safety guarantee and exploration guidance.

Continuous Control Efficient Exploration +2

Human-AI Shared Control via Policy Dissection

1 code implementation31 May 2022 Quanyi Li, Zhenghao Peng, Haibin Wu, Lan Feng, Bolei Zhou

Inspired by the neuroscience approach to investigate the motor cortex in primates, we develop a simple yet effective frequency-based approach called \textit{Policy Dissection} to align the intermediate representation of the learned neural controller with the kinematic attributes of the agent behavior.

Autonomous Driving Reinforcement Learning (RL)

Learning to Drive by Watching YouTube Videos: Action-Conditioned Contrastive Policy Pretraining

1 code implementation5 Apr 2022 Qihang Zhang, Zhenghao Peng, Bolei Zhou

Specifically, we train an inverse dynamic model with a small amount of labeled data and use it to predict action labels for all the YouTube video frames.

Autonomous Driving Imitation Learning

Efficient Learning of Safe Driving Policy via Human-AI Copilot Optimization

no code implementations ICLR 2022 Quanyi Li, Zhenghao Peng, Bolei Zhou

HACO can train agents to drive in unseen traffic scenarios with a handful of human intervention budget and achieve high safety and generalizability, outperforming both reinforcement learning and imitation learning baselines with a large margin.

Imitation Learning reinforcement-learning +1

Safe Driving via Expert Guided Policy Optimization

1 code implementation13 Oct 2021 Zhenghao Peng, Quanyi Li, Chunxiao Liu, Bolei Zhou

Offline RL technique is further used to learn from the partial demonstration generated by the expert.

Offline RL reinforcement-learning +1

MetaDrive: Composing Diverse Driving Scenarios for Generalizable Reinforcement Learning

2 code implementations26 Sep 2021 Quanyi Li, Zhenghao Peng, Lan Feng, Qihang Zhang, Zhenghai Xue, Bolei Zhou

Based on MetaDrive, we construct a variety of RL tasks and baselines in both single-agent and multi-agent settings, including benchmarking generalizability across unseen scenes, safe exploration, and learning multi-agent traffic.

Benchmarking Decision Making +5

Improving the Generalization of End-to-End Driving through Procedural Generation

2 code implementations26 Dec 2020 Quanyi Li, Zhenghao Peng, Qihang Zhang, Chunxiao Liu, Bolei Zhou

We validate that training with the increasing number of procedurally generated scenes significantly improves the generalization of the agent across scenarios of different traffic densities and road networks.

Autonomous Driving

Non-local Policy Optimization via Diversity-regularized Collaborative Exploration

no code implementations14 Jun 2020 Zhenghao Peng, Hao Sun, Bolei Zhou

Conventional Reinforcement Learning (RL) algorithms usually have one single agent learning to solve the task independently.

Reinforcement Learning (RL)

Learning with Social Influence through Interior Policy Differentiation

no code implementations25 Sep 2019 Hao Sun, Bo Dai, Jiankai Sun, Zhenghao Peng, Guodong Xu, Dahua Lin, Bolei Zhou

In this work we model the social influence into the scheme of reinforcement learning, enabling the agents to learn both from the environment and from their peers.

Reinforcement Learning (RL)

AXNet: ApproXimate computing using an end-to-end trainable neural network

2 code implementations27 Jul 2018 Zhenghao Peng, Xuyang Chen, Chengwen Xu, Naifeng Jing, Xiaoyao Liang, Cewu Lu, Li Jiang

To guarantee the approximation quality, existing works deploy two neural networks (NNs), e. g., an approximator and a predictor.

Multi-Task Learning Philosophy

Approximate Random Dropout

no code implementations23 May 2018 Zhuoran Song, Ru Wang, Dongyu Ru, Hongru Huang, Zhenghao Peng, Jing Ke, Xiaoyao Liang, Li Jiang

In this paper, we propose the Approximate Random Dropout that replaces the conventional random dropout of neurons and synapses with a regular and predefined patterns to eliminate the unnecessary computation and data access.

Cannot find the paper you are looking for? You can Submit a new open access paper.