Search Results for author: Weichao Mao

Found 12 papers, 3 papers with code

Decision Transformer as a Foundation Model for Partially Observable Continuous Control

no code implementations3 Apr 2024 Xiangyuan Zhang, Weichao Mao, Haoran Qiu, Tamer Başar

Closed-loop control of nonlinear dynamical systems with partial-state observability demands expert knowledge of a diverse, less standardized set of theoretical tools.

Continuous Control Zero-shot Generalization

Controlgym: Large-Scale Control Environments for Benchmarking Reinforcement Learning Algorithms

1 code implementation30 Nov 2023 Xiangyuan Zhang, Weichao Mao, Saviz Mowlavi, Mouhacine Benosman, Tamer Başar

This project serves the learning for dynamics & control (L4DC) community, aiming to explore key questions: the convergence of RL algorithms in learning control policies; the stability and robustness issues of learning-based controllers; and the scalability of RL algorithms to high- and potentially infinite-dimensional systems.

Benchmarking OpenAI Gym +2

Action Dynamics Task Graphs for Learning Plannable Representations of Procedural Tasks

no code implementations11 Jan 2023 Weichao Mao, Ruta Desai, Michael Louis Iuzzolino, Nitin Kamra

Given video demonstrations and paired narrations of an at-home procedural task such as changing a tire, we present an approach to extract the underlying task structure -- relevant actions and their temporal dependencies -- via action-centric task graphs.

Task 2

On Improving Model-Free Algorithms for Decentralized Multi-Agent Reinforcement Learning

no code implementations12 Oct 2021 Weichao Mao, Lin F. Yang, Kaiqing Zhang, Tamer Başar

Multi-agent reinforcement learning (MARL) algorithms often suffer from an exponential sample complexity dependence on the number of agents, a phenomenon known as \emph{the curse of multiagents}.

Multi-agent Reinforcement Learning Q-Learning +3

Provably Efficient Reinforcement Learning in Decentralized General-Sum Markov Games

no code implementations12 Oct 2021 Weichao Mao, Tamer Başar

We show that the agents can find an $\epsilon$-approximate CCE in at most $\widetilde{O}( H^6S A /\epsilon^2)$ episodes, where $S$ is the number of states, $A$ is the size of the largest individual action space, and $H$ is the length of an episode.

Multi-agent Reinforcement Learning Q-Learning +2

Decentralized Cooperative Multi-Agent Reinforcement Learning with Exploration

no code implementations29 Sep 2021 Weichao Mao, Tamer Basar, Lin Yang, Kaiqing Zhang

Many real-world applications of multi-agent reinforcement learning (RL), such as multi-robot navigation and decentralized control of cyber-physical systems, involve the cooperation of agents as a team with aligned objectives.

Multi-agent Reinforcement Learning Q-Learning +3

POLY-HOOT: Monte-Carlo Planning in Continuous Space MDPs with Non-Asymptotic Analysis

no code implementations NeurIPS 2020 Weichao Mao, Kaiqing Zhang, Qiaomin Xie, Tamer Başar

Monte-Carlo planning, as exemplified by Monte-Carlo Tree Search (MCTS), has demonstrated remarkable performance in applications with finite spaces.

Information State Embedding in Partially Observable Cooperative Multi-Agent Reinforcement Learning

1 code implementation2 Apr 2020 Weichao Mao, Kaiqing Zhang, Erik Miehling, Tamer Başar

To enable the development of tractable algorithms, we introduce the concept of an information state embedding that serves to compress agents' histories.

Multi-agent Reinforcement Learning reinforcement-learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.