no code implementations • 26 Jan 2024 • Jiachen Xi, Alfredo Garcia, Petar Momcilovic
Several successful reinforcement learning algorithms make use of regularization to promote multi-modal policies that exhibit enhanced exploration and robustness.
no code implementations • 10 Nov 2023 • Johan Engström, Ran Wei, Anthony McDonald, Alfredo Garcia, Matt O'Kelly, Leif Johnson
Understanding adaptive human driving behavior, in particular how drivers manage uncertainty, is of key importance for developing simulated human driver models that can be used in the evaluation and development of autonomous vehicles.
1 code implementation • 10 Oct 2023 • Ran Wei, Nathan Lambert, Anthony McDonald, Alfredo Garcia, Roberto Calandra
Model-based Reinforcement Learning (MBRL) aims to make agents more sample-efficient, adaptive, and explainable by learning an explicit model of the environment.
2 code implementations • 15 Sep 2023 • Ran Wei, Siliang Zeng, Chenliang Li, Alfredo Garcia, Anthony McDonald, Mingyi Hong
We consider a Bayesian approach to offline model-based inverse reinforcement learning (IRL).
no code implementations • 27 Mar 2023 • Ran Wei, Anthony D. McDonald, Alfredo Garcia, Gustav Markkula, Johan Engstrom, Matthew O'Kelly
We assessed the proposed model, the Active Inference Driving Agent (AIDA), through a benchmark analysis against the rule-based Intelligent Driver Model, and two neural network Behavior Cloning models.
1 code implementation • NeurIPS 2023 • Siliang Zeng, Chenliang Li, Alfredo Garcia, Mingyi Hong
Offline inverse reinforcement learning (Offline IRL) aims to recover the structure of rewards and environment dynamics that underlie observed actions in a fixed, finite set of demonstrations from an expert agent.
no code implementations • 4 Oct 2022 • Siliang Zeng, Mingyi Hong, Alfredo Garcia
Other approaches in the inverse reinforcement learning (IRL) literature emphasize policy estimation at the expense of reduced reward estimation accuracy.
no code implementations • 4 Oct 2022 • Siliang Zeng, Chenliang Li, Alfredo Garcia, Mingyi Hong
To reduce the computational burden of a nested loop, novel methods such as SQIL [1] and IQ-Learn [2] emphasize policy estimation at the expense of reward estimation accuracy.
no code implementations • 11 Oct 2021 • Siliang Zeng, Tianyi Chen, Alfredo Garcia, Mingyi Hong
The flexibility in our design allows the proposed MARL-CAC algorithm to be used in a {\it fully decentralized} setting, where the agents can only communicate with their neighbors, as well as a {\it federated} setting, where the agents occasionally communicate with a server while optimizing their (partially personalized) local models.
1 code implementation • 14 Feb 2021 • Shixiang Chen, Alfredo Garcia, Mingyi Hong, Shahin Shahrampour
The global function is represented as a finite sum of smooth local functions, where each local function is associated with one agent and agents communicate with each other over an undirected connected graph.
no code implementations • 22 Jan 2021 • Shixiang Chen, Alfredo Garcia, Mingyi Hong, Shahin Shahrampour
We study the convergence properties of Riemannian gradient method for solving the consensus problem (for an undirected connected graph) over the Stiefel manifold.
no code implementations • 2 Aug 2020 • Yanling Chang, Alfredo Garcia, Zhide Wang, Lu Sun
In this context, replacement decisions must be made under partial/imperfect information on the true state (i. e. condition of the equipment).
no code implementations • 28 Apr 2020 • Shixiang Chen, Alfredo Garcia, Shahin Shahrampour
In this paper, we propose a distributed implementation of the stochastic subgradient method with a theoretical guarantee.
no code implementations • 28 Oct 2019 • Lingzhou Hong, Alfredo Garcia, Ceyhun Eksin
We consider a distributed estimation method in a setting with heterogeneous streams of correlated data distributed across nodes in a network.
no code implementations • 11 Jun 2018 • Shi Pu, Alfredo Garcia
We study a distributed framework for stochastic optimization which is inspired by models of collective motion found in nature (e. g., swarming) with mild communication requirements.
no code implementations • 27 Oct 2017 • Davood Hajinezhad, Mingyi Hong, Alfredo Garcia
In this paper, we consider distributed optimization problems over a multi-agent network, where each agent can only partially evaluate the objective function, and it is allowed to exchange messages with its immediate neighbors.