Continuous Control

422 papers with code • 73 benchmarks • 10 datasets

Continuous control in the context of playing games, especially within artificial intelligence (AI) and machine learning (ML), refers to the ability to make a series of smooth, ongoing adjustments or actions to control a game or a simulation. This is in contrast to discrete control, where the actions are limited to a set of specific, distinct choices. Continuous control is crucial in environments where precision, timing, and the magnitude of actions matter, such as driving a car in a racing game, controlling a character in a simulation, or managing the flight of an aircraft in a flight simulator.

Libraries

Use these libraries to find Continuous Control models and implementations

Latest papers with no code

Probabilistic Actor-Critic: Learning to Explore with PAC-Bayes Uncertainty

no code yet • 5 Feb 2024

We introduce Probabilistic Actor-Critic (PAC), a novel reinforcement learning algorithm with improved continuous control performance thanks to its ability to mitigate the exploration-exploitation trade-off.

Frugal Actor-Critic: Sample Efficient Off-Policy Deep Reinforcement Learning Using Unique Experiences

no code yet • 5 Feb 2024

Efficient utilization of the replay buffer plays a significant role in the off-policy actor-critic reinforcement learning (RL) algorithms used for model-free control policy synthesis for complex dynamical systems.

A Strategy for Preparing Quantum Squeezed States Using Reinforcement Learning

no code yet • 29 Jan 2024

It is exemplified by the application to prepare spin-squeezed states for an open collective spin model where a linear control field is designed to govern the dynamics.

Pulse Width Modulation Method Applied to Nonlinear Model Predictive Control on an Under-actuated Small Satellite

no code yet • 21 Jan 2024

Among various satellite actuators, magnetic torquers have been widely equipped for stabilization and attitude control of small satellites.

Identifying Policy Gradient Subspaces

no code yet • 12 Jan 2024

Policy gradient methods hold great potential for solving complex continuous control tasks.

The Distributional Reward Critic Architecture for Perturbed-Reward Reinforcement Learning

no code yet • 11 Jan 2024

We study reinforcement learning in the presence of an unknown reward perturbation.

A Minimaximalist Approach to Reinforcement Learning from Human Feedback

no code yet • 8 Jan 2024

Our approach is maximalist in that it provably handles non-Markovian, intransitive, and stochastic preferences while being robust to the compounding errors that plague offline approaches to sequential prediction.

Trajectory-Oriented Policy Optimization with Sparse Rewards

no code yet • 4 Jan 2024

The proposed algorithm undergoes evaluation across extensive discrete and continuous control tasks with sparse and misleading rewards.

Adversarially Trained Actor Critic for offline CMDPs

no code yet • 1 Jan 2024

Theoretically, we demonstrate that when the actor employs a no-regret optimization oracle, SATAC achieves two guarantees: (i) For the first time in the offline RL setting, we establish that SATAC can produce a policy that outperforms the behavior policy while maintaining the same level of safety, which is critical to designing an algorithm for offline RL.

Ensemble-based Interactive Imitation Learning

no code yet • 28 Dec 2023

We study interactive imitation learning, where a learner interactively queries a demonstrating expert for action annotations, aiming to learn a policy that has performance competitive with the expert, using as few annotations as possible.