Search Results for author: Juan Fernández-Olivares

Found 6 papers, 1 papers with code

Towards a Unified Framework for Sequential Decision Making

no code implementations3 Oct 2023 Carlos Núñez-Molina, Pablo Mesejo, Juan Fernández-Olivares

In recent years, the integration of Automated Planning (AP) and Reinforcement Learning (RL) has seen a surge of interest.

Bayesian Inference Decision Making +1

On Using Admissible Bounds for Learning Forward Search Heuristics

no code implementations23 Aug 2023 Carlos Núñez-Molina, Masataro Asai, Juan Fernández-Olivares, Pablo Mesejo

This results in a different loss function from the MSE commonly employed in the literature, which implicitly models the learned heuristic as a gaussian distribution.

A Review of Symbolic, Subsymbolic and Hybrid Methods for Sequential Decision Making

no code implementations20 Apr 2023 Carlos Núñez-Molina, Pablo Mesejo, Juan Fernández-Olivares

Conversely, Reinforcement Learning (RL) proposes to learn the solution of the SDP from data, without a world model, and represent the learned knowledge subsymbolically.

Decision Making Reinforcement Learning (RL)

NeSIG: A Neuro-Symbolic Method for Learning to Generate Planning Problems

1 code implementation24 Jan 2023 Carlos Núñez-Molina, Pablo Mesejo, Juan Fernández-Olivares

In this paper we propose NeSIG, to the best of our knowledge the first domain-independent method for automatically generating planning problems that are valid, diverse and difficult to solve.

valid

Learning Numerical Action Models from Noisy Input Data

no code implementations9 Nov 2021 José Á. Segura-Muros, Juan Fernández-Olivares, Raúl Pérez

The algorithm presented here improves the learning capabilities of PlanMiner when using noisy data as input.

Goal Reasoning by Selecting Subgoals with Deep Q-Learning

no code implementations22 Dec 2020 Carlos Núñez-Molina, Vladislav Nikolov, Ignacio Vellido, Juan Fernández-Olivares

In this work we propose a goal reasoning method which learns to select subgoals with Deep Q-Learning in order to decrease the load of a planner when faced with scenarios with tight time restrictions, such as online execution systems.

Q-Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.