Search Results for author: Sheila Mcilraith

Found 8 papers, 3 papers with code

Optimal Decision Trees For Interpretable Clustering with Constraints (Extended Version)

no code implementations30 Jan 2023 Pouya Shati, Eldan Cohen, Sheila Mcilraith

In this work, we present a novel SAT-based framework for interpretable clustering that supports clustering constraints and that also provides strong theoretical guarantees on solution quality.

Constrained Clustering

You Can't Count on Luck: Why Decision Transformers and RvS Fail in Stochastic Environments

no code implementations31 May 2022 Keiran Paster, Sheila Mcilraith, Jimmy Ba

In all tested domains, ESPER achieves significantly better alignment between the target return and achieved return than simply conditioning on returns.

Offline RL Playing the Game of 2048

Efficient Multi-agent Epistemic Planning: Teaching Planners About Nested Belief

no code implementations6 Oct 2021 Christian Muise, Vaishak Belle, Paolo Felli, Sheila Mcilraith, Tim Miller, Adrian R. Pearce, Liz Sonenberg

Many AI applications involve the interaction of multiple autonomous agents, requiring those agents to reason about their own beliefs, as well as those of other agents.

LTL2Action: Generalizing LTL Instructions for Multi-Task RL

1 code implementation13 Feb 2021 Pashootan Vaezipoor, Andrew Li, Rodrigo Toro Icarte, Sheila Mcilraith

We address the problem of teaching a deep reinforcement learning (RL) agent to follow instructions in multi-task environments.

Reinforcement Learning (RL)

Learning Reward Machines for Partially Observable Reinforcement Learning

1 code implementation NeurIPS 2019 Rodrigo Toro Icarte, Ethan Waldie, Toryn Klassen, Rick Valenzano, Margarita Castro, Sheila Mcilraith

Reward Machines (RMs), originally proposed for specifying problems in Reinforcement Learning (RL), provide a structured, automata-based representation of a reward function that allows an agent to decompose problems into subproblems that can be efficiently learned using off-policy learning.

Partially Observable Reinforcement Learning Problem Decomposition +2

Using Reward Machines for High-Level Task Specification and Decomposition in Reinforcement Learning

1 code implementation ICML 2018 Rodrigo Toro Icarte, Toryn Klassen, Richard Valenzano, Sheila Mcilraith

In this paper we propose Reward Machines {—} a type of finite state machine that supports the specification of reward functions while exposing reward function structure to the learner and supporting decomposition.

Hierarchical Reinforcement Learning Q-Learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.