Search Results for author: Tomi Silander

Found 9 papers, 0 papers with code

Improving the Generalization of Visual Navigation Policies using Invariance Regularization

no code implementations ICLR 2020 Michel Aractingi, Christopher Dance, Julien Perez, Tomi Silander

The results of this method, called invariance regularization, show an improvement in the generalization of policies to environments not seen during training.

Reinforcement Learning (RL) Visual Navigation

DEEP ADVERSARIAL FORWARD MODEL

no code implementations27 Sep 2018 Morgan Funtowicz, Tomi Silander, Arnaud Sors, Julien Perez

More precisely, our forward model is trained to produce realistic observations of the future while a discriminator model is trained to distinguish between real images and the model’s prediction of the future.

Image Generation Reinforcement Learning (RL)

Contextual memory bandit for pro-active dialog engagement

no code implementations ICLR 2018 julien perez, Tomi Silander

In this paper, we propose to introduce the paradigm of contextual bandits as framework for pro-active dialog systems.

Multi-Armed Bandits

Non-Markovian Control with Gated End-to-End Memory Policy Networks

no code implementations31 May 2017 Julien Perez, Tomi Silander

In this paper, we explore the use of a recently proposed attention-based model, the Gated End-to-End Memory Network, for sequential control.

OpenAI Gym

Optimal Policies for Observing Time Series and Related Restless Bandit Problems

no code implementations29 Mar 2017 Christopher R. Dance, Tomi Silander

We discuss computation of that index, give closed-form formulae for it, and compare the performance of the associated index policy with heuristic policies.

Time Series Time Series Analysis

When are Kalman-filter restless bandits indexable?

no code implementations NeurIPS 2015 Christopher R. Dance, Tomi Silander

We study the restless bandit associated with an extremely simple scalar Kalman filter model in discrete time.

Transferring Expectations in Model-based Reinforcement Learning

no code implementations NeurIPS 2012 Trung Nguyen, Tomi Silander, Tze Y. Leong

We study how to automatically select and adapt multiple abstractions or representations of the world to support model-based reinforcement learning.

Model-based Reinforcement Learning reinforcement-learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.