Search Results for author: Nicholay Topin

Found 18 papers, 6 papers with code

Use-Case-Grounded Simulations for Explanation Evaluation

no code implementations5 Jun 2022 Valerie Chen, Nari Johnson, Nicholay Topin, Gregory Plumb, Ameet Talwalkar

SimEvals involve training algorithmic agents that take as input the information content (such as model explanations) that would be presented to each participant in a human subject study, to predict answers to the use case of interest.

counterfactual Counterfactual Reasoning

MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned

no code implementations17 Feb 2022 Anssi Kanervisto, Stephanie Milani, Karolis Ramanauskas, Nicholay Topin, Zichuan Lin, Junyou Li, Jianing Shi, Deheng Ye, Qiang Fu, Wei Yang, Weijun Hong, Zhongyue Huang, Haicheng Chen, Guangjun Zeng, Yue Lin, Vincent Micheli, Eloi Alonso, François Fleuret, Alexander Nikulin, Yury Belousov, Oleg Svidchenko, Aleksei Shpilman

With this in mind, we hosted the third edition of the MineRL ObtainDiamond competition, MineRL Diamond 2021, with a separate track in which we permitted any solution to promote the participation of newcomers.

A Survey of Explainable Reinforcement Learning

no code implementations17 Feb 2022 Stephanie Milani, Nicholay Topin, Manuela Veloso, Fei Fang

In this survey, we propose a novel taxonomy for organizing the XRL literature that prioritizes the RL setting.

Decision Making reinforcement-learning +1

The MineRL BASALT Competition on Learning from Human Feedback

no code implementations5 Jul 2021 Rohin Shah, Cody Wild, Steven H. Wang, Neel Alex, Brandon Houghton, William Guss, Sharada Mohanty, Anssi Kanervisto, Stephanie Milani, Nicholay Topin, Pieter Abbeel, Stuart Russell, Anca Dragan

Rather than training AI systems using a predefined reward function or using a labeled dataset with a predefined set of categories, we instead train the AI system using a learning signal derived from some form of human feedback, which can evolve over time as the understanding of the task changes, or as the capabilities of the AI system improve.

Imitation Learning

Iterative Bounding MDPs: Learning Interpretable Policies via Non-Interpretable Methods

no code implementations25 Feb 2021 Nicholay Topin, Stephanie Milani, Fei Fang, Manuela Veloso

Because of this decision tree equivalence, any function approximator can be used during training, including a neural network, while yielding a decision tree policy for the base MDP.

reinforcement-learning Reinforcement Learning (RL)

The MineRL 2020 Competition on Sample Efficient Reinforcement Learning using Human Priors

no code implementations26 Jan 2021 William H. Guss, Mario Ynocente Castro, Sam Devlin, Brandon Houghton, Noboru Sean Kuno, Crissman Loomis, Stephanie Milani, Sharada Mohanty, Keisuke Nakata, Ruslan Salakhutdinov, John Schulman, Shinya Shiroshita, Nicholay Topin, Avinash Ummadisingu, Oriol Vinyals

Although deep reinforcement learning has led to breakthroughs in many difficult domains, these successes have required an ever-increasing number of samples, affording only a shrinking segment of the AI community access to their development.

Decision Making Efficient Exploration +2

Guaranteeing Reproducibility in Deep Learning Competitions

no code implementations12 May 2020 Brandon Houghton, Stephanie Milani, Nicholay Topin, William Guss, Katja Hofmann, Diego Perez-Liebana, Manuela Veloso, Ruslan Salakhutdinov

To encourage the development of methods with reproducible and robust training behavior, we propose a challenge paradigm where competitors are evaluated directly on the performance of their learning procedures rather than pre-trained agents.

Retrospective Analysis of the 2019 MineRL Competition on Sample Efficient Reinforcement Learning

no code implementations10 Mar 2020 Stephanie Milani, Nicholay Topin, Brandon Houghton, William H. Guss, Sharada P. Mohanty, Keisuke Nakata, Oriol Vinyals, Noboru Sean Kuno

To facilitate research in the direction of sample efficient reinforcement learning, we held the MineRL Competition on Sample Efficient Reinforcement Learning Using Human Priors at the Thirty-third Conference on Neural Information Processing Systems (NeurIPS 2019).

Imitation Learning reinforcement-learning +1

Conservative Q-Improvement: Reinforcement Learning for an Interpretable Decision-Tree Policy

1 code implementation2 Jul 2019 Aaron M. Roth, Nicholay Topin, Pooyan Jamshidi, Manuela Veloso

There is a growing desire in the field of reinforcement learning (and machine learning in general) to move from black-box models toward more "interpretable AI."

reinforcement-learning Reinforcement Learning (RL)

Generation of Policy-Level Explanations for Reinforcement Learning

no code implementations28 May 2019 Nicholay Topin, Manuela Veloso

Though reinforcement learning has greatly benefited from the incorporation of neural networks, the inability to verify the correctness of such systems limits their use.

reinforcement-learning Reinforcement Learning (RL)

The MineRL 2019 Competition on Sample Efficient Reinforcement Learning using Human Priors

1 code implementation22 Apr 2019 William H. Guss, Cayden Codel, Katja Hofmann, Brandon Houghton, Noboru Kuno, Stephanie Milani, Sharada Mohanty, Diego Perez Liebana, Ruslan Salakhutdinov, Nicholay Topin, Manuela Veloso, Phillip Wang

To that end, we introduce: (1) the Minecraft ObtainDiamond task, a sequential decision making environment requiring long-term planning, hierarchical control, and efficient exploration methods; and (2) the MineRL-v0 dataset, a large-scale collection of over 60 million state-action pairs of human demonstrations that can be resimulated into embodied trajectories with arbitrary modifications to game state and visuals.

Decision Making Efficient Exploration +2

Super-Convergence: Very Fast Training of Residual Networks Using Large Learning Rates

no code implementations ICLR 2018 Leslie N. Smith, Nicholay Topin

In this paper, we show a phenomenon, which we named ``super-convergence'', where residual networks can be trained using an order of magnitude fewer iterations than is used with standard training methods.

Super-Convergence: Very Fast Training of Neural Networks Using Large Learning Rates

10 code implementations23 Aug 2017 Leslie N. Smith, Nicholay Topin

One of the key elements of super-convergence is training with one learning rate cycle and a large maximum learning rate.

Exploring loss function topology with cyclical learning rates

2 code implementations14 Feb 2017 Leslie N. Smith, Nicholay Topin

We present observations and discussion of previously unreported phenomena discovered while training residual networks.

Deep Convolutional Neural Network Design Patterns

1 code implementation2 Nov 2016 Leslie N. Smith, Nicholay Topin

Recent research in the deep learning field has produced a plethora of new architectures.

Cannot find the paper you are looking for? You can Submit a new open access paper.