1 code implementation • 10 Nov 2021 • Andrew Cohen, Ervin Teng, Vincent-Pierre Berges, Ruo-Ping Dong, Hunter Henry, Marwan Mattar, Alexander Zook, Sujoy Ganguly
In this work, we first demonstrate that sample complexity increases with the quantity of absorbing states in a toy supervised learning task for a fully connected network, while attention is more robust to variable size input.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 29 Aug 2021 • Ervin Teng, Bob Iannucci
Learning requires both study and curiosity.
no code implementations • 5 Feb 2019 • Ervin Teng, Bob Iannucci
We use a 3D simulation environment and deep reinforcement learning to train a curiosity agent to, in turn, train the object detection model.
3 code implementations • 4 Feb 2019 • Arthur Juliani, Ahmed Khalifa, Vincent-Pierre Berges, Jonathan Harper, Ervin Teng, Hunter Henry, Adam Crespi, Julian Togelius, Danny Lange
Unlike other benchmarks such as the Arcade Learning Environment, evaluation of agent performance in Obstacle Tower is based on an agent's ability to perform well on unseen instances of the environment.
56 code implementations • 7 Sep 2018 • Arthur Juliani, Vincent-Pierre Berges, Ervin Teng, Andrew Cohen, Jonathan Harper, Chris Elion, Chris Goy, Yuan Gao, Hunter Henry, Marwan Mattar, Danny Lange
Recent advances in artificial intelligence have been driven by the presence of increasingly realistic and complex simulated environments.
no code implementations • 27 Mar 2018 • Ervin Teng, Rui Huang, Bob Iannucci
Modern deep convolutional neural networks (CNNs) for image classification and object detection are often trained offline on large static datasets.
no code implementations • 15 Sep 2017 • Ervin Teng, João Diogo Falcão, Bob Iannucci
Today's general-purpose deep convolutional neural networks (CNN) for image classification and object detection are trained offline on large static datasets.