no code implementations • 17 Oct 2022 • Timofey Grigoryev, Polina Verezemskaya, Mikhail Krinitskiy, Nikita Anikin, Alexander Gavrikov, Ilya Trofimov, Nikita Balabin, Aleksei Shpilman, Andrei Eremchenko, Sergey Gulev, Evgeny Burnaev, Vladimir Vanovskiy
Global warming made the Arctic available for marine operations and created demand for reliable operational sea ice forecasts to make them safe.
1 code implementation • 25 May 2022 • Vladimir Egorov, Aleksei Shpilman
While in mixed environments full autonomy of the agents can be a desirable outcome, cooperative environments allow agents to share information to facilitate coordination.
Model-based Reinforcement Learning reinforcement-learning +2
1 code implementation • 31 Mar 2022 • Christian Eichenberger, Moritz Neun, Henry Martin, Pedro Herruzo, Markus Spanring, Yichao Lu, Sungbin Choi, Vsevolod Konyakhin, Nina Lukashina, Aleksei Shpilman, Nina Wiedemann, Martin Raubal, Bo wang, Hai L. Vu, Reza Mohajerpoor, Chen Cai, Inhi Kim, Luca Hermes, Andrew Melnik, Riza Velioglu, Markus Vieth, Malte Schilling, Alabi Bojesomo, Hasan Al Marzouqi, Panos Liatsis, Jay Santokhi, Dylan Hillier, Yiming Yang, Joned Sarwar, Anna Jordan, Emil Hewage, David Jonietz, Fei Tang, Aleksandra Gruca, Michael Kopp, David Kreil, Sepp Hochreiter
The IARAI Traffic4cast competitions at NeurIPS 2019 and 2020 showed that neural networks can successfully predict future traffic conditions 1 hour into the future on simply aggregated GPS probe data in time and space bins.
no code implementations • 21 Mar 2022 • Georgiy Pshikhachev, Dmitry Ivanov, Vladimir Egorov, Aleksei Shpilman
Modern LfD algorithms require meticulous tuning of hyperparameters that control the influence of demonstrations and, as we show in the paper, struggle with learning from suboptimal demonstrations.
1 code implementation • 14 Mar 2022 • Farid Bagirov, Dmitry Ivanov, Aleksei Shpilman
The former only learns from labeled positive data, whereas the latter also utilizes unlabeled data to improve the overall performance.
no code implementations • 17 Feb 2022 • Anssi Kanervisto, Stephanie Milani, Karolis Ramanauskas, Nicholay Topin, Zichuan Lin, Junyou Li, Jianing Shi, Deheng Ye, Qiang Fu, Wei Yang, Weijun Hong, Zhongyue Huang, Haicheng Chen, Guangjun Zeng, Yue Lin, Vincent Micheli, Eloi Alonso, François Fleuret, Alexander Nikulin, Yury Belousov, Oleg Svidchenko, Aleksei Shpilman
With this in mind, we hosted the third edition of the MineRL ObtainDiamond competition, MineRL Diamond 2021, with a separate track in which we permitted any solution to promote the participation of newcomers.
no code implementations • 2 Dec 2021 • Oleg Svidchenko, Aleksei Shpilman
Recent advances in reinforcement learning have demonstrated its ability to solve hard agent-environment interaction tasks on a super-human level.
1 code implementation • 20 Nov 2021 • Natalia Zenkova, Ekaterina Sedykh, Tatiana Shugaeva, Vladislav Strashko, Timofei Ermak, Aleksei Shpilman
In this work, we present an end-to-end model to predict CDR H3 loop structure, that performs on par with state-of-the-art methods in terms of accuracy but an order of magnitude faster.
1 code implementation • 5 Nov 2021 • Vsevolod Konyakhin, Nina Lukashina, Aleksei Shpilman
In this technical report, we present our solution to the Traffic4Cast 2021 Core Challenge, in which participants were asked to develop algorithms for predicting a traffic state 60 minutes ahead, based on the information from the previous hour, in 4 different cities.
no code implementations • 30 Mar 2021 • Florian Laurent, Manuel Schneider, Christian Scheller, Jeremy Watson, Jiaoyang Li, Zhe Chen, Yi Zheng, Shao-Hung Chan, Konstantin Makhnev, Oleg Svidchenko, Vladimir Egorov, Dmitry Ivanov, Aleksei Shpilman, Evgenija Spirovska, Oliver Tanevski, Aleksandar Nikov, Ramon Grunder, David Galevski, Jakov Mitrovski, Guillaume Sartoretti, Zhiyao Luo, Mehul Damani, Nilabha Bhattacharya, Shivam Agarwal, Adrian Egli, Erik Nygren, Sharada Mohanty
However, the coordination of hundreds of agents in a real-life setting like a railway network remains challenging and the Flatland environment used for the competition models these real-world properties in a simplified manner.
1 code implementation • 24 Feb 2021 • Dmitry Ivanov, Vladimir Egorov, Aleksei Shpilman
Recent reinforcement learning studies extensively explore the interplay between cooperative and competitive behaviour in mixed environments.
1 code implementation • 18 Dec 2020 • Mikita Sazanovich, Anastasiya Nikolskaya, Yury Belousov, Aleksei Shpilman
Black-box optimization is one of the vital tasks in machine learning, since it approximates real-world conditions, in that we do not always know all the properties of a given system, up to knowing almost nothing but the results.
no code implementations • 17 Dec 2020 • Aleksandra Malysheva, Daniel Kudenko, Aleksei Shpilman
Over recent years, deep reinforcement learning has shown strong successes in complex single-agent tasks, and more recently this approach has also been applied to multi-agent domains.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 17 Dec 2020 • Vladislav Belyaev, Aleksandra Malysheva, Aleksei Shpilman
The task object tracking is vital in numerous applications such as autonomous driving, intelligent surveillance, robotics, etc.
no code implementations • 16 Dec 2020 • Ivan Sosin, Daniel Kudenko, Aleksei Shpilman
Movement control of artificial limbs has made big advances in recent years.
no code implementations • 16 Dec 2020 • Anastasia Gaydashenko, Daniel Kudenko, Aleksei Shpilman
Robot navigation through crowds poses a difficult challenge to AI systems, since the methods should result in fast and efficient movement but at the same time are not allowed to compromise safety.
no code implementations • 16 Dec 2020 • Aleksandra Malysheva, Daniel Kudenko, Aleksei Shpilman
In this paper, we demonstrate how data from videos of human running (e. g. taken from YouTube) can be used to shape the reward of the humanoid learning agent to speed up the learning and produce a better result.
no code implementations • 16 Dec 2020 • Aleksei Shpilman, Dmitry Boikiy, Marina Polyakova, Daniel Kudenko, Anton Burakov, Elena Nadezhdina
Human experts find it particularly difficult to recognize the levels of chemical compound exposure of a cell.
2 code implementations • 24 Nov 2020 • Nina Lukashina, Alisa Alenicheva, Elizaveta Vlasova, Artem Kondiukov, Aigul Khakimova, Emil Magerramov, Nikita Churikov, Aleksei Shpilman
Lipophilicity is one of the factors determining the permeability of the cell membrane to a drug molecule.
1 code implementation • 8 Oct 2020 • Anna Nikiforovskaya, Nikolai Kapralov, Anna Vlasova, Oleg Shpynov, Aleksei Shpilman
In this paper, we present a method for the automatic generation of a review paper corresponding to a user-defined query.
no code implementations • 7 Jul 2020 • Mikita Sazanovich, Konstantin Chaika, Kirill Krinkin, Aleksei Shpilman
In this paper, we describe our winning approach to solving the Lane Following Challenge at the AI Driving Olympics Competition through imitation learning on a mixed set of simulation and real-world data.
1 code implementation • 7 Feb 2019 • Łukasz Kidziński, Carmichael Ong, Sharada Prasanna Mohanty, Jennifer Hicks, Sean F. Carroll, Bo Zhou, Hongsheng Zeng, Fan Wang, Rongzhong Lian, Hao Tian, Wojciech Jaśkowski, Garrett Andersen, Odd Rune Lykkebø, Nihat Engin Toklu, Pranav Shyam, Rupesh Kumar Srivastava, Sergey Kolesnikov, Oleksii Hrinchuk, Anton Pechenko, Mattias Ljungström, Zhen Wang, Xu Hu, Zehong Hu, Minghui Qiu, Jun Huang, Aleksei Shpilman, Ivan Sosin, Oleg Svidchenko, Aleksandra Malysheva, Daniel Kudenko, Lance Rane, Aditya Bhatt, Zhengfei Wang, Penghui Qi, Zeyang Yu, Peng Peng, Quan Yuan, Wenxin Li, Yunsheng Tian, Ruihan Yang, Pingchuan Ma, Shauharda Khadka, Somdeb Majumdar, Zach Dwiel, Yinyin Liu, Evren Tumer, Jeremy Watson, Marcel Salathé, Sergey Levine, Scott Delp
In the NeurIPS 2018 Artificial Intelligence for Prosthetics challenge, participants were tasked with building a controller for a musculoskeletal model with a goal of matching a given time-varying velocity vector.
1 code implementation • 30 Nov 2018 • Aleksandra Malysheva, Tegg Taekyong Sung, Chae-Bong Sohn, Daniel Kudenko, Aleksei Shpilman
Over recent years, deep reinforcement learning has shown strong successes in complex single-agent tasks, and more recently this approach has also been applied to multi-agent domains.
Multi-agent Reinforcement Learning reinforcement-learning +1