Search Results for author: Amal Rannen-Triki

Found 7 papers, 1 papers with code

Transformers for Supervised Online Continual Learning

no code implementations3 Mar 2024 Jorg Bornschein, Yazhe Li, Amal Rannen-Triki

Inspired by the in-context learning capabilities of transformers and their connection to meta-learning, we propose a method that leverages these strengths for online continual learning.

Continual Learning Few-Shot Learning +2

Towards Robust and Efficient Continual Language Learning

no code implementations11 Jul 2023 Adam Fisch, Amal Rannen-Triki, Razvan Pascanu, Jörg Bornschein, Angeliki Lazaridou, Elena Gribovskaya, Marc'Aurelio Ranzato

As the application space of language models continues to evolve, a natural question to ask is how we can quickly adapt models to new tasks.

Continual Learning

Kalman Filter for Online Classification of Non-Stationary Data

no code implementations14 Jun 2023 Michalis K. Titsias, Alexandre Galashov, Amal Rannen-Triki, Razvan Pascanu, Yee Whye Teh, Jorg Bornschein

Non-stationarity over the linear predictor weights is modelled using a parameter drift transition density, parametrized by a coefficient that quantifies forgetting.

Classification Continual Learning +1

Towards Compute-Optimal Transfer Learning

no code implementations25 Apr 2023 Massimo Caccia, Alexandre Galashov, Arthur Douillard, Amal Rannen-Triki, Dushyant Rao, Michela Paganini, Laurent Charlin, Marc'Aurelio Ranzato, Razvan Pascanu

The field of transfer learning is undergoing a significant shift with the introduction of large pretrained models which have demonstrated strong adaptability to a variety of downstream tasks.

Computational Efficiency Continual Learning +1

On the Role of Optimization in Double Descent: A Least Squares Study

no code implementations NeurIPS 2021 Ilja Kuzborskij, Csaba Szepesvári, Omar Rivasplata, Amal Rannen-Triki, Razvan Pascanu

Empirically it has been observed that the performance of deep neural networks steadily improves as we increase model size, contradicting the classical view on overfitting and generalization.

Cannot find the paper you are looking for? You can Submit a new open access paper.