no code implementations • 28 Nov 2023 • Ebtesam Almazrouei, Hamza Alobeidli, Abdulaziz Alshamsi, Alessandro Cappelli, Ruxandra Cojocaru, Mérouane Debbah, Étienne Goffinet, Daniel Hesslow, Julien Launay, Quentin Malartic, Daniele Mazzotta, Badreddine Noune, Baptiste Pannier, Guilherme Penedo
We report detailed evaluations, as well as a deep dive into the methods and custom tooling employed to pretrain Falcon.
Ranked #17 on Sentence Completion on HellaSwag
1 code implementation • 1 Jun 2023 • Guilherme Penedo, Quentin Malartic, Daniel Hesslow, Ruxandra Cojocaru, Alessandro Cappelli, Hamza Alobeidli, Baptiste Pannier, Ebtesam Almazrouei, Julien Launay
Large language models are commonly trained on a mixture of filtered web data and curated high-quality corpora, such as social media conversations, books, or technical papers.
no code implementations • 26 Oct 2022 • Matthew J. Filipovich, Alessandro Cappelli, Daniel Hesslow, Julien Launay
Alternatives to backpropagation have long been studied to better understand how biological brains may learn.
no code implementations • LREC 2022 • Julien Launay, E. L. Tommasone, Baptiste Pannier, François Boniface, Amélie Chatelain, Alessandro Cappelli, Iacopo Poli, Djamé Seddah
We fit a scaling law for compute for the French language, and compare it with its English counterpart.
no code implementations • ICML Workshop AML 2021 • Alessandro Cappelli, Julien Launay, Laurent Meunier, Ruben Ohana, Iacopo Poli
Robustness to adversarial attacks is typically obtained through expensive adversarial training with Projected Gradient Descent.
no code implementations • NeurIPS 2021 • Ruben Ohana, Hamlet J. Medina Ruiz, Julien Launay, Alessandro Cappelli, Iacopo Poli, Liva Ralaivola, Alain Rakotomamonjy
Optical Processing Units (OPUs) -- low-power photonic chips dedicated to large scale random projections -- have been used in previous work to train deep neural networks using Direct Feedback Alignment (DFA), an effective alternative to backpropagation.
no code implementations • 29 Apr 2021 • Daniel Hesslow, Alessandro Cappelli, Igor Carron, Laurent Daudet, Raphaël Lafargue, Kilian Müller, Ruben Ohana, Gustave Pariente, Iacopo Poli
Randomized Numerical Linear Algebra (RandNLA) is a powerful class of methods, widely used in High Performance Computing (HPC).
1 code implementation • 6 Jan 2021 • Alessandro Cappelli, Ruben Ohana, Julien Launay, Laurent Meunier, Iacopo Poli, Florent Krzakala
In the white-box setting, our defense works by obfuscating the parameters of the random projection.