Search Results for author: Maxime Gazeau

Found 9 papers, 2 papers with code

In-context Reinforcement Learning with Algorithm Distillation

1 code implementation25 Oct 2022 Michael Laskin, Luyu Wang, Junhyuk Oh, Emilio Parisotto, Stephen Spencer, Richie Steigerwald, DJ Strouse, Steven Hansen, Angelos Filos, Ethan Brooks, Maxime Gazeau, Himanshu Sahni, Satinder Singh, Volodymyr Mnih

We propose Algorithm Distillation (AD), a method for distilling reinforcement learning (RL) algorithms into neural networks by modeling their training histories with a causal sequence model.

reinforcement-learning

A*Net: A Scalable Path-based Reasoning Approach for Knowledge Graphs

2 code implementations NeurIPS 2023 Zhaocheng Zhu, Xinyu Yuan, Mikhail Galkin, Sophie Xhonneux, Ming Zhang, Maxime Gazeau, Jian Tang

Experiments on both transductive and inductive knowledge graph reasoning benchmarks show that A*Net achieves competitive performance with existing state-of-the-art path-based methods, while merely visiting 10% nodes and 10% edges at each iteration.

Knowledge Graphs

Higher Order Generalization Error for First Order Discretization of Langevin Diffusion

no code implementations11 Feb 2021 Mufan Bill Li, Maxime Gazeau

We propose a novel approach to analyze generalization error for discretizations of Langevin diffusion, such as the stochastic gradient Langevin dynamics (SGLD).

An Empirical Study of Large-Batch Stochastic Gradient Descent with Structured Covariance Noise

no code implementations21 Feb 2019 Yeming Wen, Kevin Luk, Maxime Gazeau, Guodong Zhang, Harris Chan, Jimmy Ba

We demonstrate that the learning performance of our method is more accurately captured by the structure of the covariance matrix of the noise rather than by the variance of gradients.

Stochastic Optimization

A general system of differential equations to model first order adaptive algorithms

no code implementations31 Oct 2018 André Belotto da Silva, Maxime Gazeau

In this paper, we derive a non-autonomous system of differential equations, which is the continuous time limit of adaptive optimization methods.

Exploring Curvature Noise in Large-Batch Stochastic Optimization

no code implementations27 Sep 2018 Yeming Wen, Kevin Luk, Maxime Gazeau, Guodong Zhang, Harris Chan, Jimmy Ba

Unfortunately, a major drawback is the so-called generalization gap: large-batch training typically leads to a degradation in generalization performance of the model as compared to small-batch training.

Stochastic Optimization

Scalable Recommender Systems through Recursive Evidence Chains

no code implementations5 Jul 2018 Elias Tragas, Calvin Luo, Maxime Gazeau, Kevin Luk, David Duvenaud

Recommender systems can be formulated as a matrix completion problem, predicting ratings from user and item parameter vectors.

Matrix Completion Recommendation Systems

Implicit Manifold Learning on Generative Adversarial Networks

no code implementations30 Oct 2017 Kry Yik Chau Lui, Yanshuai Cao, Maxime Gazeau, Kelvin Shuangjian Zhang

This paper raises an implicit manifold learning perspective in Generative Adversarial Networks (GANs), by studying how the support of the learned distribution, modelled as a submanifold $\mathcal{M}_{\theta}$, perfectly match with $\mathcal{M}_{r}$, the support of the real data distribution.

Cannot find the paper you are looking for? You can Submit a new open access paper.