Search Results for author: Grzegorz Świrszcz

Found 2 papers, 1 papers with code

Sobolev Training for Neural Networks

no code implementations NeurIPS 2017 Wojciech Marian Czarnecki, Simon Osindero, Max Jaderberg, Grzegorz Świrszcz, Razvan Pascanu

In many cases we only have access to input-output pairs from the ground truth, however it is becoming more common to have access to derivatives of the target output with respect to the input - for example when the ground truth function is itself a neural network such as in network compression or distillation.

Understanding Synthetic Gradients and Decoupled Neural Interfaces

1 code implementation ICML 2017 Wojciech Marian Czarnecki, Grzegorz Świrszcz, Max Jaderberg, Simon Osindero, Oriol Vinyals, Koray Kavukcuoglu

When training neural networks, the use of Synthetic Gradients (SG) allows layers or modules to be trained without update locking - without waiting for a true error gradient to be backpropagated - resulting in Decoupled Neural Interfaces (DNIs).

Cannot find the paper you are looking for? You can Submit a new open access paper.