Search Results for author: Friedemann Zenke

Found 15 papers, 11 papers with code

Elucidating the theoretical underpinnings of surrogate gradient learning in spiking neural networks

1 code implementation23 Apr 2024 Julia Gygax, Friedemann Zenke

We find that the latter provides the missing theoretical basis for surrogate gradients in stochastic spiking neural networks.

Dis-inhibitory neuronal circuits can control the sign of synaptic plasticity

1 code implementation NeurIPS 2023 Julian Rossbroich, Friedemann Zenke

How neuronal circuits achieve credit assignment remains a central unsolved question in systems neuroscience.

Improving equilibrium propagation without weight symmetry through Jacobian homeostasis

1 code implementation5 Sep 2023 Axel Laborieux, Friedemann Zenke

Equilibrium propagation (EP) is a compelling alternative to the backpropagation of error algorithm (BP) for computing gradients of neural networks on biological or analog neuromorphic substrates.

Implicit variance regularization in non-contrastive SSL

1 code implementation NeurIPS 2023 Manu Srinath Halvagal, Axel Laborieux, Friedemann Zenke

To gain further theoretical insight into non-contrastive SSL, we analytically study learning dynamics in conjunction with Euclidean and cosine similarity in the eigenspace of closed-form linear predictor networks.

Holomorphic Equilibrium Propagation Computes Exact Gradients Through Finite Size Oscillations

1 code implementation1 Sep 2022 Axel Laborieux, Friedemann Zenke

Equilibrium propagation (EP) is an alternative to backpropagation (BP) that allows the training of deep neural networks with local learning rules.

Fluctuation-driven initialization for spiking neural network training

1 code implementation21 Jun 2022 Julian Rossbroich, Julia Gygax, Friedemann Zenke

Thus fluctuation-driven initialization provides a practical, versatile, and easy-to-implement strategy for improving SNN training performance on diverse tasks in neuromorphic engineering and computational neuroscience.

Audio Classification

Brain-Inspired Learning on Neuromorphic Substrates

no code implementations22 Oct 2020 Friedemann Zenke, Emre O. Neftci

Neuromorphic hardware strives to emulate brain-like neural networks and thus holds the promise for scalable, low-power information processing on temporal data streams.

Finding trainable sparse networks through Neural Tangent Transfer

1 code implementation ICML 2020 Tianlin Liu, Friedemann Zenke

Deep neural networks have dramatically transformed machine learning, but their memory and energy demands are substantial.

General Classification

Surrogate gradients for analog neuromorphic computing

no code implementations12 Jun 2020 Benjamin Cramer, Sebastian Billaudelle, Simeon Kanya, Aron Leibfried, Andreas Grübl, Vitali Karasenko, Christian Pehle, Korbinian Schreiber, Yannik Stradmann, Johannes Weis, Johannes Schemmel, Friedemann Zenke

To rapidly process temporal information at a low metabolic cost, biological neurons integrate inputs as an analog sum but communicate with spikes, binary events in time.

Surrogate Gradient Learning in Spiking Neural Networks

4 code implementations28 Jan 2019 Emre O. Neftci, Hesham Mostafa, Friedemann Zenke

Spiking neural networks are nature's versatile solution to fault-tolerant and energy efficient signal processing.

SuperSpike: Supervised learning in multi-layer spiking neural networks

1 code implementation31 May 2017 Friedemann Zenke, Surya Ganguli

In summary, our results open the door to obtaining a better scientific understanding of learning and computation in spiking neural networks by advancing our ability to train them to solve nonlinear problems involving transformations between different spatiotemporal spike-time patterns.

Continual Learning Through Synaptic Intelligence

5 code implementations ICML 2017 Friedemann Zenke, Ben Poole, Surya Ganguli

While deep learning has led to remarkable advances across diverse applications, it struggles in domains where the data distribution changes over the course of learning.

Computational Efficiency Continual Learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.