no code implementations • 28 Apr 2023 • Bruno Mlodozeniec, Matthias Reisser, Christos Louizos
Well-tuned hyperparameters are crucial for obtaining good generalization behavior in neural networks.
no code implementations • 22 Jun 2022 • Kartik Gupta, Marios Fournarakis, Matthias Reisser, Christos Louizos, Markus Nagel
We perform extensive experiments on standard FL benchmarks to evaluate our proposed FedAvg variants for quantization robustness and provide a convergence analysis for our Quantization-Aware variants in FL.
no code implementations • 19 Nov 2021 • Christos Louizos, Matthias Reisser, Joseph Soriaga, Max Welling
Federated learning describes the distributed training of models across multiple clients while keeping the data private on-device.
no code implementations • 9 Nov 2021 • Aleksei Triastcyn, Matthias Reisser, Christos Louizos
Privacy and communication efficiency are important challenges in federated training of neural networks, and combining them is still an open problem.
no code implementations • 14 Jul 2021 • Matthias Reisser, Christos Louizos, Efstratios Gavves, Max Welling
Federated learning (FL) has emerged as the predominant approach for collaborative training of neural network models across multiple users, without the need to gather the data at a central location.
no code implementations • 1 Jan 2021 • Christos Louizos, Matthias Reisser, Joseph Soriaga, Max Welling
Federated averaging (FedAvg), despite its simplicity, has been the main approach in training neural networks in the federated learning setting.
1 code implementation • ICLR 2019 • Christos Louizos, Matthias Reisser, Tijmen Blankevoort, Efstratios Gavves, Max Welling
Neural network quantization has become an important research area due to its great impact on deployment of large models on resource constrained devices.