Search Results for author: Valentin Radu

Found 8 papers, 1 papers with code

Closing the Gap between Client and Global Model Performance in Heterogeneous Federated Learning

no code implementations7 Nov 2022 Hongrui Shi, Valentin Radu, Po Yang

The heterogeneity of hardware and data is a well-known and studied problem in the community of Federated Learning (FL) as running under heterogeneous settings.

Federated Learning Knowledge Distillation

Data Selection for Efficient Model Update in Federated Learning

no code implementations5 Nov 2021 Hongrui Shi, Valentin Radu

Our experiments show that only 1. 6% of the initially exchanged data can effectively transfer the characteristic of the client data to the global model in our FL approach, using split networks.

Federated Learning Transfer Learning

Optimising the Performance of Convolutional Neural Networks across Computing Systems using Transfer Learning

no code implementations20 Oct 2020 Rik Mulder, Valentin Radu, Christophe Dubach

This process requires a lengthy profiling stage, iterating over all the available primitives for each layer configuration, to measure their execution time on the target platform.

Transfer Learning

TASO: Time and Space Optimization for Memory-Constrained DNN Inference

no code implementations21 May 2020 Yuan Wen, Andrew Anderson, Valentin Radu, Michael F. P. O'Boyle, David Gregg

We optimize the trade-off between execution time and memory consumption by: 1) attempting to minimize execution time across the whole network by selecting data layouts and primitive operations to implement each layer; and 2) allocating an appropriate workspace that reflects the upper bound of memory footprint per layer.

Performance Aware Convolutional Neural Network Channel Pruning for Embedded GPUs

no code implementations20 Feb 2020 Valentin Radu, Kuba Kaszyk, Yuan Wen, Jack Turner, Jose Cano, Elliot J. Crowley, Bjorn Franke, Amos Storkey, Michael O'Boyle

We evaluate higher level libraries, which analyze the input characteristics of a convolutional layer, based on which they produce optimized OpenCL (Arm Compute Library and TVM) and CUDA (cuDNN) code.

Model Compression Network Pruning

Distilling with Performance Enhanced Students

no code implementations24 Oct 2018 Jack Turner, Elliot J. Crowley, Valentin Radu, José Cano, Amos Storkey, Michael O'Boyle

The task of accelerating large neural networks on general purpose hardware has, in recent years, prompted the use of channel pruning to reduce network size.

Model Compression

Characterising Across-Stack Optimisations for Deep Convolutional Neural Networks

1 code implementation19 Sep 2018 Jack Turner, José Cano, Valentin Radu, Elliot J. Crowley, Michael O'Boyle, Amos Storkey

Convolutional Neural Networks (CNNs) are extremely computationally demanding, presenting a large barrier to their deployment on resource-constrained devices.

Neural Network Compression

Cannot find the paper you are looking for? You can Submit a new open access paper.