Search Results for author: Blesson Varghese

Found 16 papers, 7 papers with code

Rapid Deployment of DNNs for Edge Computing via Structured Pruning at Initialization

no code implementations22 Apr 2024 Bailey J. Eccles, Leon Wong, Blesson Varghese

We develop Reconvene, a system for rapidly generating pruned models suited for edge deployments using structured PaI.

Edge-computing Model Compression +1

DRIVE: Dual Gradient-Based Rapid Iterative Pruning

no code implementations1 Apr 2024 Dhananjay Saikumar, Blesson Varghese

Pruning methods, such as iterative magnitude-based pruning (IMP) achieve up to a 90% parameter reduction while retaining accuracy comparable to the original model.

NeuroFlux: Memory-Efficient CNN Training Using Adaptive Local Learning

no code implementations21 Feb 2024 Dhananjay Saikumar, Blesson Varghese

NeuroFlux segments a CNN into blocks based on GPU memory usage and further attaches an auxiliary network to each layer in these blocks.

DNNShifter: An Efficient DNN Pruning System for Edge Computing

1 code implementation13 Sep 2023 Bailey J. Eccles, Philip Rodgers, Peter Kilpatrick, Ivor Spence, Blesson Varghese

Compared to sparse models, the pruned model variants are up to 5. 14x smaller and have a 1. 67x inference latency speedup, with no compromise to sparse model accuracy.

Edge-computing

ROMA: Run-Time Object Detection To Maximize Real-Time Accuracy

no code implementations28 Oct 2022 JunKyu Lee, Blesson Varghese, Hans Vandierendonck

This paper analyzes the effects of dynamically varying video contents and detection latency on the real-time detection accuracy of a detector and proposes a new run-time accuracy variation model, ROMA, based on the findings from the analysis.

object-detection Real-Time Object Detection

CONTINUER: Maintaining Distributed DNN Services During Edge Failures

no code implementations25 Apr 2022 Ayesha Abdul Majeed, Peter Kilpatrick, Ivor Spence, Blesson Varghese

This paper will leverage trade-offs in accuracy, end-to-end latency and downtime for selecting the best technique given user-defined objectives (accuracy, latency and downtime thresholds) when an edge node fails.

Roadmap for Edge AI: A Dagstuhl Perspective

no code implementations27 Nov 2021 Aaron Yi Ding, Ella Peltonen, Tobias Meuser, Atakan Aral, Christian Becker, Schahram Dustdar, Thomas Hiessl, Dieter Kranzlmuller, Madhusanka Liyanage, Setareh Magshudi, Nitinder Mohan, Joerg Ott, Jan S. Rellermeyer, Stefan Schulte, Henning Schulzrinne, Gurkan Solmaz, Sasu Tarkoma, Blesson Varghese, Lars Wolf

Based on the collective input of Dagstuhl Seminar (21342), this paper presents a comprehensive discussion on AI methods and capabilities in the context of edge computing, referred as Edge AI.

Edge-computing

FedFly: Towards Migration in Edge-based Distributed Federated Learning

1 code implementation2 Nov 2021 Rehmat Ullah, Di wu, Paul Harvey, Peter Kilpatrick, Ivor Spence, Blesson Varghese

Our empirical results on the CIFAR10 dataset, with both balanced and imbalanced data distribution, support our claims that FedFly can reduce training time by up to 33% when a device moves after 50% of the training is completed, and by up to 45% when 90% of the training is completed when compared to state-of-the-art offloading approach in FL.

Federated Learning Privacy Preserving

FedAdapt: Adaptive Offloading for IoT Devices in Federated Learning

1 code implementation9 Jul 2021 Di wu, Rehmat Ullah, Paul Harvey, Peter Kilpatrick, Ivor Spence, Blesson Varghese

Further, FedAdapt adopts reinforcement learning based optimization and clustering to adaptively identify which layers of the DNN should be offloaded for each individual device on to a server to tackle the challenges of computational heterogeneity and changing network bandwidth.

Federated Learning

ScissionLite: Accelerating Distributed Deep Neural Networks Using Transfer Layer

1 code implementation5 May 2021 Hyunho Ahn, Munkyu Lee, Cheol-Ho Hong, Blesson Varghese

For example, applications underpinned by deep neural networks (DNN) models can be sliced and distributed across the IIoT device and the edge of the network for improving the overall performance of inference and for enhancing privacy of the input data, such as industrial product images.

Edge-computing

AVEC: Accelerator Virtualization in Cloud-Edge Computing for Deep Learning Libraries

no code implementations8 Mar 2021 Jason Kennedy, Blesson Varghese, Carlos Reaño

This is achieved by offloading computationally intensive workloads, such as deep learning from user devices to the edge.

Edge-computing Distributed, Parallel, and Cluster Computing Networking and Internet Architecture

Scission: Performance-driven and Context-aware Cloud-Edge Distribution of Deep Neural Networks

2 code implementations8 Aug 2020 Luke Lockhart, Paul Harvey, Pierre Imai, Peter Willis, Blesson Varghese

This is because the combination of potential target hardware resources that maximizes performance and the sequence of layers of the DNN that should be distributed across the target resources needs to be determined, while accounting for user-defined objectives/constraints for partitioning.

Benchmarking Decision Making

A Case For Adaptive Deep Neural Networks in Edge Computing

1 code implementation4 Aug 2020 Francis McNamee, Schahram Dustadar, Peter Kilpatrick, Weisong Shi, Ivor Spence, Blesson Varghese

However, there is limited understanding of: (a) whether and how evolving operational conditions (increased CPU and memory utilization at the edge or reduced data transfer rates between the edge and the cloud) affect the performance of already deployed DNNs, and (b) whether a new partition configuration is required to maximize performance.

Edge-computing

DYVERSE: DYnamic VERtical Scaling in Multi-tenant Edge Environments

1 code implementation19 Sep 2018 Nan Wang, Michail Matthaiou, Dimitrios S. Nikolopoulos, Blesson Varghese

When compared to executing applications on the Edge servers without dynamic vertical scaling, static priorities and dynamic priorities reduce SLO violation rates of requests by up to 4% and 12% for the online game, respectively, and in both cases 6% for the face detection workload.

Distributed, Parallel, and Cluster Computing Systems and Control

Challenges and Opportunities in Edge Computing

no code implementations7 Sep 2016 Blesson Varghese, Nan Wang, Sakil Barbhuiya, Peter Kilpatrick, Dimitrios S. Nikolopoulos

Many cloud-based applications employ a data centre as a central server to process data that is generated by edge devices, such as smartphones, tablets and wearables.

Distributed, Parallel, and Cluster Computing

Cannot find the paper you are looking for? You can Submit a new open access paper.