Search Results for author: Andrew Anderson

Found 14 papers, 0 papers with code

Domino Saliency Metrics: Improving Existing Channel Saliency Metrics with Structural Information

no code implementations4 May 2022 Kaveena Persand, Andrew Anderson, David Gregg

The removal of these weight slices from a single layer causes mismatching number of feature maps between layers of the network.

Winograd Convolution for Deep Neural Networks: Efficient Point Selection

no code implementations25 Jan 2022 Syed Asad Alam, Andrew Anderson, Barbara Barabasz, David Gregg

The choice of points impacts the numeric accuracy of the algorithm, but the optimal set of points for small convolutions remains unknown.

Image Segmentation Object Recognition +1

TASO: Time and Space Optimization for Memory-Constrained DNN Inference

no code implementations21 May 2020 Yuan Wen, Andrew Anderson, Valentin Radu, Michael F. P. O'Boyle, David Gregg

We optimize the trade-off between execution time and memory consumption by: 1) attempting to minimize execution time across the whole network by selecting data layouts and primitive operations to implement each layer; and 2) allocating an appropriate workspace that reflects the upper bound of memory footprint per layer.

Composition of Saliency Metrics for Channel Pruning with a Myopic Oracle

no code implementations3 Apr 2020 Kaveena Persand, Andrew Anderson, David Gregg

In most cases our method finds better selections than even the best individual pruning saliency.

Network Pruning

Performance-Oriented Neural Architecture Search

no code implementations9 Jan 2020 Andrew Anderson, Jing Su, Rozenn Dahyot, David Gregg

Hardware-Software Co-Design is a highly successful strategy for improving performance of domain-specific computing systems.

Edge-computing Keyword Spotting +1

Taxonomy of Saliency Metrics for Channel Pruning

no code implementations11 Jun 2019 Kaveena Persand, Andrew Anderson, David Gregg

The result is that it is difficult to separate the effectiveness of the saliency metric from the wider pruning algorithm that surrounds it.

General Classification

Scalar Arithmetic Multiple Data: Customizable Precision for Deep Neural Networks

no code implementations27 Sep 2018 Andrew Anderson, David Gregg

Our approach yields very fast implementations of bit-precise custom DNN operations, which often match or exceed the performance of operations quantized to the sizes supported in native arithmetic.

Quantization

Optimal DNN Primitive Selection with Partitioned Boolean Quadratic Programming

no code implementations3 Oct 2017 Andrew Anderson, David Gregg

We show experimentally that significant gains are possible versus the state of the art vendor libraries by using a principled analytic solution to the problem of layout selection in the presence of data format transformations.

Low-memory GEMM-based convolution algorithms for deep neural networks

no code implementations8 Sep 2017 Andrew Anderson, Aravind Vasudevan, Cormac Keane, David Gregg

We present two novel GEMM-based algorithms that require just $O(MHW)$ and $O(KW)$ additional space respectively, where $M$ is the number of channels in the result of the convolution.

Parallel Multi Channel Convolution using General Matrix Multiplication

no code implementations6 Apr 2017 Aravind Vasudevan, Andrew Anderson, David Gregg

A common approach to implementing convolutional layers is to expand the image into a column matrix (im2col) and perform Multiple Channel Multiple Kernel (MCMK) convolution using an existing parallel General Matrix Multiplication (GEMM) library.

Cannot find the paper you are looking for? You can Submit a new open access paper.