Search Results for author: David Gregg

Found 18 papers, 0 papers with code

Domino Saliency Metrics: Improving Existing Channel Saliency Metrics with Structural Information

no code implementations4 May 2022 Kaveena Persand, Andrew Anderson, David Gregg

The removal of these weight slices from a single layer causes mismatching number of feature maps between layers of the network.

Winograd Convolution for Deep Neural Networks: Efficient Point Selection

no code implementations25 Jan 2022 Syed Asad Alam, Andrew Anderson, Barbara Barabasz, David Gregg

The choice of points impacts the numeric accuracy of the algorithm, but the optimal set of points for small convolutions remains unknown.

Image Segmentation Object Recognition +1

Low precision logarithmic number systems: Beyond base-2

no code implementations12 Feb 2021 Syed Asad Alam, James Garland, David Gregg

Second, we show that low-precision LNS addition and subtraction can be implemented efficiently in logic rather than commonly used ROM lookup tables, the complexity of which can be reduced by an appropriate choice of base.

Numerical Analysis Numerical Analysis Signal Processing 65G50 C.m; G.0

HOBFLOPS CNNs: Hardware Optimized Bitslice-Parallel Floating-Point Operations for Convolutional Neural Networks

no code implementations11 Jul 2020 James Garland, David Gregg

HOBFLOPS allows researchers to prototype different levels of custom FP precision in the arithmetic of software CNN accelerators.

Exploiting Weight Redundancy in CNNs: Beyond Pruning and Quantization

no code implementations22 Jun 2020 Yuan Wen, David Gregg

Pruning and quantization are proven methods for improving the performance and storage efficiency of convolutional neural networks (CNNs).

Quantization

TASO: Time and Space Optimization for Memory-Constrained DNN Inference

no code implementations21 May 2020 Yuan Wen, Andrew Anderson, Valentin Radu, Michael F. P. O'Boyle, David Gregg

We optimize the trade-off between execution time and memory consumption by: 1) attempting to minimize execution time across the whole network by selecting data layouts and primitive operations to implement each layer; and 2) allocating an appropriate workspace that reflects the upper bound of memory footprint per layer.

Composition of Saliency Metrics for Channel Pruning with a Myopic Oracle

no code implementations3 Apr 2020 Kaveena Persand, Andrew Anderson, David Gregg

In most cases our method finds better selections than even the best individual pruning saliency.

Network Pruning

Performance-Oriented Neural Architecture Search

no code implementations9 Jan 2020 Andrew Anderson, Jing Su, Rozenn Dahyot, David Gregg

Hardware-Software Co-Design is a highly successful strategy for improving performance of domain-specific computing systems.

Edge-computing Keyword Spotting +1

Taxonomy of Saliency Metrics for Channel Pruning

no code implementations11 Jun 2019 Kaveena Persand, Andrew Anderson, David Gregg

The result is that it is difficult to separate the effectiveness of the saliency metric from the wider pruning algorithm that surrounds it.

General Classification

Winograd Convolution for DNNs: Beyond linear polynomials

no code implementations13 May 2019 Barbara Barabasz, David Gregg

Winograd convolution is widely used in deep neural networks (DNNs).

Scalar Arithmetic Multiple Data: Customizable Precision for Deep Neural Networks

no code implementations27 Sep 2018 Andrew Anderson, David Gregg

Our approach yields very fast implementations of bit-precise custom DNN operations, which often match or exceed the performance of operations quantized to the sizes supported in native arithmetic.

Quantization

Optimal DNN Primitive Selection with Partitioned Boolean Quadratic Programming

no code implementations3 Oct 2017 Andrew Anderson, David Gregg

We show experimentally that significant gains are possible versus the state of the art vendor libraries by using a principled analytic solution to the problem of layout selection in the presence of data format transformations.

Low-memory GEMM-based convolution algorithms for deep neural networks

no code implementations8 Sep 2017 Andrew Anderson, Aravind Vasudevan, Cormac Keane, David Gregg

We present two novel GEMM-based algorithms that require just $O(MHW)$ and $O(KW)$ additional space respectively, where $M$ is the number of channels in the result of the convolution.

Parallel Multi Channel Convolution using General Matrix Multiplication

no code implementations6 Apr 2017 Aravind Vasudevan, Andrew Anderson, David Gregg

A common approach to implementing convolutional layers is to expand the image into a column matrix (im2col) and perform Multiple Channel Multiple Kernel (MCMK) convolution using an existing parallel General Matrix Multiplication (GEMM) library.

Spectral Convolution Networks

no code implementations16 Nov 2016 Maria Francesca, Arthur Hughes, David Gregg

Previous research has shown that computation of convolution in the frequency domain provides a significant speedup versus traditional convolution network implementations.

Low Complexity Multiply Accumulate Unit for Weight-Sharing Convolutional Neural Networks

no code implementations30 Aug 2016 James Garland, David Gregg

Convolutional Neural Networks (CNNs) are one of the most successful deep machine learning technologies for processing image, voice and video data.

Cannot find the paper you are looking for? You can Submit a new open access paper.