Search Results for author: Akira Nukada

Found 1 papers, 1 papers with code

Batched Sparse Matrix Multiplication for Accelerating Graph Convolutional Networks

1 code implementation27 Mar 2019 Yusuke Nagasaka, Akira Nukada, Ryosuke Kojima, Satoshi Matsuoka

We evaluated the performance of the GCNs application on TSUBAME3. 0 implementing NVIDIA Tesla P100 GPU, and our batched approach shows significant speedups of up to 1. 59x and 1. 37x in training and inference, respectively.

Distributed, Parallel, and Cluster Computing

Cannot find the paper you are looking for? You can Submit a new open access paper.