1 code implementation • 26 May 2016 • He Ma, Fei Mao, Graham W. Taylor
We develop a scalable and extendable training framework that can utilize GPUs across nodes in a cluster and accelerate the training of deep learning models based on data parallelism.
2 code implementations • 7 Dec 2014 • Weiguang Ding, Ruoyan Wang, Fei Mao, Graham Taylor
In this report, we describe a Theano-based AlexNet (Krizhevsky et al., 2012) implementation and its naive data parallelism on multiple GPUs.