no code implementations • 30 Sep 2020 • Yoonho Boo, Sungho Shin, Jungwook Choi, Wonyong Sung
In this study, we propose stochastic precision ensemble training for QDNNs (SPEQ).
no code implementations • 31 May 2020 • Yoonho Boo, Sungho Shin, Wonyong Sung
This study proposes a holistic approach for the optimization of QDNNs, which contains QDNN training methods as well as quantization-friendly architecture design.
no code implementations • 2 Feb 2020 • Sungho Shin, Yoonho Boo, Wonyong Sung
Model averaging is a promising approach for achieving the good generalization capability of DNNs, especially when the loss surface for training contains many sharp minima.
no code implementations • 4 Sep 2019 • Sungho Shin, Yoonho Boo, Wonyong Sung
Knowledge distillation (KD) is a very popular method for model size reduction.
no code implementations • NeurIPS 2018 • Jinhwan Park, Yoonho Boo, Iksoo Choi, Sungho Shin, Wonyong Sung
The RNN implementation on embedded devices can suffer from excessive DRAM accesses because the parameter size of a neural network usually exceeds that of the cache memory and the parameters are used only once for each time step.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
no code implementations • NeurIPS 2017 • Kyuhong Shim, Minjae Lee, Iksoo Choi, Yoonho Boo, Wonyong Sung
The approximate probability of each word can be estimated with only a small part of the weight matrix by using a few large singular values and the corresponding elements for most of the words.
no code implementations • 1 Jul 2017 • Yoonho Boo, Wonyong Sung
Deep neural networks (DNNs) usually demand a large amount of operations for real-time inference.
no code implementations • 27 Feb 2017 • Sungho Shin, Yoonho Boo, Wonyong Sung
Fixed-point optimization of deep neural networks plays an important role in hardware based design and low-power implementations.