1 code implementation • 17 Feb 2022 • Kento Nagatomo, Masahiro Yasuda, Kohei Yatabe, Shoichiro Saito, Yasuhiro Oikawa
Sound event localization and detection (SELD) is a combined task of identifying the sound event and its direction.
no code implementations • 16 Feb 2022 • Tomoro Tanaka, Kohei Yatabe, Masahiro Yasuda, Yasuhiro Oikawa
Still, they cannot perform well if the training data have mismatches and/or constraints in the time domain are not imposed.
no code implementations • 7 May 2021 • Tsubasa Kusano, Kohei Yatabe, Yasuhiro Oikawa
In this paper, we propose a method of estimating a sparse T-F representation using atomic norm.
no code implementations • 28 Jul 2020 • Yoshiki Masuyama, Yoshiaki Bando, Kohei Yatabe, Yoko Sasaki, Masaki Onishi, Yasuhiro Oikawa
By incorporating with the spatial information in multichannel audio signals, our method trains deep neural networks (DNNs) to distinguish multiple sound source objects.
no code implementations • 14 Feb 2020 • Yoshiki Masuyama, Kohei Yatabe, Yuma Koizumi, Yasuhiro Oikawa, Noboru Harada
In the proposed method, DNNs estimate phase derivatives instead of phase itself, which allows us to avoid the sensitivity problem.
1 code implementation • 25 Nov 2019 • Daiki Takeuchi, Kohei Yatabe, Yuma Koizumi, Yasuhiro Oikawa, Noboru Harada
Therefore, some end-to-end methods used a DNN to learn the linear T-F transform which is much easier to understand.
Audio and Speech Processing Sound
no code implementations • 10 Mar 2019 • Yoshiki Masuyama, Kohei Yatabe, Yuma Koizumi, Yasuhiro Oikawa, Noboru Harada
This paper presents a novel phase reconstruction method (only from a given amplitude spectrogram) by combining a signal-processing-based approach and a deep neural network (DNN).