1 code implementation • 14 Sep 2023 • Gašper Beguš, Thomas Lu, Alan Zhou, Peter Wu, Gopala K. Anumanchipalli
This paper introduces CiwaGAN, a model of human spoken language acquisition that combines unsupervised articulatory modeling with an unsupervised model of information exchange through the auditory modality.
1 code implementation • 27 Oct 2022 • Gašper Beguš, Alan Zhou, Peter Wu, Gopala K Anumanchipalli
Articulatory analysis suggests that the network learns to control articulators in a similar manner to humans during speech production.
no code implementations • 22 Mar 2022 • Gašper Beguš, Alan Zhou
Here, we test how encoding and decoding of lexical semantic information can emerge automatically from raw speech in unsupervised generative deep convolutional networks that combine the production and perception principles of speech.
no code implementations • 5 Oct 2021 • Gašper Beguš, Alan Zhou
We propose a technique to visualize individual convolutional layers in the classifier that yields highly informative time-series data for each convolutional layer and apply it to unobserved test data.
no code implementations • 19 Apr 2021 • Gašper Beguš, Alan Zhou
This technique allows for acoustic analysis of intermediate layers that parallels the acoustic analysis of human speech data: we can extract F0, intensity, duration, formants, and other acoustic properties from intermediate layers in order to test where and how CNNs encode various types of information.