no code implementations • 1 Dec 2023 • Abhayjeet Singh, Charu Shah, Rajashri Varadaraj, Sonakshi Chauhan, Prasanta Kumar Ghosh
Transcripts for 23 hours is generated and validated which can serve as a spontaneous speech ASR benchmark.
no code implementations • 13 Oct 2023 • Jesuraj Bandekar, Sathvik Udupa, Abhayjeet Singh, Anjali Jayakumar, Deekshitha G, Sandhya Badiger, Saurabh Kumar, Pooja VH, Prasanta Kumar Ghosh
With the advent of high-quality speech synthesis, there is a lot of interest in controlling various prosodic attributes of speech.
no code implementations • 16 Jul 2023 • Abhayjeet Singh, Arjun Singh Mehta, Ashish Khuraishi K S, Deekshitha G, Gauri Date, Jai Nanavati, Jesuraja Bandekar, Karnalius Basumatary, Karthika P, Sandhya Badiger, Sathvik Udupa, Saurabh Kumar, Savitha, Prasanta Kumar Ghosh, Prashanthi V, Priyanka Pai, Raoul Nanavati, Rohan Saxena, Sai Praneeth Reddy Mora, Srinivasa Raghavan
This is where a lot of adaptation and fine-tuning techniques can be applied to overcome the low-resource nature of the data by utilising well-resourced similar languages.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 8 Dec 2021 • Abhayjeet Singh, Achuth Rao MV, Rakesh Vaideeswaran, Chiranjeevi Yarra, Prasanta Kumar Ghosh
We observe that the sentence, speaker difficulty ratings and the WERs increase from easy to hard categories of sentences.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3
1 code implementation • 11 Apr 2021 • Sathvik Udupa, Anwesha Roy, Abhayjeet Singh, Aravind Illa, Prasanta Kumar Ghosh
Additionally, on the AAI task, we obtain 1. 5%, 3% and 3. 1% relative gain in CC on the same setups compared to the state-of-the-art baseline.
no code implementations • 4 Jun 2020 • Abhayjeet Singh, Aravind Illa, Prasanta Kumar Ghosh
As the range of articulatory motions is correlated with speaking rate, we also analyze amplitude of the transformed articulatory movements at different rates compared to their original counterparts, to examine how well the proposed AstNet predicts the extent of articulatory movements in N2F and N2S.
no code implementations • 31 Oct 2019 • Abhayjeet Singh, Aravind Illa, Prasanta Kumar Ghosh
While an attention network is used for estimating articulatory movement in the case of R2, BLSTM network is used for R1 and R3.