no code implementations • ICON 2021 • Karan Nathwani, Sunil Kumar Kopparapu
It was shown in (Raikar et al., 2020) that the measurement error in the microphone position affected the room impulse response (RIR) which in turn affected the single channel speech recognition.
no code implementations • 13 Jun 2023 • Sunil Kumar Kopparapu
The COVID-19 pandemic has led to an increased use of remote telephonic interviews, making it important to distinguish between scripted and spontaneous speech in audio recordings.
no code implementations • 22 Mar 2023 • Rajul Acharya, Ashish Panda, Sunil Kumar Kopparapu
Though resource intensive, e2e-ST systems have the inherent ability to retain para and non-linguistic characteristics of the speech unlike cascade systems.
no code implementations • 3 Oct 2022 • Swapnil Bhosale, Rupayan Chakraborty, Sunil Kumar Kopparapu
Automatic Audio Captioning (AAC) refers to the task of translating an audio sample into a natural language (NL) text that describes the audio events, source of the events and their relationships.
no code implementations • 24 Mar 2022 • Karan Nathwani, Bhavya Dixit, Sunil Kumar Kopparapu
It was shown in our earlier work that the measurement error in the microphone position affected the room impulse response (RIR) which in turn affected the single-channel close microphone and multi-channel distant microphone speech recognition.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3
no code implementations • 2 Feb 2022 • Upasana Tiwari, Rupayan Chakraborty, Sunil Kumar Kopparapu
The usefulness of these features for EEG emotion classification is investigated through extensive experiments using state-of-the-art classifiers.
no code implementations • 28 Jan 2022 • Swapnil Bhosale, Rupayan Chakraborty, Sunil Kumar Kopparapu
Automatic Audio Captioning (AAC) refers to the task of translating audio into a natural language that describes the audio events, source of the events and their relationships.
no code implementations • 24 Jan 2022 • Monisankha Pal, Aditya Raikar, Ashish Panda, Sunil Kumar Kopparapu
Furthermore, the proposed system with data augmentation outperforms the ASVspoof 2021 challenge best baseline both in the progress and evaluation phase of the LA task.
1 code implementation • 24 Mar 2021 • Ayush Tripathi, Rupayan Chakraborty, Sunil Kumar Kopparapu
In this paper, we propose a novel three step technique to address imbalanced data.
no code implementations • 10 Mar 2021 • Ayush Tripathi, Swapnil Bhosale, Sunil Kumar Kopparapu
Dysarthria is a condition which hampers the ability of an individual to control the muscles that play a major role in speech delivery.
no code implementations • 16 Feb 2021 • Swapnil Bhosale, Rupayan Chakraborty, Sunil Kumar Kopparapu
In this paper, we propose to replace the typical prototypical loss function with an Episodic Triplet Mining (ETM) technique.
no code implementations • 27 Feb 2020 • Rupayan Chakraborty, Meghna Pandharipande, Chitralekha Bhat, Sunil Kumar Kopparapu
The objective of this work is to use speech processing and machine learning techniques to automatically identify the stage of dementia such as mild cognitive impairment (MCI) or Alzheimers disease (AD).
no code implementations • 18 Dec 2019 • Sri Harsha Dumpala, Imran Sheikh, Rupayan Chakraborty, Sunil Kumar Kopparapu
Naturally introduced perturbations in audio signal, caused by emotional and physical states of the speaker, can significantly degrade the performance of Automatic Speech Recognition (ASR) systems.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 5 Nov 2019 • Bansi Shah, Sunil Kumar Kopparapu
Named Entity Recognition is one of the most important text processing requirement in many NLP tasks.
no code implementations • WS 2018 • Imran Sheikh, Sri Harsha Dumpala, Rupayan Chakraborty, Sunil Kumar Kopparapu
Multimodal sentiment classification in practical applications may have to rely on erroneous and imperfect views, namely (a) language transcription from a speech recognizer and (b) under-performing acoustic views.
Automatic Speech Recognition (ASR) General Classification +2
no code implementations • 15 Dec 2017 • Sri Harsha Dumpala, Rupayan Chakraborty, Sunil Kumar Kopparapu
Deep learning based discriminative methods, being the state-of-the-art machine learning techniques, are ill-suited for learning from lower amounts of data.
no code implementations • 12 Oct 2017 • C. Anantaram, Sunil Kumar Kopparapu
However, when such engines are used for specific domains, they may not recognize domain-specific words well, and may produce erroneous output.
no code implementations • 24 Apr 2017 • Sri Harsha Dumpala, Rupayan Chakraborty, Sunil Kumar Kopparapu
It is not immediately clear (a) how a priori temporal knowledge can be used in a FFNN architecture (b) how a FFNN performs when provided with this knowledge about temporal correlations (assuming available) during training.
no code implementations • 11 Jan 2016 • Vinod Kumar Pandey, Sunil Kumar Kopparapu
The usual mechanism to evaluate the performance of a speech solution is to do an extensive test of the system by putting it to actual people use and then evaluating the performance by analyzing the logs for successful transactions.
no code implementations • 7 Apr 2015 • Lajish V. L., Sunil Kumar Kopparapu
Identity of a vehicle is done through the vehicle license plate by traffic police in general.
no code implementations • 7 Apr 2015 • Sunil Kumar Kopparapu, Lajish VL
This paper describes a new feature set for use in the recognition of on-line handwritten Devanagari script based on Fuzzy Directional Features.
no code implementations • 7 Apr 2015 • Sunil Kumar Kopparapu
A speech based self help system ideally needs a speech recognition engine to convert spoken speech to text and in addition a language processing engine to take care of any misrecognitions by the speech recognition engine.
no code implementations • 25 Mar 2015 • Arijit De, Sunil Kumar Kopparapu
Using SMS (Short Message System), cell phones can be used to query for information about various topics.
no code implementations • 11 Jan 2015 • Lajish VL, Sunil Kumar Kopparapu
Experimental results show that the extended directional feature set performs well with about 65+% stroke level recognition accuracy for writer independent data set.
no code implementations • 25 Oct 2014 • Kiran Kumar Bhuvanagiri, Sunil Kumar Kopparapu
Mel Frequency Cepstral Coefficients (MFCCs) are the most popularly used speech features in most speech and speaker recognition applications.
no code implementations • 25 Oct 2014 • Sunil Kumar Kopparapu, Lajish V. L
The framework is based on identify- ing strokes, which in turn lead to recognition of handwritten on-line characters rather that the conventional character identification.
no code implementations • 25 Oct 2014 • Laxmi Narayana M., Sunil Kumar Kopparapu
Mel Frequency Cepstral Coefficients (MFCCs) are the most popularly used speech features in most speech and speaker recognition applications.
no code implementations • 10 Jun 2014 • Sunil Kumar Kopparapu, Meghna Pandharipande, G Sita
We first identify the distribution of these non-linear features for music only and voice only segments in the audio signal in different Mel spaced frequency bands and show that they have the ability to discriminate.
no code implementations • 5 Jun 2014 • Sunil Kumar Kopparapu, M Laxminarayana
The idea is to construct a small orthogonal set of words (basis) which can span the set of names in a given database.
no code implementations • 27 Mar 2014 • Sapna Soni, Ahmed Imran, Sunil Kumar Kopparapu
Generally audio news broadcast on radio is com- posed of music, commercials, news from correspondents and recorded statements in addition to the actual news read by the newsreader.