Search Results for author: Sunil Kumar Kopparapu

Found 33 papers, 1 papers with code

Impact of Microphone position Measurement Error on Multi Channel Distant Speech Recognition & Intelligibility

no code implementations ICON 2021 Karan Nathwani, Sunil Kumar Kopparapu

It was shown in (Raikar et al., 2020) that the measurement error in the microphone position affected the room impulse response (RIR) which in turn affected the single channel speech recognition.

Distant Speech Recognition Position +2

A Novel Scheme to classify Read and Spontaneous Speech

no code implementations13 Jun 2023 Sunil Kumar Kopparapu

The COVID-19 pandemic has led to an increased use of remote telephonic interviews, making it important to distinguish between scripted and spontaneous speech in audio recordings.

Selective Data Augmentation for Robust Speech Translation

no code implementations22 Mar 2023 Rajul Acharya, Ashish Panda, Sunil Kumar Kopparapu

Though resource intensive, e2e-ST systems have the inherent ability to retain para and non-linguistic characteristics of the speech unlike cascade systems.

Data Augmentation Machine Translation +1

Text-to-Audio Grounding Based Novel Metric for Evaluating Audio Caption Similarity

no code implementations3 Oct 2022 Swapnil Bhosale, Rupayan Chakraborty, Sunil Kumar Kopparapu

Automatic Audio Captioning (AAC) refers to the task of translating an audio sample into a natural language (NL) text that describes the audio events, source of the events and their relationships.

Audio captioning Image Captioning +2

Computing Optimal Location of Microphone for Improved Speech Recognition

no code implementations24 Mar 2022 Karan Nathwani, Bhavya Dixit, Sunil Kumar Kopparapu

It was shown in our earlier work that the measurement error in the microphone position affected the room impulse response (RIR) which in turn affected the single-channel close microphone and multi-channel distant microphone speech recognition.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +3

Spectro Temporal EEG Biomarkers For Binary Emotion Classification

no code implementations2 Feb 2022 Upasana Tiwari, Rupayan Chakraborty, Sunil Kumar Kopparapu

The usefulness of these features for EEG emotion classification is investigated through extensive experiments using state-of-the-art classifiers.

Classification EEG +1

Automatic Audio Captioning using Attention weighted Event based Embeddings

no code implementations28 Jan 2022 Swapnil Bhosale, Rupayan Chakraborty, Sunil Kumar Kopparapu

Automatic Audio Captioning (AAC) refers to the task of translating audio into a natural language that describes the audio events, source of the events and their relationships.

Audio captioning Event Detection +1

Synthetic speech detection using meta-learning with prototypical loss

no code implementations24 Jan 2022 Monisankha Pal, Aditya Raikar, Ashish Panda, Sunil Kumar Kopparapu

Furthermore, the proposed system with data augmentation outperforms the ASVspoof 2021 challenge best baseline both in the progress and evaluation phase of the LA task.

Data Augmentation Meta-Learning +2

Automatic Speaker Independent Dysarthric Speech Intelligibility Assessment System

no code implementations10 Mar 2021 Ayush Tripathi, Swapnil Bhosale, Sunil Kumar Kopparapu

Dysarthria is a condition which hampers the ability of an individual to control the muscles that play a major role in speech delivery.

Semi Supervised Learning For Few-shot Audio Classification By Episodic Triplet Mining

no code implementations16 Feb 2021 Swapnil Bhosale, Rupayan Chakraborty, Sunil Kumar Kopparapu

In this paper, we propose to replace the typical prototypical loss function with an Episodic Triplet Mining (ETM) technique.

Audio Classification Event Detection +4

Identification of Dementia Using Audio Biomarkers

no code implementations27 Feb 2020 Rupayan Chakraborty, Meghna Pandharipande, Chitralekha Bhat, Sunil Kumar Kopparapu

The objective of this work is to use speech processing and machine learning techniques to automatically identify the stage of dementia such as mild cognitive impairment (MCI) or Alzheimers disease (AD).

A Cycle-GAN Approach to Model Natural Perturbations in Speech for ASR Applications

no code implementations18 Dec 2019 Sri Harsha Dumpala, Imran Sheikh, Rupayan Chakraborty, Sunil Kumar Kopparapu

Naturally introduced perturbations in audio signal, caused by emotional and physical states of the speaker, can significantly degrade the performance of Automatic Speech Recognition (ASR) systems.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

Sentiment Analysis using Imperfect Views from Spoken Language and Acoustic Modalities

no code implementations WS 2018 Imran Sheikh, Sri Harsha Dumpala, Rupayan Chakraborty, Sunil Kumar Kopparapu

Multimodal sentiment classification in practical applications may have to rely on erroneous and imperfect views, namely (a) language transcription from a speech recognizer and (b) under-performing acoustic views.

Automatic Speech Recognition (ASR) General Classification +2

A Novel Approach for Effective Learning in Low Resourced Scenarios

no code implementations15 Dec 2017 Sri Harsha Dumpala, Rupayan Chakraborty, Sunil Kumar Kopparapu

Deep learning based discriminative methods, being the state-of-the-art machine learning techniques, are ill-suited for learning from lower amounts of data.

BIG-bench Machine Learning Classification +2

Adapting general-purpose speech recognition engine output for domain-specific natural language question answering

no code implementations12 Oct 2017 C. Anantaram, Sunil Kumar Kopparapu

However, when such engines are used for specific domains, they may not recognize domain-specific words well, and may produce erroneous output.

Question Answering Sentence +2

k-FFNN: A priori knowledge infused Feed-forward Neural Networks

no code implementations24 Apr 2017 Sri Harsha Dumpala, Rupayan Chakraborty, Sunil Kumar Kopparapu

It is not immediately clear (a) how a priori temporal knowledge can be used in a FFNN architecture (b) how a FFNN performs when provided with this knowledge about temporal correlations (assuming available) during training.

Evaluating the Performance of a Speech Recognition based System

no code implementations11 Jan 2016 Vinod Kumar Pandey, Sunil Kumar Kopparapu

The usual mechanism to evaluate the performance of a speech solution is to do an extensive test of the system by putting it to actual people use and then evaluating the performance by analyzing the logs for successful transactions.

speech-recognition Speech Recognition

On-line Handwritten Devanagari Character Recognition using Fuzzy Directional Features

no code implementations7 Apr 2015 Sunil Kumar Kopparapu, Lajish VL

This paper describes a new feature set for use in the recognition of on-line handwritten Devanagari script based on Fuzzy Directional Features.

Voice based self help System: User Experience Vs Accuracy

no code implementations7 Apr 2015 Sunil Kumar Kopparapu

A speech based self help system ideally needs a speech recognition engine to convert spoken speech to text and in addition a language processing engine to take care of any misrecognitions by the speech recognition engine.

speech-recognition Speech Recognition

A Rule-Based Short Query Intent Identification System

no code implementations25 Mar 2015 Arijit De, Sunil Kumar Kopparapu

Using SMS (Short Message System), cell phones can be used to query for information about various topics.

Online Handwritten Devanagari Stroke Recognition Using Extended Directional Features

no code implementations11 Jan 2015 Lajish VL, Sunil Kumar Kopparapu

Experimental results show that the extended directional feature set performs well with about 65+% stroke level recognition accuracy for writer independent data set.

Modified Mel Filter Bank to Compute MFCC of Subsampled Speech

no code implementations25 Oct 2014 Kiran Kumar Bhuvanagiri, Sunil Kumar Kopparapu

Mel Frequency Cepstral Coefficients (MFCCs) are the most popularly used speech features in most speech and speaker recognition applications.

Speaker Recognition

A Framework for On-Line Devanagari Handwritten Character Recognition

no code implementations25 Oct 2014 Sunil Kumar Kopparapu, Lajish V. L

The framework is based on identify- ing strokes, which in turn lead to recognition of handwritten on-line characters rather that the conventional character identification.

Choice of Mel Filter Bank in Computing MFCC of a Resampled Speech

no code implementations25 Oct 2014 Laxmi Narayana M., Sunil Kumar Kopparapu

Mel Frequency Cepstral Coefficients (MFCCs) are the most popularly used speech features in most speech and speaker recognition applications.

Speaker Recognition

Music and Vocal Separation Using Multi-Band Modulation Based Features

no code implementations10 Jun 2014 Sunil Kumar Kopparapu, Meghna Pandharipande, G Sita

We first identify the distribution of these non-linear features for music only and voice only segments in the audio signal in different Mel spaced frequency bands and show that they have the ability to discriminate.

Basis Identification for Automatic Creation of Pronunciation Lexicon for Proper Names

no code implementations5 Jun 2014 Sunil Kumar Kopparapu, M Laxminarayana

The idea is to construct a small orthogonal set of words (basis) which can span the set of names in a given database.

Automatic Segmentation of Broadcast News Audio using Self Similarity Matrix

no code implementations27 Mar 2014 Sapna Soni, Ahmed Imran, Sunil Kumar Kopparapu

Generally audio news broadcast on radio is com- posed of music, commercials, news from correspondents and recorded statements in addition to the actual news read by the newsreader.

Segmentation

Cannot find the paper you are looking for? You can Submit a new open access paper.