Search Results for author: Kenneth Church

Found 29 papers, 9 papers with code

Training on Lexical Resources

1 code implementation LREC 2022 Kenneth Church, Xingyu Cai, Yuchen Bian

We propose using lexical resources (thesaurus, VAD) to fine-tune pretrained deep nets such as BERT and ERNIE.

A Gentle Introduction to Deep Nets and Opportunities for the Future

no code implementations ACL 2022 Kenneth Church, Valia Kordoni, Gary Marcus, Ernest Davis, Yanjun Ma, Zeyu Chen

The first half of this tutorial will make deep nets more accessible to a broader audience, following “Deep Nets for Poets” and “A Gentle Introduction to Fine-Tuning.” We will also introduce GFT (general fine tuning), a little language for fine tuning deep nets with short (one line) programs that are as easy to code as regression in statistics packages such as R using glm (general linear models).

Improving Bilingual Lexicon Induction for Low Frequency Words

no code implementations EMNLP 2020 Jiaji Huang, Xingyu Cai, Kenneth Church

This paper designs a Monolingual Lexicon Induction task and observes that two factors accompany the degraded accuracy of bilingual lexicon induction for rare words.

Bilingual Lexicon Induction

Benchmarking: Past, Present and Future

1 code implementation ACL (BPPF) 2021 Kenneth Church, Mark Liberman, Valia Kordoni

There used to be more top-down leadership from government (and industry, in the case of systems, with benchmarks such as SPEC).

Benchmarking Reading Comprehension

Since the Scientific Literature Is Multilingual, Our Models Should Be Too

no code implementations27 Mar 2024 Abteen Ebrahimi, Kenneth Church

English has long been assumed the $\textit{lingua franca}$ of scientific research, and this notion is reflected in the natural language processing (NLP) research involving scientific document representation.

Position

Data-Driven Adaptive Simultaneous Machine Translation

no code implementations27 Apr 2022 Guangxu Xun, Mingbo Ma, Yuchen Bian, Xingyu Cai, Jiaji Huang, Renjie Zheng, Junkun Chen, Jiahong Yuan, Kenneth Church, Liang Huang

In simultaneous translation (SimulMT), the most widely used strategy is the wait-k policy thanks to its simplicity and effectiveness in balancing translation quality and latency.

Machine Translation Sentence +1

Efficiently Disentangle Causal Representations

1 code implementation6 Jan 2022 Yuanpeng Li, Joel Hestness, Mohamed Elhoseiny, Liang Zhao, Kenneth Church

This paper proposes an efficient approach to learning disentangled representations with causal mechanisms based on the difference of conditional probabilities in original and new distributions.

The Role of Phonetic Units in Speech Emotion Recognition

no code implementations2 Aug 2021 Jiahong Yuan, Xingyu Cai, Renjie Zheng, Liang Huang, Kenneth Church

Models of phonemes, broad phonetic classes, and syllables all significantly outperform the utterance model, demonstrating that phonetic units are helpful and should be incorporated in speech emotion recognition.

Speech Emotion Recognition speech-recognition +1

Automatic recognition of suprasegmentals in speech

no code implementations2 Aug 2021 Jiahong Yuan, Neville Ryant, Xingyu Cai, Kenneth Church, Mark Liberman

This study reports our efforts to improve automatic recognition of suprasegmentals by fine-tuning wav2vec 2. 0 with CTC, a method that has been successful in automatic speech recognition.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +1

On Attention Redundancy: A Comprehensive Study

no code implementations NAACL 2021 Yuchen Bian, Jiaji Huang, Xingyu Cai, Jiahong Yuan, Kenneth Church

(What) We define and focus the study on redundancy matrices generated from pre-trained and fine-tuned BERT-base model for GLUE datasets.

Model Compression Sentence

Better than BERT but Worse than Baseline

no code implementations12 May 2021 Boxiang Liu, Jiaji Huang, Xingyu Cai, Kenneth Church

This paper compares BERT-SQuAD and Ab3P on the Abbreviation Definition Identification (ADI) task.

Isotropy in the Contextual Embedding Space: Clusters and Manifolds

no code implementations ICLR 2021 Xingyu Cai, Jiaji Huang, Yuchen Bian, Kenneth Church

We hope the study in this paper could provide insights towards a better understanding of the deep language models.

Transferability of Compositionality

no code implementations1 Jan 2021 Yuanpeng Li, Liang Zhao, Joel Hestness, Ka Yee Lun, Kenneth Church, Mohamed Elhoseiny

To our best knowledge, this is the first work to focus on the transferability of compositionality, and it is orthogonal to existing efforts of learning compositional representations in training distribution.

Out-of-Distribution Generalization

Gradient Descent Resists Compositionality

no code implementations1 Jan 2021 Yuanpeng Li, Liang Zhao, Joel Hestness, Kenneth Church, Mohamed Elhoseiny

In this paper, we argue that gradient descent is one of the reasons that make compositionality learning hard during neural network optimization.

The Third DIHARD Diarization Challenge

3 code implementations2 Dec 2020 Neville Ryant, Prachi Singh, Venkat Krishnamohan, Rajat Varma, Kenneth Church, Christopher Cieri, Jun Du, Sriram Ganapathy, Mark Liberman

DIHARD III was the third in a series of speaker diarization challenges intended to improve the robustness of diarization systems to variability in recording equipment, noise conditions, and conversational domain.

speaker-diarization Speaker Diarization +1

Fluent and Low-latency Simultaneous Speech-to-Speech Translation with Self-adaptive Training

no code implementations Findings of the Association for Computational Linguistics 2020 Renjie Zheng, Mingbo Ma, Baigong Zheng, Kaibo Liu, Jiahong Yuan, Kenneth Church, Liang Huang

Simultaneous speech-to-speech translation is widely useful but extremely challenging, since it needs to generate target-language speech concurrently with the source-language speech, with only a few seconds delay.

Sentence Speech-to-Speech Translation +1

The Second DIHARD Diarization Challenge: Dataset, task, and baselines

1 code implementation18 Jun 2019 Neville Ryant, Kenneth Church, Christopher Cieri, Alejandrina Cristia, Jun Du, Sriram Ganapathy, Mark Liberman

This paper introduces the second DIHARD challenge, the second in a series of speaker diarization challenges intended to improve the robustness of diarization systems to variation in recording equipment, noise conditions, and conversational domain.

Action Detection Activity Detection +5

Language Modeling at Scale

no code implementations23 Oct 2018 Mostofa Patwary, Milind Chabbi, Heewoo Jun, Jiaji Huang, Gregory Diamos, Kenneth Church

We show how Zipf's Law can be used to scale up language modeling (LM) to take advantage of more training data and more GPUs.

Language Modelling Machine Translation +2

A Proposed Hierarchy of Deep Learning Tasks

no code implementations27 Sep 2018 Joel Hestness, Sharan Narang, Newsha Ardalani, Heewoo Jun, Hassan Kianinejad, Md. Mostofa Ali Patwary, Yang Yang, Yanqi Zhou, Gregory Diamos, Kenneth Church

As the pace of deep learning innovation accelerates, it becomes increasingly important to organize the space of problems by relative difficultly.

Cannot find the paper you are looking for? You can Submit a new open access paper.