no code implementations • EMNLP 2021 • Kenneth Church, Yuchen Bian
Given such estimates, WordNet’s coverage is remarkable.
1 code implementation • LREC 2022 • Kenneth Church, Xingyu Cai, Yuchen Bian
We propose using lexical resources (thesaurus, VAD) to fine-tune pretrained deep nets such as BERT and ERNIE.
no code implementations • ACL 2022 • Kenneth Church, Valia Kordoni, Gary Marcus, Ernest Davis, Yanjun Ma, Zeyu Chen
The first half of this tutorial will make deep nets more accessible to a broader audience, following “Deep Nets for Poets” and “A Gentle Introduction to Fine-Tuning.” We will also introduce GFT (general fine tuning), a little language for fine tuning deep nets with short (one line) programs that are as easy to code as regression in statistics packages such as R using glm (general linear models).
no code implementations • EMNLP 2020 • Jiaji Huang, Xingyu Cai, Kenneth Church
This paper designs a Monolingual Lexicon Induction task and observes that two factors accompany the degraded accuracy of bilingual lexicon induction for rare words.
1 code implementation • ACL (BPPF) 2021 • Kenneth Church, Mark Liberman, Valia Kordoni
There used to be more top-down leadership from government (and industry, in the case of systems, with benchmarks such as SPEC).
no code implementations • RaPID (LREC) 2022 • Jiahong Yuan, Xingyu Cai, Kenneth Church
The result represents a relative error reduction of 14% over the baseline model trained without data augmentation.
no code implementations • 27 Mar 2024 • Abteen Ebrahimi, Kenneth Church
English has long been assumed the $\textit{lingua franca}$ of scientific research, and this notion is reflected in the natural language processing (NLP) research involving scientific document representation.
no code implementations • 27 Apr 2022 • Guangxu Xun, Mingbo Ma, Yuchen Bian, Xingyu Cai, Jiaji Huang, Renjie Zheng, Junkun Chen, Jiahong Yuan, Kenneth Church, Liang Huang
In simultaneous translation (SimulMT), the most widely used strategy is the wait-k policy thanks to its simplicity and effectiveness in balancing translation quality and latency.
1 code implementation • 6 Jan 2022 • Yuanpeng Li, Joel Hestness, Mohamed Elhoseiny, Liang Zhao, Kenneth Church
This paper proposes an efficient approach to learning disentangled representations with causal mechanisms based on the difference of conditional probabilities in original and new distributions.
1 code implementation • NeurIPS 2021 • Jiaji Huang, Qiang Qiu, Kenneth Church
We model the space of tasks as a Gaussian process.
no code implementations • ICLR 2022 • Xingyu Cai, Jiahong Yuan, Yuchen Bian, Guangxu Xun, Jiaji Huang, Kenneth Church
Standard CTC computes a loss by aggregating over all possible alignment paths, that map the entire sequence to the entire label (full alignment).
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 2 Aug 2021 • Jiahong Yuan, Xingyu Cai, Renjie Zheng, Liang Huang, Kenneth Church
Models of phonemes, broad phonetic classes, and syllables all significantly outperform the utterance model, demonstrating that phonetic units are helpful and should be incorporated in speech emotion recognition.
no code implementations • 2 Aug 2021 • Jiahong Yuan, Xingyu Cai, Dongji Gao, Renjie Zheng, Liang Huang, Kenneth Church
Much of the recent literature on automatic speech recognition (ASR) is taking an end-to-end approach.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
no code implementations • 2 Aug 2021 • Jiahong Yuan, Neville Ryant, Xingyu Cai, Kenneth Church, Mark Liberman
This study reports our efforts to improve automatic recognition of suprasegmentals by fine-tuning wav2vec 2. 0 with CTC, a method that has been successful in automatic speech recognition.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
no code implementations • NAACL 2021 • Yuchen Bian, Jiaji Huang, Xingyu Cai, Jiahong Yuan, Kenneth Church
(What) We define and focus the study on redundancy matrices generated from pre-trained and fine-tuned BERT-base model for GLUE datasets.
no code implementations • 12 May 2021 • Boxiang Liu, Jiaji Huang, Xingyu Cai, Kenneth Church
This paper compares BERT-SQuAD and Ab3P on the Abbreviation Definition Identification (ADI) task.
no code implementations • ICLR 2021 • Xingyu Cai, Jiaji Huang, Yuchen Bian, Kenneth Church
We hope the study in this paper could provide insights towards a better understanding of the deep language models.
no code implementations • 1 Jan 2021 • Yuanpeng Li, Liang Zhao, Joel Hestness, Ka Yee Lun, Kenneth Church, Mohamed Elhoseiny
To our best knowledge, this is the first work to focus on the transferability of compositionality, and it is orthogonal to existing efforts of learning compositional representations in training distribution.
no code implementations • 1 Jan 2021 • Yuanpeng Li, Liang Zhao, Joel Hestness, Kenneth Church, Mohamed Elhoseiny
In this paper, we argue that gradient descent is one of the reasons that make compositionality learning hard during neural network optimization.
3 code implementations • 2 Dec 2020 • Neville Ryant, Prachi Singh, Venkat Krishnamohan, Rajat Varma, Kenneth Church, Christopher Cieri, Jun Du, Sriram Ganapathy, Mark Liberman
DIHARD III was the third in a series of speaker diarization challenges intended to improve the robustness of diarization systems to variability in recording equipment, noise conditions, and conversational domain.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Renjie Zheng, Mingbo Ma, Baigong Zheng, Kaibo Liu, Jiahong Yuan, Kenneth Church, Liang Huang
Simultaneous speech-to-speech translation is widely useful but extremely challenging, since it needs to generate target-language speech concurrently with the source-language speech, with only a few seconds delay.
1 code implementation • ICLR 2020 • Yuanpeng Li, Liang Zhao, Kenneth Church, Mohamed Elhoseiny
It also shows significant improvement in machine translation task.
3 code implementations • ICCV 2021 • Sherif Abdelkarim, Aniket Agarwal, Panos Achlioptas, Jun Chen, Jiaji Huang, Boyang Li, Kenneth Church, Mohamed Elhoseiny
We use these benchmarks to study the performance of several state-of-the-art long-tail models on the LTVRR setup.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Mingbo Ma, Baigong Zheng, Kaibo Liu, Renjie Zheng, Hairong Liu, Kainan Peng, Kenneth Church, Liang Huang
Text-to-speech synthesis (TTS) has witnessed rapid progress in recent years, where neural methods became capable of producing audios with high naturalness.
1 code implementation • ACL 2019 • Jiaji Huang, Qiang Qiu, Kenneth Church
Recent advances in BLI work by aligning the two word embedding spaces.
1 code implementation • 18 Jun 2019 • Neville Ryant, Kenneth Church, Christopher Cieri, Alejandrina Cristia, Jun Du, Sriram Ganapathy, Mark Liberman
This paper introduces the second DIHARD challenge, the second in a series of speaker diarization challenges intended to improve the robustness of diarization systems to variation in recording equipment, noise conditions, and conversational domain.
no code implementations • 23 Oct 2018 • Mostofa Patwary, Milind Chabbi, Heewoo Jun, Jiaji Huang, Gregory Diamos, Kenneth Church
We show how Zipf's Law can be used to scale up language modeling (LM) to take advantage of more training data and more GPUs.
no code implementations • 27 Sep 2018 • Joel Hestness, Sharan Narang, Newsha Ardalani, Heewoo Jun, Hassan Kianinejad, Md. Mostofa Ali Patwary, Yang Yang, Yanqi Zhou, Gregory Diamos, Kenneth Church
As the pace of deep learning innovation accelerates, it becomes increasingly important to organize the space of problems by relative difficultly.