Search Results for author: Davis Liang

Found 14 papers, 7 papers with code

RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training

1 code implementation7 Dec 2023 Jaehyung Kim, Yuning Mao, Rui Hou, Hanchao Yu, Davis Liang, Pascale Fung, Qifan Wang, Fuli Feng, Lifu Huang, Madian Khabsa

Under a unified evaluation of fine-tuned LMs by incorporating four representative perspectives of model robustness, we demonstrate the effectiveness of RoAST compared to state-of-the-art fine-tuning methods on six different types of LMs, which indicates its usefulness in practice.

Adversarial Robustness

A Study on Knowledge Distillation from Weak Teacher for Scaling Up Pre-trained Language Models

1 code implementation26 May 2023 Hayeon Lee, Rui Hou, Jongpil Kim, Davis Liang, Sung Ju Hwang, Alexander Min

Distillation from Weak Teacher (DWT) is a method of transferring knowledge from a smaller, weaker teacher model to a larger student model to improve its performance.

Knowledge Distillation

Attention-guided Generative Models for Extractive Question Answering

no code implementations12 Oct 2021 Peng Xu, Davis Liang, Zhiheng Huang, Bing Xiang

We propose a simple strategy to obtain an extractive answer span from the generative model by leveraging the decoder cross-attention patterns.

Extractive Question-Answering Hallucination +2

Multiplicative Position-aware Transformer Models for Language Understanding

no code implementations27 Sep 2021 Zhiheng Huang, Davis Liang, Peng Xu, Bing Xiang

Transformer models, which leverage architectural improvements like self-attention, perform remarkably well on Natural Language Processing (NLP) tasks.

Position

Decoding and Diversity in Machine Translation

no code implementations26 Nov 2020 Nicholas Roberts, Davis Liang, Graham Neubig, Zachary C. Lipton

This makes human-level BLEU a misleading benchmark in that modern MT systems cannot approach human-level BLEU while simultaneously maintaining human-level translation diversity.

Machine Translation NMT +1

Embedding-based Zero-shot Retrieval through Query Generation

1 code implementation22 Sep 2020 Davis Liang, Peng Xu, Siamak Shakeri, Cicero Nogueira dos Santos, Ramesh Nallapati, Zhiheng Huang, Bing Xiang

In some cases, our model trained on synthetic data can even outperform the same model trained on real data

Passage Retrieval Retrieval

TRANS-BLSTM: Transformer with Bidirectional LSTM for Language Understanding

no code implementations16 Mar 2020 Zhiheng Huang, Peng Xu, Davis Liang, Ajay Mishra, Bing Xiang

Prior to the transformer era, bidirectional Long Short-Term Memory (BLSTM) has been the dominant modeling architecture for neural machine translation and question answering.

Machine Translation Natural Language Inference +5

Masked Language Model Scoring

6 code implementations ACL 2020 Julian Salazar, Davis Liang, Toan Q. Nguyen, Katrin Kirchhoff

Instead, we evaluate MLMs out of the box via their pseudo-log-likelihood scores (PLLs), which are computed by masking tokens one by one.

Attribute Domain Adaptation +4

Learning Noise-Invariant Representations for Robust Speech Recognition

no code implementations17 Jul 2018 Davis Liang, Zhiheng Huang, Zachary C. Lipton

Despite rapid advances in speech recognition, current models remain brittle to superficial perturbations to their inputs.

Data Augmentation Representation Learning +2

Deep Automated Multi-task Learning

no code implementations IJCNLP 2017 Davis Liang, Yan Shu

Our results show that training on a primary task in parallel with a secondary automated task improves both the convergence speed and accuracy for the primary task.

Multi-Task Learning Sentiment Analysis

Cannot find the paper you are looking for? You can Submit a new open access paper.