Search Results for author: Yichao Lu

Found 25 papers, 9 papers with code

Weakly Supervised Extractive Summarization with Attention

no code implementations SIGDIAL (ACL) 2021 Yingying Zhuang, Yichao Lu, Simi Wang

Automatic summarization aims to extract important information from large amounts of textual data in order to create a shorter version of the original texts while preserving its information.

Extractive Summarization Sentence

An Efficient Two-stage Gradient Boosting Framework for Short-term Traffic State Estimation

1 code implementation21 Feb 2023 Yichao Lu

In particular, we present an efficient two-stage gradient boosting framework for short-term traffic state estimation.

Benchmarking

Learning to Transfer for Traffic Forecasting via Multi-task Learning

1 code implementation27 Nov 2021 Yichao Lu

In particular, we present a multi-task learning framework for temporal and spatio-temporal domain adaptation of traffic forecasting models.

Benchmarking Domain Adaptation +1

Pretrain-Finetune Based Training of Task-Oriented Dialogue Systems in a Real-World Setting

no code implementations NAACL 2021 Manisha Srivastava, Yichao Lu, Riley Peschon, Chenyang Li

In this work, we present a method for training retrieval-based dialogue systems using a small amount of high-quality, annotated data and a larger, unlabeled dataset.

Retrieval Task-Oriented Dialogue Systems

Unsupervised Bitext Mining and Translation via Self-trained Contextual Embeddings

no code implementations15 Oct 2020 Phillip Keung, Julian Salazar, Yichao Lu, Noah A. Smith

We then improve an XLM-based unsupervised neural MT system pre-trained on Wikipedia by supplementing it with pseudo-parallel text mined from the same corpus, boosting unsupervised translation performance by up to 3. 5 BLEU on the WMT'14 French-English and WMT'16 German-English tasks and outperforming the previous state-of-the-art.

Machine Translation Sentence +2

The Multilingual Amazon Reviews Corpus

1 code implementation EMNLP 2020 Phillip Keung, Yichao Lu, György Szarvas, Noah A. Smith

We present the Multilingual Amazon Reviews Corpus (MARC), a large-scale collection of Amazon reviews for multilingual text classification.

General Classification Multilingual text classification +4

Efficient and Information-Preserving Future Frame Prediction and Beyond

1 code implementation ICLR 2020 Wei Yu, Yichao Lu, Steve Easterbrook, Sanja Fidler

Applying resolution-preserving blocks is a common practice to maximize information preservation in video prediction, yet their high memory consumption greatly limits their application scenarios.

Computational Efficiency object-detection +3

Don't Use English Dev: On the Zero-Shot Cross-Lingual Evaluation of Contextual Embeddings

no code implementations EMNLP 2020 Phillip Keung, Yichao Lu, Julian Salazar, Vikas Bhardwaj

Multilingual contextual embeddings have demonstrated state-of-the-art performance in zero-shot cross-lingual transfer learning, where multilingual BERT is fine-tuned on one source language and evaluated on a different target language.

Model Selection Transfer Learning +2

Attentional Speech Recognition Models Misbehave on Out-of-domain Utterances

1 code implementation12 Feb 2020 Phillip Keung, Wei Niu, Yichao Lu, Julian Salazar, Vikas Bhardwaj

We discuss the problem of echographic transcription in autoregressive sequence-to-sequence attentional architectures for automatic speech recognition, where a model produces very long sequences of repetitive outputs when presented with out-of-domain utterances.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +1

CrevNet: Conditionally Reversible Video Prediction

no code implementations25 Oct 2019 Wei Yu, Yichao Lu, Steve Easterbrook, Sanja Fidler

Applying resolution-preserving blocks is a common practice to maximize information preservation in video prediction, yet their high memory consumption greatly limits their application scenarios.

Computational Efficiency Video Prediction

Goal-Oriented End-to-End Conversational Models with Profile Features in a Real-World Setting

no code implementations NAACL 2019 Yichao Lu, Manisha Srivastava, Jared Kramer, Heba Elfardy, Andrea Kahn, Song Wang, Vikas Bhardwaj

To test our models, a customer service agent handles live contacts and at each turn we present the top four model responses and allow the agent to select (and optionally edit) one of the suggestions or to type their own.

Response Generation

A neural interlingua for multilingual machine translation

no code implementations WS 2018 Yichao Lu, Phillip Keung, Faisal Ladhak, Vikas Bhardwaj, Shaonan Zhang, Jason Sun

We incorporate an explicit neural interlingua into a multilingual encoder-decoder neural machine translation (NMT) architecture.

Machine Translation NMT +3

A practical approach to dialogue response generation in closed domains

no code implementations28 Mar 2017 Yichao Lu, Phillip Keung, Shaonan Zhang, Jason Sun, Vikas Bhardwaj

We describe a prototype dialogue response generation model for the customer service domain at Amazon.

Response Generation

Finding Linear Structure in Large Datasets with Scalable Canonical Correlation Analysis

no code implementations26 Jun 2015 Zhuang Ma, Yichao Lu, Dean Foster

In this paper, we tackle the problem of large scale CCA, where classical algorithms, usually requiring computing the product of two huge matrices and huge matrix decomposition, are computationally and storage expensive.

Stochastic Optimization

Large scale canonical correlation analysis with iterative least squares

no code implementations NeurIPS 2014 Yichao Lu, Dean P. Foster

Canonical Correlation Analysis (CCA) is a widely used statistical tool with both well established theory and favorable performance for a wide range of machine learning problems.

BIG-bench Machine Learning

New Subsampling Algorithms for Fast Least Squares Regression

no code implementations NeurIPS 2013 Paramveer Dhillon, Yichao Lu, Dean P. Foster, Lyle Ungar

We address the problem of fast estimation of ordinary least squares (OLS) from large amounts of data ($n \gg p$).

regression

Faster Ridge Regression via the Subsampled Randomized Hadamard Transform

no code implementations NeurIPS 2013 Yichao Lu, Paramveer Dhillon, Dean P. Foster, Lyle Ungar

We propose a fast algorithm for ridge regression when the number of features is much larger than the number of observations ($p \gg n$).

regression

Cannot find the paper you are looking for? You can Submit a new open access paper.