Search Results for author: Yau-Shian Wang

Found 13 papers, 4 papers with code

PESCO: Prompt-enhanced Self Contrastive Learning for Zero-shot Text Classification

no code implementations24 May 2023 Yau-Shian Wang, Ta-Chung Chi, Ruohong Zhang, Yiming Yang

We present PESCO, a novel contrastive learning framework that substantially improves the performance of zero-shot text classification.

Contrastive Learning text-classification +3

Generation-driven Contrastive Self-training for Zero-shot Text Classification with Instruction-following LLM

1 code implementation24 Apr 2023 Ruohong Zhang, Yau-Shian Wang, Yiming Yang

To overcome these limitations, we introduce a novel method, namely GenCo, which leverages the strong generative power of LLMs to assist in training a smaller and more adaptable language model.

Instruction Following Language Modelling +5

Toxicity Detection with Generative Prompt-based Inference

no code implementations24 May 2022 Yau-Shian Wang, Yingshan Chang

It is a long-known risk that language models (LMs), once trained on corpus containing undesirable content, have the power to manifest biases and toxicity.

Language Modelling Prompt Engineering

Long-tailed Extreme Multi-label Text Classification with Generated Pseudo Label Descriptions

no code implementations2 Apr 2022 Ruohong Zhang, Yau-Shian Wang, Yiming Yang, Donghan Yu, Tom Vu, Likun Lei

Extreme Multi-label Text Classification (XMTC) has been a tough challenge in machine learning research and applications due to the sheer sizes of the label spaces and the severe data scarce problem associated with the long tail of rare labels in highly skewed distributions.

Multi Label Text Classification Multi-Label Text Classification +3

Exploiting Local and Global Features in Transformer-based Extreme Multi-label Text Classification

no code implementations2 Apr 2022 Ruohong Zhang, Yau-Shian Wang, Yiming Yang, Tom Vu, Likun Lei

Extreme multi-label text classification (XMTC) is the task of tagging each document with the relevant labels from a very large space of predefined categories.

Multi Label Text Classification Multi-Label Text Classification +1

Are you doing what I say? On modalities alignment in ALFRED

no code implementations12 Oct 2021 Ting-Rui Chiang, Yi-Ting Yeh, Ta-Chung Chi, Yau-Shian Wang

ALFRED is a recently proposed benchmark that requires a model to complete tasks in simulated house environments specified by instructions in natural language.

Investigation of Sentiment Controllable Chatbot

no code implementations11 Jul 2020 Hung-Yi Lee, Cheng-Hao Ho, Chien-Fu Lin, Chiung-Chih Chang, Chih-Wei Lee, Yau-Shian Wang, Tsung-Yuan Hsu, Kuan-Yu Chen

Conventional seq2seq chatbot models attempt only to find sentences with the highest probabilities conditioned on the input sequences, without considering the sentiment of the output sentences.

Chatbot reinforcement-learning +1

Tree Transformer: Integrating Tree Structures into Self-Attention

3 code implementations IJCNLP 2019 Yau-Shian Wang, Hung-Yi Lee, Yun-Nung Chen

This paper proposes Tree Transformer, which adds an extra constraint to attention heads of the bidirectional Transformer encoder in order to encourage the attention heads to follow tree structures.

Language Modelling

Learning to Encode Text as Human-Readable Summaries using Generative Adversarial Networks

1 code implementation EMNLP 2018 Yau-Shian Wang, Hung-Yi Lee

The generator encodes the input text into a shorter word sequence, and the reconstructor recovers the generator input from the generator output.

Abstractive Text Summarization

TopicGAN: Unsupervised Text Generation from Explainable Latent Topics

no code implementations27 Sep 2018 Yau-Shian Wang, Yun-Nung Chen, Hung-Yi Lee

Learning discrete representations of data and then generating data from the discovered representations have been increasingly studied because the obtained discrete representations can benefit unsupervised learning.

Image Generation Text Generation

Scalable Sentiment for Sequence-to-sequence Chatbot Response with Performance Analysis

no code implementations7 Apr 2018 Chih-Wei Lee, Yau-Shian Wang, Tsung-Yuan Hsu, Kuan-Yu Chen, Hung-Yi Lee, Lin-shan Lee

Conventional seq2seq chatbot models only try to find the sentences with the highest probabilities conditioned on the input sequences, without considering the sentiment of the output sentences.

Chatbot reinforcement-learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.