Search Results for author: Zhiliang Tian

Found 33 papers, 9 papers with code

LTOS: Layout-controllable Text-Object Synthesis via Adaptive Cross-attention Fusions

no code implementations21 Apr 2024 Xiaoran Zhao, Tianhao Wu, Yu Lai, Zhiliang Tian, Zhen Huang, Yahui Liu, Zejiang He, Dongsheng Li

Controllable text-to-image generation synthesizes visual text and objects in images with certain conditions, which are frequently applied to emoji and poster generation.

Layout-to-Image Generation Object +1

360°REA: Towards A Reusable Experience Accumulation with 360° Assessment for Multi-Agent System

no code implementations8 Apr 2024 Shen Gao, Hao Li, Zhengliang Shi, Chengrui Huang, Quan Tu, Zhiliang Tian, Minlie Huang, Shuo Shang

The framework employs a novel 360{\deg} performance assessment method for multi-perspective performance evaluation with fine-grained assessment.

Language Modelling Large Language Model

Learn to Disguise: Avoid Refusal Responses in LLM's Defense via a Multi-agent Attacker-Disguiser Game

no code implementations3 Apr 2024 Qianqiao Xu, Zhiliang Tian, Hongyan Wu, Zhen Huang, Yiping Song, Feng Liu, Dongsheng Li

In this paper, we propose a multi-agent attacker-disguiser game approach to achieve a weak defense mechanism that allows the large model to both safely reply to the attacker and hide the defense intent.

Prompt Engineering

LLM-based Privacy Data Augmentation Guided by Knowledge Distillation with a Distribution Tutor for Medical Text Classification

no code implementations26 Feb 2024 Yiping Song, Juhua Zhang, Zhiliang Tian, Yuxin Yang, Minlie Huang, Dongsheng Li

As sufficient data are not always publically accessible for model training, researchers exploit limited data with advanced learning algorithms or expand the dataset via data augmentation (DA).

Data Augmentation Knowledge Distillation +2

TFDMNet: A Novel Network Structure Combines the Time Domain and Frequency Domain Features

1 code implementation29 Jan 2024 Hengyue Pan, Yixin Chen, Zhiliang Tian, Peng Qiao, Linbo Qiao, Dongsheng Li

To get the balance between the computation complexity and memory usage, we propose a new network structure, namely Time-Frequency Domain Mixture Network (TFDMNet), which combines the advantages of both convolution layers and EMLs.

Resilient Practical Test-Time Adaptation: Soft Batch Normalization Alignment and Entropy-driven Memory Bank

no code implementations26 Jan 2024 Xingzhi Zhou, Zhiliang Tian, Ka Chun Cheung, Simon See, Nevin L. Zhang

Test-time domain adaptation effectively adjusts the source domain model to accommodate unseen domain shifts in a target domain during inference.

Test-time Adaptation

POMP: Probability-driven Meta-graph Prompter for LLMs in Low-resource Unsupervised Neural Machine Translation

no code implementations11 Jan 2024 Shilong Pan, Zhiliang Tian, Liang Ding, Zhen Huang, Zhihua Wen, Dongsheng Li

POMP involves constructing a directed acyclic meta-graph for each source language, from which we dynamically sample multiple paths to prompt LLMs to mitigate the linguistic noise and improve translations during training.

In-Context Learning Machine Translation +3

TVT: Training-Free Vision Transformer Search on Tiny Datasets

no code implementations24 Nov 2023 Zimian Wei, Hengyue Pan, Lujun Li, Peijie Dong, Zhiliang Tian, Xin Niu, Dongsheng Li

In this paper, for the first time, we investigate how to search in a training-free manner with the help of teacher models and devise an effective Training-free ViT (TVT) search framework.

GROVE: A Retrieval-augmented Complex Story Generation Framework with A Forest of Evidence

no code implementations9 Oct 2023 Zhihua Wen, Zhiliang Tian, Wei Wu, Yuxin Yang, Yanqi Shi, Zhen Huang, Dongsheng Li

Finally, we select the most fitting chains of evidence from the evidence forest and integrate them into the generated story, thereby enhancing the narrative's complexity and credibility.

Retrieval Story Generation

Recursively Summarizing Enables Long-Term Dialogue Memory in Large Language Models

no code implementations29 Aug 2023 Qingyue Wang, Liang Ding, Yanan Cao, Zhiliang Tian, Shi Wang, DaCheng Tao, Li Guo

We evaluate our method on both open and closed LLMs, and the experiments on the widely-used public dataset show that our method can generate more consistent responses in a long-context conversation.

16k 8k +1

DaMSTF: Domain Adversarial Learning Enhanced Meta Self-Training for Domain Adaptation

no code implementations5 Aug 2023 Menglong Lu, Zhen Huang, Yunxiang Zhao, Zhiliang Tian, Yang Liu, Dongsheng Li

To this end, we employ domain adversarial learning as a heuristic neural network initialization method, which can help the meta-learning module converge to a better optimal.

Domain Adaptation Meta-Learning +2

Meta-Tsallis-Entropy Minimization: A New Self-Training Approach for Domain Adaptation on Text Classification

no code implementations4 Aug 2023 Menglong Lu, Zhen Huang, Zhiliang Tian, Yunxiang Zhao, Xuanyu Fei, Dongsheng Li

Theoretically, we prove the convergence of the meta-learning algorithm in MTEM and analyze the effectiveness of MTEM in achieving domain adaptation.

Domain Adaptation Meta-Learning +2

EMQ: Evolving Training-free Proxies for Automated Mixed Precision Quantization

1 code implementation ICCV 2023 Peijie Dong, Lujun Li, Zimian Wei, Xin Niu, Zhiliang Tian, Hengyue Pan

In particular, we devise an elaborate search space involving the existing proxies and perform an evolution search to discover the best correlated MQ proxy.

Quantization

Retrieval-augmented GPT-3.5-based Text-to-SQL Framework with Sample-aware Prompting and Dynamic Revision Chain

no code implementations11 Jul 2023 Chunxi Guo, Zhiliang Tian, Jintao Tang, Shasha Li, Zhihua Wen, Kaixuan Wang, Ting Wang

Prompt learning with large language models (LLMs) has emerged as a recent approach, which designs prompts to lead LLMs to understand the input question and generate the corresponding SQL.

Retrieval Text-To-SQL

Self-Evolution Learning for Mixup: Enhance Data Augmentation on Few-Shot Text Classification Tasks

no code implementations22 May 2023 Haoqi Zheng, Qihuang Zhong, Liang Ding, Zhiliang Tian, Xin Niu, Dongsheng Li, DaCheng Tao

However, most of the mixup methods do not consider the varying degree of learning difficulty in different stages of training and generate new samples with one hot labels, resulting in the model over confidence.

Data Augmentation Few-Shot Text Classification +1

Prompting GPT-3.5 for Text-to-SQL with De-semanticization and Skeleton Retrieval

no code implementations26 Apr 2023 Chunxi Guo, Zhiliang Tian, Jintao Tang, Pancheng Wang, Zhihua Wen, Kang Yang, Ting Wang

Text-to-SQL is a task that converts a natural language question into a structured query language (SQL) to retrieve information from a database.

Informativeness Retrieval +2

Hashtag-Guided Low-Resource Tweet Classification

1 code implementation20 Feb 2023 Shizhe Diao, Sedrick Scott Keh, Liangming Pan, Zhiliang Tian, Yan Song, Tong Zhang

Social media classification tasks (e. g., tweet sentiment analysis, tweet stance detection) are challenging because social media posts are typically short, informal, and ambiguous.

Classification Sentiment Analysis +1

RD-NAS: Enhancing One-shot Supernet Ranking Ability via Ranking Distillation from Zero-cost Proxies

1 code implementation24 Jan 2023 Peijie Dong, Xin Niu, Lujun Li, Zhiliang Tian, Xiaodong Wang, Zimian Wei, Hengyue Pan, Dongsheng Li

In this paper, we propose Ranking Distillation one-shot NAS (RD-NAS) to enhance ranking consistency, which utilizes zero-cost proxies as the cheap teacher and adopts the margin ranking loss to distill the ranking knowledge.

Computational Efficiency Neural Architecture Search

Progressive Meta-Pooling Learning for Lightweight Image Classification Model

no code implementations24 Jan 2023 Peijie Dong, Xin Niu, Zhiliang Tian, Lujun Li, Xiaodong Wang, Zimian Wei, Hengyue Pan, Dongsheng Li

Practical networks for edge devices adopt shallow depth and small convolutional kernels to save memory and computational cost, which leads to a restricted receptive field.

Classification Image Classification

StyleFlow: Disentangle Latent Representations via Normalizing Flow for Unsupervised Text Style Transfer

no code implementations19 Dec 2022 Kangchen Zhu, Zhiliang Tian, Ruifeng Luo, Xiaoguang Mao

Since cycle construction helps to improve the style transfer ability of the model by rebuilding transferred sentences back to original-style sentences, it brings about a content loss in unsupervised text style transfer tasks.

Data Augmentation Disentanglement +4

Semi-Supervised Lifelong Language Learning

1 code implementation23 Nov 2022 Yingxiu Zhao, Yinhe Zheng, Bowen Yu, Zhiliang Tian, Dongkyu Lee, Jian Sun, Haiyang Yu, Yongbin Li, Nevin L. Zhang

In this paper, we explore a novel setting, semi-supervised lifelong language learning (SSLL), where a model learns sequentially arriving language tasks with both labeled and unlabeled data.

Transfer Learning

Hard Gate Knowledge Distillation -- Leverage Calibration for Robust and Reliable Language Model

no code implementations22 Oct 2022 Dongkyu Lee, Zhiliang Tian, Yingxiu Zhao, Ka Chun Cheung, Nevin L. Zhang

The question is answered in our work with the concept of model calibration; we view a teacher model not only as a source of knowledge but also as a gauge to detect miscalibration of a student.

Knowledge Distillation Language Modelling +2

SeqPATE: Differentially Private Text Generation via Knowledge Distillation

no code implementations29 Sep 2021 Zhiliang Tian, Yingxiu Zhao, Ziyue Huang, Yu-Xiang Wang, Nevin Zhang, He He

Differentially private (DP) learning algorithms provide guarantees on identifying the existence of a training sample from model outputs.

Knowledge Distillation Sentence +2

Enhancing Content Preservation in Text Style Transfer Using Reverse Attention and Conditional Layer Normalization

1 code implementation ACL 2021 Dongkyu Lee, Zhiliang Tian, Lanqing Xue, Nevin L. Zhang

A common approach is to map a given sentence to content representation that is free of style, and the content representation is fed to a decoder with a target style.

Sentence Style Transfer +1

Learning from My Friends: Few-Shot Personalized Conversation Systems via Social Networks

no code implementations21 May 2021 Zhiliang Tian, Wei Bi, Zihan Zhang, Dongkyu Lee, Yiping Song, Nevin L. Zhang

The task requires models to generate personalized responses for a speaker given a few conversations from the speaker and a social network.

Meta-Learning

Towards Efficiently Diversifying Dialogue Generation via Embedding Augmentation

1 code implementation2 Mar 2021 Yu Cao, Liang Ding, Zhiliang Tian, Meng Fang

Dialogue generation models face the challenge of producing generic and repetitive responses.

Dialogue Generation

Response-Anticipated Memory for On-Demand Knowledge Integration in Response Generation

no code implementations ACL 2020 Zhiliang Tian, Wei Bi, Dongkyu Lee, Lanqing Xue, Yiping Song, Xiaojiang Liu, Nevin L. Zhang

In previous work, the external document is utilized by (1) creating a context-aware document memory that integrates information from the document and the conversational context, and then (2) generating responses referring to the memory.

Informativeness Response Generation

Learning to Abstract for Memory-augmented Conversational Response Generation

1 code implementation ACL 2019 Zhiliang Tian, Wei Bi, Xiaopeng Li, Nevin L. Zhang

In this work, we propose a memory-augmented generative model, which learns to abstract from the training corpus and saves the useful information to the memory to assist the response generation.

Conversational Response Generation Informativeness +2

Diversifying Neural Conversation Model with Maximal Marginal Relevance

no code implementations IJCNLP 2017 Yiping Song, Zhiliang Tian, Dongyan Zhao, Ming Zhang, Rui Yan

However, traditional seq2seq suffer from a severe weakness: during beam search decoding, they tend to rank universal replies at the top of the candidate list, resulting in the lack of diversity among candidate replies.

Document Summarization Information Retrieval +1

Cannot find the paper you are looking for? You can Submit a new open access paper.