Search Results for author: Yanzhe Zhang

Found 10 papers, 8 papers with code

Best Practices and Lessons Learned on Synthetic Data for Language Models

no code implementations11 Apr 2024 Ruibo Liu, Jerry Wei, Fangyu Liu, Chenglei Si, Yanzhe Zhang, Jinmeng Rao, Steven Zheng, Daiyi Peng, Diyi Yang, Denny Zhou, Andrew M. Dai

The success of AI models relies on the availability of large, diverse, and high-quality datasets, which can be challenging to obtain due to data scarcity, privacy concerns, and high costs.

Dynamic LLM-Agent Network: An LLM-agent Collaboration Framework with Agent Team Optimization

1 code implementation3 Oct 2023 Zijun Liu, Yanzhe Zhang, Peng Li, Yang Liu, Diyi Yang

We further design an automatic agent team optimization algorithm based on an unsupervised metric termed $\textit{Agent Importance Score}$, enabling the selection of best agents based on the contribution each agent makes.

Code Generation Language Modelling +2

LLaVAR: Enhanced Visual Instruction Tuning for Text-Rich Image Understanding

1 code implementation29 Jun 2023 Yanzhe Zhang, Ruiyi Zhang, Jiuxiang Gu, Yufan Zhou, Nedim Lipka, Diyi Yang, Tong Sun

Instruction tuning unlocks the superior capability of Large Language Models (LLM) to interact with humans.

16k Image Captioning +3

Auditing Gender Presentation Differences in Text-to-Image Models

1 code implementation7 Feb 2023 Yanzhe Zhang, Lu Jiang, Greg Turk, Diyi Yang

Text-to-image models, which can generate high-quality images based on textual input, have recently enabled various content-creation tools.

Robustness of Demonstration-based Learning Under Limited Data Scenario

1 code implementation19 Oct 2022 Hongxin Zhang, Yanzhe Zhang, Ruiyi Zhang, Diyi Yang

Demonstration-based learning has shown great potential in stimulating pretrained language models' ability under limited data scenario.

Few-shot NER

Continual Sequence Generation with Adaptive Compositional Modules

2 code implementations ACL 2022 Yanzhe Zhang, Xuezhi Wang, Diyi Yang

Continual learning is essential for real-world deployment when there is a need to quickly adapt the model to new tasks without forgetting knowledge of old tasks.

Continual Learning Transfer Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.