no code implementations • 19 Apr 2024 • Pablo Biedma, Xiaoyuan Yi, Linus Huang, Maosong Sun, Xing Xie
Recent advancements in Large Language Models (LLMs) have revolutionized the AI field but also pose potential safety and ethical risks.
no code implementations • 7 Mar 2024 • Xinpeng Wang, Shitong Duan, Xiaoyuan Yi, Jing Yao, Shanlin Zhou, Zhihua Wei, Peng Zhang, Dongkuan Xu, Maosong Sun, Xing Xie
Big models have achieved revolutionary breakthroughs in the field of AI, but they might also pose potential concerns.
1 code implementation • 6 Mar 2024 • Shitong Duan, Xiaoyuan Yi, Peng Zhang, Tun Lu, Xing Xie, Ning Gu
Large language models (LLMs) have revolutionized the role of AI, yet also pose potential risks of propagating unethical content.
no code implementations • 29 Feb 2024 • Xukun Liu, Zhiyuan Peng, Xiaoyuan Yi, Xing Xie, Lirong Xiang, Yuchen Liu, Dongkuan Xu
While achieving remarkable progress in a broad range of tasks, large language models (LLMs) remain significantly limited in properly using massive external tools.
no code implementations • 27 Jan 2024 • Pengjie Liu, Zhenghao Liu, Xiaoyuan Yi, Liner Yang, Shuo Wang, Yu Gu, Ge Yu, Xing Xie, Shuang-Hua Yang
It proposes a dual-view legal clue reasoning mechanism, which derives from two reasoning chains of judges: 1) Law Case Reasoning, which makes legal judgments according to the judgment experiences learned from analogy/confusing legal cases; 2) Legal Ground Reasoning, which lies in matching the legal clues between criminal cases and legal decisions.
1 code implementation • 13 Dec 2023 • Xinpeng Wang, Xiaoyuan Yi, Han Jiang, Shanlin Zhou, Zhihua Wei, Xing Xie
Warning: this paper includes model outputs showing offensive content.
1 code implementation • 28 Nov 2023 • Yuhang Wang, Yanxu Zhu, Chao Kong, Shuyu Wei, Xiaoyuan Yi, Xing Xie, Jitao Sang
This benchmark serves as a valuable resource for cultural studies in LLMs, paving the way for more culturally aware and sensitive models.
no code implementations • 16 Nov 2023 • Jing Yao, Wei Xu, Jianxun Lian, Xiting Wang, Xiaoyuan Yi, Xing Xie
In this paper, we propose a general paradigm that augments LLMs with DOmain-specific KnowledgE to enhance their performance on practical applications, namely DOKE.
no code implementations • 15 Nov 2023 • Jing Yao, Xiaoyuan Yi, Xiting Wang, Yifan Gong, Xing Xie
The rapid advancement of Large Language Models (LLMs) has attracted much attention to value alignment for their responsible development.
no code implementations • 26 Oct 2023 • Xiaoyuan Yi, Jing Yao, Xiting Wang, Xing Xie
Big models have greatly advanced AI's ability to understand, generate, and manipulate information and content, enabling numerous applications.
no code implementations • 17 Oct 2023 • Shitong Duan, Xiaoyuan Yi, Peng Zhang, Tun Lu, Xing Xie, Ning Gu
We discovered that most models are essentially misaligned, necessitating further ethical value alignment.
no code implementations • 23 Aug 2023 • Jing Yao, Xiaoyuan Yi, Xiting Wang, Jindong Wang, Xing Xie
Big models, exemplified by Large Language Models (LLMs), are models typically pre-trained on massive data and comprised of enormous parameters, which not only obtain significantly improved performance across diverse tasks but also present emergent capabilities absent in smaller models.
1 code implementation • 6 Jul 2023 • Yupeng Chang, Xu Wang, Jindong Wang, Yuan Wu, Linyi Yang, Kaijie Zhu, Hao Chen, Xiaoyuan Yi, Cunxiang Wang, Yidong Wang, Wei Ye, Yue Zhang, Yi Chang, Philip S. Yu, Qiang Yang, Xing Xie
Large language models (LLMs) are gaining increasing popularity in both academia and industry, owing to their unprecedented performance in various applications.
1 code implementation • 17 Jun 2023 • Yuxi Feng, Xiaoyuan Yi, Laks V. S. Lakshmanan, Xing Xie
Self-training (ST) has come to fruition in language understanding tasks by producing pseudo labels, which reduces the labeling bottleneck of language model fine-tuning.
no code implementations • 19 May 2023 • Jinyi Hu, Xu Han, Xiaoyuan Yi, Yutong Chen, Wenhao Li, Zhiyuan Liu, Maosong Sun
IAP optimizes only a separate Chinese text encoder with all other parameters fixed to align Chinese semantics space to the English one in CLIP.
1 code implementation • 16 Dec 2022 • Yuxi Feng, Xiaoyuan Yi, Xiting Wang, Laks V. S. Lakshmanan, Xing Xie
Augmented by only self-generated pseudo text, generation models over-emphasize exploitation of the previously learned space, suffering from a constrained generalization boundary.
1 code implementation • 14 Nov 2022 • Wenhao Li, Xiaoyuan Yi, Jinyi Hu, Maosong Sun, Xing Xie
In this work, we dig into the intrinsic mechanism of this problem and found that sparser attention values in Transformer could improve diversity.
no code implementations • 22 Oct 2022 • Jinyi Hu, Xiaoyuan Yi, Wenhao Li, Maosong Sun, Xing Xie
We demonstrate that TRACE could enhance the entanglement of each segment and preceding latent variables and deduce a non-zero lower bound of the KL term, providing a theoretical guarantee of generation diversity.
1 code implementation • 13 Oct 2022 • Seungeon Lee, Xiting Wang, Sungwon Han, Xiaoyuan Yi, Xing Xie, Meeyoung Cha
We present SELOR, a framework for integrating self-explaining capabilities into a given deep model to achieve both high prediction performance and human precision.
no code implementations • 10 Oct 2022 • Zonghan Yang, Xiaoyuan Yi, Peng Li, Yang Liu, Xing Xie
Warning: this paper contains model outputs exhibiting offensiveness and biases.
1 code implementation • NAACL 2022 • Jinyi Hu, Xiaoyuan Yi, Wenhao Li, Maosong Sun, Xing Xie
The past several years have witnessed Variational Auto-Encoder's superiority in various text generation tasks.
1 code implementation • 3 Jun 2021 • Wenhao Li, Fanchao Qi, Maosong Sun, Xiaoyuan Yi, Jiarui Zhang
We hope this dataset can further enhance the study on incorporating deep semantics into the understanding and generation system of Chinese classical poetry.
1 code implementation • NAACL 2021 • Zhenghao Liu, Xiaoyuan Yi, Maosong Sun, Liner Yang, Tat-Seng Chua
Grammatical Error Correction (GEC) aims to correct writing errors and help language learners improve their writing skills.
Ranked #1 on Grammatical Error Detection on FCE
no code implementations • 13 Mar 2020 • Xiaoyuan Yi, Ruoyu Li, Cheng Yang, Wenhao Li, Maosong Sun
Though recent neural models make prominent progress in some criteria of poetry quality, generated poems still suffer from the problem of poor diversity.
no code implementations • ACL 2019 • Guo Zhipeng, Xiaoyuan Yi, Maosong Sun, Wenhao Li, Cheng Yang, Jiannan Liang, Huimin Chen, Yuhui Zhang, Ruoyu Li
By exposing the options of poetry genres, styles and revision modes, Jiuge, acting as a professional assistant, allows constant and active participation of users in poetic creation.
no code implementations • EMNLP 2018 • Cheng Yang, Maosong Sun, Xiaoyuan Yi, Wenhao Li
The ability to write diverse poems in different styles under the same poetic imagery is an important characteristic of human poetry writing.
no code implementations • EMNLP 2018 • Xiaoyuan Yi, Maosong Sun, Ruoyu Li, Wenhao Li
Human experts evaluate poetry in terms of some specific criteria, instead of word-level likelihood.
no code implementations • CONLL 2018 • Xiaoyuan Yi, Ruoyu Li, Maosong Sun
As a precious part of the human cultural heritage, Chinese poetry has influenced people for generations.
1 code implementation • 12 Sep 2018 • Xiaoyuan Yi, Maosong Sun, Ruoyu Li, Zonghan Yang
Different from previous methods, our model explicitly maintains topics and informative limited history in a neural memory.
no code implementations • 6 Apr 2016 • Xiaoyuan Yi, Ruoyu Li, Maosong Sun
We take the generation of Chinese classical poem lines as a sequence-to-sequence learning problem, and build a novel system based on the RNN Encoder-Decoder structure to generate quatrains (Jueju in Chinese), with a topic word as input.