1 code implementation • EACL (BEA) 2021 • Zheng Yuan, Christopher Bryant
Document-level context can provide valuable information in grammatical error correction (GEC), which is crucial for correcting certain errors and resolving inconsistencies.
1 code implementation • EMNLP 2021 • Zheng Yuan, Shiva Taslimipoor, Christopher Davis, Christopher Bryant
In this paper, we show how a multi-class grammatical error detection (GED) system can be used to improve grammatical error correction (GEC) for English.
no code implementations • LREC (MWE) 2022 • Shiva Taslimipoor, Christopher Bryant, Zheng Yuan
Grammatical error correction (GEC) is the task of automatically correcting errors in text.
1 code implementation • 18 Apr 2024 • Kelvin Wey Han Chan, Christopher Bryant, Li Nguyen, Andrew Caines, Zheng Yuan
Through this exploration, we propose a novel method of generating synthetic CSW GEC datasets by translating different spans of text within existing GEC corpora.
no code implementations • 18 Apr 2024 • Siya Qi, Yulan He, Zheng Yuan
Hallucination in Natural Language Generation (NLG) is like the elephant in the room, obvious but often overlooked until recent achievements significantly improved the fluency and grammatical accuracy of generated text.
no code implementations • 16 Apr 2024 • Zheng Yuan, Štefan Beňuš, Alessandro D'Ausilio
This study explores F0 entrainment in second language (L2) English speech imitation during an Alternating Reading Task (ART).
no code implementations • 3 Apr 2024 • Zheng Yuan, Dorina De Jong, Štefan Beňuš, Noël Nguyen, Ruitao Feng, Róbert Sabo, Luciano Fadiga, Alessandro D`Ausilio
We introduce the Alternating Reading Task (ART) Corpus, a collection of dyadic sentence reading for studying the entrainment and imitation behaviour in speech communication.
no code implementations • 28 Feb 2024 • Qiao Wang, Zheng Yuan
In this study, we evaluated the performance of the state-of-the-art sequence tagging grammar error detection and correction model (SeqTagger) using Japanese university students' writing samples.
no code implementations • 22 Feb 2024 • Keren Tan, Kangyang Luo, Yunshi Lan, Zheng Yuan, Jinlong Shu
Lexical Simplification (LS) aims to simplify text at the lexical level.
no code implementations • 18 Feb 2024 • Zijin Hong, Zheng Yuan, Hao Chen, Qinggang Zhang, Feiran Huang, Xiao Huang
Generating accurate SQL for user queries (text-to-SQL) is a long-standing problem since the generation of the SQL requires comprehending the query and database and retrieving the accurate data from the database accordingly.
no code implementations • 12 Feb 2024 • Yijie Zhang, Yuanchen Bei, Hao Chen, Qijie Shen, Zheng Yuan, Huan Gong, Senzhang Wang, Feiran Huang, Xiao Huang
POG defines the partial order relation of multiple behaviors and models behavior combinations as weighted edges to merge separate behavior graphs into a joint POG.
no code implementations • 15 Jan 2024 • Christopher Davis, Andrew Caines, Øistein Andersen, Shiva Taslimipoor, Helen Yannakoudakis, Zheng Yuan, Christopher Bryant, Marek Rei, Paula Buttery
Thanks to recent advances in generative AI, we are able to prompt large language models (LLMs) to produce texts which are fluent and grammatical.
1 code implementation • 9 Jan 2024 • Sibo Wang, Jie Zhang, Zheng Yuan, Shiguang Shan
Specifically, PMG-AFT minimizes the distance between the features of adversarial examples in the target model and those in the pre-trained model, aiming to preserve the generalization features already captured by the pre-trained model.
no code implementations • 3 Jan 2024 • Zheng Yuan, Jie Zhang, Yude Wang, Shiguang Shan, Xilin Chen
The attention mechanism has been proven effective on various visual tasks in recent years.
no code implementations • 3 Jan 2024 • Zheng Yuan, Jie Zhang, Shiguang Shan
In recent years, the Vision Transformer (ViT) model has gradually become mainstream in various computer vision tasks, and the robustness of the model has received increasing attention.
no code implementations • 15 Nov 2023 • Hongyi Yuan, Keming Lu, Fei Huang, Zheng Yuan, Chang Zhou
Large language models~(LLMs) exhibit exceptional performance in language tasks, yet their auto-regressive inference is limited due to high computational requirements and is sub-optimal due to the exposure bias.
no code implementations • 15 Nov 2023 • Keming Lu, Hongyi Yuan, Runji Lin, Junyang Lin, Zheng Yuan, Chang Zhou, Jingren Zhou
Zooter shows computation efficiency in inference as it introduces only a minor computation overhead of a routing function compared with reward model ranking methods.
1 code implementation • 7 Nov 2023 • Enhong Liu, Joseph Suarez, Chenhui You, Bo Wu, BingCheng Chen, Jun Hu, Jiaxin Chen, Xiaolong Zhu, Clare Zhu, Julian Togelius, Sharada Mohanty, Weijun Hong, Rui Du, Yibing Zhang, Qinwen Wang, Xinhang Li, Zheng Yuan, Xiang Li, Yuejia Huang, Kun Zhang, Hanhui Yang, Shiqi Tang, Phillip Isola
In this paper, we present the results of the NeurIPS-2022 Neural MMO Challenge, which attracted 500 participants and received over 1, 600 submissions.
1 code implementation • 25 Oct 2023 • Mingfeng Xue, Dayiheng Liu, Kexin Yang, Guanting Dong, Wenqiang Lei, Zheng Yuan, Chang Zhou, Jingren Zhou
Furthermore, we assemble three test sets for comprehensive evaluation, an occu-test set covering 25 occupational categories, an estate set focusing on real estate, and an occu-quora set containing real-world questions from Quora.
1 code implementation • 20 Oct 2023 • Andrea Sottana, Bin Liang, Kai Zou, Zheng Yuan
Large Language Models (LLMs) evaluation is a patchy and inconsistent landscape, and it is becoming clear that the quality of automatic evaluation metrics is not keeping up with the pace of development of generative models.
2 code implementations • 9 Oct 2023 • Guanting Dong, Hongyi Yuan, Keming Lu, Chengpeng Li, Mingfeng Xue, Dayiheng Liu, Wei Wang, Zheng Yuan, Chang Zhou, Jingren Zhou
We propose four intriguing research questions to explore the association between model performance and various factors including data amount, composition ratio, model size and SFT strategies.
1 code implementation • 9 Oct 2023 • Chengpeng Li, Zheng Yuan, Hongyi Yuan, Guanting Dong, Keming Lu, Jiancan Wu, Chuanqi Tan, Xiang Wang, Chang Zhou
In this paper, we conduct an investigation for such data augmentation in math reasoning and are intended to answer: (1) What strategies of data augmentation are more effective; (2) What is the scaling relationship between the amount of augmented data and model performance; and (3) Can data augmentation incentivize generalization to out-of-domain mathematical reasoning tasks?
Ranked #50 on Math Word Problem Solving on MATH (using extra training data)
2 code implementations • 28 Sep 2023 • Jinze Bai, Shuai Bai, Yunfei Chu, Zeyu Cui, Kai Dang, Xiaodong Deng, Yang Fan, Wenbin Ge, Yu Han, Fei Huang, Binyuan Hui, Luo Ji, Mei Li, Junyang Lin, Runji Lin, Dayiheng Liu, Gao Liu, Chengqiang Lu, Keming Lu, Jianxin Ma, Rui Men, Xingzhang Ren, Xuancheng Ren, Chuanqi Tan, Sinan Tan, Jianhong Tu, Peng Wang, Shijie Wang, Wei Wang, Shengguang Wu, Benfeng Xu, Jin Xu, An Yang, Hao Yang, Jian Yang, Shusheng Yang, Yang Yao, Bowen Yu, Hongyi Yuan, Zheng Yuan, Jianwei Zhang, Xingxuan Zhang, Yichang Zhang, Zhenru Zhang, Chang Zhou, Jingren Zhou, Xiaohuan Zhou, Tianhang Zhu
Large language models (LLMs) have revolutionized the field of artificial intelligence, enabling natural language processing tasks that were previously thought to be exclusive to humans.
Ranked #3 on Multi-Label Text Classification on CC3M-TagMask
2 code implementations • 14 Sep 2023 • JiaQi Zhang, Yu Cheng, Yongxin Ni, Yunzhu Pan, Zheng Yuan, Junchen Fu, Youhua Li, Jie Wang, Fajie Yuan
The development of TransRec has encountered multiple challenges, among which the lack of large-scale, high-quality transfer learning recommendation dataset and benchmark suites is one of the biggest obstacles.
1 code implementation • 14 Aug 2023 • Keming Lu, Hongyi Yuan, Zheng Yuan, Runji Lin, Junyang Lin, Chuanqi Tan, Chang Zhou, Jingren Zhou
Based on this observation, we propose a data selector based on InsTag to select 6K diverse and complex samples from open-source datasets and fine-tune models on InsTag-selected data.
1 code implementation • 3 Aug 2023 • Zheng Yuan, Hongyi Yuan, Chengpeng Li, Guanting Dong, Keming Lu, Chuanqi Tan, Chang Zhou, Jingren Zhou
We find with augmented samples containing more distinct reasoning paths, RFT improves mathematical reasoning performance more for LLMs.
Ranked #100 on Arithmetic Reasoning on GSM8K (using extra training data)
no code implementations • 17 Jul 2023 • Andrew Caines, Luca Benedetto, Shiva Taslimipoor, Christopher Davis, Yuan Gao, Oeistein Andersen, Zheng Yuan, Mark Elliott, Russell Moore, Christopher Bryant, Marek Rei, Helen Yannakoudakis, Andrew Mullooly, Diane Nicholls, Paula Buttery
The recent release of very large language models such as PaLM and GPT-4 has made an unprecedented impact in the popular media and public consciousness, giving rise to a mixture of excitement and fear as to their capabilities and potential uses, and shining a light on natural language processing research which had not previously received so much attention.
no code implementations • 1 Jul 2023 • Bryan Cai, Sihang Zeng, Yucong Lin, Zheng Yuan, Doudou Zhou, Lu Tian
Electronic health records (EHR) contain narrative notes that provide extensive details on the medical condition and management of patients.
no code implementations • 24 Jun 2023 • Lei Huang, Zheng Yuan, Huihui Yan, Rong Sheng, Linjing Liu, Fuzhou Wang, Weidun Xie, Nanjun Chen, Fei Huang, Songfang Huang, Ka-Chun Wong, Yaoyun Zhang
However, molecule generation targeted for dual protein targets still faces formidable challenges including protein 3D structure data requisition for model training, auto-regressive sampling, and model generalization for unseen targets.
no code implementations • 12 Jun 2023 • Anaïs Tack, Ekaterina Kochmar, Zheng Yuan, Serge Bibauw, Chris Piech
This paper describes the results of the first shared task on the generation of teacher responses in educational dialogues.
no code implementations • 8 Jun 2023 • Adaeze Adigwe, Zheng Yuan
This paper presents the ADAIO team's system entry in the Building Educational Applications (BEA) 2023 Shared Task on Generating AI Teacher Responses in Educational Dialogues.
1 code implementation • 8 Jun 2023 • Xinhang Li, Yiying Yang, Zheng Yuan, Zhe Wang, Qinwen Wang, Chen Xu, Lei LI, Jianhua He, Lin Zhang
For the more challenging problem of pursuing multiple evading vehicles, these algorithms typically select a fixed target evading vehicle for pursuing vehicles without considering dynamic traffic situation, which significantly reduces pursuing success rate.
1 code implementation • 8 Jun 2023 • Zheng Yuan, Aldo Pastore, Dorina De Jong, Hao Xu, Luciano Fadiga, Alessandro D'Ausilio
Phonetic convergence describes the automatic and unconscious speech adaptation of two interlocutors in a conversation.
1 code implementation • 24 May 2023 • Junchen Fu, Fajie Yuan, Yu Song, Zheng Yuan, Mingyue Cheng, Shenghui Cheng, JiaQi Zhang, Jie Wang, Yunzhu Pan
If yes, we benchmark these existing adapters, which have been shown to be effective in NLP and CV tasks, in item recommendation tasks.
no code implementations • 19 May 2023 • Ruyu Li, Wenhao Deng, Yu Cheng, Zheng Yuan, JiaQi Zhang, Fajie Yuan
Furthermore, we compare the performance of the TCF paradigm utilizing the most powerful LMs to the currently dominant ID embedding-based paradigm and investigate the transferability of this TCF paradigm.
no code implementations • 24 Apr 2023 • Chen Zhao, Wei-Ling Cai, Zheng Yuan, Cheng-Wei Hu
Recently, unsupervised image-to-image translation methods based on contrastive learning have achieved state-of-the-art results in many tasks.
1 code implementation • 22 Apr 2023 • Chen Zhao, Wei-Ling Cai, Zheng Yuan
Similarly, we design a style contrastive loss to improve the global texture information of the generated images.
1 code implementation • 11 Apr 2023 • Zheng Yuan, Hongyi Yuan, Chuanqi Tan, Wei Wang, Songfang Huang, Fei Huang
Reinforcement Learning from Human Feedback (RLHF) facilitates the alignment of large language models with human preferences, significantly enhancing the quality of interactions between humans and models.
1 code implementation • 24 Mar 2023 • Zheng Yuan, Fajie Yuan, Yu Song, Youhua Li, Junchen Fu, Fei Yang, Yunzhu Pan, Yongxin Ni
In fact, this question was answered ten years ago when IDRec beats MoRec by a strong margin in both recommendation accuracy and efficiency.
1 code implementation • 18 Mar 2023 • Hongyi Yuan, Keming Lu, Zheng Yuan
Biomedical entity linking (EL) consists of named entity recognition (NER) and named entity disambiguation (NED).
1 code implementation • 16 Mar 2023 • Zheng Yuan, Hongyi Yuan, Chuanqi Tan, Wei Wang, Songfang Huang
Large language models have emerged abilities including chain-of-thought to answer math word problems step by step.
1 code implementation • 1 Mar 2023 • Zheng Yuan, Qiao Jin, Chuanqi Tan, Zhengyun Zhao, Hongyi Yuan, Fei Huang, Songfang Huang
We propose to retrieve similar image-text pairs based on ITC from pretraining datasets and introduce a novel retrieval-attention module to fuse the representation of the image and the question with the retrieved images and texts.
2 code implementations • 12 Feb 2023 • Stuart Mesham, Christopher Bryant, Marek Rei, Zheng Yuan
We extend a current sequence-tagging approach to Grammatical Error Correction (GEC) by introducing specialised tags for spelling correction and morphological inflection using the SymSpell and LemmInflect algorithms.
no code implementations • 2 Feb 2023 • Zheng Yuan, Yaoyun Zhang, Chuanqi Tan, Wei Wang, Fei Huang, Songfang Huang
To alleviate this limitation, we propose Moleformer, a novel Transformer architecture that takes nodes (atoms) and edges (bonds and nonbonding atom pairs) as inputs and models the interactions among them using rotational and translational invariant geometry-aware spatial encoding.
1 code implementation • 20 Dec 2022 • Hongyi Yuan, Zheng Yuan, Chuanqi Tan, Fei Huang, Songfang Huang
We propose SeqDiffuSeq, a text diffusion model for sequence-to-sequence generation.
1 code implementation • 17 Dec 2022 • Hongyi Yuan, Zheng Yuan, Chuanqi Tan, Fei Huang, Songfang Huang
Unlike previous works that only add noise to inputs or parameters, we argue that the hidden representations of Transformers layers convey more diverse and meaningful language information.
no code implementations • 9 Nov 2022 • Christopher Bryant, Zheng Yuan, Muhammad Reza Qorib, Hannan Cao, Hwee Tou Ng, Ted Briscoe
Grammatical Error Correction (GEC) is the task of automatically detecting and correcting errors in text.
1 code implementation • 24 Oct 2022 • Yiying Yang, Xinhang Li, Zheng Yuan, Qinwen Wang, Chen Xu, Lin Zhang
However, existing works on MVP pay little attention to the importance of information exchange and cooperation among pursuing vehicles under the complex urban traffic environment.
1 code implementation • NAACL 2022 • Hongyi Yuan, Zheng Yuan, Sheng Yu
Entities lie in the heart of biomedical natural language understanding, and the biomedical entity linking (EL) task remains challenging due to the fine-grained and diversiform concept names.
1 code implementation • 10 Apr 2022 • Xinhang Li, Zihao Li, Nan Yang, Zheng Yuan, Qinwen Wang, Yiying Yang, Yupeng Huang, Xuri Song, Lei LI, Lin Zhang
The expansion of renewable energy could help realizing the goals of peaking carbon dioxide emissions and carbon neutralization.
1 code implementation • BioNLP (ACL) 2022 • Hongyi Yuan, Zheng Yuan, Ruyi Gan, Jiaxing Zhang, Yutao Xie, Sheng Yu
Furthermore, we conduct ablation studies on the pretraining tasks for BioBART and find that sentence permutation has negative effects on downstream tasks.
Ranked #3 on Entity Linking on MedMentions
1 code implementation • BioNLP (ACL) 2022 • Sihang Zeng, Zheng Yuan, Sheng Yu
Term clustering is important in biomedical knowledge graph construction.
no code implementations • 18 Mar 2022 • Sheng Yu, Zheng Yuan, Jun Xia, Shengxuan Luo, Huaiyuan Ying, Sihang Zeng, Jingyi Ren, Hongyi Yuan, Zhengyun Zhao, Yucong Lin, Keming Lu, Jing Wang, Yutao Xie, Heung-Yeung Shum
For decades, these knowledge graphs have been developed via expert curation; however, this method can no longer keep up with today's AI development, and a transition to algorithmically generated BioMedKGs is necessary.
1 code implementation • ACL 2022 • Zheng Yuan, Chuanqi Tan, Songfang Huang
Automatic ICD coding is defined as assigning disease codes to electronic medical records (EMRs).
Ranked #5 on Medical Code Prediction on MIMIC-III
1 code implementation • 1 Mar 2022 • Zheng Yuan, Tianhao Wu, Qinwen Wang, Yiying Yang, Lei LI, Lin Zhang
Although there are some achievements in the field of MVP in the open space environment, the urban area brings complicated road structures and restricted moving spaces as challenges to the resolution of MVP games.
2 code implementations • 27 Nov 2021 • Zheng Yuan, Jie Zhang, Shiguang Shan
Adversarial attacks provide a good way to study the robustness of deep learning models.
no code implementations • 27 Nov 2021 • Zheng Yuan, Jie Zhang, Zhaoyan Jiang, Liangliang Li, Shiguang Shan
Instead of using the sign function, we propose to directly utilize the exact gradient direction with a scaling factor for generating adversarial perturbations, which improves the attack success rates of adversarial examples even with fewer perturbations.
1 code implementation • Findings (ACL) 2022 • Zheng Yuan, Chuanqi Tan, Songfang Huang, Fei Huang
To fuse these heterogeneous factors, we propose a novel triaffine mechanism including triaffine attention and scoring.
Ranked #1 on Nested Named Entity Recognition on TAC-KBP 2017
no code implementations • 29 Sep 2021 • Zheng Yuan, Andre Esteva, ran Xu
We also curate a histopathology meta dataset - a benchmark dataset for training and validating models on out-of-distribution performance across a range of cancer types.
1 code implementation • ICCV 2021 • Zheng Yuan, Jie Zhang, Yunpei Jia, Chuanqi Tan, Tao Xue, Shiguang Shan
In recent years, research on adversarial attacks has become a hot spot.
no code implementations • SEMEVAL 2021 • Zheng Yuan, David Strohmaier
This paper describes the system of the Cambridge team submitted to the SemEval-2021 shared task on Multilingual and Cross-lingual Word-in-Context Disambiguation.
no code implementations • SEMEVAL 2021 • Zheng Yuan, Gladys Tyen, David Strohmaier
This paper describes our submission to the SemEval-2021 shared task on Lexical Complexity Prediction.
2 code implementations • ACL 2022 • Ningyu Zhang, Mosha Chen, Zhen Bi, Xiaozhuan Liang, Lei LI, Xin Shang, Kangping Yin, Chuanqi Tan, Jian Xu, Fei Huang, Luo Si, Yuan Ni, Guotong Xie, Zhifang Sui, Baobao Chang, Hui Zong, Zheng Yuan, Linfeng Li, Jun Yan, Hongying Zan, Kunli Zhang, Buzhou Tang, Qingcai Chen
Artificial Intelligence (AI), along with the recent progress in biomedical language understanding, is gradually changing medical practice.
Ranked #1 on Semantic Similarity on CHIP-STS
1 code implementation • NAACL (BioNLP) 2021 • Zheng Yuan, Yijia Liu, Chuanqi Tan, Songfang Huang, Fei Huang
To this end, we propose KeBioLM, a biomedical pretrained language model that explicitly leverages knowledge from the UMLS knowledge bases.
Ranked #1 on Named Entity Recognition (NER) on JNLPBA
no code implementations • 10 Feb 2021 • Qiao Jin, Zheng Yuan, Guangzhi Xiong, Qianlan Yu, Huaiyuan Ying, Chuanqi Tan, Mosha Chen, Songfang Huang, Xiaozhong Liu, Sheng Yu
Automatic Question Answering (QA) has been successfully applied in various domains such as search engines and chatbots.
1 code implementation • 3 Dec 2020 • Zheng Yuan, Jie Zhang, Shiguang Shan, Xilin Chen
Recent studies have shown remarkable success in face image generations.
1 code implementation • 5 Nov 2020 • Zheng Yuan, Zhengyun Zhao, Haixia Sun, Jiao Li, Fei Wang, Sheng Yu
This paper proposes CODER: contrastive learning on knowledge graphs for cross-lingual medical term representation.
no code implementations • WS 2019 • Zheng Yuan, Felix Stahlberg, Marek Rei, Bill Byrne, Helen Yannakoudakis
In this paper, we describe our submission to the BEA 2019 shared task on grammatical error correction.
no code implementations • WS 2018 • Zheng Yuan
This paper describes our use of two recurrent neural network sequence models: sequence labelling and sequence-to-sequence models, for the prediction of future learner errors in our submission to the 2018 Duolingo Shared Task on Second Language Acquisition Modeling (SLAM).
no code implementations • NAACL 2018 • Waleed Ammar, Dirk Groeneveld, Chandra Bhagavatula, Iz Beltagy, Miles Crawford, Doug Downey, Jason Dunkelberger, Ahmed Elgohary, Sergey Feldman, Vu Ha, Rodney Kinney, Sebastian Kohlmeier, Kyle Lo, Tyler Murray, Hsu-Han Ooi, Matthew Peters, Joanna Power, Sam Skjonsberg, Lucy Lu Wang, Chris Wilhelm, Zheng Yuan, Madeleine van Zuylen, Oren Etzioni
We describe a deployed scalable system for organizing published scientific literature into a heterogeneous graph to facilitate algorithmic manipulation and discovery.
no code implementations • EMNLP 2017 • Helen Yannakoudakis, Marek Rei, {\O}istein E. Andersen, Zheng Yuan
We propose an approach to N-best list reranking using neural sequence-labelling models.
no code implementations • WS 2017 • Marek Rei, Mariano Felice, Zheng Yuan, Ted Briscoe
Shortage of available training data is holding back progress in the area of automated error detection.
Ranked #3 on Grammatical Error Detection on FCE