Search Results for author: Yiming Cui

Found 66 papers, 36 papers with code

Rethinking LLM Language Adaptation: A Case Study on Chinese Mixtral

1 code implementation4 Mar 2024 Yiming Cui, Xin Yao

Mixtral, a representative sparse mixture of experts (SMoE) language model, has received significant attention due to its unique model design and superior performance.

Language Modelling

Facing the Elephant in the Room: Visual Prompt Tuning or Full Finetuning?

no code implementations23 Jan 2024 Cheng Han, Qifan Wang, Yiming Cui, Wenguan Wang, Lifu Huang, Siyuan Qi, Dongfang Liu

As the scale of vision models continues to grow, the emergence of Visual Prompt Tuning (VPT) as a parameter-efficient transfer learning technique has gained attention due to its superior performance compared to traditional full-finetuning.

Transfer Learning Visual Prompt Tuning

ClusterFormer: Clustering As A Universal Visual Learner

1 code implementation22 Sep 2023 James C. Liang, Yiming Cui, Qifan Wang, Tong Geng, Wenguan Wang, Dongfang Liu

This paper presents CLUSTERFORMER, a universal vision model that is based on the CLUSTERing paradigm with TransFORMER.

Clustering Image Classification +7

E^2VPT: An Effective and Efficient Approach for Visual Prompt Tuning

1 code implementation ICCV 2023 Cheng Han, Qifan Wang, Yiming Cui, Zhiwen Cao, Wenguan Wang, Siyuan Qi, Dongfang Liu

Specifically, we introduce a set of learnable key-value prompts and visual prompts into self-attention and input layers, respectively, to improve the effectiveness of model fine-tuning.

Visual Prompt Tuning

Learning Dynamic Query Combinations for Transformer-based Object Detection and Segmentation

1 code implementation23 Jul 2023 Yiming Cui, Linjie Yang, Haichao Yu

Transformer-based detection and segmentation methods use a list of learned detection queries to retrieve information from the transformer network and learn to predict the location and category of one specific object from each query.

Instance Segmentation Object +5

IDOL: Indicator-oriented Logic Pre-training for Logical Reasoning

1 code implementation27 Jun 2023 Zihang Xu, Ziqing Yang, Yiming Cui, Shijin Wang

IDOL achieves state-of-the-art performance on ReClor and LogiQA, the two most representative benchmarks in logical reasoning MRC, and is proven to be capable of generalizing to different pre-trained models and other types of MRC benchmarks like RACE and SQuAD 2. 0 while keeping competitive general language understanding ability through testing on tasks in GLUE.

Logical Reasoning Machine Reading Comprehension

Efficient and Effective Text Encoding for Chinese LLaMA and Alpaca

5 code implementations17 Apr 2023 Yiming Cui, Ziqing Yang, Xin Yao

While several large language models, such as LLaMA, have been open-sourced by the community, these predominantly focus on English corpora, limiting their usefulness for other languages.

Communication-efficient Personalized Federated Edge Learning for Massive MIMO CSI Feedback

no code implementations24 Mar 2023 Yiming Cui, Jiajia Guo, Chao-Kai Wen, Shi Jin

Additionally, since the heterogeneity of CSI datasets in different UEs can degrade the performance of the FEEL-based framework, we introduce a personalization strategy to improve feedback performance.

FAQ: Feature Aggregated Queries for Transformer-based Video Object Detectors

1 code implementation15 Mar 2023 Yiming Cui, Linjie Yang

With Transformerbased object detectors getting a better performance on the image domain tasks, recent works began to extend those methods to video object detection.

Object object-detection +1

One Neuron Saved Is One Neuron Earned: On Parametric Efficiency of Quadratic Networks

1 code implementation11 Mar 2023 Feng-Lei Fan, Hang-Cheng Dong, Zhongming Wu, Lecheng Ruan, Tieyong Zeng, Yiming Cui, Jing-Xiao Liao

In this paper, with theoretical and empirical studies, we show that quadratic networks enjoy parametric efficiency, thereby confirming that the superior performance of quadratic networks is due to the intrinsic expressive capability.

Feature Aggregated Queries for Transformer-Based Video Object Detectors

1 code implementation CVPR 2023 Yiming Cui

With Transformer-based object detectors getting a better performance on the image domain tasks, recent works began to extend those methods to video object detection.

Object object-detection +1

Gradient-based Intra-attention Pruning on Pre-trained Language Models

1 code implementation15 Dec 2022 Ziqing Yang, Yiming Cui, Xin Yao, Shijin Wang

In this work, we propose a structured pruning method GRAIN (Gradient-based Intra-attention pruning), which performs task-specific pruning with knowledge distillation and yields highly effective models.

Knowledge Distillation

LERT: A Linguistically-motivated Pre-trained Language Model

1 code implementation10 Nov 2022 Yiming Cui, Wanxiang Che, Shijin Wang, Ting Liu

We propose LERT, a pre-trained language model that is trained on three types of linguistic features along with the original MLM pre-training task, using a linguistically-informed pre-training (LIP) strategy.

Language Modelling Stock Market Prediction +1

Lightweight Neural Network with Knowledge Distillation for CSI Feedback

no code implementations31 Oct 2022 Yiming Cui, Jiajia Guo, Zheng Cao, Huaze Tang, Chao-Kai Wen, Shi Jin, Xin Wang, Xiaolin Hou

Firstly, an autoencoder KD-based method is introduced by training a student autoencoder to mimic the reconstructed CSI of a pretrained teacher autoencoder.

Knowledge Distillation

DFA: Dynamic Feature Aggregation for Efficient Video Object Detection

no code implementations2 Oct 2022 Yiming Cui

Video object detection is a fundamental yet challenging task in computer vision.

Object object-detection +1

Dynamic Proposals for Efficient Object Detection

no code implementations12 Jul 2022 Yiming Cui, Linjie Yang, Ding Liu

Object detection is a basic computer vision task to loccalize and categorize objects in a given image.

Object object-detection +1

GL-RG: Global-Local Representation Granularity for Video Captioning

1 code implementation22 May 2022 Liqi Yan, Qifan Wang, Yiming Cui, Fuli Feng, Xiaojun Quan, Xiangyu Zhang, Dongfang Liu

Video captioning is a challenging task as it needs to accurately transform visual understanding into natural language description.

Caption Generation Descriptive +1

TextPruner: A Model Pruning Toolkit for Pre-Trained Language Models

no code implementations ACL 2022 Ziqing Yang, Yiming Cui, Zhigang Chen

Pre-trained language models have been prevailed in natural language processing and become the backbones of many NLP tasks, but the demands for computational resources have limited their applications.

Model Compression

PERT: Pre-training BERT with Permuted Language Model

1 code implementation14 Mar 2022 Yiming Cui, Ziqing Yang, Ting Liu

We permute a proportion of the input text, and the training objective is to predict the position of the original token.

Language Modelling Natural Language Understanding +1

Cross-Lingual Text Classification with Multilingual Distillation and Zero-Shot-Aware Training

no code implementations28 Feb 2022 Ziqing Yang, Yiming Cui, Zhigang Chen, Shijin Wang

In this paper, we aim to improve the multilingual model's supervised and zero-shot performance simultaneously only with the resources from supervised languages.

Language Modelling text-classification +1

DG-Labeler and DGL-MOTS Dataset: Boost the Autonomous Driving Perception

no code implementations15 Oct 2021 Yiming Cui, Zhiwen Cao, Yixin Xie, Xingyu Jiang, Feng Tao, Yingjie Chen, Lin Li, Dongfang Liu

The existing MOTS studies face two critical challenges: 1) the published datasets inadequately capture the real-world complexity for network training to address various driving settings; 2) the working pipeline annotation tool is under-studied in the literature to improve the quality of MOTS learning examples.

Autonomous Driving Multi-Object Tracking +1

Multilingual Multi-Aspect Explainability Analyses on Machine Reading Comprehension Models

no code implementations26 Aug 2021 Yiming Cui, Wei-Nan Zhang, Wanxiang Che, Ting Liu, Zhigang Chen, Shijin Wang

Achieving human-level performance on some of the Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained Language Models (PLMs).

Machine Reading Comprehension Question Answering +1

TF-Blender: Temporal Feature Blender for Video Object Detection

1 code implementation ICCV 2021 Yiming Cui, Liqi Yan, Zhiwen Cao, Dongfang Liu

One of the popular solutions is to exploit the temporal information and enhance per-frame representation through aggregating features from neighboring frames.

Object object-detection +1

ExpMRC: Explainability Evaluation for Machine Reading Comprehension

1 code implementation10 May 2021 Yiming Cui, Ting Liu, Wanxiang Che, Zhigang Chen, Shijin Wang

Achieving human-level performance on some of Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained Language Models (PLMs).

Machine Reading Comprehension Multi-Choice MRC +2

Hierarchical Attention Fusion for Geo-Localization

1 code implementation18 Feb 2021 Liqi Yan, Yiming Cui, Yingjie Chen, Dongfang Liu

We extract the hierarchical feature maps from a convolutional neural network (CNN) and organically fuse the extracted features for image representations.

Image Retrieval Retrieval

Memory Augmented Sequential Paragraph Retrieval for Multi-hop Question Answering

no code implementations7 Feb 2021 Nan Shao, Yiming Cui, Ting Liu, Shijin Wang, Guoping Hu

To deal with this challenge, most of the existing works consider paragraphs as nodes in a graph and propose graph-based methods to retrieve them.

Information Retrieval Multi-hop Question Answering +2

Unsupervised Explanation Generation for Machine Reading Comprehension

no code implementations13 Nov 2020 Yiming Cui, Ting Liu, Shijin Wang, Guoping Hu

With the blooming of various Pre-trained Language Models (PLMs), Machine Reading Comprehension (MRC) has embraced significant improvements on various benchmarks and even surpass human performances.

Explanation Generation Machine Reading Comprehension +1

CharBERT: Character-aware Pre-trained Language Model

1 code implementation COLING 2020 Wentao Ma, Yiming Cui, Chenglei Si, Ting Liu, Shijin Wang, Guoping Hu

Most pre-trained language models (PLMs) construct word representations at subword level with Byte-Pair Encoding (BPE) or its variations, by which OOV (out-of-vocab) words are almost avoidable.

Language Modelling Question Answering +3

Revisiting Pre-Trained Models for Chinese Natural Language Processing

6 code implementations Findings of the Association for Computational Linguistics 2020 Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Shijin Wang, Guoping Hu

Bidirectional Encoder Representations from Transformers (BERT) has shown marvelous improvements across various NLP tasks, and consecutive variants have been proposed to further improve the performance of the pre-trained language models.

Language Modelling Stock Market Prediction

Benchmarking Robustness of Machine Reading Comprehension Models

1 code implementation Findings (ACL) 2021 Chenglei Si, Ziqing Yang, Yiming Cui, Wentao Ma, Ting Liu, Shijin Wang

To fill this important gap, we construct AdvRACE (Adversarial RACE), a new model-agnostic benchmark for evaluating the robustness of MRC models under four different types of adversarial attacks, including our novel distractor extraction and generation attacks.

Benchmarking Machine Reading Comprehension +1

Is Graph Structure Necessary for Multi-hop Question Answering?

no code implementations EMNLP 2020 Nan Shao, Yiming Cui, Ting Liu, Shijin Wang, Guoping Hu

We construct a strong baseline model to establish that, with the proper use of pre-trained models, graph structure may not be necessary for multi-hop question answering.

Graph Attention Multi-hop Question Answering +1

A Sentence Cloze Dataset for Chinese Machine Reading Comprehension

1 code implementation COLING 2020 Yiming Cui, Ting Liu, Ziqing Yang, Zhipeng Chen, Wentao Ma, Wanxiang Che, Shijin Wang, Guoping Hu

To add diversity in this area, in this paper, we propose a new task called Sentence Cloze-style Machine Reading Comprehension (SC-MRC).

Machine Reading Comprehension Sentence

Discriminative Sentence Modeling for Story Ending Prediction

no code implementations19 Dec 2019 Yiming Cui, Wanxiang Che, Wei-Nan Zhang, Ting Liu, Shijin Wang, Guoping Hu

Story Ending Prediction is a task that needs to select an appropriate ending for the given story, which requires the machine to understand the story and sometimes needs commonsense knowledge.

Cloze Test Sentence

CJRC: A Reliable Human-Annotated Benchmark DataSet for Chinese Judicial Reading Comprehension

no code implementations19 Dec 2019 Xingyi Duan, Baoxin Wang, Ziyue Wang, Wentao Ma, Yiming Cui, Dayong Wu, Shijin Wang, Ting Liu, Tianxiang Huo, Zhen Hu, Heng Wang, Zhiyuan Liu

We present a Chinese judicial reading comprehension (CJRC) dataset which contains approximately 10K documents and almost 50K questions with answers.

Machine Reading Comprehension

Contextual Recurrent Units for Cloze-style Reading Comprehension

no code implementations14 Nov 2019 Yiming Cui, Wei-Nan Zhang, Wanxiang Che, Ting Liu, Zhipeng Chen, Shijin Wang, Guoping Hu

Recurrent Neural Networks (RNN) are known as powerful models for handling sequential data, and especially widely utilized in various natural language processing tasks.

Reading Comprehension Sentence +2

Improving Machine Reading Comprehension via Adversarial Training

no code implementations9 Nov 2019 Ziqing Yang, Yiming Cui, Wanxiang Che, Ting Liu, Shijin Wang, Guoping Hu

With virtual adversarial training (VAT), we explore the possibility of improving the RC models with semi-supervised learning and prove that examples from a different task are also beneficial.

General Classification Image Classification +3

Pre-Training with Whole Word Masking for Chinese BERT

2 code implementations19 Jun 2019 Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang

To demonstrate the effectiveness of these models, we create a series of Chinese pre-trained language models as our baselines, including BERT, RoBERTa, ELECTRA, RBT, etc.

Document Classification General Classification +5

Exploiting Persona Information for Diverse Generation of Conversational Responses

1 code implementation29 May 2019 Haoyu Song, Wei-Nan Zhang, Yiming Cui, Dong Wang, Ting Liu

Giving conversational context with persona information to a chatbot, how to exploit the information to generate diverse and sustainable conversations is still a non-trivial task.

Chatbot

Convolutional Spatial Attention Model for Reading Comprehension with Multiple-Choice Questions

no code implementations21 Nov 2018 Zhipeng Chen, Yiming Cui, Wentao Ma, Shijin Wang, Guoping Hu

Machine Reading Comprehension (MRC) with multiple-choice questions requires the machine to read given passage and select the correct answer among several candidates.

Machine Reading Comprehension Multiple-choice

Context-Sensitive Generation of Open-Domain Conversational Responses

no code implementations COLING 2018 Wei-Nan Zhang, Yiming Cui, Yifa Wang, Qingfu Zhu, Lingzhi Li, Lianqiang Zhou, Ting Liu

Despite the success of existing works on single-turn conversation generation, taking the coherence in consideration, human conversing is actually a context-sensitive process.

Information Retrieval Machine Translation

HFL-RC System at SemEval-2018 Task 11: Hybrid Multi-Aspects Model for Commonsense Reading Comprehension

no code implementations15 Mar 2018 Zhipeng Chen, Yiming Cui, Wentao Ma, Shijin Wang, Ting Liu, Guoping Hu

This paper describes the system which got the state-of-the-art results at SemEval-2018 Task 11: Machine Comprehension using Commonsense Knowledge.

Multiple-choice Reading Comprehension

Generating and Exploiting Large-scale Pseudo Training Data for Zero Pronoun Resolution

no code implementations ACL 2017 Ting Liu, Yiming Cui, Qingyu Yin, Wei-Nan Zhang, Shijin Wang, Guoping Hu

Most existing approaches for zero pronoun resolution are heavily relying on annotated data, which is often released by shared task organizers.

Reading Comprehension

Augmenting Phrase Table by Employing Lexicons for Pivot-based SMT

no code implementations1 Dec 2015 Yiming Cui, Conghui Zhu, Xiaoning Zhu, Tiejun Zhao

Pivot language is employed as a way to solve the data sparseness problem in machine translation, especially when the data for a particular language pair does not exist.

Machine Translation Translation

LSTM Neural Reordering Feature for Statistical Machine Translation

no code implementations NAACL 2016 Yiming Cui, Shijin Wang, Jianfeng Li

Artificial neural networks are powerful models, which have been widely applied into many aspects of machine translation, such as language modeling and translation modeling.

Language Modelling Machine Translation +1

Cannot find the paper you are looking for? You can Submit a new open access paper.