Search Results for author: Biqing Qi

Found 8 papers, 2 papers with code

On Large Language Models' Hallucination with Regard to Known Facts

no code implementations29 Mar 2024 Che Jiang, Biqing Qi, Xiangyu Hong, Dayuan Fu, Yang Cheng, Fandong Meng, Mo Yu, BoWen Zhou, Jie zhou

In hallucinated cases, the output token's information rarely demonstrates abrupt increases and consistent superiority in the later stages of the model.

Hallucination

Contrastive Augmented Graph2Graph Memory Interaction for Few Shot Continual Learning

no code implementations7 Mar 2024 Biqing Qi, Junqi Gao, Xingquan Chen, Dong Li, Jianxing Liu, Ligang Wu, BoWen Zhou

However, current EM-based methods retrieves memory globally by performing Vector-to-Vector (V2V) interaction between features corresponding to the input and prototypes stored in EM, neglecting the geometric structure of local features.

Few-Shot Class-Incremental Learning Few-Shot Learning +1

Interactive Continual Learning: Fast and Slow Thinking

1 code implementation5 Mar 2024 Biqing Qi, Xingquan Chen, Junqi Gao, Dong Li, Jianxing Liu, Ligang Wu, BoWen Zhou

Drawing on Complementary Learning System theory, this paper presents a novel Interactive Continual Learning (ICL) framework, enabled by collaborative interactions among models of various sizes.

Continual Learning Outlier Detection +1

CoGenesis: A Framework Collaborating Large and Small Language Models for Secure Context-Aware Instruction Following

no code implementations5 Mar 2024 Kaiyan Zhang, Jianyu Wang, Ermo Hua, Biqing Qi, Ning Ding, BoWen Zhou

With the advancement of language models (LMs), their exposure to private data is increasingly inevitable, and their deployment (especially for smaller ones) on personal devices, such as PCs and smartphones, has become a prevailing trend.

Instruction Following

Investigating Deep Watermark Security: An Adversarial Transferability Perspective

no code implementations26 Feb 2024 Biqing Qi, Junqi Gao, Yiang Luo, Jianxing Liu, Ligang Wu, BoWen Zhou

The rise of generative neural networks has triggered an increased demand for intellectual property (IP) protection in generated content.

Large Language Models are Zero Shot Hypothesis Proposers

no code implementations10 Nov 2023 Biqing Qi, Kaiyan Zhang, Haoxiang Li, Kai Tian, Sihang Zeng, Zhang-Ren Chen, BoWen Zhou

We subsequently evaluate the hypothesis generation capabilities of various top-tier instructed models in zero-shot, few-shot, and fine-tuning settings, including both closed and open-source LLMs.

CRaSh: Clustering, Removing, and Sharing Enhance Fine-tuning without Full Large Language Model

no code implementations24 Oct 2023 Kaiyan Zhang, Ning Ding, Biqing Qi, Xuekai Zhu, Xinwei Long, BoWen Zhou

Instruction tuning has recently been recognized as an effective way of aligning Large Language Models (LLMs) to enhance their generalization ability across various tasks.

Clustering Language Modelling +1

PaD: Program-aided Distillation Can Teach Small Models Reasoning Better than Chain-of-thought Fine-tuning

1 code implementation23 May 2023 Xuekai Zhu, Biqing Qi, Kaiyan Zhang, Xinwei Long, Zhouhan Lin, BoWen Zhou

While large language models (LLMs) excel in various natural language processing tasks, their huge size and the inaccessibility of parameters present challenges for practical deployment.

Arithmetic Reasoning GSM8K +1

Cannot find the paper you are looking for? You can Submit a new open access paper.