Search Results for author: Yixuan Zhang

Found 36 papers, 12 papers with code

MathVC: An LLM-Simulated Multi-Character Virtual Classroom for Mathematics Education

no code implementations10 Apr 2024 Murong Yue, Wijdane Mifdal, Yixuan Zhang, Jennifer Suh, Ziyu Yao

Mathematical modeling (MM) is considered a fundamental skill for students in STEM disciplines.

Math

Prelimit Coupling and Steady-State Convergence of Constant-stepsize Nonsmooth Contractive SA

no code implementations9 Apr 2024 Yixuan Zhang, Dongyan Huo, Yudong Chen, Qiaomin Xie

Motivated by Q-learning, we study nonsmooth contractive stochastic approximation (SA) with constant stepsize.

Q-Learning

Against The Achilles' Heel: A Survey on Red Teaming for Generative Models

no code implementations31 Mar 2024 Lizhi Lin, Honglin Mu, Zenan Zhai, Minghan Wang, Yuxia Wang, Renxi Wang, Junjie Gao, Yixuan Zhang, Wanxiang Che, Timothy Baldwin, Xudong Han, Haonan Li

Generative models are rapidly gaining popularity and being integrated into everyday applications, raising concerns over their safety issues as various vulnerabilities are exposed.

Bias Mitigation in Fine-tuning Pre-trained Models for Enhanced Fairness and Efficiency

no code implementations1 Mar 2024 Yixuan Zhang, Feng Zhou

Fine-tuning pre-trained models is a widely employed technique in numerous real-world applications.

Fairness Transfer Learning

Confidence Matters: Revisiting Intrinsic Self-Correction Capabilities of Large Language Models

1 code implementation19 Feb 2024 Loka Li, Guangyi Chen, Yusheng Su, Zhenhao Chen, Yixuan Zhang, Eric Xing, Kun Zhang

We have experimentally observed that LLMs possess the capability to understand the "confidence" in their own responses.

Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language Models as Agents

1 code implementation18 Feb 2024 Renxi Wang, Haonan Li, Xudong Han, Yixuan Zhang, Timothy Baldwin

However, LLMs are optimized for language generation instead of tool use during training or alignment, limiting their effectiveness as agents.

Mathematical Reasoning Multi-hop Question Answering +2

Deep Equilibrium Models are Almost Equivalent to Not-so-deep Explicit Models for High-dimensional Gaussian Mixtures

no code implementations5 Feb 2024 Zenan Ling, Longbo Li, Zhanbo Feng, Yixuan Zhang, Feng Zhou, Robert C. Qiu, Zhenyu Liao

Deep equilibrium models (DEQs), as a typical implicit neural network, have demonstrated remarkable success on various tasks.

Constant Stepsize Q-learning: Distributional Convergence, Bias and Extrapolation

no code implementations25 Jan 2024 Yixuan Zhang, Qiaomin Xie

By connecting the constant stepsize Q-learning to a time-homogeneous Markov chain, we show the distributional convergence of the iterates in Wasserstein distance and establish its exponential convergence rate.

Q-Learning Reinforcement Learning (RL)

The Good, The Bad, and Why: Unveiling Emotions in Generative AI

no code implementations18 Dec 2023 Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Xinyi Wang, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie

Through extensive experiments involving language and multi-modal models on semantic understanding, logical reasoning, and generation tasks, we demonstrate that both textual and visual EmotionPrompt can boost the performance of AI models while EmotionAttack can hinder it.

Logical Reasoning

Mitigating Label Bias in Machine Learning: Fairness through Confident Learning

no code implementations14 Dec 2023 Yixuan Zhang, Boyu Li, Zenan Ling, Feng Zhou

In this paper, we demonstrate that despite only having access to the biased labels, it is possible to eliminate bias by filtering the fairest instances within the framework of confident learning.

Fairness

Leveraging Laryngograph Data for Robust Voicing Detection in Speech

1 code implementation5 Dec 2023 Yixuan Zhang, Heming Wang, DeLiang Wang

Accurately detecting voiced intervals in speech signals is a critical step in pitch tracking and has numerous applications.

Out-of-Distribution Knowledge Distillation via Confidence Amendment

1 code implementation14 Nov 2023 Zhilin Zhao, Longbing Cao, Yixuan Zhang

This paper introduces OOD knowledge distillation, a pioneering learning framework applicable whether or not training ID data is available, given a standard network.

Knowledge Distillation Out of Distribution (OOD) Detection

CompeteAI: Understanding the Competition Behaviors in Large Language Model-based Agents

no code implementations26 Oct 2023 Qinlin Zhao, Jindong Wang, Yixuan Zhang, Yiqiao Jin, Kaijie Zhu, Hao Chen, Xing Xie

Large language models (LLMs) have been widely used as agents to complete different tasks, such as personal assistance or event planning.

Language Modelling Large Language Model

Can Large Language Model Comprehend Ancient Chinese? A Preliminary Test on ACLUE

1 code implementation14 Oct 2023 Yixuan Zhang, Haonan Li

To bridge this gap, we present ACLUE, an evaluation benchmark designed to assess the capability of language models in comprehending ancient Chinese.

Language Modelling Large Language Model

MetaAgents: Simulating Interactions of Human Behaviors for LLM-based Task-oriented Coordination via Collaborative Generative Agents

1 code implementation10 Oct 2023 Yuan Li, Yixuan Zhang, Lichao Sun

We propose a novel framework that equips collaborative generative agents with human-like reasoning abilities and specialized skills.

Neural Network Augmented Kalman Filter for Robust Acoustic Howling Suppression

no code implementations27 Sep 2023 Yixuan Zhang, Hao Zhang, Meng Yu, Dong Yu

Acoustic howling suppression (AHS) is a critical challenge in audio communication systems.

Advancing Acoustic Howling Suppression through Recursive Training of Neural Networks

no code implementations27 Sep 2023 Hao Zhang, Yixuan Zhang, Meng Yu, Dong Yu

In this paper, we introduce a novel training framework designed to comprehensively address the acoustic howling issue by examining its fundamental formation process.

Acoustic echo cancellation

Large Language Models Understand and Can be Enhanced by Emotional Stimuli

no code implementations14 Jul 2023 Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie

In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts.

Emotional Intelligence Informativeness

CMMLU: Measuring massive multitask language understanding in Chinese

1 code implementation15 Jun 2023 Haonan Li, Yixuan Zhang, Fajri Koto, Yifei Yang, Hai Zhao, Yeyun Gong, Nan Duan, Timothy Baldwin

As the capabilities of large language models (LLMs) continue to advance, evaluating their performance becomes increasingly crucial and challenging.

Large Language Model

Hierarchical Optimization-Derived Learning

no code implementations11 Feb 2023 Risheng Liu, Xuan Liu, Shangzhi Zeng, Jin Zhang, Yixuan Zhang

In recent years, by utilizing optimization techniques to formulate the propagation of deep model, a variety of so-called Optimization-Derived Learning (ODL) approaches have been proposed to address diverse learning and vision tasks.

NeuralKalman: A Learnable Kalman Filter for Acoustic Echo Cancellation

no code implementations29 Jan 2023 Yixuan Zhang, Meng Yu, Hao Zhang, Dong Yu, DeLiang Wang

The robustness of the Kalman filter to double talk and its rapid convergence make it a popular approach for addressing acoustic echo cancellation (AEC) challenges.

Acoustic echo cancellation

De-biased Representation Learning for Fairness with Unreliable Labels

no code implementations1 Aug 2022 Yixuan Zhang, Feng Zhou, Zhidong Li, Yang Wang, Fang Chen

In other words, the fair pre-processing methods ignore the discrimination encoded in the labels either during the learning procedure or the evaluation stage.

Fairness Representation Learning

Optimization-Derived Learning with Essential Convergence Analysis of Training and Hyper-training

no code implementations16 Jun 2022 Risheng Liu, Xuan Liu, Shangzhi Zeng, Jin Zhang, Yixuan Zhang

Recently, Optimization-Derived Learning (ODL) has attracted attention from learning and vision areas, which designs learning models from the perspective of optimization.

Image Deconvolution

Continuous Speech Separation with Recurrent Selective Attention Network

no code implementations28 Oct 2021 Yixuan Zhang, Zhuo Chen, Jian Wu, Takuya Yoshioka, Peidong Wang, Zhong Meng, Jinyu Li

In this paper, we propose to apply recurrent selective attention network (RSAN) to CSS, which generates a variable number of output channels based on active speaker counting.

speech-recognition Speech Recognition +1

Value-Function-based Sequential Minimization for Bi-level Optimization

1 code implementation11 Oct 2021 Risheng Liu, Xuan Liu, Shangzhi Zeng, Jin Zhang, Yixuan Zhang

We also extend BVFSM to address BLO with additional functional constraints.

Bias-Tolerant Fair Classification

no code implementations7 Jul 2021 Yixuan Zhang, Feng Zhou, Zhidong Li, Yang Wang, Fang Chen

Therefore, we propose a Bias-TolerantFAirRegularizedLoss (B-FARL), which tries to regain the benefits using data affected by label bias and selection bias.

Classification Fairness +2

Nonlinear Hawkes Processes in Time-Varying System

no code implementations9 Jun 2021 Feng Zhou, Quyu Kong, Yixuan Zhang, Cheng Feng, Jun Zhu

Hawkes processes are a class of point processes that have the ability to model the self- and mutual-exciting phenomena.

Bayesian Inference Point Processes +1

Efficient Inference of Flexible Interaction in Spiking-neuron Networks

no code implementations ICLR 2021 Feng Zhou, Yixuan Zhang, Jun Zhu

Hawkes process provides an effective statistical framework for analyzing the time-dependent interaction of neuronal spiking activities.

Digital Collaborator: Augmenting Task Abstraction in Visualization Design with Artificial Intelligence

no code implementations3 Mar 2020 Aditeya Pandey, Yixuan Zhang, John A. Guerra-Gomez, Andrea G. Parker, Michelle A. Borkin

In the task abstraction phase of the visualization design process, including in "design studies", a practitioner maps the observed domain goals to generalizable abstract tasks using visualization theory in order to better understand and address the users needs.

Fine-grained Image-to-Image Transformation towards Visual Recognition

no code implementations CVPR 2020 Wei Xiong, Yutong He, Yixuan Zhang, Wenhan Luo, Lin Ma, Jiebo Luo

In this paper, we aim at transforming an image with a fine-grained category to synthesize new images that preserve the identity of the input image, which can thereby benefit the subsequent fine-grained image recognition and few-shot learning tasks.

Few-Shot Learning Fine-Grained Image Recognition

Fashion Editing with Adversarial Parsing Learning

no code implementations CVPR 2020 Haoye Dong, Xiaodan Liang, Yixuan Zhang, Xujie Zhang, Zhenyu Xie, Bowen Wu, Ziqi Zhang, Xiaohui Shen, Jian Yin

Interactive fashion image manipulation, which enables users to edit images with sketches and color strokes, is an interesting research problem with great application value.

Generative Adversarial Network Human Parsing +1

Human-Centered Emotion Recognition in Animated GIFs

1 code implementation27 Apr 2019 Zhengyuan Yang, Yixuan Zhang, Jiebo Luo

The framework consists of a facial attention module and a hierarchical segment temporal module.

Emotion Recognition

End-to-end Multi-Modal Multi-Task Vehicle Control for Self-Driving Cars with Visual Perception

1 code implementation20 Jan 2018 Zhengyuan Yang, Yixuan Zhang, Jerry Yu, Junjie Cai, Jiebo Luo

In this work, we propose a multi-task learning framework to predict the steering angle and speed control simultaneously in an end-to-end manner.

Autonomous Driving Multi-Task Learning +2

Boundary-based Image Forgery Detection by Fast Shallow CNN

1 code implementation20 Jan 2018 Zhongping Zhang, Yixuan Zhang, Zheng Zhou, Jiebo Luo

In this paper, we substantiate that Fast SCNN can detect drastic change of chroma and saturation.

Demosaicking Image Forgery Detection

Cannot find the paper you are looking for? You can Submit a new open access paper.