Search Results for author: Zhengmian Hu

Found 11 papers, 3 papers with code

Token-Level Adversarial Prompt Detection Based on Perplexity Measures and Contextual Information

no code implementations20 Nov 2023 Zhengmian Hu, Gang Wu, Saayan Mitra, Ruiyi Zhang, Tong Sun, Heng Huang, Viswanathan Swaminathan

Our work aims to address this concern by introducing a novel approach to detecting adversarial prompts at a token level, leveraging the LLM's capability to predict the next token's probability.

GPT-4 Vision on Medical Image Classification -- A Case Study on COVID-19 Dataset

no code implementations27 Oct 2023 Ruibo Chen, Tianyi Xiong, Yihan Wu, Guodong Liu, Zhengmian Hu, Lichang Chen, Yanshuo Chen, Chenxi Liu, Heng Huang

This technical report delves into the application of GPT-4 Vision (GPT-4V) in the nuanced realm of COVID-19 image classification, leveraging the transformative potential of in-context learning to enhance diagnostic processes.

GPT-4 Image Classification +2

DiPmark: A Stealthy, Efficient and Resilient Watermark for Large Language Models

no code implementations11 Oct 2023 Yihan Wu, Zhengmian Hu, Hongyang Zhang, Heng Huang

Watermarking techniques offer a promising way to secure data via embedding covert information into the data.

Language Modelling

Solving a Class of Non-Convex Minimax Optimization in Federated Learning

1 code implementation NeurIPS 2023 Xidong Wu, Jianhui Sun, Zhengmian Hu, Aidong Zhang, Heng Huang

We propose FL algorithms (FedSGDA+ and FedSGDA-M) and reduce existing complexity results for the most common minimax problems.

Federated Learning

Serverless Federated AUPRC Optimization for Multi-Party Collaborative Imbalanced Data Mining

1 code implementation6 Aug 2023 Xidong Wu, Zhengmian Hu, Jian Pei, Heng Huang

To address the above challenge, we study the serverless multi-party collaborative AUPRC maximization problem since serverless multi-party collaborative training can cut down the communications cost by avoiding the server node bottleneck, and reformulate it as a conditional stochastic optimization problem in a serverless multi-party collaborative learning setting and propose a new ServerLess biAsed sTochastic gradiEnt (SLATE) algorithm to directly optimize the AUPRC.

Federated Learning Stochastic Optimization

Decentralized Riemannian Algorithm for Nonconvex Minimax Problems

no code implementations8 Feb 2023 Xidong Wu, Zhengmian Hu, Heng Huang

The minimax optimization over Riemannian manifolds (possibly nonconvex constraints) has been actively applied to solve many problems, such as robust dimensionality reduction and deep neural networks with orthogonal weights (Stiefel manifold).

Dimensionality Reduction

Faster Adaptive Federated Learning

no code implementations2 Dec 2022 Xidong Wu, Feihu Huang, Zhengmian Hu, Heng Huang

Federated learning has attracted increasing attention with the emergence of distributed data.

Federated Learning Image Classification +1

Optimal Underdamped Langevin MCMC Method

no code implementations NeurIPS 2021 Zhengmian Hu, Feihu Huang, Heng Huang

In the paper, we study the underdamped Langevin diffusion (ULD) with strongly-convex potential consisting of finite summation of $N$ smooth components, and propose an efficient discretization method, which requires $O(N+d^\frac{1}{3}N^\frac{2}{3}/\varepsilon^\frac{2}{3})$ gradient evaluations to achieve $\varepsilon$-error (in $\sqrt{\mathbb{E}{\lVert{\cdot}\rVert_2^2}}$ distance) for approximating $d$-dimensional ULD.

Fast and Scalable Adversarial Training of Kernel SVM via Doubly Stochastic Gradients

1 code implementation21 Jul 2021 Huimin Wu, Zhengmian Hu, Bin Gu

Although a wide range of researches have been done in recent years to improve the adversarial robustness of learning models, but most of them are limited to deep neural networks (DNNs) and the work for kernel SVM is still vacant.

Adversarial Robustness

AdaGDA: Faster Adaptive Gradient Descent Ascent Methods for Minimax Optimization

no code implementations30 Jun 2021 Feihu Huang, Xidong Wu, Zhengmian Hu

Specifically, we propose a fast Adaptive Gradient Descent Ascent (AdaGDA) method based on the basic momentum technique, which reaches a lower gradient complexity of $\tilde{O}(\kappa^4\epsilon^{-4})$ for finding an $\epsilon$-stationary point without large batches, which improves the existing results of the adaptive GDA methods by a factor of $O(\sqrt{\kappa})$.

A New Framework for Variance-Reduced Hamiltonian Monte Carlo

no code implementations9 Feb 2021 Zhengmian Hu, Feihu Huang, Heng Huang

Moreover, our HMC methods with biased gradient estimators, such as SARAH and SARGE, require $\tilde{O}(N+\sqrt{N} \kappa^2 d^{\frac{1}{2}} \varepsilon^{-1})$ gradient complexity, which has the same dependency on condition number $\kappa$ and dimension $d$ as full gradient method, but improves the dependency of sample size $N$ for a factor of $N^\frac{1}{2}$.

Cannot find the paper you are looking for? You can Submit a new open access paper.