Search Results for author: Yajie Wang

Found 10 papers, 2 papers with code

Teeth-SEG: An Efficient Instance Segmentation Framework for Orthodontic Treatment based on Anthropic Prior Knowledge

no code implementations1 Apr 2024 Bo Zou, Shaofeng Wang, Hao liu, Gaoyue Sun, Yajie Wang, FeiFei Zuo, Chengbin Quan, Youjian Zhao

Teeth localization, segmentation, and labeling in 2D images have great potential in modern dentistry to enhance dental diagnostics, treatment planning, and population-based studies on oral health.

Image Segmentation Instance Segmentation +2

Towards Transferable Adversarial Attacks with Centralized Perturbation

no code implementations11 Dec 2023 Shangbo Wu, Yu-an Tan, Yajie Wang, Ruinan Ma, Wencong Ma, Yuanzhang Li

To this end, we propose a transferable adversarial attack with fine-grained perturbation optimization in the frequency domain, creating centralized perturbation.

Adversarial Attack

Unified High-binding Watermark for Unconditional Image Generation Models

no code implementations14 Oct 2023 Ruinan Ma, Yu-an Tan, Shangbo Wu, Tian Chen, Yajie Wang, Yuanzhang Li

In the first stage, we use an encoder to invisibly write the watermark image into the output images of the original AIGC tool, and reversely extract the watermark image through the corresponding decoder.

Image Generation Unconditional Image Generation

Enhancing Clean Label Backdoor Attack with Two-phase Specific Triggers

no code implementations10 Jun 2022 Nan Luo, Yuanzhang Li, Yajie Wang, Shangbo Wu, Yu-an Tan, Quanxin Zhang

Clean-label settings make the attack more stealthy due to the correct image-label pairs, but some problems still exist: first, traditional methods for poisoning training data are ineffective; second, traditional triggers are not stealthy which are still perceptible.

Backdoor Attack backdoor defense +1

Improving the Transferability of Adversarial Examples with Restructure Embedded Patches

1 code implementation27 Apr 2022 Huipeng Zhou, Yu-an Tan, Yajie Wang, Haoran Lyu, Shangbo Wu, Yuanzhang Li

We attack the unique self-attention mechanism in ViTs by restructuring the embedded patches of the input.

Specificity

Boosting Adversarial Transferability of MLP-Mixer

no code implementations26 Apr 2022 Haoran Lyu, Yajie Wang, Yu-an Tan, Huipeng Zhou, Yuhang Zhao, Quanxin Zhang

Our method can mask the part input of the Mixer layer, avoid overfitting of the adversarial examples to the source model, and improve the transferability of cross-architecture.

Adversarial Attack

Cannot find the paper you are looking for? You can Submit a new open access paper.