Search Results for author: Yanyan Liang

Found 15 papers, 6 papers with code

CFPL-FAS: Class Free Prompt Learning for Generalizable Face Anti-spoofing

no code implementations21 Mar 2024 Ajian Liu, Shuai Xue, Jianwen Gan, Jun Wan, Yanyan Liang, Jiankang Deng, Sergio Escalera, Zhen Lei

Specifically, we propose a novel Class Free Prompt Learning (CFPL) paradigm for DG FAS, which utilizes two lightweight transformers, namely Content Q-Former (CQF) and Style Q-Former (SQF), to learn the different semantic prompts conditioned on content and style features by using a set of learnable query vectors, respectively.

Domain Generalization Face Anti-Spoofing

Gloss-free Sign Language Translation: Improving from Visual-Language Pretraining

1 code implementation ICCV 2023 Benjia Zhou, Zhigang Chen, Albert Clapés, Jun Wan, Yanyan Liang, Sergio Escalera, Zhen Lei, Du Zhang

Many previous methods employ an intermediate representation, i. e., gloss sequences, to facilitate SLT, thus transforming it into a two-stage task of sign language recognition (SLR) followed by sign language translation (SLT).

Gloss-free Sign Language Translation Self-Supervised Learning +3

MA-ViT: Modality-Agnostic Vision Transformers for Face Anti-Spoofing

no code implementations15 Apr 2023 Ajian Liu, Yanyan Liang

The existing multi-modal face anti-spoofing (FAS) frameworks are designed based on two strategies: halfway and late fusion.

Face Anti-Spoofing

A Unified Multimodal De- and Re-coupling Framework for RGB-D Motion Recognition

1 code implementation16 Nov 2022 Benjia Zhou, Pichao Wang, Jun Wan, Yanyan Liang, Fan Wang

Although improving motion recognition to some extent, these methods still face sub-optimal situations in the following aspects: (i) Data augmentation, i. e., the scale of the RGB-D datasets is still limited, and few efforts have been made to explore novel data augmentation strategies for videos; (ii) Optimization mechanism, i. e., the tightly space-time-entangled network structure brings more challenges to spatiotemporal information modeling; And (iii) cross-modal knowledge fusion, i. e., the high similarity between multimodal representations caused to insufficient late fusion.

Action Recognition Data Augmentation +2

Effective Vision Transformer Training: A Data-Centric Perspective

no code implementations29 Sep 2022 Benjia Zhou, Pichao Wang, Jun Wan, Yanyan Liang, Fan Wang

To achieve these two purposes, we propose a novel data-centric ViT training framework to dynamically measure the ``difficulty'' of training samples and generate ``effective'' samples for models at different training stages.

MxPool: Multiplex Pooling for Hierarchical Graph Representation Learning

no code implementations ICLR 2020 Yanyan Liang, Yanfeng Zhang, Dechao Gao, Qian Xu

This motivates us to use a multiplex structure in a diverse way and utilize a priori properties of graphs to guide the learning.

Clustering General Classification +3

CASIA-SURF: A Large-scale Multi-modal Benchmark for Face Anti-spoofing

no code implementations28 Aug 2019 Shifeng Zhang, Ajian Liu, Jun Wan, Yanyan Liang, Guogong Guo, Sergio Escalera, Hugo Jair Escalante, Stan Z. Li

To facilitate face anti-spoofing research, we introduce a large-scale multi-modal dataset, namely CASIA-SURF, which is the largest publicly available dataset for face anti-spoofing in terms of both subjects and modalities.

Face Anti-Spoofing Face Recognition

Cannot find the paper you are looking for? You can Submit a new open access paper.