Search Results for author: Suchen Wang

Found 8 papers, 5 papers with code

Top-Down Framework for Weakly-supervised Grounded Image Captioning

no code implementations13 Jun 2023 Chen Cai, Suchen Wang, Kim-Hui Yap, Yi Wang

Weakly-supervised grounded image captioning (WSGIC) aims to generate the caption and ground (localize) predicted object words in the input image without using bounding box supervision.

Image Captioning Multi-Label Classification +2

VLT: Vision-Language Transformer and Query Generation for Referring Segmentation

1 code implementation28 Oct 2022 Henghui Ding, Chang Liu, Suchen Wang, Xudong Jiang

We propose a Vision-Language Transformer (VLT) framework for referring segmentation to facilitate deep interactions among multi-modal information and enhance the holistic understanding to vision-language features.

Referring Expression Segmentation Referring Video Object Segmentation

Learning Transferable Human-Object Interaction Detector With Natural Language Supervision

1 code implementation CVPR 2022 Suchen Wang, Yueqi Duan, Henghui Ding, Yap-Peng Tan, Kim-Hui Yap, Junsong Yuan

More specifically, we propose a new HOI visual encoder to detect the interacting humans and objects, and map them to a joint feature space to perform interaction recognition.

Human-Object Interaction Detection

Joint Representative Selection and Feature Learning: A Semi-Supervised Approach

no code implementations CVPR 2019 Suchen Wang, Jingjing Meng, Junsong Yuan, Yap-Peng Tan

Given labeled source data and big unlabeled target data, we aim to find representatives in the target data, which can not only represent and associate data points belonging to each labeled category, but also discover novel categories in the target data, if any.

Cannot find the paper you are looking for? You can Submit a new open access paper.