Search Results for author: Hezheng Lin

Found 4 papers, 3 papers with code

Generation-Guided Multi-Level Unified Network for Video Grounding

no code implementations14 Mar 2023 Xing Cheng, Xiangyu Wu, Dong Shen, Hezheng Lin, Fan Yang

Video grounding aims to locate the timestamps best matching the query description within an untrimmed video.

Video Grounding

Improving Video-Text Retrieval by Multi-Stream Corpus Alignment and Dual Softmax Loss

2 code implementations9 Sep 2021 Xing Cheng, Hezheng Lin, Xiangyu Wu, Fan Yang, Dong Shen

In this paper, we propose a multi-stream Corpus Alignment network with single gate Mixture-of-Experts (CAMoE) and a novel Dual Softmax Loss (DSL) to solve the two heterogeneity.

Ranked #9 on Video Retrieval on MSVD (using extra training data)

Retrieval Text Retrieval +1

CAT: Cross Attention in Vision Transformer

1 code implementation10 Jun 2021 Hezheng Lin, Xing Cheng, Xiangyu Wu, Fan Yang, Dong Shen, Zhongyuan Wang, Qing Song, Wei Yuan

In this paper, we propose a new attention mechanism in Transformer termed Cross Attention, which alternates attention inner the image patch instead of the whole image to capture local information and apply attention between image patches which are divided from single-channel feature maps capture global information.

Cannot find the paper you are looking for? You can Submit a new open access paper.