Search Results for author: Ruijin Liu

Found 5 papers, 2 papers with code

Align before Adapt: Leveraging Entity-to-Region Alignments for Generalizable Video Action Recognition

no code implementations27 Nov 2023 Yifei Chen, Dapeng Chen, Ruijin Liu, Sai Zhou, Wenyuan Xue, Wei Peng

With the aligned entities, we feed their text embeddings to a transformer-based video adapter as the queries, which can help extract the semantics of the most important entities from a video to a vector.

Action Recognition Representation Learning +1

PBFormer: Capturing Complex Scene Text Shape with Polynomial Band Transformer

no code implementations29 Aug 2023 Ruijin Liu, Ning Lu, Dapeng Chen, Cheng Li, Zejian yuan, Wei Peng

We present PBFormer, an efficient yet powerful scene text detector that unifies the transformer with a novel text shape representation Polynomial Band (PB).

Video Action Recognition with Attentive Semantic Units

no code implementations ICCV 2023 Yifei Chen, Dapeng Chen, Ruijin Liu, Hao Li, Wei Peng

Supervised by the semantics of action labels, recent works adapt the visual branch of VLMs to learn video representations.

Action Recognition Temporal Action Localization +1

Learning to Predict 3D Lane Shape and Camera Pose from a Single Image via Geometry Constraints

1 code implementation31 Dec 2021 Ruijin Liu, Dapeng Chen, Tie Liu, Zhiliang Xiong, Zejian yuan

In this task, the correct camera pose is the key to generating accurate lanes, which can transform an image from perspective-view to the top-view.

3D Lane Detection Autonomous Vehicles +2

End-to-end Lane Shape Prediction with Transformers

2 code implementations9 Nov 2020 Ruijin Liu, Zejian yuan, Tie Liu, Zhiliang Xiong

To tackle these issues, we propose an end-to-end method that directly outputs parameters of a lane shape model, using a network built with a transformer to learn richer structures and context.

Lane Detection

Cannot find the paper you are looking for? You can Submit a new open access paper.