Visual Relationship Detection

36 papers with code • 5 benchmarks • 5 datasets

Visual relationship detection (VRD) is one newly developed computer vision task aiming to recognize relations or interactions between objects in an image. It is a further learning task after object recognition and is essential for fully understanding images, even the visual world.

Groupwise Query Specialization and Quality-Aware Multi-Assignment for Transformer-based Visual Relationship Detection

mlvlab/speaq 26 Mar 2024

Groupwise Query Specialization trains a specialized query by dividing queries and relations into disjoint groups and directing a query in a specific query group solely toward relations in the corresponding relation group.

8
26 Mar 2024

Video Relationship Detection Using Mixture of Experts

shibshib/Moe-VRD IEEE Access 2023

Secondly, classifiers trained by a single, monolithic neural network often lack stability and generalization.

2
06 Mar 2024

Self-Supervised Learning for Visual Relationship Detection through Masked Bounding Box Reconstruction

deeplab-ai/selfsupervisedvrd 8 Nov 2023

We present a novel self-supervised approach for representation learning, particularly for the task of Visual Relationship Detection (VRD).

0
08 Nov 2023

STUPD: A Synthetic Dataset for Spatial and Temporal Relation Reasoning

palaashagrawal/stupd 13 Sep 2023

In addition to spatial relations, we also propose 50K visual depictions across 10 temporal relations, consisting of videos depicting event/time-point interactions.

0
13 Sep 2023

Unified Visual Relationship Detection with Vision and Language Models

google-research/scenic ICCV 2023

To address this challenge, we propose UniVRD, a novel bottom-up method for Unified Visual Relationship Detection by leveraging vision and language models (VLMs).

2,991
16 Mar 2023

Distance-Aware Occlusion Detection with Focused Attention

yang-li-2000/distance-aware-occlusion-detection-with-focused-attention 23 Aug 2022

In this work, (1) we propose a novel three-decoder architecture as the infrastructure for focused attention; 2) we use the generalized intersection box prediction task to effectively guide our model to focus on occlusion-specific regions; 3) our model achieves a new state-of-the-art performance on distance-aware relationship detection.

40
23 Aug 2022

Neural Message Passing for Visual Relationship Detection

phyllish/nmp 8 Aug 2022

Visual relationship detection aims to detect the interactions between objects in an image; however, this task suffers from combinatorial explosion due to the variety of objects and interactions.

30
08 Aug 2022

PEVL: Position-enhanced Pre-training and Prompt Tuning for Vision-language Models

thunlp/pevl 23 May 2022

We show that PEVL enables state-of-the-art performance of detector-free VLP models on position-sensitive tasks such as referring expression comprehension and phrase grounding, and also improves the performance on position-insensitive tasks with grounded inputs.

43
23 May 2022

Representing Prior Knowledge Using Randomly, Weighted Feature Networks for Visual Relationship Detection

pavliclab/aaai2022-clear2022-visual_relationship_detection-rwfn AAAI Workshop CLeaR 2022

Furthermore, background knowledge represented by RWFNs can be used to alleviate the incompleteness of training sets even though the space complexity of RWFNs is much smaller than LTNs (1:27 ratio).

3
20 Nov 2021

Image Scene Graph Generation (SGG) Benchmark

microsoft/scene_graph_benchmark 27 Jul 2021

There is a surge of interest in image scene graph generation (object, attribute and relationship detection) due to the need of building fine-grained image understanding models that go beyond object detection.

375
27 Jul 2021