IcoCap: Improving Video Captioning by Compounding Images

Video captioning is a more challenging task compared to image captioning, primarily due to differences in content density. Video data contains redundant visual content, making it difficult for captioners to generalize diverse content and avoid being misled by irrelevant elements. Moreover, redundant content is not well-trimmed to match the corresponding visual semantics in the ground truth, further increasing the difficulty of video captioning. Current research in video captioning predominantly focuses on captioner design, neglecting the impact of content density on captioner performance. Considering the differences between videos and images, there exists another line to improve video captioning by leveraging concise and easily-learned image samples to further diversify video samples. This modification to content density compels the captioner to learn more effectively against redundancy and ambiguity. In this paper, we propose a novel approach called Image-Compounded learning for video Captioners (IcoCap) to facilitate better learning of complex video semantics. IcoCap comprises two components: the Image-Video Compounding Strategy (ICS) and Visual-Semantic Guided Captioning (VGC). ICS compounds easily-learned image semantics into video semantics, further diversifying video content and prompting the network to generalize contents in a more diverse sample. Besides, learning with the sample compounded with image contents, the captioner is compelled to better extract valuable video cues in the presence of straightforward image semantics. This helps the captioner further focus on relevant information while filtering out extraneous content. Then, VGC guides the network in flexibly learning ground truth captions based on the compounded samples, helping to mitigate the mismatch between the ground truth and ambiguous semantics in video samples. Our experimental results demonstrate the effectiveness of IcoCap in improving the learning of video captioners. Applied to the widely-used MSVD, MSR-VTT, and VATEX datasets, our approach achieves competitive or superior results compared to state-of-the-art methods, illustrating its capacity to handle redundant and ambiguous video data.

PDF

Results from the Paper


Ranked #5 on Video Captioning on VATEX (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Benchmark
Video Captioning MSR-VTT IcoCap (ViT-B/32) CIDEr 59.1 # 15
METEOR 30.3 # 11
ROUGE-L 64.3 # 11
BLEU-4 46.1 # 14
Video Captioning MSR-VTT IcoCap (ViT-B/16) CIDEr 60.2 # 13
METEOR 31.1 # 7
ROUGE-L 64.9 # 8
BLEU-4 47.0 # 12
Video Captioning MSVD IcoCap (ViT-B/16) CIDEr 110.3 # 12
BLEU-4 59.1 # 9
METEOR 39.5 # 8
ROUGE-L 76.5 # 8
Video Captioning MSVD IcoCap (ViT-B/32) CIDEr 103.8 # 14
BLEU-4 56.3 # 10
METEOR 38.9 # 10
ROUGE-L 75.0 # 9
Video Captioning VATEX IcoCap (ViT-B/16) BLEU-4 37.4 # 5
CIDEr 67.8 # 5
METEOR 25.7 # 2
ROUGE-L 53.1 # 3
Video Captioning VATEX IcoCap (ViT-B/32) BLEU-4 36.9 # 6
CIDEr 63.4 # 8
METEOR 24.6 # 5
ROUGE-L 52.5 # 4

Methods