Image Captioning

618 papers with code • 32 benchmarks • 65 datasets

Image Captioning is the task of describing the content of an image in words. This task lies at the intersection of computer vision and natural language processing. Most image captioning systems use an encoder-decoder framework, where an input image is encoded into an intermediate representation of the information in the image, and then decoded into a descriptive text sequence. The most popular benchmarks are nocaps and COCO, and models are typically evaluated according to a BLEU or CIDER metric.

( Image credit: Reflective Decoding Network for Image Captioning, ICCV'19)

Libraries

Use these libraries to find Image Captioning models and implementations
4 papers
8,732
3 papers
2,324
See all 8 libraries.

Latest papers with no code

LocCa: Visual Pretraining with Location-aware Captioners

no code yet • 28 Mar 2024

In this paper, we propose a simple visual pretraining method with location-aware captioners (LocCa).

Text Data-Centric Image Captioning with Interactive Prompts

no code yet • 28 Mar 2024

Among them, the mainstream solution is to project image embeddings into the text embedding space with the assistance of consistent representations between image-text pairs from the CLIP model.

A Review of Multi-Modal Large Language and Vision Models

no code yet • 28 Mar 2024

Large Language Models (LLMs) have recently emerged as a focal point of research and application, driven by their unprecedented ability to understand and generate text with human-like quality.

A Survey on Large Language Models from Concept to Implementation

no code yet • 27 Mar 2024

Recent advancements in Large Language Models (LLMs), particularly those built on Transformer architectures, have significantly broadened the scope of natural language processing (NLP) applications, transcending their initial use in chatbot technology.

The Solution for the ICCV 2023 1st Scientific Figure Captioning Challenge

no code yet • 26 Mar 2024

In this paper, we propose a solution for improving the quality of captions generated for figures in papers.

Visual Hallucination: Definition, Quantification, and Prescriptive Remediations

no code yet • 26 Mar 2024

The troubling rise of hallucination presents perhaps the most significant impediment to the advancement of responsible AI.

Semi-Supervised Image Captioning Considering Wasserstein Graph Matching

no code yet • 26 Mar 2024

Image captioning can automatically generate captions for the given images, and the key challenge is to learn a mapping function from visual features to natural language features.

Automated Report Generation for Lung Cytological Images Using a CNN Vision Classifier and Multiple-Transformer Text Decoders: Preliminary Study

no code yet • 26 Mar 2024

Independent text decoders for benign and malignant cells are prepared for text generation, and the text decoder switches according to the CNN classification results.

Image Captioning in news report scenario

no code yet • 24 Mar 2024

Image captioning strives to generate pertinent captions for specified images, situating itself at the crossroads of Computer Vision (CV) and Natural Language Processing (NLP).

Cognitive resilience: Unraveling the proficiency of image-captioning models to interpret masked visual content

no code yet • 23 Mar 2024

This study explores the ability of Image Captioning (IC) models to decode masked visual content sourced from diverse datasets.