Zero-Shot Learning

562 papers with code • 18 benchmarks • 29 datasets

Zero-shot learning (ZSL) is a model's ability to detect classes never seen during training. The condition is that the classes are not known during supervised learning.

Earlier work in zero-shot learning use attributes in a two-step approach to infer unknown classes. In the computer vision context, more recent advances learn mappings from image feature space to semantic space. Other approaches learn non-linear multimodal embeddings. In the modern NLP context, language models can be evaluated on downstream tasks without fine tuning.

Benchmark datasets for zero-shot learning include aPY, AwA, and CUB, among others.

( Image credit: Prototypical Networks for Few shot Learning in PyTorch )

Further readings:

Libraries

Use these libraries to find Zero-Shot Learning models and implementations

The Devil is in the Few Shots: Iterative Visual Knowledge Completion for Few-shot Learning

mark-sky/kcl 15 Apr 2024

Few-shot learning aims to further enhance the transfer capability of CLIP by giving few images in each class, aka 'few shots'.

8
15 Apr 2024

CREST: Cross-modal Resonance through Evidential Deep Learning for Enhanced Zero-Shot Learning

JethroJames/CREST 15 Apr 2024

Zero-shot learning (ZSL) enables the recognition of novel classes by leveraging semantic knowledge transfer from known to unknown categories.

5
15 Apr 2024

Knowledge-enhanced Visual-Language Pretraining for Computational Pathology

magic-ai4med/kep 15 Apr 2024

In this paper, we consider the problem of visual representation learning for computational pathology, by exploiting large-scale image-text pairs gathered from public resources, along with the domain specific knowledge in pathology.

5
15 Apr 2024

Audio-Visual Generalized Zero-Shot Learning using Pre-Trained Large Multi-Modal Models

faceonlive/ai-research 9 Apr 2024

However, existing benchmarks predate the popularization of large multi-modal models, such as CLIP and CLAP.

152
09 Apr 2024

Forget NLI, Use a Dictionary: Zero-Shot Topic Classification for Low-Resource Languages with Application to Luxembourgish

faceonlive/ai-research 5 Apr 2024

A common method for ZSC is to fine-tune a language model on a Natural Language Inference (NLI) dataset and then use it to infer the entailment between the input document and the target labels.

152
05 Apr 2024

Label Propagation for Zero-shot Classification with Vision-Language Models

faceonlive/ai-research 5 Apr 2024

We leverage the graph structure of the unlabeled data and introduce ZLaP, a method based on label propagation (LP) that utilizes geodesic distances for classification.

152
05 Apr 2024

Emergent Abilities in Reduced-Scale Generative Language Models

text-machine-lab/mini_gpt 2 Apr 2024

Large language models can solve new tasks without task-specific fine-tuning.

2
02 Apr 2024

X-MIC: Cross-Modal Instance Conditioning for Egocentric Action Generalization

annusha/xmic 28 Mar 2024

Lately, there has been growing interest in adapting vision-language models (VLMs) to image and third-person video classification due to their success in zero-shot recognition.

4
28 Mar 2024

VLM-CPL: Consensus Pseudo Labels from Vision-Language Models for Human Annotation-Free Pathological Image Classification

lanfz2000/vlm-cpl 23 Mar 2024

To address this issue, we introduce VLM-CPL, a novel approach based on consensus pseudo labels that integrates two noisy label filtering techniques with a semi-supervised learning strategy.

3
23 Mar 2024

Long-CLIP: Unlocking the Long-Text Capability of CLIP

beichenzbc/long-clip 22 Mar 2024

Contrastive Language-Image Pre-training (CLIP) has been the cornerstone for zero-shot classification, text-image retrieval, and text-image generation by aligning image and text modalities.

307
22 Mar 2024