Zero-Shot Learning

531 papers with code • 18 benchmarks • 29 datasets

Zero-shot learning (ZSL) is a model's ability to detect classes never seen during training. The condition is that the classes are not known during supervised learning.

Earlier work in zero-shot learning use attributes in a two-step approach to infer unknown classes. In the computer vision context, more recent advances learn mappings from image feature space to semantic space. Other approaches learn non-linear multimodal embeddings. In the modern NLP context, language models can be evaluated on downstream tasks without fine tuning.

Benchmark datasets for zero-shot learning include aPY, AwA, and CUB, among others.

( Image credit: Prototypical Networks for Few shot Learning in PyTorch )

Further readings:

Libraries

Use these libraries to find Zero-Shot Learning models and implementations

Most implemented papers

Learning Transferable Visual Models From Natural Language Supervision

openai/CLIP 26 Feb 2021

State-of-the-art computer vision systems are trained to predict a fixed set of predetermined object categories.

Language Models are Few-Shot Learners

openai/gpt-3 NeurIPS 2020

By contrast, humans can generally perform a new language task from only a few examples or from simple instructions - something which current NLP systems still largely struggle to do.

Prototypical Networks for Few-shot Learning

jakesnell/prototypical-networks NeurIPS 2017

We propose prototypical networks for the problem of few-shot classification, where a classifier must generalize to new classes not seen in the training set, given only a small number of examples of each new class.

Learning to Compare: Relation Network for Few-Shot Learning

floodsung/LearningToCompare_FSL CVPR 2018

Once trained, a RN is able to classify images of new classes by computing relation scores between query images and the few examples of each new class without further updating the network.

Learning Deep Representations of Fine-grained Visual Descriptions

hanzhanggit/StackGAN-v2 CVPR 2016

State-of-the-art methods for zero-shot visual recognition formulate learning as a joint embedding problem of images and side information.

Zero-Shot Learning -- A Comprehensive Evaluation of the Good, the Bad and the Ugly

chichilicious/embarrsingly-simple-zero-shot-learning 3 Jul 2017

Due to the importance of zero-shot learning, i. e. classifying images where there is a lack of labeled training data, the number of proposed approaches has recently increased steadily.

Zero-shot User Intent Detection via Capsule Neural Networks

congyingxia/ZeroShotCapsule EMNLP 2018

User intent detection plays a critical role in question-answering and dialog systems.

CPM: A Large-scale Generative Chinese Pre-trained Language Model

TsinghuaAI/CPM-Generate 1 Dec 2020

However, applying GPT-3 to address Chinese NLP tasks is still challenging, as the training corpus of GPT-3 is primarily English, and the parameters are not publicly available.

Sampling Matters in Deep Embedding Learning

CompVis/metric-learning-divide-and-conquer ICCV 2017

In addition, we show that a simple margin based loss is sufficient to outperform all other loss functions.

Finetuned Language Models Are Zero-Shot Learners

google-research/flan ICLR 2022

We show that instruction tuning -- finetuning language models on a collection of tasks described via instructions -- substantially improves zero-shot performance on unseen tasks.