Few-Shot Learning

1041 papers with code • 22 benchmarks • 41 datasets

Few-Shot Learning is an example of meta-learning, where a learner is trained on several related tasks, during the meta-training phase, so that it can generalize well to unseen (but related) tasks with just few examples, during the meta-testing phase. An effective approach to the Few-Shot Learning problem is to learn a common representation for various tasks and train task specific classifiers on top of this representation.

Source: Penalty Method for Inversion-Free Deep Bilevel Optimization

Libraries

Use these libraries to find Few-Shot Learning models and implementations

Enhancing Information Maximization with Distance-Aware Contrastive Learning for Source-Free Cross-Domain Few-Shot Learning

xuhuali-mxj/im-dcl 4 Mar 2024

For this reason, this paper explores a Source-Free CDFSL (SF-CDFSL) problem, in which CDFSL is addressed through the use of existing pretrained models instead of training a model with source data, avoiding accessing source data.

1
04 Mar 2024

STAR: Constraint LoRA with Dynamic Active Learning for Data-Efficient Fine-Tuning of Large Language Models

callanwu/star 2 Mar 2024

For poor model calibration, we incorporate the regularization method during LoRA training to keep the model from being over-confident, and the Monte-Carlo dropout mechanism is employed to enhance the uncertainty estimation.

3
02 Mar 2024

Parameter-efficient Prompt Learning for 3D Point Cloud Understanding

auniquesun/ppt 24 Feb 2024

Finally, a lightweight PointAdapter module is arranged near target tasks to enhance prompt tuning for 3D point cloud understanding.

11
24 Feb 2024

Me LLaMA: Foundation Large Language Models for Medical Applications

bids-xu-lab/me-llama 20 Feb 2024

In response to this challenge, this study introduces Me-LLaMA, a novel medical LLM family that includes foundation models - Me-LLaMA 13/70B, along with their chat-enhanced versions - Me-LLaMA 13/70B-chat, developed through continual pre-training and instruction tuning of LLaMA2 using large medical datasets.

44
20 Feb 2024

Spatio-Temporal Few-Shot Learning via Diffusive Neural Network Generation

tsinghua-fib-lab/gpd 19 Feb 2024

Spatio-temporal modeling is foundational for smart city applications, yet it is often hindered by data scarcity in many cities and regions.

27
19 Feb 2024

Modularized Networks for Few-shot Hateful Meme Detection

social-ai-studio/mod_hate 19 Feb 2024

We then use the few available annotated samples to train a module composer, which assigns weights to the LoRA modules based on their relevance.

3
19 Feb 2024

Vision Superalignment: Weak-to-Strong Generalization for Vision Foundation Models

ggjy/vision_weak_to_strong 6 Feb 2024

Recent advancements in large language models have sparked interest in their extraordinary and near-superhuman capabilities, leading researchers to explore methods for evaluating and optimizing these abilities, which is called superalignment.

33
06 Feb 2024

Large Language Models to Enhance Bayesian Optimization

tennisonliu/llambo 6 Feb 2024

Bayesian optimization (BO) is a powerful approach for optimizing complex and expensive-to-evaluate black-box functions.

14
06 Feb 2024

BECLR: Batch Enhanced Contrastive Few-Shot Learning

stypoumic/beclr ICLR 2024

Learning quickly from very few labeled samples is a fundamental attribute that separates machines and humans in the era of deep representation learning.

10
04 Feb 2024

HyperPlanes: Hypernetwork Approach to Rapid NeRF Adaptation

gmum/hyperplanes 2 Feb 2024

Neural radiance fields (NeRFs) are a widely accepted standard for synthesizing new 3D object views from a small number of base images.

2
02 Feb 2024