Self-Supervised Learning

1731 papers with code • 10 benchmarks • 41 datasets

Self-Supervised Learning is proposed for utilizing unlabeled data with the success of supervised learning. Producing a dataset with good labels is expensive, while unlabeled data is being generated all the time. The motivation of Self-Supervised Learning is to make use of the large amount of unlabeled data. The main idea of Self-Supervised Learning is to generate the labels from unlabeled data, according to the structure or characteristics of the data itself, and then train on this unsupervised data in a supervised manner. Self-Supervised Learning is wildly used in representation learning to make a model learn the latent features of the data. This technique is often employed in computer vision, video processing and robot control.

Source: Self-supervised Point Set Local Descriptors for Point Cloud Registration

Image source: LeCun

Libraries

Use these libraries to find Self-Supervised Learning models and implementations
14 papers
2,743
11 papers
1,355
See all 10 libraries.

Latest papers with no code

OPTiML: Dense Semantic Invariance Using Optimal Transport for Self-Supervised Medical Image Representation

no code yet • 18 Apr 2024

In response to these constraints, we introduce a novel SSL framework OPTiML, employing optimal transport (OT), to capture the dense semantic invariance and fine-grained details, thereby enhancing the overall effectiveness of SSL in medical image representation learning.

Learning to Solve the Constrained Most Probable Explanation Task in Probabilistic Graphical Models

no code yet • 17 Apr 2024

Given an assignment $\mathbf{x}$ to all variables in $\mathbf{X}$ (evidence) and a real number $q$, the constrained most-probable explanation (CMPE) task seeks to find an assignment $\mathbf{y}$ to all variables in $\mathbf{Y}$ such that $f(\mathbf{x}, \mathbf{y})$ is maximized and $g(\mathbf{x}, \mathbf{y})\leq q$.

Spatial Context-based Self-Supervised Learning for Handwritten Text Recognition

no code yet • 17 Apr 2024

Handwritten Text Recognition (HTR) is a relevant problem in computer vision, and implies unique challenges owing to its inherent variability and the rich contextualization required for its interpretation.

Deep Pattern Network for Click-Through Rate Prediction

no code yet • 17 Apr 2024

These patterns harbor substantial potential to significantly enhance CTR prediction performance.

When are Foundation Models Effective? Understanding the Suitability for Pixel-Level Classification Using Multispectral Imagery

no code yet • 17 Apr 2024

This work aims to enhance the understanding of the status and suitability of foundation models for pixel-level classification using multispectral imagery at moderate resolution, through comparisons with traditional machine learning (ML) and regular-size deep learning models.

Pretraining Billion-scale Geospatial Foundational Models on Frontier

no code yet • 17 Apr 2024

Although large FMs have demonstrated significant impact in natural language processing and computer vision, efforts toward FMs for geospatial applications have been restricted to smaller size models, as pretraining larger models requires very large computing resources equipped with state-of-the-art hardware accelerators.

Integration of Self-Supervised BYOL in Semi-Supervised Medical Image Recognition

no code yet • 16 Apr 2024

Image recognition techniques heavily rely on abundant labeled data, particularly in medical contexts.

Self-Supervised Learning Featuring Small-Scale Image Dataset for Treatable Retinal Diseases Classification

no code yet • 15 Apr 2024

The proposed SSL model achieves the state-of-art accuracy of 98. 84% using only 4, 000 training images.

Label-free Anomaly Detection in Aerial Agricultural Images with Masked Image Modeling

no code yet • 13 Apr 2024

Hence, this is posed as an anomaly detection task in agricultural images.

MMA-DFER: MultiModal Adaptation of unimodal models for Dynamic Facial Expression Recognition in-the-wild

no code yet • 13 Apr 2024

Within the field of multimodal DFER, recent methods have focused on exploiting advances of self-supervised learning (SSL) for pre-training of strong multimodal encoders.