no code implementations • 6 Jun 2024 • Benedikt Alkin, Maximilian Beck, Korbinian Pöppel, Sepp Hochreiter, Johannes Brandstetter
Transformers are widely used as generic backbones in computer vision, despite initially introduced for natural language processing.
1 code implementation • 19 Feb 2024 • Benedikt Alkin, Andreas Fürst, Simon Schmid, Lukas Gruber, Markus Holzleitner, Johannes Brandstetter
This is of special interest since, akin to their numerical counterparts, different techniques are used across applications, even if the underlying dynamics of the systems are similar.
1 code implementation • 15 Feb 2024 • Benedikt Alkin, Lukas Miklautz, Sepp Hochreiter, Johannes Brandstetter
We introduce MIM (Masked Image Modeling)-Refiner, a contrastive learning boost for pre-trained MIM models.
Ranked #1 on Image Clustering on ImageNet
1 code implementation • 20 Apr 2023 • Johannes Lehner, Benedikt Alkin, Andreas Fürst, Elisabeth Rumetshofer, Lukas Miklautz, Sepp Hochreiter
In this work, we study how to combine the efficiency and scalability of MIM with the ability of ID to perform downstream classification in the absence of large amounts of labeled data.
Ranked #1 on Image Clustering on Imagenet-dog-15 (using extra training data)