no code implementations • 15 Nov 2023 • Muhammad Waleed Gondal, Jochen Gast, Inigo Alonso Ruiz, Richard Droste, Tommaso Macri, Suren Kumar, Luitpold Staudigl
Large vision-language representation learning models like CLIP have demonstrated impressive performance for zero-shot transfer to downstream tasks while largely benefiting from inter-modal (image-text) alignment via contrastive objectives.
1 code implementation • 26 Sep 2019 • Jochen Gast, Stefan Roth
In contrast to these involved models, we found that a simple baseline CNN can perform astonishingly well when particular care is taken w. r. t.
4 code implementations • CVPR 2018 • Jochen Gast, Stefan Roth
Even though probabilistic treatments of neural networks have a long history, they have not found widespread use in practice.
no code implementations • CVPR 2016 • Jochen Gast, Anita Sellent, Stefan Roth
A two-stage pipeline, first in derivative space and then in image space, allows to estimate both parametric object motion as well as a motion segmentation from a single image alone.