2 code implementations • 11 Apr 2024 • Aleksander Nagaj, Zenjie Li, Dim P. Papadopoulos, Kamal Nasrollahi
During training, pixel-based transformations are applied to segmented objects, and the models are then evaluated on raw images without segmentation.
no code implementations • 11 Nov 2023 • Jens Parslov, Erik Riise, Dim P. Papadopoulos
For part segmentation, we show that the segmentation models trained on a combination of real data and our synthetic data outperform all models trained only on real data.
no code implementations • 8 Nov 2023 • Thanos Delatolas, Vicky Kalogeiton, Dim P. Papadopoulos
To reduce this annotation cost, in this paper, we propose EVA-VOS, a human-in-the-loop annotation framework for video object segmentation.
no code implementations • CVPR 2022 • Dim P. Papadopoulos, Enrique Mora, Nadiia Chepurko, Kuan Wei Huang, Ferda Ofli, Antonio Torralba
To validate our idea, we crowdsource programs for cooking recipes and show that: (a) projecting the image-recipe embeddings into programs leads to better cross-modal retrieval results; (b) generating programs from images leads to better recognition results compared to predicting raw cooking instructions; and (c) we can generate food images by manipulating programs via optimizing the latent code of a GAN.
1 code implementation • 11 Jan 2022 • Ethan Weber, Dim P. Papadopoulos, Agata Lapedriza, Ferda Ofli, Muhammad Imran, Antonio Torralba
In this work, we present the Incidents1M Dataset, a large-scale multi-label dataset which contains 977, 088 images, with 43 incident and 49 place categories.
no code implementations • ICCV 2021 • Dim P. Papadopoulos, Ethan Weber, Antonio Torralba
Through a large-scale experiment to populate 1M unlabeled images with object segmentation masks for 80 object classes, we show that (1) we obtain 1M object segmentation masks with an total annotation time of only 290 hours; (2) we reduce annotation time by 76x compared to manual annotation; (3) the segmentation quality of our masks is on par with those from manually annotated datasets.
1 code implementation • ECCV 2020 • Ethan Weber, Nuria Marzo, Dim P. Papadopoulos, Aritro Biswas, Agata Lapedriza, Ferda Ofli, Muhammad Imran, Antonio Torralba
While most studies on social media are limited to text, images offer more information for understanding disaster and incident scenes.
no code implementations • CVPR 2019 • Dim P. Papadopoulos, Youssef Tamaazousti, Ferda Ofli, Ingmar Weber, Antonio Torralba
From a visual perspective, every instruction step can be seen as a way to change the visual appearance of the dish by adding extra objects (e. g., adding an ingredient) or changing the appearance of the existing ones (e. g., cooking the dish).
no code implementations • ICCV 2017 • Dim P. Papadopoulos, Jasper R. R. Uijlings, Frank Keller, Vittorio Ferrari
We crowd-source extreme point annotations for PASCAL VOC 2007 and 2012 and show that (1) annotation time is only 7s per box, 5x faster than the traditional way of drawing boxes [62]; (2) the quality of the boxes is as good as the original ground-truth drawn the traditional way; (3) detectors trained on our annotations are as accurate as those trained on the original ground-truth.
no code implementations • CVPR 2016 • Radu Tudor Ionescu, Bogdan Alexe, Marius Leordeanu, Marius Popescu, Dim P. Papadopoulos, Vittorio Ferrari
We address the problem of estimating image difficulty defined as the human response time for solving a visual search task.
no code implementations • CVPR 2017 • Dim P. Papadopoulos, Jasper R. R. Uijlings, Frank Keller, Vittorio Ferrari
Training object class detectors typically requires a large set of images with objects annotated by bounding boxes.
1 code implementation • CVPR 2016 • Dim P. Papadopoulos, Jasper R. R. Uijlings, Frank Keller, Vittorio Ferrari
Training object class detectors typically requires a large set of images in which objects are annotated by bounding-boxes.