no code implementations • 25 Apr 2024 • Gabriela Kadlecová, Jovita Lukasik, Martin Pilát, Petra Vidnerová, Mahmoud Safari, Roman Neruda, Frank Hutter
Performance prediction has been a key part of the neural architecture search (NAS) process, allowing to speed up NAS algorithms by avoiding resource-consuming network training.
1 code implementation • 14 Mar 2024 • Paul Gavrikov, Jovita Lukasik, Steffen Jung, Robert Geirhos, Bianca Lamm, Muhammad Jehanzeb Mirza, Margret Keuper, Janis Keuper
If text does indeed influence visual biases, this suggests that we may be able to steer visual biases not just through visual input but also through language: a hypothesis that we confirm through extensive experiments.
no code implementations • 18 Jul 2023 • Jovita Lukasik, Michael Moeller, Margret Keuper
We are interested in the single prediction task for robustness and the joint multi-objective of clean and robust accuracy.
no code implementations • 11 Jun 2023 • Steffen Jung, Jovita Lukasik, Margret Keuper
We evaluate all these networks on a range of common adversarial attacks and corruption types and introduce a database on neural architecture design and robustness evaluations.
1 code implementation • 16 Mar 2022 • Jovita Lukasik, Steffen Jung, Margret Keuper
the optimization of architectures for highest classification accuracy but also in the context of hardware constraints and outperform state-of-the-art methods on several NAS benchmarks for single and multiple objectives.
no code implementations • NeurIPS Workshop Deep_Invers 2021 • Jonas Geiping, Jovita Lukasik, Margret Keuper, Michael Moeller
Differentiable architecture search (DARTS) is a widely researched tool for neural architecture search, due to its promising results for image classification.
no code implementations • 12 Aug 2021 • Jonas Geiping, Jovita Lukasik, Margret Keuper, Michael Moeller
In this work, we investigate DAS in a systematic case study of inverse problems, which allows us to analyze these potential benefits in a controlled manner.
no code implementations • 19 Oct 2020 • Jovita Lukasik, David Friede, Heiner Stuckenschmidt, Margret Keuper
In computer vision research, the process of automating architecture engineering, Neural Architecture Search (NAS), has gained substantial interest.
2 code implementations • 9 Oct 2020 • Jovita Lukasik, David Friede, Arber Zela, Frank Hutter, Margret Keuper
We evaluate the proposed approach on neural architectures defined by the ENAS approach, the NAS-Bench-101 and the NAS-Bench-201 search space and show that our smooth embedding space allows to directly extrapolate the performance prediction to architectures outside the seen domain (e. g. with more operations).
1 code implementation • ICLR 2022 • Arber Zela, Julien Siems, Lucas Zimmer, Jovita Lukasik, Margret Keuper, Frank Hutter
We show that surrogate NAS benchmarks can model the true performance of architectures better than tabular benchmarks (at a small fraction of the cost), that they lead to faithful estimates of how well different NAS methods work on the original non-surrogate benchmark, and that they can generate new scientific insight.
1 code implementation • 11 Dec 2019 • David Friede, Jovita Lukasik, Heiner Stuckenschmidt, Margret Keuper
In computer vision research, the process of automating architecture engineering, Neural Architecture Search (NAS), has gained substantial interest.
no code implementations • 15 Feb 2019 • Margret Keuper, Jovita Lukasik, Maneesh Singh, Julian Yarkony
We tackle the problem of graph partitioning for image segmentation using correlation clustering (CC), which we treat as an integer linear program (ILP).