no code implementations • 29 Nov 2023 • Martin Wistuba, Prabhu Teja Sivaprasad, Lukas Balles, Giovanni Zappella
Recent work using pretrained transformers has shown impressive performance when fine-tuned with data from the downstream problem of interest.
no code implementations • 1 Nov 2023 • Evann Courdier, Prabhu Teja Sivaprasad, François Fleuret
We study the problem of improving the efficiency of segmentation transformers by using disparate amounts of computation for different parts of the image.
1 code implementation • 19 Oct 2021 • Prabhu Teja Sivaprasad, François Fleuret
Data samples generated by several real world processes are dynamic in nature \textit{i. e.}, their characteristics vary with time.
no code implementations • ICML 2020 • Prabhu Teja Sivaprasad, Florian Mai, Thijs Vogels, Martin Jaggi, François Fleuret
The performance of optimizers, particularly in deep learning, depends considerably on their chosen hyperparameter configuration.