Search Results for author: Samuel Schulter

Found 31 papers, 7 papers with code

Shuffle and Attend: Video Domain Adaptation

no code implementations ECCV 2020 Jinwoo Choi, Gaurav Sharma, Samuel Schulter, Jia-Bin Huang

As the first novelty, we propose an attention mechanism which focuses on more discriminative clips and directly optimizes for video-level (cf.

Action Recognition Temporal Action Localization +1

Self-Training Large Language Models for Improved Visual Program Synthesis With Visual Reinforcement

no code implementations6 Apr 2024 Zaid Khan, Vijay Kumar BG, Samuel Schulter, Yun Fu, Manmohan Chandraker

We propose a method where we exploit existing annotations for a vision-language task to improvise a coarse reward signal for that task, treat the LLM as a policy, and apply reinforced self-training to improve the visual program synthesis ability of the LLM for that task.

object-detection Object Detection +4

Generating Enhanced Negatives for Training Language-Based Object Detectors

1 code implementation29 Dec 2023 Shiyu Zhao, Long Zhao, Vijay Kumar B. G, Yumin Suh, Dimitris N. Metaxas, Manmohan Chandraker, Samuel Schulter

The recent progress in language-based open-vocabulary object detection can be largely attributed to finding better ways of leveraging large-scale data with free-form text annotations.

Object object-detection +1

Efficient Controllable Multi-Task Architectures

no code implementations ICCV 2023 Abhishek Aich, Samuel Schulter, Amit K. Roy-Chowdhury, Manmohan Chandraker, Yumin Suh

Further, we present a simple but effective search algorithm that translates user constraints to runtime width configurations of both the shared encoder and task decoders, for sampling the sub-architectures.

Knowledge Distillation

Taming Self-Training for Open-Vocabulary Object Detection

2 code implementations11 Aug 2023 Shiyu Zhao, Samuel Schulter, Long Zhao, Zhixing Zhang, Vijay Kumar B. G, Yumin Suh, Manmohan Chandraker, Dimitris N. Metaxas

This work identifies two challenges of using self-training in OVD: noisy PLs from VLMs and frequent distribution changes of PLs.

Object object-detection +1

OmniLabel: A Challenging Benchmark for Language-Based Object Detection

no code implementations ICCV 2023 Samuel Schulter, Vijay Kumar B G, Yumin Suh, Konstantinos M. Dafnis, Zhixing Zhang, Shiyu Zhao, Dimitris Metaxas

With more than 28K unique object descriptions on over 25K images, OmniLabel provides a challenging benchmark with diverse and complex object descriptions in a naturally open-vocabulary setting.

Object object-detection +1

Exploiting Unlabeled Data with Vision and Language Models for Object Detection

1 code implementation18 Jul 2022 Shiyu Zhao, Zhixing Zhang, Samuel Schulter, Long Zhao, Vijay Kumar B. G, Anastasis Stathopoulos, Manmohan Chandraker, Dimitris Metaxas

We propose a novel method that leverages the rich semantics available in recent vision and language models to localize and classify objects in unlabeled images, effectively generating pseudo labels for object detection.

Ranked #15 on Open Vocabulary Object Detection on MSCOCO (using extra training data)

Object object-detection +3

Controllable Dynamic Multi-Task Architectures

no code implementations CVPR 2022 Dripta S. Raychaudhuri, Yumin Suh, Samuel Schulter, Xiang Yu, Masoud Faraki, Amit K. Roy-Chowdhury, Manmohan Chandraker

In contrast to the existing dynamic multi-task approaches that adjust only the weights within a fixed architecture, our approach affords the flexibility to dynamically control the total computational cost and match the user-preferred task importance better.

Multi-Task Learning

Single-Stream Multi-Level Alignment for Vision-Language Pretraining

1 code implementation27 Mar 2022 Zaid Khan, Vijay Kumar BG, Xiang Yu, Samuel Schulter, Manmohan Chandraker, Yun Fu

Self-supervised vision-language pretraining from pure images and text with a contrastive loss is effective, but ignores fine-grained alignment due to a dual-stream architecture that aligns image and text representations only on a global level.

Question Answering Referring Expression +4

On Generalizing Beyond Domains in Cross-Domain Continual Learning

no code implementations CVPR 2022 Christian Simon, Masoud Faraki, Yi-Hsuan Tsai, Xiang Yu, Samuel Schulter, Yumin Suh, Mehrtash Harandi, Manmohan Chandraker

Humans have the ability to accumulate knowledge of new tasks in varying conditions, but deep neural networks often suffer from catastrophic forgetting of previously learned knowledge after learning a new task.

Continual Learning Knowledge Distillation

Object Detection with a Unified Label Space from Multiple Datasets

no code implementations ECCV 2020 Xiangyun Zhao, Samuel Schulter, Gaurav Sharma, Yi-Hsuan Tsai, Manmohan Chandraker, Ying Wu

To address this challenge, we design a framework which works with such partial annotations, and we exploit a pseudo labeling approach that we adapt for our specific case.

Object object-detection +1

Domain Adaptive Semantic Segmentation Using Weak Labels

no code implementations ECCV 2020 Sujoy Paul, Yi-Hsuan Tsai, Samuel Schulter, Amit K. Roy-Chowdhury, Manmohan Chandraker

In this work, we propose a novel framework for domain adaptation in semantic segmentation with image-level weak labels in the target domain.

Segmentation Semantic Segmentation +1

Domain Adaptation for Structured Output via Disentangled Patch Representations

no code implementations ICLR 2019 Yi-Hsuan Tsai, Kihyuk Sohn, Samuel Schulter, Manmohan Chandraker

To this end, we propose to learn discriminative feature representations of patches based on label histograms in the source domain, through the construction of a disentangled space.

Domain Adaptation Semantic Segmentation

Domain Adaptation for Structured Output via Discriminative Patch Representations

8 code implementations ICCV 2019 Yi-Hsuan Tsai, Kihyuk Sohn, Samuel Schulter, Manmohan Chandraker

Predicting structured outputs such as semantic segmentation relies on expensive per-pixel annotations to learn supervised models like convolutional neural networks.

Domain Adaptation Segmentation +2

Learning To Simulate

no code implementations ICLR 2019 Nataniel Ruiz, Samuel Schulter, Manmohan Chandraker

Simulation is a useful tool in situations where training data for machine learning models is costly to annotate or even hard to acquire.

Learning to Look around Objects for Top-View Representations of Outdoor Scenes

no code implementations ECCV 2018 Samuel Schulter, Menghua Zhai, Nathan Jacobs, Manmohan Chandraker

Given a single RGB image of a complex outdoor road scene in the perspective view, we address the novel problem of estimating an occlusion-reasoned semantic scene layout in the top-view.

Semantic Segmentation

Memory Warps for Learning Long-Term Online Video Representations

no code implementations28 Mar 2018 Tuan-Hung Vu, Wongun Choi, Samuel Schulter, Manmohan Chandraker

This paper proposes a novel memory-based online video representation that is efficient, accurate and predictive.

object-detection Object Detection

Deep Network Flow for Multi-Object Tracking

no code implementations CVPR 2017 Samuel Schulter, Paul Vernaza, Wongun Choi, Manmohan Chandraker

In this work, we demonstrate that it is possible to learn features for network-flow-based data association via backpropagation, by expressing the optimum of a smoothed network flow problem as a differentiable function of the pairwise association costs.

Graph Matching Multi-Object Tracking +1

Accurate Object Detection with Joint Classification-Regression Random Forests

no code implementations CVPR 2014 Samuel Schulter, Christian Leistner, Paul Wohlhart, Peter M. Roth, Horst Bischof

In this way, we can simultaneously predict the object probability of a window in a sliding window approach as well as regress its aspect ratio with a single model.

Classification General Classification +5

Alternating Decision Forests

no code implementations CVPR 2013 Samuel Schulter, Paul Wohlhart, Christian Leistner, Amir Saffari, Peter M. Roth, Horst Bischof

Contrary to Boosted Trees, in our method the loss minimization is an inherent part of the tree growing process, thus allowing to keep the benefits of common Random Forests, such as, parallel processing.

object-detection Object Detection

Cannot find the paper you are looking for? You can Submit a new open access paper.