Search Results for author: Jae-Pil Heo

Found 24 papers, 15 papers with code

VarSR: Variational Super-Resolution Network for Very Low Resolution Images

no code implementations ECCV 2020 Sangeek Hyun, Jae-Pil Heo

As is well known, single image super-resolution (SR) is an ill-posed problem where multiple high resolution (HR) images can be matched to one low resolution (LR) image due to the difference of their representation capabilities.

Image Super-Resolution

Diversity-aware Channel Pruning for StyleGAN Compression

1 code implementation20 Mar 2024 Jiwoo Chung, Sangeek Hyun, Sang-Heon Shim, Jae-Pil Heo

Specifically, by assessing channel importance based on their sensitivities to latent vector perturbations, our method enhances the diversity of samples in the compressed model.

Image Generation Unconditional Image Generation

Noise-free Optimization in Early Training Steps for Image Super-Resolution

1 code implementation29 Dec 2023 MinKyu Lee, Jae-Pil Heo

To tackle this issue, we propose a novel optimization method that can effectively remove the inherent noise term in the early steps of vanilla training by estimating the optimal centroid and directly optimizing toward the estimation.

Image Super-Resolution

VLCounter: Text-aware Visual Representation for Zero-Shot Object Counting

1 code implementation27 Dec 2023 Seunggu Kang, WonJun Moon, Euiyeon Kim, Jae-Pil Heo

Zero-Shot Object Counting (ZSOC) aims to count referred instances of arbitrary classes in a query image without human-annotated exemplars.

Object Counting Zero-Shot Counting

Towards Squeezing-Averse Virtual Try-On via Sequential Deformation

1 code implementation26 Dec 2023 Sang-Heon Shim, Jiwoo Chung, Jae-Pil Heo

In this paper, we first investigate a visual quality degradation problem observed in recent high-resolution virtual try-on approach.

Virtual Try-on

Task-Disruptive Background Suppression for Few-Shot Segmentation

1 code implementation26 Dec 2023 Suho Park, SuBeen Lee, Sangeek Hyun, Hyun Seok Seong, Jae-Pil Heo

Based on these two scores, we define a query background relevant score that captures the similarity between the backgrounds of the query and the support, and utilize it to scale support background features to adaptively restrict the impact of disruptive support backgrounds.

Style Injection in Diffusion: A Training-free Approach for Adapting Large-scale Diffusion Models for Style Transfer

1 code implementation11 Dec 2023 Jiwoo Chung, Sangeek Hyun, Jae-Pil Heo

Despite the impressive generative capabilities of diffusion models, existing diffusion model-based style transfer methods require inference-stage optimization (e. g. fine-tuning or textual inversion of style) which is time-consuming, or fails to leverage the generative ability of large-scale diffusion models.

Style Transfer

Correlation-guided Query-Dependency Calibration in Video Representation Learning for Temporal Grounding

2 code implementations15 Nov 2023 WonJun Moon, Sangeek Hyun, SuBeen Lee, Jae-Pil Heo

Dummy tokens conditioned by text query take portions of the attention weights, preventing irrelevant video clips from being represented by the text query.

Highlight Detection Moment Retrieval +3

Self-Feedback DETR for Temporal Action Detection

no code implementations ICCV 2023 JiHwan Kim, Miso Lee, Jae-Pil Heo

In this paper, we point out the problem in the self-attention of DETR for TAD; the attention modules focus on a few key elements, called temporal collapse problem.

Action Detection

Task-Oriented Channel Attention for Fine-Grained Few-Shot Classification

1 code implementation28 Jul 2023 SuBeen Lee, WonJun Moon, Hyun Seok Seong, Jae-Pil Heo

While TDM influences high-level feature maps by task-adaptive calibration of channel-wise importance, we further introduce Instance Attention Module (IAM) operating in intermediate layers of feature extractors to instance-wisely highlight object-relevant channels, by extending QAM.

Cross-Domain Few-Shot Fine-Grained Image Classification

Leveraging Hidden Positives for Unsupervised Semantic Segmentation

1 code implementation CVPR 2023 Hyun Seok Seong, WonJun Moon, SuBeen Lee, Jae-Pil Heo

Specifically, we add the loss propagating to local hidden positives, semantically similar nearby patches, in proportion to the predefined similarity scores.

Contrastive Learning Unsupervised Semantic Segmentation

Query-Dependent Video Representation for Moment Retrieval and Highlight Detection

1 code implementation CVPR 2023 WonJun Moon, Sangeek Hyun, Sanguk Park, Dongchan Park, Jae-Pil Heo

As we observe the insignificant role of a given query in transformer architectures, our encoding module starts with cross-attention layers to explicitly inject the context of text query into video representation.

Highlight Detection Moment Retrieval +4

Minority-Oriented Vicinity Expansion with Attentive Aggregation for Video Long-Tailed Recognition

1 code implementation24 Nov 2022 WonJun Moon, Hyun Seok Seong, Jae-Pil Heo

A dramatic increase in real-world video volume with extremely diverse and emerging topics naturally forms a long-tailed video distribution in terms of their categories, and it spotlights the need for Video Long-Tailed Recognition (VLTR).

Tailoring Self-Supervision for Supervised Learning

1 code implementation20 Jul 2022 WonJun Moon, Ji-Hwan Kim, Jae-Pil Heo

Our exhaustive experiments validate the merits of LoRot as a pretext task tailored for supervised learning in terms of robustness and generalization capability.

Adversarial Robustness Data Augmentation +3

Difficulty-Aware Simulator for Open Set Recognition

1 code implementation20 Jul 2022 WonJun Moon, Junho Park, Hyun Seok Seong, Cheol-Ho Cho, Jae-Pil Heo

Furthermore, moderate- and easy-difficulty samples are also yielded by our modified GAN and Copycat, respectively.

Generative Adversarial Network Open Set Learning

Task Discrepancy Maximization for Fine-grained Few-Shot Classification

1 code implementation CVPR 2022 SuBeen Lee, WonJun Moon, Jae-Pil Heo

Specifically, TDM learns task-specific channel weights based on two novel components: Support Attention Module (SAM) and Query Attention Module (QAM).

Ranked #9 on Few-Shot Image Classification on CUB 200 5-way 5-shot (using extra training data)

Classification Few-Shot Image Classification

Local Attention Pyramid for Scene Image Generation

no code implementations CVPR 2022 Sang-Heon Shim, Sangeek Hyun, DaeHyun Bae, Jae-Pil Heo

To address this, we propose a novel attention module, Local Attention Pyramid (LAP) module tailored for scene image synthesis, that encourages GANs to generate diverse object classes in a high quality by explicit spread of high attention scores to local regions, since objects in scene images are scattered over the entire images.

Image Generation Object

Pixel-Level Kernel Estimation for Blind Super-Resolution

1 code implementation IEEE Access 2021 Jaihyun Lew, Euiyeon Kim, Jae-Pil Heo

Furthermore, based on this assumption, there have been attempts to estimate the blur kernel of a given LR image, since correct kernel priors are known to be helpful in super-resolution.

Blind Super-Resolution Super-Resolution

Self-Supervised Video GANs: Learning for Appearance Consistency and Motion Coherency

no code implementations CVPR 2021 Sangeek Hyun, JiHwan Kim, Jae-Pil Heo

The proposed tasks enable the discriminators to learn representations of appearance and temporal context, and force the generator to synthesize videos with consistent appearance and natural flow of motions.

Contrastive Learning

Product Quantizer Aware Inverted Index for Scalable Nearest Neighbor Search

no code implementations ICCV 2021 Haechan Noh, TaeHo Kim, Jae-Pil Heo

To address the raised question, we suggest a joint optimization of the coarse and fine quantizers by substituting the original objective of the coarse quantizer to end-to-end quantization distortion.

Quantization

Collaborative Learning With Disentangled Features for Zero-Shot Domain Adaptation

no code implementations ICCV 2021 Won Young Jhoo, Jae-Pil Heo

In such a case, we can capture the domain shift between the source domain and target domain from an unseen task and transfer it to the task of interest, which is known as zero-shot domain adaptation (ZSDA).

Domain Adaptation

Shortlist Selection With Residual-Aware Distance Estimator for K-Nearest Neighbor Search

no code implementations CVPR 2016 Jae-Pil Heo, Zhe Lin, Xiaohui Shen, Jonathan Brandt, Sung-Eui Yoon

We have tested the proposed method with the inverted index and multi-index on a diverse set of benchmarks including up to one billion data points with varying dimensions, and found that our method robustly improves the accuracy of shortlists (up to 127% relatively higher) over the state-of-the-art techniques with a comparable or even faster computational cost.

Quantization

Distance Encoded Product Quantization

no code implementations CVPR 2014 Jae-Pil Heo, Zhe Lin, Sung-Eui Yoon

This result is achieved mainly because our method accurately estimates distances between two data points with the new binary codes and distance metric.

Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.