no code implementations • 8 Jun 2023 • Yunsung Lee, Jin-Young Kim, Hyojun Go, Myeongho Jeong, Shinhyeok Oh, Seungtaek Choi
In this paper, we address the performance degradation of efficient diffusion models by introducing Multi-architecturE Multi-Expert diffusion models (MEME).
no code implementations • 7 Jun 2023 • Jin-Young Kim, Soonwoo Kwon, Hyojun Go, Yunsung Lee, Seungtaek Choi
Self-supervised contrastive learning (CL) has achieved state-of-the-art performance in representation learning by minimizing the distance between positive pairs while maximizing that of negative ones.
1 code implementation • NeurIPS 2023 • Hyojun Go, Jinyoung Kim, Yunsung Lee, SeungHyun Lee, Shinhyeok Oh, Hyeongdon Moon, Seungtaek Choi
Through this, our approach addresses the issue of negative transfer in diffusion models by allowing for efficient computation of MTL methods.
no code implementations • 30 May 2023 • Hyun Seung Lee, Seungtaek Choi, Yunsung Lee, Hyeongdon Moon, Shinhyeok Oh, Myeongho Jeong, Hyojun Go, Christian Wallraven
To mitigate these issues, here we propose a novel retrieval approach CEAA that provides effective learning in educational text classification.
no code implementations • 26 May 2023 • Shinhyeok Oh, Hyojun Go, Hyeongdon Moon, Yunsung Lee, Myeongho Jeong, Hyun Seung Lee, Seungtaek Choi
To this end, we propose to paraphrase the reference question for a more robust QG evaluation.
1 code implementation • CVPR 2023 • Hyojun Go, Yunsung Lee, Jin-Young Kim, SeungHyun Lee, Myeongho Jeong, Hyun Seung Lee, Seungtaek Choi
For that, the existing practice is to fine-tune the guidance models with labeled data corrupted with noises.
no code implementations • 4 Oct 2022 • Yunsung Lee, Gyuseong Lee, Kwangrok Ryoo, Hyojun Go, JiHye Park, Seungryong Kim
In addition, through Fourier analysis of feature maps, the model's response patterns according to signal frequency changes, we observe which inductive bias is advantageous for each data scale.
no code implementations • 21 Nov 2021 • Yunsung Lee, Teakgyu Hong, Han-Cheol Cho, Junbum Cha, Seungryong Kim
Compared to previous works, our method shows better or comparable performance on dense prediction fine-tuning tasks.
1 code implementation • NeurIPS 2021 • Seokju Cho, Sunghwan Hong, Sangryul Jeon, Yunsung Lee, Kwanghoon Sohn, Seungryong Kim
We propose a novel cost aggregation network, called Cost Aggregation Transformers (CATs), to find dense correspondences between semantically similar images with additional challenges posed by large intra-class appearance and geometric variations.
Ranked #5 on Semantic correspondence on PF-WILLOW
4 code implementations • NeurIPS 2021 • Junbum Cha, Sanghyuk Chun, Kyungjae Lee, Han-Cheol Cho, Seunghyun Park, Yunsung Lee, Sungrae Park
Domain generalization (DG) methods aim to achieve generalizability to an unseen target domain by using only training data from the source domains.
Ranked #17 on Domain Generalization on TerraIncognita
1 code implementation • 11 Aug 2020 • Seokeon Choi, Junhyun Lee, Yunsung Lee, Alexander Hauptmann
We propose an improved discriminative model prediction method for robust long-term tracking based on a pre-trained short-term tracker.
no code implementations • CVPR 2020 • Junsoo Lee, Eungyeup Kim, Yunsung Lee, Dongjun Kim, Jaehyuk Chang, Jaegul Choo
However, it is difficult to prepare for a training data set that has a sufficient amount of semantically meaningful pairs of images as well as the ground truth for a colored image reflecting a given reference (e. g., coloring a sketch of an originally blue car given a reference green car).