no code implementations • ECCV 2020 • Yunjae Jung, Donghyeon Cho, Sanghyun Woo, In So Kweon
In order to summarize a content video properly, it is important to grasp the sequential structure of video as well as the long-term dependency between frames.
no code implementations • 21 Oct 2021 • Dong-Jin Kim, Jae Won Cho, Jinsoo Choi, Yunjae Jung, In So Kweon
In this work, we address Active Learning in the multi-modal setting of Visual Question Answering (VQA).
1 code implementation • 23 Jul 2021 • Jae Won Cho, Dong-Jin Kim, Yunjae Jung, In So Kweon
Recent state-of-the-art active learning methods have mostly leveraged Generative Adversarial Networks (GAN) for sample acquisition; however, GAN is usually known to suffer from instability and sensitivity to hyper-parameters.
no code implementations • 13 Apr 2021 • Jae Won Cho, Dong-Jin Kim, Jinsoo Choi, Yunjae Jung, In So Kweon
In this work, we address the issues of missing modalities that have arisen from the Visual Question Answer-Difference prediction task and find a novel method to solve the task at hand.
no code implementations • 3 Feb 2020 • Yunjae Jung, Dahun Kim, Sanghyun Woo, Kyung-Su Kim, Sungjin Kim, In So Kweon
In this paper, we propose to explicitly learn to imagine a storyline that bridges the visual gap.
Ranked #7 on Visual Storytelling on VIST
1 code implementation • 24 Nov 2018 • Yunjae Jung, Donghyeon Cho, Dahun Kim, Sanghyun Woo, In So Kweon
The proposed variance loss allows a network to predict output scores for each frame with high discrepancy which enables effective feature learning and significantly improves model performance.
Ranked #3 on Unsupervised Video Summarization on SumMe
Supervised Video Summarization Unsupervised Video Summarization