Search Results for author: Hyunji Lee

Found 12 papers, 8 papers with code

Semiparametric Token-Sequence Co-Supervision

1 code implementation14 Mar 2024 Hyunji Lee, Doyoung Kim, Jihoon Jun, Sejune Joo, Joel Jang, Kyoung-Woon On, Minjoon Seo

Especially, the robustness of parametric token space which is established during the pretraining step tends to effectively enhance the stability of nonparametric sequence embedding space, a new space established by another language model.

Language Modelling

INSTRUCTIR: A Benchmark for Instruction Following of Information Retrieval Models

1 code implementation22 Feb 2024 Hanseok Oh, Hyunji Lee, Seonghyeon Ye, Haebin Shin, Hansol Jang, Changwook Jun, Minjoon Seo

Enhancing the capability of retrievers to understand intentions and preferences of users, akin to language model instructions, has the potential to yield more aligned search targets.

Information Retrieval Instruction Following +2

Back to Basics: A Simple Recipe for Improving Out-of-Domain Retrieval in Dense Encoders

1 code implementation16 Nov 2023 Hyunji Lee, Luca Soldaini, Arman Cohan, Minjoon Seo, Kyle Lo

Prevailing research practice today often relies on training dense retrievers on existing large datasets such as MSMARCO and then experimenting with ways to improve zero-shot generalization capabilities to unseen domains.

Data Augmentation Domain Generalization +2

How Well Do Large Language Models Truly Ground?

1 code implementation15 Nov 2023 Hyunji Lee, Sejune Joo, Chaeeun Kim, Joel Jang, Doyoung Kim, Kyoung-Woon On, Minjoon Seo

Reliance on the inherent knowledge of Large Language Models (LLMs) can cause issues such as hallucinations, lack of control, and difficulties in integrating variable knowledge.

KTRL+F: Knowledge-Augmented In-Document Search

2 code implementations14 Nov 2023 Hanseok Oh, Haebin Shin, Miyoung Ko, Hyunji Lee, Minjoon Seo

We introduce a new problem KTRL+F, a knowledge-augmented in-document search task that necessitates real-time identification of all semantic targets within a document with the awareness of external sources through a single natural query.

Retrieval

Zero-Shot Dense Video Captioning by Jointly Optimizing Text and Moment

no code implementations5 Jul 2023 Yongrae Jo, Seongyun Lee, Aiden SJ Lee, Hyunji Lee, Hanseok Oh, Minjoon Seo

This is accomplished by introducing a soft moment mask that represents a temporal segment in the video and jointly optimizing it with the prefix parameters of a language model.

Language Modelling Text Generation +1

Local 3D Editing via 3D Distillation of CLIP Knowledge

no code implementations CVPR 2023 Junha Hyung, Sungwon Hwang, Daejin Kim, Hyunji Lee, Jaegul Choo

Specifically, we present three add-on modules of LENeRF, the Latent Residual Mapper, the Attention Field Network, and the Deformation Network, which are jointly used for local manipulations of 3D features by estimating a 3D attention field.

Exploring the Practicality of Generative Retrieval on Dynamic Corpora

no code implementations27 May 2023 Soyoung Yoon, Chaeeun Kim, Hyunji Lee, Joel Jang, Sohee Yang, Minjoon Seo

Benchmarking the performance of information retrieval (IR) methods are mostly conducted with a fixed set of documents (static corpora); in realistic scenarios, this is rarely the case and the document to be retrieved are constantly updated and added.

Benchmarking Information Retrieval +1

Improving Probability-based Prompt Selection Through Unified Evaluation and Analysis

1 code implementation24 May 2023 Sohee Yang, Jonghyeon Kim, Joel Jang, Seonghyeon Ye, Hyunji Lee, Minjoon Seo

Previous works in prompt engineering for large language models have introduced different gradient-free probability-based prompt selection methods that aim to choose the optimal prompt among the candidates for a given task but have failed to provide a comprehensive and fair comparison between each other.

Prompt Engineering

Nonparametric Decoding for Generative Retrieval

1 code implementation5 Oct 2022 Hyunji Lee, Jaeyoung Kim, Hoyeon Chang, Hanseok Oh, Sohee Yang, Vlad Karpukhin, Yi Lu, Minjoon Seo

The generative retrieval model depends solely on the information encoded in its model parameters without external memory, its information capacity is limited and fixed.

Language Modelling Retrieval +1

Generative Multi-hop Retrieval

1 code implementation27 Apr 2022 Hyunji Lee, Sohee Yang, Hanseok Oh, Minjoon Seo

A common practice for text retrieval is to use an encoder to map the documents and the query to a common vector space and perform a nearest neighbor search (NNS); multi-hop retrieval also often adopts the same paradigm, usually with a modification of iteratively reformulating the query vector so that it can retrieve different documents at each hop.

Retrieval Text Retrieval

Cost-effective End-to-end Information Extraction for Semi-structured Document Images

no code implementations EMNLP 2021 Wonseok Hwang, Hyunji Lee, Jinyeong Yim, Geewook Kim, Minjoon Seo

A real-world information extraction (IE) system for semi-structured document images often involves a long pipeline of multiple modules, whose complexity dramatically increases its development and maintenance cost.

Cannot find the paper you are looking for? You can Submit a new open access paper.