Search Results for author: Sophia Althammer

Found 13 papers, 10 papers with code

Annotating Data for Fine-Tuning a Neural Ranker? Current Active Learning Strategies are not Better than Random Selection

no code implementations12 Sep 2023 Sophia Althammer, Guido Zuccon, Sebastian Hofstätter, Suzan Verberne, Allan Hanbury

We further find that gains provided by AL strategies come at the expense of more assessments (thus higher annotation costs) and AL strategies underperform random selection when comparing effectiveness given a fixed annotation cost.

Active Learning Domain Adaptation

Ranger: A Toolkit for Effect-Size Based Multi-Task Evaluation

1 code implementation24 May 2023 Mete Sertkan, Sophia Althammer, Sebastian Hofstätter

In this paper, we introduce Ranger - a toolkit to facilitate the easy use of effect-size-based meta-analysis for multi-task evaluation in NLP and IR.

Introducing Neural Bag of Whole-Words with ColBERTer: Contextualized Late Interactions using Enhanced Reduction

no code implementations24 Mar 2022 Sebastian Hofstätter, Omar Khattab, Sophia Althammer, Mete Sertkan, Allan Hanbury

Recent progress in neural information retrieval has demonstrated large gains in effectiveness, while often sacrificing the efficiency and interpretability of the neural model compared to classical approaches.

Information Retrieval Retrieval

PARM: A Paragraph Aggregation Retrieval Model for Dense Document-to-Document Retrieval

1 code implementation5 Jan 2022 Sophia Althammer, Sebastian Hofstätter, Mete Sertkan, Suzan Verberne, Allan Hanbury

However in the web domain we are in a setting with large amounts of training data and a query-to-passage or a query-to-document retrieval task.

Passage Retrieval Retrieval

Establishing Strong Baselines for TripClick Health Retrieval

2 code implementations2 Jan 2022 Sebastian Hofstätter, Sophia Althammer, Mete Sertkan, Allan Hanbury

We present strong Transformer-based re-ranking and dense retrieval baselines for the recently released TripClick health ad-hoc retrieval collection.

Re-Ranking Retrieval

A Time-Optimized Content Creation Workflow for Remote Teaching

1 code implementation11 Oct 2021 Sebastian Hofstätter, Sophia Althammer, Mete Sertkan, Allan Hanbury

We describe our workflow to create an engaging remote learning experience for a university course, while minimizing the post-production time of the educators.

Description-based Label Attention Classifier for Explainable ICD-9 Classification

no code implementations WNUT (ACL) 2021 Malte Feucht, Zhiliang Wu, Sophia Althammer, Volker Tresp

ICD-9 coding is a relevant clinical billing task, where unstructured texts with information about a patient's diagnosis and treatments are annotated with multiple ICD-9 codes.

Classification

DoSSIER@COLIEE 2021: Leveraging dense retrieval and summarization-based re-ranking for case law retrieval

1 code implementation9 Aug 2021 Sophia Althammer, Arian Askari, Suzan Verberne, Allan Hanbury

We address this challenge by combining lexical and dense retrieval methods on the paragraph-level of the cases for the first stage retrieval.

Passage Retrieval Re-Ranking +1

Linguistically Informed Masking for Representation Learning in the Patent Domain

1 code implementation10 Jun 2021 Sophia Althammer, Mark Buckley, Sebastian Hofstätter, Allan Hanbury

Domain-specific contextualized language models have demonstrated substantial effectiveness gains for domain-specific downstream tasks, like similarity matching, entity recognition or information retrieval.

Domain Adaptation Information Retrieval +2

Mitigating the Position Bias of Transformer Models in Passage Re-Ranking

1 code implementation18 Jan 2021 Sebastian Hofstätter, Aldo Lipani, Sophia Althammer, Markus Zlabinger, Allan Hanbury

In this work we analyze position bias on datasets, the contextualized representations, and their effect on retrieval results.

Passage Re-Ranking Position +4

Cross-domain Retrieval in the Legal and Patent Domains: a Reproducibility Study

1 code implementation21 Dec 2020 Sophia Althammer, Sebastian Hofstätter, Allan Hanbury

For reproducibility and transparency as well as to benefit the community we make our source code and the trained models publicly available.

Information Retrieval Language Modelling +1

Improving Efficient Neural Ranking Models with Cross-Architecture Knowledge Distillation

1 code implementation6 Oct 2020 Sebastian Hofstätter, Sophia Althammer, Michael Schröder, Mete Sertkan, Allan Hanbury

Based on this finding, we propose a cross-architecture training procedure with a margin focused loss (Margin-MSE), that adapts knowledge distillation to the varying score output distributions of different BERT and non-BERT passage ranking architectures.

Knowledge Distillation Passage Ranking +3

Cannot find the paper you are looking for? You can Submit a new open access paper.