Rad-ReStruct: A Novel VQA Benchmark and Method for Structured Radiology Reporting

11 Jul 2023  ·  Chantal Pellegrini, Matthias Keicher, Ege Özsoy, Nassir Navab ·

Radiology reporting is a crucial part of the communication between radiologists and other medical professionals, but it can be time-consuming and error-prone. One approach to alleviate this is structured reporting, which saves time and enables a more accurate evaluation than free-text reports. However, there is limited research on automating structured reporting, and no public benchmark is available for evaluating and comparing different methods. To close this gap, we introduce Rad-ReStruct, a new benchmark dataset that provides fine-grained, hierarchically ordered annotations in the form of structured reports for X-Ray images. We model the structured reporting task as hierarchical visual question answering (VQA) and propose hi-VQA, a novel method that considers prior context in the form of previously asked questions and answers for populating a structured radiology report. Our experiments show that hi-VQA achieves competitive performance to the state-of-the-art on the medical VQA benchmark VQARad while performing best among methods without domain-specific vision-language pretraining and provides a strong baseline on Rad-ReStruct. Our work represents a significant step towards the automated population of structured radiology reports and provides a valuable first benchmark for future research in this area. Our dataset and code is available at https://github.com/ChantalMP/Rad-ReStruct.

PDF Abstract

Datasets


Introduced in the Paper:

Rad-ReStruct

Used in the Paper:

VQA-RAD
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Structured Report Generation Rad-ReStruct hi-VQA Report Accuracy 30.2 # 1
F1 32.0 # 1
Medical Visual Question Answering VQA-RAD hi-VQA Overall Accuracy 76.3 # 6

Methods


No methods listed for this paper. Add relevant methods here