Search Results for author: Johan W. Verjans

Found 10 papers, 8 papers with code

CAPE: CAM as a Probabilistic Ensemble for Enhanced DNN Interpretation

1 code implementation3 Apr 2024 Townim Faisal Chowdhury, Kewen Liao, Vu Minh Hieu Phan, Minh-Son To, Yutong Xie, Kevin Hung, David Ross, Anton Van Den Hengel, Johan W. Verjans, Zhibin Liao

Deep Neural Networks (DNNs) are widely used for visual classification tasks, but their complex computation process and black-box nature hinder decision transparency and interpretability.

Decision Making

Decomposing Disease Descriptions for Enhanced Pathology Detection: A Multi-Aspect Vision-Language Pre-training Framework

1 code implementation12 Mar 2024 Vu Minh Hieu Phan, Yutong Xie, Yuankai Qi, Lingqiao Liu, Liyang Liu, BoWen Zhang, Zhibin Liao, Qi Wu, Minh-Son To, Johan W. Verjans

Medical vision language pre-training (VLP) has emerged as a frontier of research, enabling zero-shot pathological recognition by comparing the query image with the textual descriptions for each disease.

Language Modelling Large Language Model

Structure-Preserving Synthesis: MaskGAN for Unpaired MR-CT Translation

1 code implementation30 Jul 2023 Minh Hieu Phan, Zhibin Liao, Johan W. Verjans, Minh-Son To

Extensive experiments demonstrate that MaskGAN outperforms state-of-the-art synthesis methods on a challenging pediatric dataset, where MR and CT scans are heavily misaligned due to rapid growth in children.

Anatomy Image Generation +1

CNN Attention Guidance for Improved Orthopedics Radiographic Fracture Classification

1 code implementation21 Mar 2022 Zhibin Liao, Kewen Liao, Haifeng Shen, Marouska F. van Boxel, Jasper Prijs, Ruurd L. Jaarsma, Job N. Doornberg, Anton Van Den Hengel, Johan W. Verjans

Convolutional neural networks (CNNs) have gained significant popularity in orthopedic imaging in recent years due to their ability to solve fracture classification problems.

Classification

Self-supervised Pseudo Multi-class Pre-training for Unsupervised Anomaly Detection and Segmentation in Medical Images

2 code implementations3 Sep 2021 Yu Tian, Fengbei Liu, Guansong Pang, Yuanhong Chen, Yuyuan Liu, Johan W. Verjans, Rajvinder Singh, Gustavo Carneiro

Pre-training UAD methods with self-supervised learning, based on computer vision techniques, can mitigate this challenge, but they are sub-optimal because they do not explore domain knowledge for designing the pretext tasks, and their contrastive learning losses do not try to cluster the normal training images, which may result in a sparse distribution of normal images that is ineffective for anomaly detection.

Contrastive Learning Data Augmentation +2

Constrained Contrastive Distribution Learning for Unsupervised Anomaly Detection and Localisation in Medical Images

1 code implementation5 Mar 2021 Yu Tian, Guansong Pang, Fengbei Liu, Yuanhong Chen, Seon Ho Shin, Johan W. Verjans, Rajvinder Singh, Gustavo Carneiro

Unsupervised anomaly detection (UAD) learns one-class classifiers exclusively with normal (i. e., healthy) images to detect any abnormal (i. e., unhealthy) samples that do not conform to the expected normal patterns.

Contrastive Learning Representation Learning +1

Weakly-supervised Video Anomaly Detection with Robust Temporal Feature Magnitude Learning

3 code implementations ICCV 2021 Yu Tian, Guansong Pang, Yuanhong Chen, Rajvinder Singh, Johan W. Verjans, Gustavo Carneiro

To address this issue, we introduce a novel and theoretically sound method, named Robust Temporal Feature Magnitude learning (RTFM), which trains a feature magnitude learning function to effectively recognise the positive instances, substantially improving the robustness of the MIL approach to the negative instances from abnormal videos.

Anomaly Detection In Surveillance Videos Contrastive Learning +2

Few-Shot Anomaly Detection for Polyp Frames from Colonoscopy

1 code implementation26 Jun 2020 Yu Tian, Gabriel Maicas, Leonardo Zorron Cheng Tao Pu, Rajvinder Singh, Johan W. Verjans, Gustavo Carneiro

Anomaly detection methods generally target the learning of a normal image distribution (i. e., inliers showing healthy cases) and during testing, samples relatively far from the learned distribution are classified as anomalies (i. e., outliers showing disease cases).

Anomaly Detection

Cannot find the paper you are looking for? You can Submit a new open access paper.