Sep 3 – 4, 2025
Hörsaalgebäude, Campus Poppelsdorf, Universität Bonn
Europe/Berlin timezone

Adversarial Perturbations Improve Generalization of Confidence Prediction in Medical Image Segmentation

TAI.2.1
Sep 3, 2025, 2:00 PM
1h 30m
Open Space (first floor)

Open Space (first floor)

Board: TAI.2
Poster Trustworthy AI Poster Session

Speaker

Jonathan Lennartz

Description

Trustworthy methods for medical image segmentation should come with a reliable mechanism to estimate the quality of their results. Training a separate component for confidence prediction is relatively fast, and can easily be adapted to different quality metrics. However, the resulting estimates are usually not sufficiently reliable under domain shifts, for example when images are taken with different devices. We introduce a novel adversarial strategy for training confidence predictors for the widely used U-Net architecture that greatly improves such generalization. It is based on creating adversarial image perturbations, aimed at substantially decreasing segmentation quality, via the gradients of the confidence predictor, leading to images outside of the original training distribution. We observe that these perturbations initially have little effect on segmentation quality. However, including them in the training gradually improves the confidence predictor's understanding of what actually affects segmentation quality when moving outside of the training distribution. On two medical image segmentation tasks, we demonstrate that this strategy significantly improves direct quality estimates and outperforms a more computationally intensive state-of-the-art method - which only produces relative, rather than absolute, scores - on volumetric and surface Dice for out-of-distribution images.

Authors

Presentation materials

There are no materials yet.