Closing the Domain Gap in Biomedical Imaging by In-Context Control Samples

2026-04-22Machine Learning

Machine Learning
AI summary

The authors address a big problem in biomedical imaging called batch effects, which are unwanted technical differences that mess up deep learning models when applied to new data batches. They introduce a new method called CS-ARM-BN that uses special control samples—unchanged reference images always included in experiments—to help models adapt better. Tested on a drug discovery task with large data, their approach significantly maintains accuracy on new batches, unlike standard methods. Their work shows that using these control samples for adaptation can effectively reduce batch effects and improve model reliability.

Batch effectsBiomedical imagingDeep learningMeta-learningBatch normalizationControl samplesMechanism-of-Action classificationDomain adaptationJUMP-CP datasetExperimental reproducibility
Authors
Ana Sanchez-Fernandez, Thomas Pinetz, Werner Zellinger, Günter Klambauer
Abstract
The central problem in biomedical imaging are batch effects: systematic technical variations unrelated to the biological signal of interest. These batch effects critically undermine experimental reproducibility and are the primary cause of failure of deep learning systems on new experimental batches, preventing their practical use in the real world. Despite years of research, no method has succeeded in closing this performance gap for deep learning models. We propose Control-Stabilized Adaptive Risk Minimization via Batch Normalization (CS-ARM-BN), a meta-learning adaptation method that exploits negative control samples. Such unperturbed reference images are present in every experimental batch by design and serve as stable context for adaptation. We validate our novel method on Mechanism-of-Action (MoA) classification, a crucial task for drug discovery, on the large-scale JUMP-CP dataset. The accuracy of standard ResNets drops from 0.939 $\pm$ 0.005, on the training domain, to 0.862 $\pm$ 0.060 on data from new experimental batches. Foundation models, even after Typical Variation Normalization, fail to close this gap. We are the first to show that meta-learning approaches close the domain gap by achieving 0.935 $\pm$ 0.018. If the new experimental batches exhibit strong domain shifts, such as being generated in a different lab, meta-learning approaches can be stabilized with control samples, which are always available in biomedical experiments. Our work shows that batch effects in bioimaging data can be effectively neutralized through principled in-context adaptation, which also makes them practically usable and efficient.