Jiayi WuZikai ChenYiyi ChenZiteng CuiShaoyi DuXiaofeng YangJingmin Xin
Abstract Background Medical image segmentation is vital for clinical diagnosis, yet deep learning models face domain shift challenges when test data distributions differ from training data. Single‐source domain generalization (DG) has emerged as a solution to these limitations of deep learning models by training models on a single source domain to generalize to unseen target domains. Current single‐source DG methods employ domain randomization (e.g., input/feature‐space perturbations) to simulate unseen target domains, but face two key limitations: (1) Texture bias in CNN‐based architectures due to their local processing nature, leading to overfitting and poor generalization; (2) Restricted augmented style diversity caused by solely source‐dependent feature perturbations and input‐ or feature‐only augmentation, resulting in insufficient coverage of target domain distributions and degraded model's generalization ability. Purpose Motivated by the Fourier transform's capability to model global dependency and decouple domain invariant features, we propose Random Domain Generalization (RandDG), a frequency‐aware domain randomization method for single‐source DG in medical image segmentation. This method enhances generalization ability through coordinated input and feature spaces perturbations while maintaining parameter efficiency via lightweight architectural design in the frequency domain. Methods The proposed RandDG incorporates a novel Global U‐Shape Network (GUNet) segmentation architecture for efficient long‐range dependency modeling and a Uniform Low Frequency spectrum Transform (ULoFT) filter for feature‐space perturbation to enhance domain generalization ability, facilitating the learning of global domain invariant structural information through consistency constraints within the dual‐space randomization framework. More specifically, first, the GUNet architecture combines a lightweight encoder with frequency‐domain global filtering and forms a U‐shape structure through a skip‐connected decoder, utilizing Fourier transforms to capture long‐range dependencies while minimizing parameters. Second, the ULoFT filter perturbs feature‐space styles by Bernoulli mixing statistics of the source low‐frequency amplitude spectrum with uniformly sampled values, thereby reducing sole reliance on source statistics and broadening the coverage of potential target domains. Third, a dual‐space randomization framework simultaneously applies input‐space style augmentation via global intensity nonlinear augmentation (GIN) filters and feature‐space amplitude spectrum perturbation via ULoFT filter, regularized by a consistency loss to enforce domain invariant representation learning. Results Experiments were conducted on an abdominal CT‐MRI dataset for abdominal multi‐organ segmentation from CT to T2 spectral presaturation with inversion recovery (T2‐SPIR) MRI and a cross‐center prostate dataset for prostate segmentation from one center to others. The proposed RandDG method demonstrated superior generalization ability, significantly improving robustness against domain shifts compared to competitive methods. When tested on unseen domains, the proposed method achieved an average DSC of 87.96% and an average HD of 4.82 mm on the abdominal CT‐MRI dataset, as well as an average DSC of 75.95% and an average HD of 8.36 mm on the cross‐center prostate dataset. Compared to the UNet baseline, the proposed method achieved an improvement of 10.65% in average DSC and a reduction of 5.23 mm in average HD on the abdominal CT‐MRI dataset, as well as 19.36% higher average DSC and 3.39 mm lower average HD on the cross‐center prostate dataset. Ablation studies confirmed the effectiveness of each component of the proposed RandDG method. Conclusions The proposed RandDG method effectively addresses the critical challenges of texture bias and limited style diversity in single‐source DG for medical image segmentation through frequency‐aware dual‐space randomization framework. This method shows promise for practical deployment in clinical settings, where target domain data may not always be available.
Zikai ChenJiayi WuJincheng LiWeiliang ZuoChenzdong LiJingmin XinNanning Zheng
Ziyang ChenYongsheng PanYiwen YeHengfei CuiYong Xia