Abstract
Recently, medical image synthesis has attracted the attention of an increasing number of researchers. However, most of current approaches suffer from the loss of multi-modal complementary information and thus fail to preserve the property of each modality, resulting in image distortion and texture detail loss. To alleviate this issue, a multi-modal magnetic resonance (MR) image synthesis algorithm based on a mixed attention fusion module in hybrid generative adversarial network is proposed. Firstly, a novel mixed attention fusion (MAF) module aggregating an adaptive fusion strategy (AFS) and a soft attention module is proposed to fuse the high-level semantic information and the low-level fine-grained feature at different scales between different layers to exploit rich representative complementary information adaptively. Subsequently, Resnet-bottlenect attention mechanism (Res-BAM) is designed to perform adaptive optimization and exploit mutual information while preserving the original property of each modality. Thereafter, the attention weight is inferred by a 1D channel feature map and a 2D spatial feature map, and multiplied with the original feature map in order to get the adaptive feature map, which is integrated with the original feature map in a residual connection to preserve the original property of each modality and prevent network degradation. Finally, the structural similarity (SSIM) and \({\text{L}}_{1}\)-norm are point-wise combined by an optimal weighting impact factor to preserve the high frequency information, brightness, color and SSIM, which are viewed as the original property of each modality. The experimental results demonstrate the superiority of our model on the state of the art in quantitative measures, reasonable visual quality and clinic significance.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Chartsias A, Joyce T, Dharmakumar R, Tsaftaris SA (2017) Adversarial image synthesis for unpaired multi-modal cardiac data. International workshop on simulation and synthesis in medical imaging. Springer, pp 3–13
Lee D, Kim J, Moon W-J, Ye JC (2019) CollaGAN: collaborative GAN for missing image data imputation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2487–2496
Gaur L, Bhatia U, Jhanjhi NZ, Muhammad G, Masud M (2021) Medical image-based detection of COVID-19 using Deep Convolution Neural Networks. Springer, Berlin Heidelberg
Biswas M, Gaur L, Alenezi F, Santosh KC, Mahbub MK (2022) Deep features to detect pulmonary abnormalities in chest X-rays due to infectious diseaseX: Covid-19, pneumonia, and tuberculosis. Inform Sci Int J 592:592
Yang H, Wang L, Xu Y, Liu X (2023) CovidViT: a novel neural network with self-attention mechanism to detect Covid-19 through X-ray images. Int J Mach Learn Cybern 14(3):973–987
Degen J, Heinrich MP (2016) Multi-atlas based pseudo-ct synthesis using multimodal image registration and local atlas fusion strategies. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 160–168
Burgos N et al (2014) Attenuation correction synthesis for hybrid PET-MR scanners: application to brain studies. IEEE Trans Med Imaging 33(12):2332–2341
Chen M, Carass A, Jog A, Lee J, Roy S, Prince JLJ (2017) Cross contrast multi-channel image registration using image synthesis for MR brain images. Med Image Anal 36:2–14
Jog A, Carass A, Roy S, Pham DL, Prince JL (2017) Random forest regression for magnetic resonance image synthesis. Med Image Anal 35:475–488
Wang Y et al (2016) Semisupervised tripled dictionary learning for standard-dose PET image prediction using low-dose PET and multimodal MRI. IEEE Trans Biomed Eng 64(3):569–579
Qu X, Hou Y, Lam F, Guo D, Zhong J, Chen ZJ (2014) Magnetic resonance image reconstruction from undersampled measurements using a patch-based nonlocal operator. Med Image Anal 18(6):843–856
Ye DH, Zikic D, Glocker B, Criminisi A, Konukoglu E (2013) Modality propagation: coherent synthesis of subject-specific scans with data-driven regularization. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, pp. 606-613
Huynh T et al (2015) Estimating CT image from MRI data using structured random forest and auto-context model. IEEE Trans Med Imaging 35(1):174–183
LeCun Y, Bottou L, Bengio Y, Haffner PJ (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324
Sharma DK, Gaur L, Okunbor D (2007) Image compression and feature extraction using Kohonen's self-organizing map neural network. J Strategic E-commerce 5(1/2):25
Li R et al (2014) Deep learning based imaging data completion for improved brain disease diagnosis. International conference on medical image computing and computer-assisted intervention. Springer, pp 305–312
Dong C, Loy CC, He K, Tang X (2015) Image super-resolution using deep convolutional networks. IEEE Trans Pattern Anal Mach Intell 38(2):295–307
Gao F, Wu T, Chu X, Yoon H, Xu Y, Patel B (2019) Deep residual inception encoder–decoder network for medical imaging synthesis. IEEE J Biomed Health Inform 24(1):39–49
Creswell A, White T, Dumoulin V, Arulkumaran K, Sengupta B, Bharath AA (2018) Generative adversarial networks: An overview. IEEE Signal Process Mag 35(1):53–65
Yi X, Walia E, Babyn P (2019) Generative adversarial network in medical imaging: a review. Med Image Anal 58:101552
Ben-Cohen A et al (2019) Cross-modality synthesis from CT to PET using FCN and GAN networks for improved automated lesion detection. Eng Appl Artif Intell 78:186–194
Armanious K et al (2020) MedGAN: Medical image translation using GANs. Comput Med Imaging Graph 79:101684
Dar SU, Yurt M, Karacan L, Erdem A, Erdem E, Çukur T (2019) Image synthesis in multi-contrast MRI with conditional generative adversarial networks. IEEE Trans Med Imaging 38(10):2375–2388
Nie D et al (2017) Medical image synthesis with context-aware generative adversarial networks. International conference on medical image computing and computer-assisted intervention. Springer, pp 417–425
Wang Y et al (2018) 3D auto-context-based locality adaptive multi-modality GANs for PET synthesis. IEEE Trans Med Imaging 38(6):1328–1339
Huang Y, Shao L, Frangi AF (2017) Simultaneous super-resolution and cross-modality synthesis of 3D medical images using weakly-supervised joint convolutional sparse coding. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 6070–6079
Du G, Zhou L, Li Z, Wang L, Lü K (2023) Neighbor-aware deep multi-view clustering via graph convolutional network. Inf Fusion 93:330–343
Du G, Zhou L, Yang Y, Lü K, Wang L (2021) Deep multiple auto-encoder-based multi-view clustering. Data Sci Eng 6(3):323–338
Olut S, Sahin YH, Demir U, Unal G (2018) Generative adversarial training for MRA image synthesis using multi-contrast MRI. In: International workshop on predictive intelligence in medicine, Springer, pp. 147-154
A. Sharma and G. J. I. t. o. m. i. Hamarneh, "Missing MRI pulse sequence synthesis using multi-modal generative adversarial network," vol. 39, no. 4, pp. 1170–1183, 2019.
Reaungamornrat S, Sari H, Catana C, Kamen A (2022) Multimodal image synthesis based on disentanglement representations of anatomical and modality specific features, learned using uncooperative relativistic GAN. Med Image Anal 80:102514
Yi D, Lei Z, Li SZ (2015) Shared representation learning for heterogenous face recognition. In: 2015 11th IEEE international conference and workshops on automatic face and gesture recognition (FG), vol. 1, IEEE, pp. 1–7
Chen G, Srihari SN (2015) Generalized K-fan multimodal deep model with shared representations. arXiv preprint arXiv:1503.07906
Ngiam J, Khosla A, Kim M, Nam J, Lee H, Ng AY (2011) Multimodal deep learning. In: Proceedings of the 28th international conference on machine learning (ICML-11), pp 689–696
Jog A, Roy S, Carass A, Prince JL (2013) Magnetic resonance image synthesis through patch regression. In: 2013 IEEE 10th International Symposium on Biomedical Imaging, IEEE, pp. 350–353
Wen C, Huai T, Zhang Q, Song Z, Cao F (2022) A new rotation forest ensemble algorithm. Int J Mach Learn Cybern 13(11):3569–3576
Huang Y, Shao L, Frangi AF (2017) Cross-modality image synthesis via weakly coupled and geometry co-regularized joint dictionary learning. IEEE Trans Med Imaging 37(3):815–827
Oksuz I (2022) Dictionary learning for medical image synthesis. Biomedical Image Synthesis and Simulation. Elsevier, pp. 79–89
Lee J, Carass A, Jog A, Zhao C, Prince JL (2017) Multi-atlas-based CT synthesis from conventional MRI with patch-based refinement for MRI-based radiotherapy planning. In: Medical Imaging 2017: Image Processing, vol. 10133: SPIE, pp. 434–439
Jog A, Carass A, Roy S, Pham DL, Prince JL (2015) MR image synthesis by contrast learning on neighborhood ensembles. Med Image Anal 24(1):63–76
Zhao C, Carass A, Lee J, He Y, Prince JL (2017) Whole brain segmentation and labeling from CT using synthetic MR images. In: International Workshop on Machine Learning in Medical Imaging, Springer, pp. 291-298
Miller MI, Christensen GE, Amit Y, Grenander U (1993) Mathematical textbook of deformable neuroanatomies. Proc Natl Acad Sci 90(24):11944–11948
Guibas JT, Virdi TS, Li PS (2017) Synthetic medical images from dual generative adversarial networks. arXiv preprint arXiv:1709.01872
Zhang T et al (2019) SkrGAN: sketching-rendering unconditional generative adversarial networks for medical image synthesis. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, pp. 777-785
You C et al (2019) CT super-resolution GAN constrained by the identical, residual, and cycle learning ensemble (GAN-CIRCLE). IEEE Trans Med Imaging 39(1):188–203
Isola P, Zhu J-Y, Zhou T, Efros AA (2017) Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1125–1134
Sangkloy P, Lu J, Fang C, Yu F, Hays J (2017) Scribbler: controlling deep image synthesis with sketch and color. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5400–5409
Pathak D, Krahenbuhl P, Donahue J, Darrell T, Efros AA (2016) Context encoders: feature learning by inpainting. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2536–2544
Zhang H; Xu T; Li H; Zhang S; Wang X; Huang X; Metaxas DN (2017) StackGAN: text to photo-realistic image synthesis with stacked generative adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 5907–5915
Mathesul S, Bhutkar G, Rambhad A (2022) AttnGAN: realistic text-to-image synthesis with attentional generative adversarial networks. In: IFIP Conference on Human-Computer Interaction
Yin G, Liu B, Sheng L, Yu N, Wang X, Shao J (2019) Semantics disentangling for text-to-image generation. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
Wang F et al (2017) Residual attention network for image classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3156–3164
Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
Woo S, Park J, Lee J-Y, Kweon IS (2018) Cbam: convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV), pp. 3–19
Hardoon DR, Szedmak S, Shawe-Taylor JJ (2004) Canonical correlation analysis: An overview with application to learning methods. Neural Comput 16(12):2639–2664
Nazarpour A, Adibi PJ (2015) Two-stage multiple kernel learning for supervised dimensionality reduction. Pattern Recogn 48(5):1854–1862
Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. International Conference on Medical image computing and computer-assisted intervention. Springer, pp 234–241
Zhang H, Goodfellow I, Metaxas D, Odena A (2019) Self-attention generative adversarial networks. In: International conference on machine learning PMLR, pp. 7354–7363
Oktay O, Schlemper J, Folgoc LL, Lee M, Heinrich M, Misawa K, Mori K, McDonagh S, Hammerla NY, Kainz B, Glocker B, Rueckert D (2018) Attention U-Net: learning where to look for the pancreas. arXiv preprint. arXiv:1804.03999
Li S, Liu J, Song Z (2022) Brain tumor segmentation based on region of interest-aided localization and segmentation U-Net. Int J Mach Learn Cybern 13(9):2435–2445
Zhu J-Y, Park T, Isola P, Efros AA (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp. 2223–2232
Menze BH et al (2014) The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans Med Imaging 34(10):1993–2024
Zhou T, Fu H, Chen G, Shen J, Shao LJ (2020) Hi-net: hybrid-fusion network for multi-modal MR image synthesis. IEEE Transact Med Imaging 39(9):2772–2781
Acknowledgements
This research was supported by “Famous teacher of teaching” of Yunnan 10000 Talents Program, the National Nature Science Foundation of China under Grants 62266049 and 62066047, the Program of Yunnan Key Laboratory of Intelligent Systems and Computing under grants 202205AG070003, the Postgraduate Research and Innovation Foundation of Yunnan University 2021Z075 and 2021Y256.
Author information
Authors and Affiliations
Contributions
HL: conceptualization, investigation, writing—review and editing. YH: validation, formal analysis, visualization, software, writing—original draft. JC: writing—review and editing. LZ: resources, writing—review & editing.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Li, H., Han, Y., Chang, J. et al. Hybrid generative adversarial network based on a mixed attention fusion module for multi-modal MR image synthesis algorithm. Int. J. Mach. Learn. & Cyber. 15, 2111–2130 (2024). https://doi.org/10.1007/s13042-023-02019-w
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13042-023-02019-w