Nothing Special   »   [go: up one dir, main page]

Skip to main content

Can SegFormer be a True Competitor to U-Net for Medical Image Segmentation?

  • Conference paper
  • First Online:
Medical Image Understanding and Analysis (MIUA 2023)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 14122))

Included in the following conference series:

  • 719 Accesses

Abstract

The U-Net model, introduced in 2015, is established as the state-of-the-art architecture for medical image segmentation, along with its variants UNet++, nnU-Net, V-Net, etc. Vision transformers made a breakthrough in the computer vision world in 2021. Since then, many transformer based architectures or hybrid architectures (combining convolutional blocks and transformer blocks) have been proposed for image segmentation, that are challenging the predominance of U-Net. In this paper, we ask the question whether transformers could overtake U-Net for medical image segmentation. We compare SegFormer, one of the most popular transformer architectures for segmentation, to U-Net using three publicly available medical image datasets that include various modalities and organs: segmentation of cardiac structures in ultrasound images from the CAMUS challenge, segmentation of polyp in endoscopy images and segmentation of instrument in colonoscopy images from the MedAI challenge. We compare them in the light of various metrics (segmentation performance, training time) and show that SegFormer can be a true competitor to U-Net and should be carefully considered for future tasks in medical image segmentation.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 59.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 74.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Azad, R., et al.: Advances in medical image analysis with vision transformers: a comprehensive review (2023). https://doi.org/10.48550/arXiv.2301.03505, arXiv:2301.03505

  2. Chen, J., et al.: TransUNet: transformers make strong encoders for medical image segmentation. arXiv preprint arXiv:2102.04306 (2021)

  3. Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. CoRR abs/2010.11929 (2020). https://arxiv.org/abs/2010.11929

  4. Galdran, A., Anjos, A., Dolz, J., et al.: State-of-the-art retinal vessel segmentation with minimalistic models. Sci. Rep. 12, 6174 (2022). https://doi.org/10.1038/s41598-022-09675-y

  5. Hatamizadeh, A., et al.: UNETR: transformers for 3d medical image segmentation. In: WACV, pp. 1748–1758 (2022)

    Google Scholar 

  6. He, K., et al.: Transformers in medical image analysis. Intell. Med. 3(1), 59–78 (2023). https://doi.org/10.1016/j.imed.2022.07.002, https://www.sciencedirect.com/science/article/pii/S2667102622000717

  7. Isensee, F., et al.: nnU-Net: Self-adapting framework for u-net-based medical image segmentation. CoRR abs/1809.10486 (2018). http://arxiv.org/abs/1809.10486

  8. Kirillov, A., et al.: Segment anything (2023)

    Google Scholar 

  9. Leclerc, S., et al.: Deep learning for segmentation using an open large-scale dataset in 2d echocardiography. IEEE Trans. Med. Imaging 38(9), 2198–2210 (2019). https://doi.org/10.1109/tmi.2019.2900516

    Article  Google Scholar 

  10. Li, H., Hu, D., Liu, H., Wang, J., Oguz, I.: Cats: complementary CNN and transformer encoders for segmentation. In: 2022 IEEE 19th International Symposium on Biomedical Imaging (ISBI), pp. 1–5 (2022). https://doi.org/10.1109/ISBI52829.2022.9761596

  11. Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28

    Chapter  Google Scholar 

  12. Wang, W., et al.: Pyramid vision transformer: a versatile backbone for dense prediction without convolutions. CoRR abs/2102.12122 (2021). https://arxiv.org/abs/2102.12122

  13. Xie, E., Wang, W., Yu, Z., Anandkumar, A., Alvarez, J.M., Luo, P.: SegFormer: simple and efficient design for semantic segmentation with transformers. CoRR abs/2105.15203 (2021). https://arxiv.org/abs/2105.15203

  14. Zheng, S., et al.: Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. CoRR abs/2012.15840 (2020). https://arxiv.org/abs/2012.15840

Download references

Acknowledgments

The authors acknowledge the support of the French Agence Nationale de la Recherche (ANR), under grant Project-ANR-21-CE23-0013 (project MediSEG).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Caroline Petitjean .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Sourget, T., Hasany, S.N., Mériaudeau, F., Petitjean, C. (2024). Can SegFormer be a True Competitor to U-Net for Medical Image Segmentation?. In: Waiter, G., Lambrou, T., Leontidis, G., Oren, N., Morris, T., Gordon, S. (eds) Medical Image Understanding and Analysis. MIUA 2023. Lecture Notes in Computer Science, vol 14122. Springer, Cham. https://doi.org/10.1007/978-3-031-48593-0_8

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-48593-0_8

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-48592-3

  • Online ISBN: 978-3-031-48593-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics