Abstract
In the field of diagnosis and treatment planning of Coronavirus disease 2019 (COVID-19), accurate infected area segmentation is challenging due to the significant variations in the COVID-19 lesion size, shape, and position, boundary ambiguity, as well as complex structure. To bridge these gaps, this study presents a robust deep learning model based on a novel multi-scale contextual information fusion strategy, called Multi-Level Context Attentional Feature Fusion (MLCA2F), which consists of the Multi-Scale Context-Attention Network (MSCA-Net) blocks for segmenting COVID-19 lesions from Computed Tomography (CT) images. Unlike the previous classical deep learning models, the MSCA-Net integrates Multi-Scale Contextual Feature Fusion (MC2F) and Multi-Context Attentional Feature (MCAF) to learn more lesion details and guide the model to estimate the position of the boundary of infected regions, respectively. Practically, extensive experiments are performed on the Kaggle CT dataset to explore the optimal structure of MLCA2F. In comparison with the current state-of-the-art methods, the experiments show that the proposed methodology provides efficient results. Therefore, we can conclude that the MLCA2F framework has the potential to dramatically improve the conventional segmentation methods for assisting clinical decision-making.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
According to the recent survey [1], COVID-19 is ranked among the largest global public health crisis. Due to the influence of the virus, in the past few months, a high number of mortalities have been triggered over the world [2]. Currently, many recent clinical studies have reported that COVID-19 lesion segmentation from CT images is an important step in the clinical treatment assessment [3, 4] and can be beneficial in this context due to the presence of some prominent features [5]. However, finding an accurate method for COVID-19 lesion segmentation from CT images is a very challenging task due to the large variations in the COVID-19 lesion size, shape, and position, boundary ambiguity, as well as complex structure [6, 7].
More recently, convolutional neural network (CNN) architectures have also been extended for medical image segmentation [8, 9]. They have shown remarkable performance in segmentation tasks of medical images due to high representation power, filter sharing properties, and fast interference. However, research on COVID-19 lesion segmentation using CNN is still relatively scarce and often suffers from false positives. Following the above discussion, the main intention of this paper is to implement a segmentation algorithm for clinical decision-making based on CNN architectures to cope with this pandemic issue. The main contributions of this study can be summarized as follows:
-
(1)
It introduces a new framework, called MLCA2F, that integrates three fused MSCA-Net units for an effective COVID-19 lesion segmentation with a high focus on small and crucial lesion boundaries.
-
(2)
It proposes an effective MC2F paradigm to capture more specific and effective contextual information.
-
(3)
It designs a novel MCAF module that contributes to improving the performance of our segmentation system by forcing the MSCA-Net model to focus on COVID-19 lesions in CT images.
-
(4)
It presents a series of comparative experiments to validate the effectiveness of MLCA2F.
The remainder of this paper is organized as follows: In Sect. 2, we describe the novel methods applied for COVID-19 lesion segmentation. In Sect. 3, we detail the proposed approach. Then, we present the results of experiments realized on the Kaggle CT dataset in Sect. 4 and we discuss in detail the results of our proposed method in Sect. 5. Finally, we conclude this paper in Sect. 6.
2 Related works
Currently, deep learning models have been applied in various image processing tasks including medical imaging fields and showed promising performance [10]. For instance, pixel-wise classification-based CNN methods have achieved state-of-the-art performances in COVID-19 image segmentation [8]. In most cases, this is done by the U-Net algorithm which has been extensively studied over the past few years [11]. Khalifa et al. [12] proposed a standard U-Net architecture for COVID-19 lesions detection in chest CT scans. The same idea has been explored in [13], by introducing a U-Net architecture to realize accurate COVID-19 region localization in CT images. Chen et al. [5] proposed an variant of U-Net, called U-Net++ [14]. Unlike U-Net, U-Net++ consists of an encoder and decoder connecting through a series of nested dense convolutional blocks for detecting COVID-19 lesions on CT scans. Inspired by the success of attention-based CNN approaches in computer vision [8], researchers have proposed various methods to extend the use of the attention mechanism to automatically identify COVID-19-infected areas from chest CT images. Kuchana et al. [6] proposed the attention gates integrated into standard U-Net architecture for semantic segmentation of COVID-19 anomalies present in the CT scans of patients. Additionally, Zhou et al. [7] modeled the spatial and channel attention modules for automatic COVID-19 CT segmentation using U-Net to capture rich contextual relationships. Moreover, COVID-19 lesion segmentation from CT images has been also approached by Fan et al. [3] where the parallel partial decoder combined with a reverse attention module was adopted to establish the relationship between areas and boundary cues. Recently, region-based CNN methods have gained a lot of interest in medical imaging [9]. Mask R-CNN [15], which is considered as an instance of region-based CNNs, represents a significant potential in COVID-19 structure delineation. Ter-Sarkisov [2] performed COVID-19 lesion segmentation in CT images using Mask R-CNN to preserve the true region within the bounding box. Lately, multi-scale CNN-based approaches are broadly used for medical image segmentation [16] which can capture complementary contextual information [17]. In the COVID-19 lesion segmentation context, multi-scale CNN schemes have currently been addressed by only a few research works. Zheng et al. [4] presented a new multi-scale discriminative approach, called MSD-Net. In the MSD-Net, the authors proposed a pyramid convolution block, channel attention block, and residual refinement block for the accurate segmentation of COVID-19 lesions from CT images into three categories.
As a matter of fact, multi-scale contextual information learning is a crucial factor in effectively recovering accurate boundaries, extracting complex regions, and fulfilling better semantic segmentation performance. However, in the literature, only a restricted number of researches are available using multi-scale CNN for COVID-19 lesion segmentation from CT images. In view of the scarce existing studies on this domain, a detailed plan will be described in a fair amount of detail to capture fine details without losing any spatial information.
3 Proposed approach
As illustrated in Fig. 1, in this section, we first introduce the whole network architecture and then present the detail of each block. The proposed MLCA2F is developed in three levels each containing one MSCA-Net fused with the original input to improve the segmentation performance. Every MSCA-Net block is built upon the six modules of MC2F and MCAF to better delineate the COVID-19 regions. The specific ideas of the proposed framework are described as follows.
3.1 Data pre-processing
Pre-processing of the digital CT data is performed in two stages, including data size normalization and intensity correction to facilitate the application of MLCA2F for COVID-19 lesion segmentation. The resolution of CT images varies from \(401\times 630\) to \(630\times 630\), which requires a high cost of computation. To overcome this limitation, it is necessary to rescale the CT images. We first cropped the center area of the lung region and then adopt the pyramidal down-sampling process as indicated in [18] to resize all images to \(256\times 256\) pixels. The second pre-processing technique is based on adaptive bi-histogram equalization [19] algorithm to boost image brightness.
3.2 Multi-scale context-attention network (MSCA-Net)
The main component of our proposed framework is the MSCA-Net block that localizes the area of COVID-19 lesions in CT images. One of the key advantages of MSCA-Net is that it can be trained to directly focus on the target regions, resolve object ambiguities, and capture contextual information. It adopts the encoder \({E}_{{i}}\) and decoder \({D}_{{i}}\) structures, where i \(\in \) {1,2,3,4,5,6}, for semantic segmentation, equipped with MC2F and MCAF modules. The encoder and decoder components consist of six convolution blocks, \({C}_{{i}}\), each containing two \(3\times 3\) convolution layers followed by a ReLU activation function, group normalization [20], while the seventh convolution block is a transitional block between encoder and decoder. The output shape of each convolutional group is shown in Table 1. To progressively reduce the resolution of feature maps, the group normalization output of the encoder component is fed into the \(2\times 2\) max-pooling layer. Broadly, the encoder is the process of capturing the latent feature representation from input images, whereas the decoder path is designed in a way to upsample the feature maps and force the model to focus on infected areas in CT images through the cross-connections [21]. For accurate COVID-19 lesion segmentation, our framework mainly has three innovative mechanisms, the first is integrating the MC2F component to learn more COVID-19 lesion details and capture contextual information, the second is using the MCAF module to highlight the salient features, while the third is applying the fused MSCA-Net blocks through the element-wise maximum layer to significantly boost the performance.
3.3 Multi-scale contextual feature fusion (MC2F)
Considering that a single-scale feature model cannot highlight the contextual information and learn more COVID-19 lesion details, due to the large variations in the COVID-19 lesion size and shape, boundary ambiguity, as well as complex structure, we design a Multi-Scale Contextual Feature Fusion (MC2F). Taking advantage of the fact that the multi-scale features provide semantically different information, MC2F captures the detailed features from different receptive fields that are useful for locating boundaries. Specifically, in each MC2F block, four scales of convolutional layers with the size of [\(3\times 3\)], [\(5\times 5\)], [\(7\times 7\)], and [\(9\times 9\)] are involved. Within each scale structure, we integrate three convolutional layers such that the output of each layer is passed to the Scaled Exponential Linear Unit (SELU) activation function [22]. After that, each feature is fed to a Local Response Normalization (LRN) [23]. The output of the three LRN layers is later fused through the element-wise multiplication layer [24]. The four parallel scale paths are connected via a global element-wise multiplication layer. Finally, the output of the element-wise multiplication layer is fed into a \(1\times 1\) convolution layer to get the same MC2F input depth. The arrangement and description of the MC2F block are shown in Fig. 2. The constructed scale blocks are implemented in a similar structure and built by the same number of kernels (n) and stride (S), and they differ essentially from the kernel shape \([\hbox {width}\times \hbox {height}]\times n\), padding (P), LRN and SELU parameters. Table 2 shows the parameters specification of MC2F block.
3.4 Multi-context attentional feature (MCAF)
Since the current COVID-19 lesion segmentation methods are facing challenges posed by varying shapes, sizes, and positions of the infected area [6, 7], we propose a multi-context attentional feature mechanism, called MCAF, which guides the MLCA2F model to learn more discriminative features for separating the COVID-19 and non-COVID-19 pixels, and then, to estimate the position of the more complex regions in the whole CT image. As shown in Fig. 3, the proposed MCAF adopts three typical fusion layers, namely \({L}_{2}\) normalization fusion approach [25], and element-wise multiplication and addition layers [24]. Specifically, in the first, the output of \({E}_{{i}}\) is fed to the first convolutional layer. To generate the layers with the same \({E}_{{i}}\) depth, the \({D}_{{i}+1}\) goes through \(1\times 1\) convolution layer. Moreover, a bilinear interpolation [26] is introduced to upsample the decoded features, \({D}_{{i}+1}\), and obtain the same size resolution as the output of \({E}_{{i}}\). The output of the bilinear interpolation layer is passed to the second convolutional layer. The fused \({E}_{{i}}\) and upsampled \({D}_{{i}+1}\) features fed into the third convolutional layer. Note that, to keep the same predefined depth, the convolutional layers are used with the same kernel shape as \({C}_{{i}}\). After that, the obtained features are directly passed through an element-wise sigmoid function to generate multi-contextual attentive feature maps. At the end of MCAF, the second \({L}_{2}\) normalization fusion concatenates the output of the multi-contextual attentive features with the feature maps extracted from the upsampled decoding path to recover the COVID-19 lesion details.
4 Experimental results
In this section, we validate the proposed method on a publicly available CT dataset. First, we introduce the CT images used to assess the segmentation performance and the implementation details. Then, we highlight the details of the findings and report the performance comparison with the state-of-the-art methods.
4.1 CT image acquisition
In this study, the chest axial CT dataset is collected from Kaggle [27] which is a combination of [28, 29] and [30]. It consists of 20 full-chest 3D CT scans from different patients diagnosed with COVID-19. We extracted 3520 2D CT axial slices from the 3D volumes. All CT slices came with a corresponding ground truth defined by the radiologists and physicians. The CT dataset was divided into 60% for training, 20% for validation, and 20% for testing.
4.2 Implementation details
The proposed MLCA2F system was trained using Adaptive Moment Estimation (Adam) algorithm [31] with learning rate \( \alpha =10^{-5}\), first moment-decay \( \beta _{1}=0.9 \) and second moment-decay \( \beta _{2}=0.999 \) using he_normal algorithm [32] for weight initialization. The optimization ran for 200 epochs with a batch size of 64. All experiments were conducted on a desktop with an Intel Xeon E5-1603 v4 Quad-Core, 2.80 GHz processor base frequency, 10 MB cache, 16 GB RAM, and a single NVIDIA GTX Titan X GPU with 12 GB of installed memory. Our network was implemented using the TensorFlow framework and Keras as backend with Python 3 programming language. The training took 21 h and 46 min. For testing, on average, it took 6 h and 33 min.
4.3 Segmentation results
In this section, a series of comparative experiments are conducted to select the optimal network structure. To evaluate the segmentation results, we follow previous works [33, 34]. Then, we validated the MLCA2F framework through the following metrics: Accuracy (Acc), Sensitivity (Sen), Specificity (Spe), Dice Coefficient (DC), also called F1-measure, and Jaccard score (JC) compared with the corresponding ground truths. The experiments are mainly divided into two parts. The first part determines the best quality segmentation network by the absence of the MC2F module, indicated by the sign “-”, while the second part is evaluating the proposed approach by integrating the MC2F block, which is denoted by the sign “+”. Each part is evaluated as a function of the \(\hbox {Level}_{{n}}\), where n \(\in \) {1,2,3}. During this process, the use of the MC2F blocks improves the segmentation performance, as shown in Table 3. The gain of the MLCA2F at \(\hbox {Level}_{3}\) with MC2F blocks compared to the segmentation results of the model without the MC2F modules at \(\hbox {Level}_{1}\) is 8.20%, 3.70%, 12.70%, 7.72%, and 14.04% in terms of Acc, Sen, Spe, DC, and JC, respectively. Thus, the MC2F block makes an important contribution to the proposed method. Moreover, we also notice that the segmentation results of the MLCA2F at \(\hbox {Level}_{3}\) are more precise than the MLCA2F at \(\hbox {Level}_{1}\) and \(\hbox {Level}_{2}\). To evaluate the visual results, we displayed in Fig. 4 the COVID-19 lesion segmentation results of our approach on three representative CT images. It is remarkable that the network learned the detailed features from different receptive fields in the MLCA2F framework. Hence, we can conclude that these three cooperative MSCA-Net models are performing better in segmenting COVID-19 lesions.
4.4 Comparison with the state-of-the-art
Since the research on the Kaggle CT dataset is still relatively scarce, we have implemented the recent state-of-the-art image segmentation methods to demonstrate the effectiveness of our solution. Therefore, we compare our method against six segmentation approaches, namely U-Net [11], U-Net++ [14], Attention U-Net (AU-Net) [8], Mask R-CNN [15], Multi-Scale Discriminative Network (MSD-Net)[4], and Multi-scale Context-attention Network (MC-Net) [16]. In Table 4, we provide the quantitative results with respect to the five evaluation metrics. We observe from the results that MLCA2F outperforms the recent state-of-the-art methods by a large margin over all five metrics. Figure 5 shows visual segmentation results on the lung dataset obtained by MLCA2F and the other state-of-the-art methods. We can see from the figure that the proposed MLCA2F framework is more precise than the other methods, which can capture detailed boundary information.
5 Discussion
In this paper, we present an optimal data representation framework for COVID-19 lesion segmentation from CT images. Motivated by the outstanding performance of multi-scale CNN-based methods, the MLCA2F architecture is proposed to cope with the large variations in the COVID-19 lesion size and shape, differences in infected area position, boundary ambiguity, and complex structure problem. To better demonstrate the effectiveness of the proposed framework, two main study axes are presented. In the first axis, we have analyzed the performance results with the change in the number of MSCA-Net blocks, denoted by \(\hbox {Lavel}_{{n}}\). The second study axis is conducted to demonstrate the potentiality of the MC2F module. As shown in Table 3, the MC2F approach has demonstrated outstanding performance for the COVID-19 lesion segmentation. In addition, by integrating the MC2F module, as the number of MSCA-Net blocks increases, the improvements of MLCA2F increment significantly.
To confirm the quality of the MLCA2F algorithm, the experimental results are compared with six previous works. These methods can be broadly categorized into four classes, namely pixel-wise classification-based CNN methods [11, 14], attention-based CNNs [8], region-based CNN models [15], and multi-scale CNN methods [4, 16]. As clearly shown in Table 4, the proposed MLCA2F achieves the best overall performance and consistently outperforms all the other approaches on the Kaggle CT dataset by a large margin. Furthermore, as can be seen from Fig. 5, the COVID-19 lesions segmented by U-Net [11] and U-Net++ [14] have a poor segmentation effect and the small regions are not identified, while the CT images segmented by AU-Net [8] possess an incomplete shape of the lesion. Moreover, it can be observed that Mask R-CNN [15] completely fails to segment small tissues. MSD-Net [4] and MC-Net [16] can detect most COVID-19 areas. However, the protruded parts in the boundaries of the COVID-19 regions are not covered correctly and some fine details are missed. By contrast, our approach successfully captures fine structures, extracts more complex regions, and detects the complete shapes.
According to these experimental outcomes, we gain three significant reports. Firstly, it is noticeable that the single MSCA-Net stream without fusion is not sufficient for accurate infected area segmentation. Secondly, these findings reaffirm the superiority of the MC2F module which is more appropriate to preserve information. Finally, the experiments on the CT dataset reveal that using multi-level fusion combined with multi-scale contextual features can automatically focus on critical information, harvest discriminatory features, and strengthen the discriminative representation for COVID-19 lesion segmentation compared with the existing conventional segmentation frameworks.
6 Conclusion
In this study, we propose a novel deep learning model, called MLCA2F, which integrates three fused MSCA-Net units for COVID-19 lesion segmentation from CT scans. The MSCA-Net consists of two robust blocks: MC2F and MCAF. The MC2F module is implemented to capture smooth lesion details, while the MCAF block is modeled to force the MSCA-Net model to focus on COVID-19 lesions in CT images. Furthermore, the repetitive MSCA-Net blocks are introduced to handle the complex structure of COVID-19 lesions and significantly boost the performance. To demonstrate the effectiveness of the proposed MLCA2F framework, several experiments are conducted on the Kaggle CT dataset. The results of the MLCA2F segmentation experiment show that our method has better semantically consistent segmentation than the existing methods. In conclusion, the proposed segmentation method is beneficial for the automatic COVID-19 lesion segmentation study. In the future, we will further collect more challenging images, improve the proposed method by taking into consideration multi-modal analysis, and evaluate our approach on more datasets.
Data Availability
The datasets analyzed during the current study are available in the Kaggle repository, https://www.kaggle.com/datasets/andrewmvd/covid19-ct-scans.
References
Murakami, M., Miura, F., Kitajima, M., Fujii, K., Yasutaka, T., Iwasaki, Y., Ono, K., Shimazu, Y., Sorano, S., Okuda, T., Ozaki, A., Katayama, K., Nishikawa, Y., Kobashi, Y., Sawano, T., Abe, T., Saito, M., Tsubokura, M., Naito, W., Imoto, S.: COVID-19 risk assessment at the opening ceremony of the Tokyo 2020 Olympic Games. Microb. Risk Anal. 19, 00162 (2021)
Ter-Sarkisov, A.: One shot model for the prediction of COVID-19 and lesions segmentation in chest CT scans through the affinity among lesion mask features. Appl. Soft Comput. 116, 108261 (2022)
Fan, D., Zhou, T., Ji, G., Zhou, Y., Chen, G., Fu, H., Shen, J., Shao, L.: Inf-Net: automatic COVID-19 lung infection segmentation from CT images. IEEE Trans. Med. Imaging 39, 2626–2637 (2020)
Zheng, B., Liu, Y., Zhu, Y., Yu, F., Jiang, T., Yang, D., Xu, T.: MSD-Net: multi-scale discriminative network for COVID-19 lung infection segmentation on CT. IEEE Access. 8, 185786–185795 (2020)
Chen, J., Wu, L., Zhang, J., Zhang, L., Gong, D., Zhao, Y., Chen, Q., Huang, S., Yang, M., Yang, X., Hu, S., Wang, Y., Hu, X., Zheng, B., Zhang, K., Wu, H., Dong, Z., Xu, Y., Zhu, Y., Chen, X., Zhang, M., Yu, L., Cheng, F., Yu, H.: Deep learning-based model for detecting 2019 novel coronavirus pneumonia on high-resolution computed tomography. Sci. Rep. 10, 1–11 (2020)
Kuchana, M., Srivastava, A., Das, R., Mathew, J., Mishra, A., Khatter, K.: AI aiding in diagnosing, tracking recovery of COVID-19 using deep learning on Chest CT scans. Multimed. Tools Appl. 80, 9161–9175 (2020)
Zhou, T., Canu, S., Ruan, S.: Automatic COVID-19 CT segmentation using U-Net integrated spatial and channel attention mechanism. Int. J. Imaging Syst. Technol. 31, 16–27 (2020)
Oktay, O., Schlemper, J., Folgoc, L.L., Lee, M.C.H., Heinrich, M.P., Misawa, K., Mori, K., Mcdonagh, S., Hammerla, N., Kainz, B.: Attention U-Net: learning where to look for the pancreas. arXiv:1804.03999 (2018)
Liu, Z., Yang, C., Huang, J., Liu, S., Zhuo, Y., Lu, X.: Deep learning framework based on integration of S-Mask R-CNN and Inception-v3 for ultrasound image-aided diagnosis of prostate cancer. Futur. Gener. Comput. Syst. 114, 358–367 (2021)
Alalwan, N., Abozeid, A., ElHabshy, A., Alzahrani, A.: Efficient 3D deep learning model for medical image semantic segmentation. Alex. Eng. J. 60, 1231–1239 (2021)
Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 234–241 (2015)
Khalifa, N., Manogaran, G., Taha, M., Loey, M.: A deep learning semantic segmentation architecture for COVID-19 lesions discovery in limited chest CT datasets. Expert Syst. 39, e12742 (2021)
Peyvandi, A., Majidi, B., Peyvandi, S., Patra, J.: Computer-aided-diagnosis as a service on decentralized medical cloud for efficient and rapid emergency response intelligence. New Gener. Comput. 39, 677–700 (2021)
Zhou, Z., Rahman Siddiquee, M., Tajbakhsh, N., Liang, J.: UNet++: a nested U-Net architecture for medical image segmentation. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, pp. 3–11 (2018)
He, K., Gkioxari, G., Dollar, P., Girshick, R.: Mask R-CNN. In: Proceedings of 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2980–2988 (2017)
Xia, H., Ma, M., Li, H., Song, S.: MC-Net: multi-scale context-attention network for medical CT image segmentation. Appl. Intell. 52, 1508–1519 (2021)
Sang, H., Wang, Q., Zhao, Y.: Multi-scale context attention network for stereo matching. IEEE Access. 7, 15152–15161 (2019)
Bakkouri, I., Afdel, K.: Multi-scale CNN based on region proposals for efficient breast abnormality recognition. Multimed. Tools Appl. 78, 12939–12960 (2018)
Tang, J., Mat Isa, N.: Adaptive image enhancement based on bi-histogram equalization with a clipping limit. Comput. Electr. Eng. 40, 86–103 (2014)
Crytzer, T., Keramati, M., Anthony, S., Cheng, Y., Robertson, R., Dicianno, B.: Exercise prescription using a group-normalized rating of perceived exertion in adolescents and adults with spina bifida. PM &R 10, 738–747 (2018)
Trebing, K., Sta\({\grave{\text{n}}}\)czyk, T., Mehrkanoon, S.: SmaAt-UNet: precipitation nowcasting using a small attention-UNet architecture. Pattern Recogn. Lett. 145, 178–186 (2021)
Klambauer, G., Unterthiner, T., Mayr, A., Hochreiter, S.: Self-normalizing neural networks. In: Proceedings of the 31st International Conference on Neural Information Processing Systems 2017, pp. 972–981 (2017)
Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional neural networks. Commun. ACM 60, 84–90 (2017)
Ren, S., Han, C., Yang, X., Han, G., He, S.: TENet: triple excitation network for video salient object detection. In: European Conference on Computer Vision. Computer Vision: ECCV 2020, pp. 212–228 (2020)
Jin, X., Jiang, Q., Chu, X., Lang, X., Yao, S., Li, K., Zhou, W.: Brain medical image fusion using L2-norm-based features and fuzzy-weighted measurements in 2-D Littlewood–Paley EWT domain. IEEE Trans. Instrum. Meas. 69, 5900–5913 (2020)
Zhou, R., Cheng, Y., Liu, D.: Quantum image scaling based on bilinear interpolation with arbitrary scaling ratio. Quantum Inf. Process. 18, 1–19 (2019)
COVID-19 CT scans in Kaggle. https://www.kaggle.com/andrewmvd/covid19-ct-scans. Accessed 19 June 2020
Paiva, O.: CT scans of patients with COVID-19 from Wenzhou Medical University. https://coronacases.org/. Accessed 19 June 2020
Glick, Y.: COVID-19 Pneumonia. https://radiopaedia.org/playlists/25887. Accessed 19 June 2020
Jun, M., Cheng, G., Yixin, W., Xingle, A., Jiantao, G., Ziqi, Y., Minqing, Z., Xin, L., Xueyuan, D., Shucheng, C., Hao, W., Sen, M., Xiaoyu, Y., Ziwei, N., Chen, L., Lu, T., Yuntao, Z., Qiongjie, Z., Guoqiang, D., Jian, H.: COVID-19 CT lung and infection segmentation dataset. https://zenodo.org/record/3757476. Accessed 19 June 2020
Kingma, D., Ba, J.: Adam: a method for stochastic optimization. arXiv:1412.6980
He, K., Zhang, X., Ren, S., Sun, J.: Delving deep into rectifiers: surpassing human-level performance on ImageNet classification. In: 2015 IEEE International Conference on Computer Vision (ICCV) (2015)
Zhao, X., Wang, S., Zhao, J., Wei, H., Xiao, M., Ta, N.: Application of an attention U-Net incorporating transfer learning for optic disc and cup segmentation. Signal Image Video Process. 15, 913–921 (2020)
Al-Shamasneh, A., Jalab, H., Shivakumara, P., Ibrahim, R., Obaidellah, U.: Kidney segmentation in MR images using active contour model driven by fractional-based energy minimization. SIViP 14, 1361–1368 (2020)
Acknowledgements
This work is supported in part by the PPR2-2015 project under grant number 14UIZ2015, in part by the Scientific and Technological Research Support Program related to COVID-19 under grant number COV/2020/73, and in part by the Al Khawarizmi project under grant number ALKHAWARIZMI/2020/02 financed by the Moroccan government through the CNRST funding program.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Bakkouri, I., Afdel, K. MLCA2F: Multi-Level Context Attentional Feature Fusion for COVID-19 lesion segmentation from CT scans. SIViP 17, 1181–1188 (2023). https://doi.org/10.1007/s11760-022-02325-w
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11760-022-02325-w