Nothing Special   »   [go: up one dir, main page]

Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2021 Aug 1.
Published in final edited form as: Comput Biol Med. 2020 Jul 16;123:103914. doi: 10.1016/j.compbiomed.2020.103914

Spatially Localized Sparse Representations for Breast Lesion Characterization

Keni Zheng a, Chelsea Harris a, Predrag Bakic b, Sokratis Makrogiannis a
PMCID: PMC7416513  NIHMSID: NIHMS1613625  PMID: 32768050

Abstract

Rationale

The topic of sparse representation of samples in high dimensional spaces has attracted growing interest during the past decade. In this work, we develop sparse representation-based methods for classification of radiological imaging patterns of breast lesions into benign and malignant states.

Methods

We propose a spatial block decomposition method to address irregularities of the approximation problem and to build an ensemble of classifiers (CL) that we expect to yield more accurate numerical solutions than conventional whole-region of interest (ROI) sparse analyses. We introduce two classification decision strategies based on maximum a posteriori probability (BBMAP-S), or a log likelihood function (BBLL-S).

Results

To evaluate the performance of the proposed approach we used cross-validation techniques on imaging datasets with disease class labels. We utilized the proposed approach for separation of breast lesions into benign and malignant categories in mammograms. The level of difficulty is high in this application and the accuracy may depend on the lesion size. Our results indicate that the proposed integrative sparse analysis addresses the ill-posedness of the approximation problem, producing AUC (area under the receiver operating curve) value of 89.1% for randomized 30-fold cross-validation.

Conclusions

Furthermore, our comparative experiments showed that the BBLL-S decision function may yield more accurate classification than BBMAP-S because BBLL-S accounts for possible estimation bias.

Keywords: sparse analysis, breast lesion characterization, CAD/CADx

1. INTRODUCTION

Early detection and diagnosis of disease is an area of great importance and can ultimately extend the human life span. Breast cancer diagnosis is a significant area of concern and because of its significance, automated detection and diagnosis of breast cancer is a popular field of research [111]. Early diagnosis has been shown to reduce mortality and significantly improve quality of life. To achieve this, mammograms are used to help detect breast cancer at an early stage. Detection and diagnosis of breast cancer is not an easy task even for highly trained radiologists, making automated diagnosis a very useful tool.

In medical imaging classification applications, the development of machine learning classification models begins with selecting or assembling medical imaging datasets. Various breast imaging databases have been used in classification applications to classify not only mass types, but microcalifications, architectural distortions, and asymmetric densities. Breast imaging databases come in various sizes and may contain different imaging modalities ranging from mammograms and digital tomosynthesis to MRI, PET, and CT [12, 13]. Two of the most widely used datasets in breast imaging classification research are the MIAS (Mammographic Image Analysis Society digital mammographic) and DDSM (Digital Database in Screening Mammography) datasets [13]. In the following section we provide a brief summary of the subsequent steps in medical imaging classification models and a survey of classification techniques that use these datasets. Result comparisons of our method with existing techniques will be discussed in the Discussion section of this paper.

Feature extraction and selection is a critical step in medical imaging classification systems. The features that are extracted and selected at this stage will be used to classify images. Conventional classification models such as those used in [1419] use specific rules to craft features. Texture, shape and intensity features were extracted in [15]. Among the extracted features genetic algorithm (GA) selected the most appropriate features. Zernike moments have also been used to extract features due to their ability to well describe the shape of objects [1921]. Generalized pseudo-zernike and pseudo-zernike moments were used in [19] as texture feature descriptors.

However, in more recent years, state of the art techniques using neural networks have been used to extract and select features [12, 22]. Among these techniques, Convolutional Neural Nets (CNNs) are a popular choice. Although convolutional network architectures were introduced in the late 1980’s [23], they have gained significant popularity in the last 10 years, driven by algorithmic improvements and technological advances. Key advances in design and application of CNNs were presented in [24] that significantly improved the state of the art in object recognition for the imagenet dataset. In the recent years, intense research in the deep learning field has yielded impressive results in classification, recognition and segmentation tasks on natural images. These techniques have been successfully applied to medical imaging domain [12]. Because of the limited sizes of medical imaging datasets, the research community has transerred solutions for using networks that were originally pre-trained on extensive databases of natural images, to medical imaging data. A commonly used training strategy for training neural networks for medical imaging classification applications is transfer learning. This is a learning strategy that transfers knowledge from a pre-trained network for applications outside of its original training application. In [25], pretrained VGG16, ResNet50, and Inception v3 networks were customized to be applied on different datasets.

The final major stage in medical imaging classification techniques is the learning model. At this stage, machine learning is performed through supervised, unsupervised, or semi-supervised techniques to make a classification decision. Following the training of the classifier model, the performance of the model is evaluated on unlabeled samples.

Our work focuses on computer aided diagnosis (CAD) of breast lesions using sparse representation classification and texture-based classification. We applied our method of Integrative Ensemble Sparse Representation on the digitized mammographic dataset, MIAS dataset. The MIAS dataset consists of 322 digitized MLO images; 66 benign lesion subjects, 51 malignant lesion subjects and 203 normal subjects. In our Integrative Ensemble Sparse Analysis method we compare the results of classification experiments using a ROI size of 64×64 with 10, 20, and 30-fold cross validation (CV). We explore the use of sparse modeling in the classification of benign and malignant masses.

This paper highlights the advantages of the use of spatial information in a sparse analysis approach for classification of benign and malignant breast masses. In the Materials and Methods section we overview conventional sparse representation classification (SRC), our Integrative Ensemble Sparse Analysis technique and the decision functions. In the Results section and the Discussion section, we describe the data and classification experiments using Integrative Ensemble Sparse Analysis in comparison with conventional SRC and CNN-based classification and we discuss their results. We perform analysis for the ROI size of 64 × 64 and explore the relationship of block-size and cross validation with classification accuracy.

2. Materials and Methods

2.1. Sparse Representation Classification

Conventional sparse representations classification is a classification technique that utilizes sparse matrices (matrices with many zero entries). The nonzero entries of sparse matrices are of great importance as they represent the more relevant features of an image. Sparse representation classification has been utilized in the areas of image processing [2628], face recognition [2931], digit recognition [32], texture classification [33] and biomedical image classification [11, 34, 35]. Sparse representation techniques typically follow a two stage system, 1: dictionary learning (learning the dictionary to find approximations of signals) and 2: sparse coding (computing the sparse representation coefficients of a test signal). SRC first forms a dictionary matrix D from the training set.

D=[u1,1,u1,2,,uk,s]l×s (1)

with si data samples for the ith class, l is the dimension of the sample vector, and k is the number of classes. The total number of samples is s such that s = Σi si.

Given a dictionary of patterns D, corresponding to the object classes, sparse coding aims to solve for an unknown vector x0 that represents the sample y as a linear combination of the atoms in D:

y=Dx0. (2)

The solution vector x0 is assumed to be sparse; the sparsity of x0 ultimately ensures that signal recovery is possible. Sparsity is measured by the l0 norm. If the solution x0 is sparse enough, it is approximated by the solution x^ of the l1-minimization problem:

(l1):x^=arg minx1s.t.Dx=y. (3)

2.2. Integrative Ensemble Sparse Analysis

Our Integrative Ensemble Sparse Analysis technique regularizes the mathematical optimization problem of sparse representation by reducing the dimensionality and improving the spatial localization. Conventional SRC techniques generally cannot handle patterns in high dimensions and a small number of training samples. We divide each ROI into nonoverlapping blocks and solve a sparse coding problem for each block. At the final stage, we combine the output of all blocks to make a joint decision. We display an outline of our method in Figure 1 and detail the main stages next.

Figure 1:

Figure 1:

Outline of our spatially localized sparse analysis method.

A test sample y is first divided into blocks. For each block, say the jth block, a solution xj is found. The solution xj is determined from the l1-minimization problem

xj^=arg min xj1       subject    to       Djxyj2ϵ (4)

where NB is the number of blocks and j = 1, 2, … , NB. Then the test sample block yj is assigned to a class ωij that yields the minimal approximation error. The class assignment is determined by

ωi=arg miniri(x^)=arg miniyy^i2 (5)

Our method then uses an ensemble learning approach for classifier predictions. The block-based maximum a posteriori decision function (BBMAP) provides a majority voting and the block-based log likelihood approximation decision function (BBLL-S) is an ensemble of log-likelihood discriminants.

2.2.1. Block Based Maximum a Posterior Decision Function using Sparsity Criterion (BBMAP-S)

We determine class labels for each test sample by voting over the ensemble of Bayesian block-based classifiers, where the predicted class label is given by

ω^BBMAP=FBBMAP(x^)=arg max ipr(ωix^) (6)

x^ is the composite extracted feature from the test sample given by the solution of equation (4). The probability that the test sample x^ belongs to class ωi is given by the indicator function mean over all blocks NB. The indicator function NDij is the binary decision map indicating that block j belongs to class i (1 if x^jωi, 0 otherwise).

2.2.2. Block Based Log-likelihood Decision Function using Sparsity Criterion (BBLL-S)

This likelihood score is based on the relative sparsity scores δmxj^1, δnxj^1 calculated as part of the solution of equation (4) for each classifier

LLSj(x^)=log(δmxj^1)(δnxj^1), (7)

We use the expectation of LLS(x^), denoted by ELLS(x^)=E[LLS] over all the localized classifiers to determine the class of each test sample.

We apply a sigmoid function ς(.) to determine the state ω^ using a decision threshold τLLS

ω^BBLL=FBBLL(x^)ς(ELLS(x^)τLLS). (8)

2.3. Data Description

A widely adopted method for diagnosis and early prediction of breast cancer is the X-ray mammographic test [36, 37]. In general, there are two views for each breast: the craniocaudal (CC) view, this is view from top to bottom of breast; another view is mediolateral oblique (MLO) view, ML is from middle to side and LM is from side to middle view. Mammograms show the masses, calcifications, architectural distortion of breast tissue, and symmetries [6].

We evaluated our CAD techniques for separation of breast lesions into two classes: malignant and benign. The training and testing data were obtained from the Mammographic Image Analysis Society (MIAS) database that is available online [2, 5]. The resolution of the mammograms is 200 micron pixel edge that corresponds to about 0.7559 pixel resolution resulting in about 264.58 μm pixel size, and the size of each image is 1024 × 1024 px after clipping/padding. MIAS contains 322 MLO scans from 161 subjects. The data is categorized into groups of normal subjects, subjects with benign lesions, and subjects with malignant lesions. Figure 2 shows an example of a malignant and a benign lesion sample used from the MIAS dataset. Our goal is to characterize the lesion type, therefore we utilized 68 benign and 51 malignant mammograms for performance evaluation.

Figure 2:

Figure 2:

Examples of a malignant lesion (left) and a benign lesion (right) from the MIAS dataset.

ROI selection is applied first, in order to prepare the data for block decomposition. We need to ensure that the majority of the blocks cover the lesion to improve the accuracy. Hence, we designed our system so that the lesion ROI sizes are greater than or equal to the analysis ROI size. Our method reads-in the centroid and radius of each mass from the provided radiological readings. It uses these two values to automatically determine a minimum bounding square ROI and to select the masses that satisfy the size criterion. We trained and tested all classifiers on these ROI patches centered at the mass centroid. In order to evaluate the classification performance with respect to the lesion size, we performed validation experiments on variable minimum ROI sizes, i.e., we selected subsets of the dataset that met the minimum lesion radius criteria described above. We used an ROI size criterion of 64×64, resulting in 36 benign and 37 malignant lesions. These ROIs contain sufficient visual information, while preserving a big part of the data samples. We performed 10-, 20- and 30-fold cross-validation to study the effect of the size of folds on performance. We expect that the classification performance will improve as the number of folds increases, because more samples will be available for training. This effect is more noticeable on small datasets such as MIAS.

3. Results

In this section we describe our experiments and report results produced by traditional methods based on texture features [38], and results of widely used CNNs [24, 39] after applying transfer learning, and then we present the results produced by our integrative ensemble sparse analysis.

3.1. Texture-based Classification

The goal of our first experiment was to separate benign from malignant masses using conventional texture-based classification. The texture feature set consists of fractal dimension, local binary patterns, discrete wavelet frames, Gabor filters, discrete Fourier and Cosine Transforms, statistical co-occurrence indices, edge histogram, and Law’s energy maps [38]. Next, we employed feature selection methods (FSM) using a correlation-based functional criterion optimized using a best-first search strategy (CFS-BF), or genetic algorithm optimization (CFS-GA). For classification we employed bagging methods using fast decision tree learners, Random Forests (RF), Bayes Network (BN), or Naïve Bayes (NB) techniques.

In Table 1 we display texture-based classification results computed for lesions with 64 × 64 pixels minimum ROI size that performed better than the other ROI sizes. The feature dimensionality in this experiment is 451. The best accuracy with 10-fold cross-validation was 71.2% and corresponding area under the curve was 69.8% for 64 × 64 ROI size, obtained by no feature reduction and Random Forest classification.

Table 1:

ROI images of size 64 × 64 classification performance (10, 20, 30-fold cross-validation) of NB, BN, RF, and Bagging classifiers

FSM CL k-Fold CV TPR TNR ACC AUC
No NB 10 63.9 37.8 50.7 56.3
BN 58.3 24.3 41.1 41.7
RF 72.2 70.3 71.2 69.8
Bagging 61.1 48.6 54.8 51.8
CFS-BF NB 10 63.9 32.4 47.9 43.2
BN 61.1 24.3 42.5 44.2
RF 58.3 45.9 52.1 47.5
Bagging 58.3 48.6 53.4 47.7
CFS-GA NB 10 69.4 45.9 57.5 55.6
BN 38.9 59.5 49.3 48.9
RF 50.0 67.6 58.9 62.0
Bagging 61.1 62.2 61.6 55.9
No NB 20 33.3 50.0 45.0 39.3
BN 25.0 25.0 25.0 33.3
RF 44.4 27.3 35.0 32.3
Bagging 50.0 66.7 60.0 66.7
CFS-BF NB 20 33.3 50.0 45.0 66.0
BN 12.5 58.3 40.0 40.6
RF 42.9 69.2 60.0 56.0
Bagging 33.3 35.7 35.0 27.4
CFS-GA NB 20 62.5 33.3 45.0 38.5
BN 50.0 35.7 40.0 44.0
RF 33.3 27.3 30.0 31.3
Bagging 45.5 77.8 60.0 60.6
No NB 30 16.7 37.5 33.3 29.9
BN 50.0 55.6 53.3 53.7
RF 66.7 60.0 63.3 58.7
Bagging 44.4 50.0 57.1 47.2
CFS-BF NB 30 50.0 58.3 53.3 62.5
BN 55.6 47.3 50.0 50.6
RF 30.0 65.0 53.3 45.0
Bagging 30.0 45.0 40.0 33.5
CFS-GA NB 30 50.0 63.6 60.0 57.4
BN 18.2 47.4 36.7 41.1
RF 33.3 46.7 40.0 36.4
Bagging 47.4 54.5 50.0 59.3

3.2. Convolutional Neural Networks (CNNs)

In this part of our experiments we implemented CNN classification using the AlexNet [24] and Googlenet [39] architectures with transfer learning. Both networks were pre-trained on Imagenet that is a database of 1.2 million natural images.

We applied transfer learning to each network in slightly different ways. To adjust Alexnet to our data, we replaced the pre-trained fully connected layers with three new fully connected layers. We set the learning rates of the pre-trained layers to 0 to keep the network weights fixed, and we trained the new fully connected layers only. In the case of Googlenet, we set the learning rates of the bottom 10 layers to 0, we replaced the top fully connected layer with a new fully connected layer, and we assigned a greater learning rate factor for the new layer.

To provide the networks with additional training examples, we employed data resampling using randomly-centered patches, followed by data augmentation by rotation, scaling, horizontal flipping, and vertical flipping. Finally, we applied hyperparameter tuning using Bayesian optimization to find the optimal learning rate, mini-batch size and number of epochs.

We first applied 10-, 20-, and 30-fold cross-validation to 64 × 64px ROIs. Because deep networks are able to extract information from the edges of the lesion, we also decided to test our method on 256 × 256px ROIs of all lesions (66 benign and 51 malignant) to improve the classification performance. We report the results of our cross-validation experiments in Table 2. We note that Alexnet produces the top ACC of 67.65% for 30-fold cross-validation and for all ROIs.

Table 2:

Classification performance for breast lesion characterization using convolutional neural network classifiers (ROI size: 64 × 64)

Method k-Fold CV Block Size TPR (%) TNR (%) ACC (%) AUC (%)
Alexnet 10 64 × 64 50.0 58.33 54.17 47.69
All 56.86 72.55 64.71 62.19
20 64 × 64 44.44 69.44 56.94 52.55
All 47.06 84.31 65.69 60.7
30 64 × 64 38.89 72.22 55.56 53.97
All 58.82 64.71 61.77 60.29
Googlenet 10 64 × 64 25.0 83.33 54.17 47.21
All 64.71 58.82 61.77 57.86
20 64 × 64 58.33 50.0 54.17 50.22
All 62.75 62.75 62.75 61.5
30 64 × 64 58.33 50.00 54.17 50.96
All 66.67 68.63 67.65 63.04

3.3. Integrative Ensemble Sparse Analysis

In this experiment we validated our block-based ensemble classification system. We performed 10-, 20- and 30-fold cross-validation on these samples. We present results on characterization of lesions with minimum ROI size of 64 × 64 pixels that forms a dataset of 36 benign and 37 malignant lesions.

Table 3 contains the classification rates produced by our cross-validation experiments for multiple block sizes. In the first row of this table, the results were obtained from a single block that is equivalent to conventional SRC analysis [28]. We note that the accuracy increases when the number of folds increases for the same ROI size. The highest accuracy by using 10-fold cross-validation is 70.00% for 16×16 block size using BBMAP and 8 8 block size using BBLL. The largest area under the curve for 10-fold CV ×is 71.42% for 8×8 block size using BBLL. For 20-fold cross-validation, the best accuracy is 76.67% and AUC is 81.20% for 8×8 block size. The best overall performance is obtained for 30-fold cross validation. The highest accuracy is 86.67% for 16 × 16 block size, and the largest area under the curve is 89.10% for 8 × 8 block size. There are 2 or 3 test samples in each fold when k = 30. Fig. 3 displays the ROC graphs. These graphs lead to the same observations that we made from Table 3.

Table 3:

Classification performance for breast lesion characterization using ensembles of block-based sparse classifiers (ROI size: 64 × 64)

Method k-Fold CV Block Size TPR (%) TNR (%) ACC (%) AUC (%)
BBMAP-S 10 64 × 64 48.65 66.67 57.14 54.55
32 × 32 40.54 90.91 64.29 64.29
16 × 16 59.46 81.82 70.00 69.70
8 × 8 48.65 81.82 64.29 63.64
BBLL-S 10 64 × 64 89.19 27.27 60.00 62.33
32 × 32 67.57 60.61 64.29 60.44
16 × 16 72.97 63.64 68.57 70.84
8 × 8 62.16 78.79 70.00 71.42
BBMAP-S 20 64 × 64 29.03 65.52 46.67 42.38
32 × 32 19.35 100.00 58.33 54.83
16 × 16 61.29 89.66 75.00 76.64
8 × 8 70.97 79.31 75.00 74.64
BBLL-S 20 64 × 64 87.10 37.93 63.33 53.73
32 × 32 70.97 48.00 60.00 55.95
16 × 16 67.74 86.21 76.67 78.87
8 × 8 74.19 79.31 76.67 81.20
BBMAP-S 30 64 × 64 32.26 74.41 51.67 46.50
32 × 32 9.68 100.00 53.33 48.83
16 × 16 70.97 100.00 85.00 85.65
8 × 8 74.19 93.10 83.33 82.09
BBLL-S 30 64 × 64 87.10 41.38 65.00 58.62
32 × 32 74.19 62.00 68.33 65.41
16 × 16 93.55 79.00 86.67 88.21
8 × 8 93.55 72.41 83.33 89.10

Figure 3:

Figure 3:

ROC curves for 64 × 64 ROI size breast lesion characterization using the proposed block-based ensemble method with BBMAP-S (left), and BBLL-S (right) decision functions with 10-fold (top row), 20-fold (second row) and 30-fold (bottom row) cross-validation.

We estimated the AUC values of BBLL with optimized threshold τLLS* and BBMAP by applying DeLong’s statistical test between the ROCs produced by BBMAP and BBLL as well for k-fold cross-validation. The p-values for block sizes of 32 × 32, 16 × 16, 8 × 8 and 4 × 4 were 0.78, 0.49, 0.24 and 0.21 respectively for 10-fold cross-validation, and p-values for 30-fold cross-validation were 0.72, 0.54, 0.16 and 0.0086 respectively.

4. Discussion

Texture-based Classification.

Our experiments on MIAS 64 × 64 ROIs have shown the greatest classification accuracy with 10 fold CV using RF classifiers. The best classification accuracy achieved was 71.2% and the top area under the curve was 69.8%. Our Integrative Ensembles Sparse Analysis method outperforms texture based classification methods for CV folds of 20 and 30 and achieves similar best performance for 10-fold CV.

Convolutional Neural Networks.

The CNN models show moderate classification performances for this dataset despite extensive work on data resampling and augmentation and applying multiple transfer learning techniques. The small size of this dataset is the main limiting factor for CNN approaches. This is mainly because there are not enough training examples for the network to learn, especially when the ROIs are small. This effect may explain the improved performance of all ROIs relatively to the 64 × 64px ROIs.

Integrative Sparse Classification.

We note that the BBLL method using 8 × 8 block size improved ACC by 28.1% on average, relative to the traditional SRC method. Traditional SRC is applied when the block size and the ROI size are equal. This indicates that the block decomposition and sampling combined with classifier decision fusion yields more accurate solutions than SRC. We selected block sizes for our experimentation starting with a block size of 64 (the entire ROI) and decreasing block size by a factor of 2 until reaching a block size of 8×8. Block sizes smaller than 8×8 would contain a small amount of image information, thus making it more challenging to produce accurate classification predictions in addition to requiring more computational time. Figure 3 compares the ROC curves for both decision functions for block size of 32×32, 16×16, and 8×8. There is an observable increase in AUC between the 32×32 block size and the smaller block sizes of 16×16 and 8×8.

5. Conclusions

We introduced an integrative ensemble sparse analysis approach to separating benign from malignant breast masses. Our method uses block decomposition and to find localized sparse representations of lesions. We propose a method for combining individual classification decisions and an approach to tuning the ensembles to minimize bias. We observe that localization regularizes the constrained optimization problem for small datasets and leads to feasible solutions. In addition, the integration of localized decisions improves the classification accuracy. Our experiments have shown that smaller block sizes and increased cross validation folds yield higher classification accuracy. The highest overall classification accuracy was 86.67% with the BBLL-S decision function, using 30-fold CV and block size of 16 × 16 and the highest AUC was 89.1% for 8 × 8 blocks.

Highlights.

  • Integrative sparse analysis method based on localized image patches for breast lesion characterization

  • Method produces more sparse and regularized solutions than conventional sparse representation

  • Extensive cross-validation experiments conducted for performance evaluation

  • Method outperforms texture based classifiers and conventional sparse representation methods

  • Application to diagnosis of other diseases and other imaging modalities is straightforward

Acknowledgements

The authors acknowledge the support by the National Institute of General Medical Sciences of the National Institutes of Health under Award Number SC3GM113754. We also acknowledge the support by the Federal Research and Development Matching Grant Program of the Delaware Economic Development Office (DEDO - award #105).

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

Conflicts of Interest: None Declared

References

  • [1].Heath M, Bowyer K, Kopans DB, K. P Jr., Moore RH, Chang K, Munishkumaran S, Current status of the digital database for screening mammography., in: Karssemeijer N, Thijssen M, Hendriks JHCL, van Erning L (Eds.), Digital Mammography / IWDM, Vol. 13 of Computational Imaging and Vision, Springer, 1998, pp. 457–460. [Google Scholar]
  • [2].Oliver A, Freixenet J, Marti J, Perez E, Pont J, Denton ER, Zwiggelaar R, A review of automatic mass detection and segmentation in mammographic images, Medical image analysis 14 (2) (2010) 87–110. [DOI] [PubMed] [Google Scholar]
  • [3].Oliver A, Lladó X, Pérez E, Pont J, Denton ERE, Freixenet J, Martí J, A statistical approach for breast density segmentation, Journal of Digital Imaging 23 (5) (2010) 527–537. doi: 10.1007/s10278-009-9217-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [4].Verma B, McLeod P, Klevansky A, Classification of benign and malignant patterns in digital mammograms for the diagnosis of breast cancer, Expert systems with applications 37 (4) (2010) 3344–3351. [Google Scholar]
  • [5].Matheus BRN, Schiabel H, Online mammographic images database for development and comparison of cad schemes, Journal of digital imaging 24 (3) (2011) 500–506. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [6].Moreira IC, Amaral I, Domingues I, Cardoso A, Cardoso MJ, Cardoso JS, Inbreast: toward a full-field digital mammographic database, Academic radiology 19 (2) (2012) 236–248. [DOI] [PubMed] [Google Scholar]
  • [7].Kulkarni P, Stranieri A, Kulkarni S, Ugon J, Mittal M, Hybrid technique based on ngram and neural networks for classification of mammographic images, in: Second International Conference on Signal, Image Processing and Pattern Recognition, 2014, pp. 297–306. [Google Scholar]
  • [8].Pereira DC, Ramos RP, Do Nascimento MZ, Segmentation and detection of breast cancer in mammograms combining wavelet analysis and genetic algorithm, Computer methods and programs in biomedicine 114 (1) (2014) 88–101. [DOI] [PubMed] [Google Scholar]
  • [9].Benjamin MLG Huynh Q, Li Hui, Digital mammographic tumor classification using transfer learning from deep convolutional neural networks, Journal of Medical Imaging 3 (3) (2016) 1–5–5. doi: 10.1117/1.JMI.3.3.034501. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [10].Sudarshan VK, Mookiah MRK, Acharya UR, Chandran V, Molinari F, Fujita H, Ng KH, Application of wavelet techniques for cancer diagnosis using ultrasound images: A review, Computers in Biology and Medicine 69 (2016) 97–111. doi: 10.1016/j.compbiomed.2015.12.006. [DOI] [PubMed] [Google Scholar]
  • [11].Nagarajan R, Upreti M, An ensemble predictive modeling framework for breast cancer classification, Methods 131 (2017) 128–34, systems Approaches for Identifying Disease Genes and Drug Targets. doi: 10.1016/j.ymeth.2017.07.011. [DOI] [PubMed] [Google Scholar]
  • [12].Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M, Van Der Laak JA, Van Ginneken B, Sánchez CI, A survey on deep learning in medical image analysis, Medical image analysis 42 (2017) 60–88. [DOI] [PubMed] [Google Scholar]
  • [13].Nahid A-A, Kong Y, Involvement of machine learning for breast cancer image classification: a survey, Computational and mathematical methods in medicine 2017 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [14].Beura S, Majhi B, Dash R, Mammogram classification using two dimensional discrete wavelet transform and gray-level co-occurrence matrix for detection of breast cancer, Neurocomputing 154 (2015) 1–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [15].Rouhi R, Jafari M, Kasaei S, Keshavarzian P, Benign and malignant breast tumors classification based on region growing and cnn segmentation, Expert Systems with Applications 42 (3) (2015) 990–1002. [Google Scholar]
  • [16].George M, Chen Z, Zwiggelaar R, Multiscale connected chain topological modelling for microcalcification classification., Computers in biology and medicine 114 (2019) 103422. doi: 10.1016/j.compbiomed.2019.103422. [DOI] [PubMed] [Google Scholar]
  • [17].Rabidas R, Midya A, Chakraborty J, Neighborhood structural similarity mapping for the classification of masses in mammograms, IEEE journal of biomedical and health informatics 22 (3) (2017) 826–834. [DOI] [PubMed] [Google Scholar]
  • [18].Narváez F, Alvarez J, Garcia-Arteaga JD, Tarquino J, Romero E, Characterizing architectural distortion in mammograms by linear saliency., Journal of medical systems 41 (2017) 26. doi: 10.1007/s10916-016-0672-5. [DOI] [PubMed] [Google Scholar]
  • [19].Singh SP, Urooj S, An improved cad system for breast cancer diagnosis based on generalized pseudo-zernike moment and ada-dewnn classifier., Journal of medical systems 40 (2016) 105. doi: 10.1007/s10916-016-0454-0. [DOI] [PubMed] [Google Scholar]
  • [20].Oliver A, Torrent A, Llado X, Marti J, Automatic diagnosis of masses by using level set segmentation and shape description, in: 2010 20th International Conference on Pattern Recognition, IEEE, 2010, pp. 2528–2531. [Google Scholar]
  • [21].Sharma MK, Jas M, Karale V, Sadhu A, Mukhopadhyay S, Mammogram segmentation using multi-atlas deformable registration., Computers in biology and medicine 110 (2019) 244–253. doi: 10.1016/j.compbiomed.2019.06.001. [DOI] [PubMed] [Google Scholar]
  • [22].Higham CF, Higham DJ, Deep learning: An introduction for applied mathematicians, SIAM Review 61 (4) (2019) 860–891. [Google Scholar]
  • [23].LeCun Y, Boser B, Denker JS, Henderson D, Howard RE, Hubbard W, Jackel LD, Backpropagation applied to handwritten zip code recognition, Neural computation 1 (4) (1989) 541–551. [Google Scholar]
  • [24].Krizhevsky A, Sutskever I, Hinton GE, Imagenet classification with deep convolutional neural networks, in: Advances in neural information processing systems, 2012, pp. 1097–1105. [Google Scholar]
  • [25].Chougrad H, Zouaki H, Alheyane O, Deep convolutional neural networks for breast cancer screening., Computer methods and programs in biomedicine 157 (2018) 19–30. doi: 10.1016/j.cmpb.2018.01.011. [DOI] [PubMed] [Google Scholar]
  • [26].Mairal J, Elad M, Sapiro G, Sparse representation for color image restoration, Trans. Img. Proc 17 (1) (2008) 53–69. doi: 10.1109/TIP.2007.911828. [DOI] [PubMed] [Google Scholar]
  • [27].Yang J, Wright J, Huang TS, Ma Y, Image super-resolution as sparse representation of raw image patches, Computer Vision and Pattern Recognition, 2008. CVPR 2008. IEEE Conference on (2008) 1–8 doi: 10.1109/CVPR.2008.4587647. [DOI] [Google Scholar]
  • [28].Wright J, Ma Y, Mairal J, Sapiro G, Huang TS, Yan S, Sparse representation for computer vision and pattern recognition., Proceedings of the IEEE 98 (6) (2010) 1031–1044. [Google Scholar]
  • [29].Wright J, Yang AY, Ganesh A, Sastry SS, Ma Y, Robust face recognition via sparse representation, IEEE Transactions on Pattern Analysis and Machine Intelligence 31 (2) (2009) 210–227. doi: 10.1109/TPAMI.2008.79. [DOI] [PubMed] [Google Scholar]
  • [30].Qiao L, Chen S, Tan X, Sparsity preserving projections with applications to face recognition, Pattern Recognition 43 (1) (2010) 331–341. doi: 10.1016/j.patcog.2009.05.005. [DOI] [Google Scholar]
  • [31].Zhao S-H, Hu Z-P, Occluded face recognition based on block-label and residual, International Journal on Artificial Intelligence Tools 25 (03) (2016) 1650019 arXiv:http://www.worldscientific.com/doi/pdf/10.1142/S0218213016500196, doi: 10.1142/S0218213016500196. [DOI] [Google Scholar]
  • [32].Bishop CM, Pattern Recognition and Machine Learning, Springer, 2006. [Google Scholar]
  • [33].Plenge E, Klein SS, Niessen WJ, Meijering E, Multiple sparse representations classification, PLos ONE (Jul. 2015). doi: 10.1371/journal.pone.0131968. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [34].Zhao W, Xu R, Hirano Y, Tachibana R, Kido S, A sparse representation based method to classify pulmonary patterns of diffuse lung diseases, Computational and Mathematical Methods in Medicine 2015 (2015). doi: 10.1155/2015/567932. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [35].Zheng K, Jennane R, Makrogiannis S, Ensembles of sparse classifiers for osteoporosis characterization in digital radiographs, in: Medical Imaging 2019: Computer-Aided Diagnosis, Vol. 10950, International Society for Optics and Photonics, 2019, p. 1095024. [Google Scholar]
  • [36].Boyle P, Levin B, et al. , World cancer report 2008., IARC Press, International Agency for Research on Cancer, 2008. [Google Scholar]
  • [37].Misra S, Solomon NL, Moat FL, Koniaris LG, Screening criteria for breast cancer, Advances in surgery 44 (1) (2010) 87–100. [DOI] [PubMed] [Google Scholar]
  • [38].Zheng K, Makrogiannis S, Bone texture characterization for osteoporosis diagnosis using digital radiography, in: 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), 2016, pp. 1034–1037. doi: 10.1109/EMBC.2016.7590879. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [39].Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A, Going deeper with convolutions, in: Computer Vision and Pattern Recognition (CVPR), 2015. URL http://arxiv.org/abs/1409.4842 [Google Scholar]

RESOURCES