Machine Learning in Medical Imaging: Maryellen L. Giger, PHD
Machine Learning in Medical Imaging: Maryellen L. Giger, PHD
Machine Learning in Medical Imaging: Maryellen L. Giger, PHD
Abstract
Advances in both imaging and computers have synergistically led to a rapid rise in the potential use of artificial intelligence in various
radiological imaging tasks, such as risk assessment, detection, diagnosis, prognosis, and therapy response, as well as in multi-omics disease
discovery. A brief overview of the field is given here, allowing the reader to recognize the terminology, the various subfields, and
components of machine learning, as well as the clinical potential. Radiomics, an expansion of computer-aided diagnosis, has been
defined as the conversion of images to minable data. The ultimate benefit of quantitative radiomics is to (1) yield predictive image-based
phenotypes of disease for precision medicine or (2) yield quantitative image-based phenotypes for data mining with other -omics for
discovery (ie, imaging genomics). For deep learning in radiology to succeed, note that well-annotated large data sets are needed since
deep networks are complex, computer software and hardware are evolving constantly, and subtle differences in disease states are more
difficult to perceive than differences in everyday objects. In the future, machine learning in radiology is expected to have a substantial
clinical impact with imaging examinations being routinely obtained in clinical practice, providing an opportunity to improve decision
support in medical image interpretation. The term of note is decision support, indicating that computers will augment human decision
making, making it more effective and efficient. The clinical impact of having computers in the routine clinical practice may allow
radiologists to further integrate their knowledge with their clinical colleagues in other medical specialties and allow for precision
medicine.
Key Words: Machine learning, deep learning, radiomics, computer-aided diagnosis, computer-assisted decision support
J Am Coll Radiol 2018;15:512-520. Copyright 2018 Published by Elsevier Inc. on behalf of American College of Radiology
Advances in both imaging and computers have synergisti- at the time of interpretation (eg, clinical history, laboratory
cally led to a rapid rise in the potential use of artificial in- data, prior examinations).
telligence in various radiological imaging tasks, such as risk A brief overview of the field is given here, allowing the
assessment, detection, diagnosis, prognosis, and therapy reader to recognize the terminology, the various subfields,
response, as well as in multi-omics disease discovery. and components of machine learning, as well as the clinical
Although computer-aided detection (CADe) has been potential. Figure 1 shows the number of publication
proposed, developed, and clinically used since 1966, espe- counts in PubMed for searches on computer-aided diag-
cially in thoracic and breast imaging [1-5], the widespread nosis (CADx) in radiology, machine learning, and deep
progress in multiple clinical decision-making tasks and learning from 1972 to middle of 2017. Note that in each
multiple disease sites has only advanced in the past decades of these areas, there are numerous review publications;
with the corresponding access to large computational re- however, the aim of this article is to elucidate the concepts
sources, including computer power, storage, and digital and generalities. The range in presentation of various subtle
imaging, as well as increased electronic access to information disease states, the need for large annotated clinical data sets,
and the complex structure of many machine learning
Department of Radiology, The University of Chicago, Chicago, Illinois. methods signify much need for continued research and
Corresponding author and reprints: Maryellen L. Giger, PhD, University of development before full clinical incorporation and use.
Chicago, Department of Radiology, MC 2026, 5841 S Maryland Ave,
Chicago, IL 60637; e-mail: m-giger@uchicago.edu.
Funded in parts by NIH U01CA195564, U01CA189240, and CADe, CADx, AND DECISION SUPPORT
R01CA166945. M.L.G. is a stockholder in R2/Hologic, cofounder and equity
holder in Quantitative Insights, and shareholder in QView and receives roy- Medical image interpretation is the main undertaking of
alties from Hologic, GE Medical Systems, MEDIAN Technologies, Riverain radiologists, with the tasks requiring both good image
Medical, Mitsubishi, and Toshiba. It is the University of Chicago Conflict of quality and good image interpretation. Image interpretation
Interest Policy that investigators disclose publicly actual or potential significant
financial interest that would reasonably seem to be directly and significantly by humans is limited by the presence of structure noise
affected by the research activities. (camouflaging normal anatomical background), incomplete
3500 Diagnosis
has been defined as the conversion of images to minable
3000
2500
Machine Learning
data [13-15]. Obtaining radiomic data may involve
PubMed
Fig 3. Radiomics heat map. (a) Unsupervised clustering of lung cancer patients (Lung1 set, n.422) on the y axis and radiomic
feature expression (n.440) on the x axis revealed clusters of patients with similar radiomic expression patterns. (b) Clinical
patient parameters for showing significant association of the radiomic expression patterns with primary tumor stage (T-stage;
Po1_10_20, w2 test), overall stage (P.3.4_10_3, w2 test), and histology (P.0.019, w2 test). (c) Correspondence of radiomic feature
groups with the clustered expression patterns. Reprinted with permission [35].
random forests, and neural networks. Reviews of machine identify the optimal signature. That is, a computer-
learning have been written over the past many years derived tumor signature needs to both perform well in
including those that serve as tutorials to new investigators its specific task and be generalizable across cases.
into the field [15,38].
Given the ever-increasing variations of computer-
extracted features, both handcrafted and deep-learned, DEEP LEARNING
appropriate feature selection techniques are important. Deep learning is a subcategory of machine learning in
Various studies have been conducted in which in- which multiple-layered networks are used to assess com-
vestigators, using moderately large data sets, have evalu- plex patterns within the raw imaging input data. Most
ated the combination of feature selection and recently, deep learning has been conducted using deep
classification methods [39-41]. Such analyses have taken convolutional neural networks (CNNs). Just as radiolo-
into account both performance (such as the area under gists learn, during residency and beyond, by repeatedly
the receiver operating characteristic curve for a correlating their visual interpretation of radiological images
particular clinical task) and variability as a way to to actual clinical truth, so can machines. Although CNNs
Localization
ion of Tumor Localization
ion of Tumor
Computerized Tumor
umo Segmentation
Deep Learning
ning Algorithm
Computerized, Quantitative,
ative Analytically-Extracted (CNNs)
Tumor Features
Classifier
assi Classifier
Co
Combined Output for Decision
on
Support and/or Discovery
Fig 5. Schematic demonstrating the comparison of conventional hand-crafted computer-aided diagnosis and radiomic features,
convolutional neural network (CNN)-extracted features, and an ensemble technique in the task of distinguishing between lesion
type as used in Antropova et al [37] and Huynh et al [42].
Fig 7. Receiver operating characteristic curves showing statistically significant improvement in diagnostic classification of breast
lesions on FFDM, ultrasound, and breast MRI when output from conventional CADx and deep learning are combined [37]. AUC,
area under the curve; CAD, computer-aided diagnosis; CNN, convolutional neural network; DCE ¼ dynamic contrast-enhanced
MRI; FFDM ¼ full field digital mammography; US ¼ ultrasound. Reprinted with permission [37].