Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/2390803.2390808acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
research-article

Ground truth generation in medical imaging: a crowdsourcing-based iterative approach

Published: 29 October 2012 Publication History

Abstract

As in many other scientific domains where computer--based tools need to be evaluated, also medical imaging often requires the expensive generation of manual ground truth. For some specific tasks medical doctors can be required to guarantee high quality and valid results, whereas other tasks such as the image modality classification described in this text can in sufficiently high quality be performed with simple domain experts. Crowdsourcing has received much attention in many domains recently as volunteers perform so--called human intelligence tasks for often small amounts of money, allowing to reduce the cost of creating manually annotated data sets and ground truth in evaluation tasks. On the other hand there has often been a discussion on the quality when using unknown experts. Controlling task quality has remained one of the main challenges in crowdsourcing approaches as potentially the persons performing the tasks may not be interested in results quality but rather their payment.
On the other hand several crowdsourcing platforms such as Crowdflower that we used allow creating interfaces and sharing them with only a limited number of known persons. The text describes the interfaces developed and the quality obtained through manual annotation of several domain experts and one medical doctor. Particularly the feedback loop of semi--automatic tools is explained. The results of an initial crowdsourcing round classifying medical images into a set of image categories were manually controlled by domain experts and then used to train an automatic system that visually classified these images. The automatic classification results were then used to manually confirm or refuse the automatic classes, reducing the time for the initial tasks.
Crowdsourcing platforms allow creating a large variety of interfaces for judgements. Whether used among known experts or paying for unknown persons, they allow increasing the speed of ground truth creation and limit the amount of money to be paid.

References

[1]
C. Akgül, D. Rubin, S. Napel, C. Beaulieu, H. Greenspan, and B. Acar. Content-based image retrieval in radiology: Current status and future directions. Journal of Digital Imaging, 24(2):208--222, 2011.
[2]
O. Alonso and S. Mizzaro. Can we get rid of TREC assessors? Using Mechanical Turk for relevance assessment. In SIGIR 2009 Workshop on The Future of IR Evaluation, 2009.
[3]
O. Alonso, D. E. Rose, and B. Stewart. Crowdsourcing for relevance evaluation. SIGIR Forum, 42(2):9--15, December 2008.
[4]
B. Caputo, H. Müller, T. S. Mahmood, J. Kalpathy-Cramer, F. Wang, and J. Duncan. Editorial of miccai workshop proceedings on medical content-based retrieval for clinical decision support. In Proceedings on MICCAI Workshop on Medical Content-based Retrieval for Clinical Decision Support, volume 5853 of Lecture Notes in Computer Science (LNCS). Springer, 2009.
[5]
A. Depeursinge, H. Greenspan, T. Syeda Mahmood, and H. Müller. Overview of the second workshop on medical content-based retrieval for clinical decision support. In H. Greenspan, H. Müller, and T. Syeda Mahmood, editors, Medical Content-based Retrieval for Clinical Decision Support, MCBR--CDS 2011. Lecture Notes in Computer Sciences (LNCS), Sept. 2011.
[6]
J. Howe. Crowdsourcing. http://www.crowdsourcing.com/cs, 2012. {Online; accessed 29-Jun-2012}.
[7]
J. Kalpathy-Cramer, H. Müller, S. Bedrick, I. Eggel, A. G. Seco de Herrera, and T. Tsikrika. The CLEF 2011 medical image retrieval and classification tasks. In Working Notes of CLEF 2011 (Cross Language Evaluation Forum), September 2011.
[8]
H. Müller, J. Kalpathy-Cramer, D. Demner-Fushman, and S. Antani. Creating a classification of image types in the medical literature for visual categorization. In SPIE medical imaging, 2012.
[9]
H. Müller, N. Michoux, D. Bandon, and A. Geissbuhler. A review of content-based image retrieval systems in medicine--clinical benefits and future directions. International Journal of Medical Informatics, 73(1):1--23, 2004.
[10]
T. B. Nguyen, S. Wang, V. Anugu, N. Rose, M. McKenna, N. Petrick, J. E. Burns, and R. M. Summers. Distributed human intelligence for colonic polyp classification in computer--aided detection for CT colonography. Radiology, 262(3):824--828, March 2012.
[11]
S. Nowak and S. Rüger. How reliable are annotations via crowdsourcing: a study about inter-annotator agreement for multi-label image annotation. In Proceedings of the international conference on Multimedia information retrieval, MIR '10, pages 557--566, New York, NY, USA, 2010. ACM.

Cited By

View all
  • (2024) Development of the quantitative PET prostate phantom (Q3P) for improved quality assurance of 18 F‐PSMA PET imaging in metastatic prostate cancer Medical Physics10.1002/mp.1697751:6(4311-4323)Online publication date: 13-Feb-2024
  • (2023)Gamified Crowdsourcing as a Novel Approach to Lung Ultrasound Dataset Labeling: A Prospective Analysis (Preprint)Journal of Medical Internet Research10.2196/51397Online publication date: 4-Aug-2023
  • (2023)Crowdsourcing Utilizing Subgroup Structure of Latent Factor ModelingJournal of the American Statistical Association10.1080/01621459.2023.2178925119:546(1192-1204)Online publication date: 16-Mar-2023
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
CrowdMM '12: Proceedings of the ACM multimedia 2012 workshop on Crowdsourcing for multimedia
October 2012
60 pages
ISBN:9781450315890
DOI:10.1145/2390803
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 29 October 2012

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. crowdsourcing
  2. ground truth
  3. information classification and retrieval

Qualifiers

  • Research-article

Conference

MM '12
Sponsor:
MM '12: ACM Multimedia Conference
October 29, 2012
Nara, Japan

Acceptance Rates

Overall Acceptance Rate 16 of 42 submissions, 38%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)17
  • Downloads (Last 6 weeks)3
Reflects downloads up to 16 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2024) Development of the quantitative PET prostate phantom (Q3P) for improved quality assurance of 18 F‐PSMA PET imaging in metastatic prostate cancer Medical Physics10.1002/mp.1697751:6(4311-4323)Online publication date: 13-Feb-2024
  • (2023)Gamified Crowdsourcing as a Novel Approach to Lung Ultrasound Dataset Labeling: A Prospective Analysis (Preprint)Journal of Medical Internet Research10.2196/51397Online publication date: 4-Aug-2023
  • (2023)Crowdsourcing Utilizing Subgroup Structure of Latent Factor ModelingJournal of the American Statistical Association10.1080/01621459.2023.2178925119:546(1192-1204)Online publication date: 16-Mar-2023
  • (2022)COVLIAS 1.0Lesion vs. MedSeg: An Artificial Intelligence Framework for Automated Lesion Segmentation in COVID-19 Lung Computed Tomography ScansDiagnostics10.3390/diagnostics1205128312:5(1283)Online publication date: 21-May-2022
  • (2022)It's All Relative! A Method to Counter Human Bias in Crowdsourced Stance Detection of News ArticlesProceedings of the ACM on Human-Computer Interaction10.1145/35556366:CSCW2(1-25)Online publication date: 11-Nov-2022
  • (2022)BioLumin: An Immersive Mixed Reality Experience for Interactive Microscopic Visualization and Biomedical Research AnnotationACM Transactions on Computing for Healthcare10.1145/35487773:4(1-28)Online publication date: 3-Nov-2022
  • (2022)Can Masses of Non-Experts Train Highly Accurate Image Classifiers?Medical Image Computing and Computer-Assisted Intervention – MICCAI 201410.1007/978-3-319-10470-6_55(438-445)Online publication date: 10-Mar-2022
  • (2022)Crowdsourcing for Reference Correspondence Generation in Endoscopic ImagesMedical Image Computing and Computer-Assisted Intervention – MICCAI 201410.1007/978-3-319-10470-6_44(349-356)Online publication date: 10-Mar-2022
  • (2022)Development of scalable lymphatic system in the 4D XCAT phantom: Application to quantitative evaluation of lymphoma PET segmentationsMedical Physics10.1002/mp.1596349:11(6871-6884)Online publication date: 18-Sep-2022
  • (2021)Data Tagging in Medical Images: A Survey of the State-of-ArtCurrent Medical Imaging Formerly Current Medical Imaging Reviews10.2174/157340561666620021813004316:10(1214-1228)Online publication date: 12-Jan-2021
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media