Rationale and Objectives
Determine whether there are patterns of lesion recall among breast imaging subspecialists
interpreting screening mammography, and if so, whether recall patterns correlate to
morphologies of screen-detected cancers.
Materials and Methods
This Institutional Review Board-approved, retrospective review included all screening
examinations January 3, 2012–October 1, 2018 interpreted by fifteen breast imaging
subspecialists at a large academic medical center and two outpatient imaging centers.
Natural language processing identified radiologist recalls by lesion type (mass, calcifications,
asymmetry, architectural distortion); proportions of callbacks by lesion types were
calculated per radiologist. Hierarchical cluster analysis grouped radiologists based
on recall patterns. Groups were compared to overall practice and each other by proportions
of lesion types recalled, and overall and lesion-specific positive predictive value-1
(PPV1).
Results
Among 161,859 screening mammograms with 13,086 (8.1%) recalls, Hierarchical cluster
analysis grouped 15 radiologists into five groups. There was substantial variation
in proportions of lesions recalled: calcifications 13%–18% (Chi-square 45.69, p < 0.00001); mass 16%–44% (Chi-square 498.42, p < 0.00001); asymmetry 13%–47% (Chi-square 660.93, p < 0.00001) architectural distortion 6%–20% (Chi-square 283.81, p < 0.00001). Radiologist groups differed significantly in overall PPV1 (range 5.6%–8.8%;
Chi-square 17.065, p = 0.0019). PPV1 by lesion type varied among groups: calcifications 9.2%–15.4% (Chi-square
2.56, p = 0.6339); mass 5.6%–8.5% (Chi-square 1.31, p = 0.8597); asymmetry 3.4%–5.9% (Chi-square 2.225, p = 0.6945); architectural distortion 5.6%–10.8% (Chi-square 5.810, p = 0.2138). Proportions of recalled lesions did not consistently correlate to proportions
of screen-detected cancer.
Conclusion
Breast imaging subspecialists have patterns for screening mammography recalls, suggesting
differential weighting of imaging findings for perceived malignant potential. Radiologist
recall patterns are not always predictive of screen-detected cancers nor lesion-specific
PPV1s.
Key Words
Abbreviations:
CDR (cancer detection rate), DBT (digital breast tomosynthesis), FFDM (full field digital mammogram), HCA (hierarchical cluster analysis), NLP (natural language processing), PPV1 (positive predictive value of recalled screening examinations]), US (United States)To read this article in full you will need to make a payment
Purchase one-time access:
Academic & Personal: 24 hour online accessCorporate R&D Professionals: 24 hour online accessOne-time access price info
- For academic or personal research use, select 'Academic and Personal'
- For corporate R&D use, select 'Corporate R&D Professionals'
Subscribe:
Subscribe to Academic RadiologyAlready a print subscriber? Claim online access
Already an online subscriber? Sign in
Register: Create an account
Institutional Access: Sign in to ScienceDirect
REFERENCES
- ACR BI-RADS follow-up and outcome monitoring 2013.in: D'Orsi CJ Sickles EA Mendelson EB ACR BI-RADS Atlas Breast Imaging Reporting and Data System. 5th ed. American College of Radiology, Reston, VA2013
- National performance benchmarks for modern screening digital mammography: update from the breast cancer surveillance consortium.Radiology. 2017; 283: 49-58https://doi.org/10.1148/radiol.2016161174
- BI-RADS mammography.in: D'Orsi CJ Mendelson EB Ikeda DM Breast imaging and reporting data system: ACR BI-RADS-breast imaging atlas. 4th ed. American College of Radiology, Reston, VA2003
- Screening mammograms by community radiologists: variability in false-positive rates.J Natl Cancer Inst. 2002; 94: 1373-1380https://doi.org/10.1093/jnci/94.18.1373
- Accuracy of screening mammography interpretation by characteristics of radiologists.J Natl Cancer Inst. 2004; 96: 1840-1850https://doi.org/10.1093/jnci/djh333
- Physician predictors of mammographic accuracy.J Natl Cancer Inst. 2005; 97: 358-367https://doi.org/10.1093/jnci/dji060
- Variation in false-positive rates of mammography reading among 1067 radiologists: a population-based assessment.Breast Cancer Res Treat. 2006; 100: 309-318https://doi.org/10.1007/s10549-006-9252-6
- Influence of annual interpretive volume on screening mammography performance in the United States.Radiology. 2011; 259: 72-83https://doi.org/10.1148/radiol.10101698
- Factors associated with rates of false-positive and false-negative results from digital mammography screening: an analysis of registry data.Ann Intern Med. 2016; 164: 226-235https://doi.org/10.7326/M15-0971
- Five consecutive years of screening with digital breast tomosynthesis: outcomes by screening year and round.Radiology. 2020; 295: 285-293https://doi.org/10.1148/radiol.2020191751
- Changes in recall type and patient treatment following implementation of screening digital breast tomosynthesis.Radiology. 2015; 274: 337-342https://doi.org/10.1148/radiol.14140317
- Early clinical experience with digital breast tomosynthesis for screening mammography.Radiology. 2015; 274: 85-92https://doi.org/10.1148/radiol.14131319
- Comparing diagnostic performance of digital breast tomosynthesis and full-field digital mammography in a hybrid imaging environment.AJR Am J Roentgenol. 2017; 209: 929-934
- Patient, radiologist, and examination characteristics affecting screening mammography recall rates in a large academic practice.J Am Coll Radiol. 2019; 16: 411-418https://doi.org/10.1016/j.jacr.2018.06.016
- Strategies for decreasing screening mammography recall rates while maintaining performance metrics.Acad Radiol. 2017; 24: 1556-1560https://doi.org/10.1016/j.acra.2017.06.009
- ACR BI-RADS Mammography.ACR BI-RADS atlas, breast imaging reporting and data system. American College of Radiology, Reston, VA2013
- Evaluation of an automated information extraction tool for imaging data elements to populate a breast cancer screening registry.J Digit Imaging. 2015; 28: 567-575https://doi.org/10.1007/s10278-014-9762-4
- Hierarchical cluster analysis in clinical research with heterogenous study population: highlighting its visualization with R.Ann Transl Med. 2017; 5: 75https://doi.org/10.21037/atm.2017.02.05
- Positive predictive value of specific mammographic findings according to reader and patient variables.Radiology. 2009; 250: 648-657https://doi.org/10.1148/radiol.2503080541
- Quantifying the benefits and harms of screening mammography.JAMA Intern Med. 2014; 174: 448-454https://doi.org/10.1001/jamainternmed.2013
- Screening for breast cancer: U.S. Preventive Services Task Force recommendations statement.Ann Intern Med. 2016; 164: 279-296https://doi.org/10.7326/M15-2886
- Is maximum positive predictive value a good indicator of an optimal screening mammography practice?.AJR Am J Roentgenol. 2005; 184: 1505-1507https://doi.org/10.2214/ajr.184.5.01841505
- Criteria for identifying radiologists with acceptable screening mammography interpretive performance based on multiple performance measures.AJR Am J Roentgenol. 2015; 204: W486-W491https://doi.org/10.2214/AJR.13.12313
- Bias in radiology: the how and why of misses and misinterpretations.RadioGraphics. 2018; 38: 236-247https://doi.org/10.1148/rg.2018170107
- Quality and variability in diagnostic radiology.J Am Coll Radiol. 2004; 1: 127-132https://doi.org/10.1016/j.jacr.2003
- International evaluation of an AI system for breast cancer screening.Nature. 2020; 577: 89-94https://doi.org/10.1038/s41586-019-1799-6
- Variation in follow-up imaging recommendations in radiology reports: patient, modality, and radiologist predictors.Radiology. 2019; 291: 700-707
Article info
Publication history
Published online: July 05, 2022
Accepted:
June 8,
2022
Received in revised form:
May 22,
2022
Received:
April 21,
2022
Identification
Copyright
© 2022 The Association of University Radiologists. Published by Elsevier Inc. All rights reserved.