Understanding diagnostic variability in breast pathology: lessons learned from an expert consensus review panel
Corresponding Author
Kimberly H Allison
Department of Pathology, University of Washington Medical Center, Seattle, WA, USA
Address for correspondence: K H Allison, MD, Department of Pathology, Stanford University Medical Center, 300 Pasteur Drive, Lane L235, Stanford, CA 94305, USA. e-mail: [email protected]Search for more papers by this authorLisa M Reisch
Department of Internal Medicine, University of Washington School of Medicine, Seattle, WA, USA
Search for more papers by this authorPatricia A Carney
Departments of Family Medicine and Public Health & Preventive Medicine, Oregon Health & Science University, Portland, OR, USA
Search for more papers by this authorDonald L Weaver
Department of Pathology, University of Vermont College of Medicine and Vermont Cancer Center, Burlington, VT, USA
Search for more papers by this authorStuart J Schnitt
Department of Pathology, Beth Israel Deaconess Medical Center, Boston, MA, USA
Search for more papers by this authorFrances P O'Malley
Department of Laboratory Medicine, Keenan Research Centre of the Li Ka Shing Knowledge Institute, St Michael's Hospital, University of Toronto, Toronto, ON, Canada
Search for more papers by this authorBerta M Geller
Department of Family Medicine, Health Promotion Research, University of Vermont, Burlington, VT, USA
Search for more papers by this authorJoann G Elmore
Department of Internal Medicine, University of Washington School of Medicine, Seattle, WA, USA
Search for more papers by this authorCorresponding Author
Kimberly H Allison
Department of Pathology, University of Washington Medical Center, Seattle, WA, USA
Address for correspondence: K H Allison, MD, Department of Pathology, Stanford University Medical Center, 300 Pasteur Drive, Lane L235, Stanford, CA 94305, USA. e-mail: [email protected]Search for more papers by this authorLisa M Reisch
Department of Internal Medicine, University of Washington School of Medicine, Seattle, WA, USA
Search for more papers by this authorPatricia A Carney
Departments of Family Medicine and Public Health & Preventive Medicine, Oregon Health & Science University, Portland, OR, USA
Search for more papers by this authorDonald L Weaver
Department of Pathology, University of Vermont College of Medicine and Vermont Cancer Center, Burlington, VT, USA
Search for more papers by this authorStuart J Schnitt
Department of Pathology, Beth Israel Deaconess Medical Center, Boston, MA, USA
Search for more papers by this authorFrances P O'Malley
Department of Laboratory Medicine, Keenan Research Centre of the Li Ka Shing Knowledge Institute, St Michael's Hospital, University of Toronto, Toronto, ON, Canada
Search for more papers by this authorBerta M Geller
Department of Family Medicine, Health Promotion Research, University of Vermont, Burlington, VT, USA
Search for more papers by this authorJoann G Elmore
Department of Internal Medicine, University of Washington School of Medicine, Seattle, WA, USA
Search for more papers by this authorAbstract
Aims
To gain a better understanding of the reasons for diagnostic variability, with the aim of reducing the phenomenon.
Methods and results
In preparation for a study on the interpretation of breast specimens (B-PATH), a panel of three experienced breast pathologists reviewed 336 cases to develop consensus reference diagnoses. After independent assessment, cases coded as diagnostically discordant were discussed at consensus meetings. By the use of qualitative data analysis techniques, transcripts of 16 h of consensus meetings for a subset of 201 cases were analysed. Diagnostic variability could be attributed to three overall root causes: (i) pathologist-related; (ii) diagnostic coding/study methodology-related; and (iii) specimen-related. Most pathologist-related root causes were attributable to professional differences in pathologists' opinions about whether the diagnostic criteria for a specific diagnosis were met, most frequently in cases of atypia. Diagnostic coding/study methodology-related root causes were primarily miscategorizations of descriptive text diagnoses, which led to the development of a standardized electronic diagnostic form (BPATH-Dx). Specimen-related root causes included artefacts, limited diagnostic material, and poor slide quality. After re-review and discussion, a consensus diagnosis could be assigned in all cases.
Conclusions
Diagnostic variability is related to multiple factors, but consensus conferences, standardized electronic reporting formats and comments on suboptimal specimen quality can be used to reduce diagnostic variability.
References
- 1Abdollahi A, Meysamie A, Sheikhbahaei S et al. Inter/intra-observer reproducibility of Gleason scoring in prostate adenocarcinoma in Iranian pathologists. Urol. J. 2012; 9; 486–490.
- 2Barnhill RL, Argenyi Z, Berwick M et al. Atypical cellular blue nevi (cellular blue nevi with atypical features): lack of consensus for diagnosis and distinction from cellular blue nevi and malignant melanoma (‘malignant blue nevus’). Am. J. Surg. Pathol. 2008; 32; 36–44.
- 3Bean SM, Meara RS, Vollmer RT et al. p16 Improves interobserver agreement in diagnosis of anal intraepithelial neoplasia. J. Low. Genit. Tract Dis. 2009; 13; 145–153.
- 4Coco DP, Goldblum JR, Hornick JL et al. Interobserver variability in the diagnosis of crypt dysplasia in Barrett esophagus. Am. J. Surg. Pathol. 2011; 35; 45–54.
- 5El-Zimaity HM, Wotherspoon A, de Jong D. Interobserver variation in the histopathological assessment of MALT/MALT lymphoma: towards a consensus. Blood Cells Mol. Dis. 2005; 34; 6–16.
- 6Fadare O, Parkash V, Dupont WD et al. The diagnosis of endometrial carcinomas with clear cells by gynecologic pathologists: an assessment of interobserver variability and associated morphologic features. Am. J. Surg. Pathol. 2012; 36; 1107–1118.
- 7Fleskens SA, Bergshoeff VE, Voogd AC et al. Interobserver variability of laryngeal mucosal premalignant lesions: a histopathological evaluation. Mod. Pathol. 2011; 24; 892–898.
- 8Foss FA, Milkins S, McGregor AH. Inter-observer variability in the histological assessment of colorectal polyps detected through the NHS Bowel Cancer Screening Programme. Histopathology 2012; 61; 47–52.
- 9Gerhard R, da Cunha Santos G. Inter- and intraobserver reproducibility of thyroid fine needle aspiration cytology: an analysis of discrepant cases. Cytopathology 2007; 18; 105–111.
- 10Gupta T, Nair V, Epari S, Pietsch T, Jalali R. Concordance between local, institutional, and central pathology review in glioblastoma: implications for research and practice: a pilot study. Neurol. India 2012; 60; 61–65.
- 11Kerkhof M, van Dekken H, Steyerberg EW et al. Grading of dysplasia in Barrett's oesophagus: substantial interobserver variation between general and gastrointestinal pathologists. Histopathology 2007; 50; 920–927.
- 12Longnecker DS, Adsay NV, Fernandez-del Castillo C et al. Histopathological diagnosis of pancreatic intraepithelial neoplasia and intraductal papillary-mucinous neoplasms: interobserver agreement. Pancreas 2005; 31; 344–349.
- 13Parkash V, Bifulco C, Feinn R, Concato J, Jain D. To count and how to count, that is the question: interobserver and intraobserver variability among pathologists in lymph node counting. Am. J. Clin. Pathol. 2010; 134; 42–49.
- 14Puppa G, Senore C, Sheahan K et al. Diagnostic reproducibility of tumour budding in colorectal cancer: a multicentre, multinational study using virtual microscopy. Histopathology 2012; 61; 562–575.
- 15Stang A, Trocchi P, Ruschke K et al. Factors influencing the agreement on histopathological assessments of breast biopsies among pathologists. Histopathology 2011; 59; 939–949.
- 16Anderson TJ, Sufi F, Ellis IO, Sloane JP, Moss S. Implications of pathologist concordance for breast cancer assessments in mammography screening from age 40 years. Hum. Pathol. 2002; 33; 365–371.
- 17Beck JS. Observer variability in reporting of breast lesions. J. Clin. Pathol. 1985; 38; 1358–1365.
- 18Bianchi S, Palli D, Galli M et al. Reproducibility of histological diagnoses and diagnostic accuracy of non palpable breast lesions. Pathol. Res. Pract. 1994; 190; 69–76.
- 19Bodian CA, Perzin KH, Lattes R, Hoffmann P. Reproducibility and validity of pathologic classifications of benign breast disease and implications for clinical applications. Cancer 1993; 71; 3908–3913.
10.1002/1097-0142(19930615)71:12<3908::AID-CNCR2820711218>3.0.CO;2-F CAS PubMed Web of Science® Google Scholar
- 20Collins LC, Connolly JL, Page DL et al. Diagnostic agreement in the evaluation of image-guided breast core needle biopsies: results from a randomized clinical trial. Am. J. Surg. Pathol. 2004; 28; 126–131.
- 21Elston CW, Sloane JP, Amendoeira I et al. Causes of inconsistency in diagnosing and classifying intraductal proliferations of the breast. European Commission Working Group on Breast Screening Pathology. Eur. J. Cancer 2000; 36; 1769–1772.
- 22Ghofrani M, Tapia B, Tavassoli FA. Discrepancies in the diagnosis of intraductal proliferative lesions of the breast and its management implications: results of a multinational survey. Virchows Arch. 2006; 449; 609–616.
- 23Palazzo JP, Hyslop T. Hyperplastic ductal and lobular lesions and carcinomas in situ of the breast: reproducibility of current diagnostic criteria among community- and academic-based pathologists. Breast J. 1998; 4; 230–237.
- 24Palli D, Galli M, Bianchi S et al. Reproducibility of histological diagnosis of breast lesions: results of a panel in Italy. Eur. J. Cancer 1996; 32A; 603–607.
- 25Rosai J. Borderline epithelial lesions of the breast. Am. J. Surg. Pathol. 1991; 15; 209–221.
- 26Schnitt SJ, Connolly JL, Tavassoli FA et al. Interobserver reproducibility in the diagnosis of ductal proliferative breast lesions using standardized criteria. Am. J. Surg. Pathol. 1992; 16; 1133–1143.
- 27Sloane JP, Ellman R, Anderson TJ et al. Consistency of histopathological reporting of breast lesions detected by screening: findings of the UK National External Quality Assessment (EQA) Scheme. UK National Coordinating Group for Breast Screening Pathology. Eur. J. Cancer 1994; 30A; 1414–1419.
- 28Wells WA, Carney PA, Eliassen MS, Tosteson AN, Greenberg ER. Statewide study of diagnostic agreement in breast pathology. J. Natl Cancer Inst. 1998; 90; 142–145.
- 29Jain RK, Mehta R, Dimitrov R et al. Atypical ductal hyperplasia: interobserver and intraobserver variability. Mod. Pathol. 2011; 24; 917–923.
- 30Perkins C, Balma D, Garcia R. Why current breast pathology practices must be evaluated. A Susan G. Komen for the Cure white paper: June 2006. Breast J. 2007; 13; 443–447.
- 31Saul S. Prone to error: earliest steps to find cancer. New York Times 2010; 1.
- 32Sloane JP, Amendoeira I, Apostolikas N et al. Consistency achieved by 23 European pathologists in categorizing ductal carcinoma in situ of the breast using five classifications. European Commission Working Group on Breast Screening Pathology. Hum. Pathol. 1998; 29; 1056–1062.
- 33Masood S, Rosa M. Borderline breast lesions: diagnostic challenges and clinical implications. Adv. Anat. Pathol. 2011; 18; 190–198.
- 34Oster NV, Carney PA, Allison KH et al. Development of a diagnostic test set to assess agreement in breast pathology: practical application of the Guidelines for Reporting Reliability and Agreement Studies (GRRAS). BMC Women's Health 2013; 13; 3.
- 35Hsu CC, Sanford BA. The Delphi technique: making sense of consensus. Pract. Assess. Res. Eval. 2007; 12; 1–8.
- 36Borkan J. Immersian/crystallization. In BF Crabtree, WL Miller eds. Doing qualitative research. Thousand Oaks, CA: Sage Publications, 1999; 179–194.
- 37Miller W, Crabtree BF. Qualitative analysis: how to begin making sense. Fam. Pract. Res. J. 1994; 14; 289–297.
- 38Miller W, Crabtree B, Denzin N, Lincoln Y. Handbook of qualitative research. Newbury Park, CA: Sage Publications, 1994.
- 39Bauer MW. Classical content analysis: a review. In MW Bauer, G Gaskell eds. Qualitative researching with text, image and sound. New York, NY: Sage Publishing, 2000; 131–151.
- 40Lakhani SR, Collins N, Stratton MR, Sloane JP. Atypical ductal hyperplasia of the breast: clonal proliferation with loss of heterozygosity on chromosomes 16q and 17p. J. Clin. Pathol. 1995; 48; 611–615.
- 41Allison KH, Reed SD, Voight LF, Jordan CD, Newton KM, Garcia RL. Diagnosing endometrial hyperplasia: why is it so difficult to agree? Am. J. Surg. Pathol. 2008; 32; 691–698.
- 42Elmore JG, Feinstein AR. A bibliography of publications on observer variability (Final Installment). J. Clin. Epidemiol. 1992; 45; 567–580.
- 43Nagurney JT, Brown DF, Chae C et al. Disagreement between formal and medical record criteria for diagnosis of acute coronary syndrome. Acad. Emerg. Med. 2005; 12; 446–452.
- 44Lester SC, Bose S, Chen YY et al. Protocol for the examination of specimens from patients with invasive carcinoma of the breast. Arch. Pathol. Lab. Med. 2009; 133; 1515–1538.
- 45Pagni F, Bosisio FM, Salvioni D, Colombo P, Leone BE, Di Bella C. Application of the British National Health Service Breast Cancer Screening Programme classification in 226 breast core needle biopsies: correlation with resected specimens. Ann. Diagn. Pathol. 2012; 16; 112–118.
- 46Ellis IO, Humphreys S, Michell M, Pinder SE, Wells CA, Zakhour HD. Best Practice No. 179. Guidelines for breast needle core biopsy handling and reporting in breast screening assessment. J. Clin. Pathol. 2004; 57; 897–902.