#PAGE_PARAMS# #ADS_HEAD_SCRIPTS# #MICRODATA#

Institutional differences in USMLE Step 1 and 2 CK performance: Cross-sectional study of 89 US allopathic medical schools


Authors: Jesse Burk-Rafel aff001;  Ricardo W. Pulido aff002;  Yousef Elfanagely aff003;  Joseph C. Kolars aff004
Authors place of work: Department of Internal Medicine, New York University Langone Health, New York, NY, United States of America aff001;  Department of Otolaryngology–Head and Neck Surgery, University of Washington, Seattle, WA, United States of America aff002;  Department of Internal Medicine, Brown University, Providence, RI, United States of America aff003;  Department of Internal Medicine, University of Michigan Medical School, Ann Arbor, MI, United States of America aff004
Published in the journal: PLoS ONE 14(11)
Category: Research Article
doi: https://doi.org/10.1371/journal.pone.0224675

Summary

Introduction

The United States Medical Licensing Examination (USMLE) Step 1 and Step 2 Clinical Knowledge (CK) are important for trainee medical knowledge assessment and licensure, medical school program assessment, and residency program applicant screening. Little is known about how USMLE performance varies between institutions. This observational study attempts to identify institutions with above-predicted USMLE performance, which may indicate educational programs successful at promoting students’ medical knowledge.

Methods

Self-reported institution-level data was tabulated from publicly available US News and World Report and Association of American Medical Colleges publications for 131 US allopathic medical schools from 2012–2014. Bivariate and multiple linear regression were performed. The primary outcome was institutional mean USMLE Step 1 and Step 2 CK scores outside a 95% prediction interval (≥2 standard deviations above or below predicted) based on multiple regression accounting for students’ prior academic performance.

Results

Eighty-nine US medical schools (54 public, 35 private) reported complete USMLE scores over the three-year study period, representing over 39,000 examinees. Institutional mean grade point average (GPA) and Medical College Admission Test score (MCAT) achieved an adjusted R2 of 72% for Step 1 (standardized βMCAT 0.7, βGPA 0.2) and 41% for Step 2 CK (standardized βMCAT 0.5, βGPA 0.3) in multiple regression. Using this regression model, 5 institutions were identified with above-predicted institutional USMLE performance, while 3 institutions had below-predicted performance.

Conclusions

This exploratory study identified several US allopathic medical schools with significant above- or below-predicted USMLE performance. Although limited by self-reported data, the findings raise questions about inter-institutional USMLE performance parity, and thus, educational parity. Additional work is needed to determine the etiology and robustness of the observed performance differences.

Keywords:

Schools – Medical education – United States – Standardized tests – Undergraduates – Primary care – Academic skills – Observational studies

Introduction

The United States Medical Licensing Examination (USMLE) is a 3-step examination required for medical licensure in the United States. The first two exams, USMLE Step 1 and Step 2 Clinical Knowledge (CK), assess medical students’ mastery of basic biomedical principles and their clinical applications [1,2]. About 40,000 trainees take each exam annually, of which over 35% are non-US/Canadian medical students [3]. Both exams are high-stakes parameters of medical student performance critical for advancement [4], residency applicant screening and selection [5,6], and future board certification [7]. Multiple studies have demonstrated correlations between individual factors–including Medical College Admission Test (MCAT) score [8], undergraduate grade point average (GPA) [9], and study behaviors [10]–and USMLE performance. However, little is known about institutional USMLE performance variation. One group analyzing data from the 1990s demonstrated that institutional variables, including curricular differences, did not predict USMLE performance [11,12]. A recent study using one year of national data found some evidence of inter-institutional USMLE performance differences, but the short study duration precludes definitive conclusions [13].

In this exploratory, institution-level study, we analyze institutional variation in USMLE Step 1 and Step 2 CK performance relative to mean matriculant GPA and MCAT. Our primary objective was to identify institutions with above-predicted USMLE performance, which may indicate educational programs successful at promoting students’ medical knowledge.

Methods

This observational study was conducted in accordance with the STROBE guidelines for observational studies in epidemiology [14].

Data sources

We manually tabulated self-reported institutional data–aggregate percentages and means representing yearly medical student cohorts at single institutions–from the annual US News and World Report “Best Graduate Schools” publication (2008–2016 editions) [15] and the Association of American Medical Colleges (AAMC) Medical School Admission Requirements publication (2008–2012 editions) [16] for all 131 US allopathic medical schools. Osteopathic institutions were excluded from this study, as osteopathic students typically take the COMLEX licensing examination rather than the USMLE and very few US osteopathic institutions reported USMLE performance data. A sample size calculation was not performed because we obtained available data for a census of US allopathic medical schools during the study period. National averages for all allopathic matriculants and examinees were obtained from official AAMC [17] and USMLE sources [18,19]. Institutional Review Board approval was not required as no human subjects or identifiable data were involved.

Primary outcome measures and predictor variables

The primary outcome measures were institutional mean USMLE Step 1 and 2 CK scores, averaged over the 3-year study period 2012–2014. Predictor variables included students’ prior academic performance (institutional mean undergraduate GPA and MCAT, averaged over 3 years) and demographics (percentage non-traditional students, minority students, undergraduate biological sciences or humanities majors), and medical school factors (acceptance rate, public/private status, faculty-to-student ratio, National Institutes of Health research funding, graduates entering primary care). MCAT scores represented total scores computed as the sum of the average institutional scores on all 3 sections (biological sciences, physical sciences, verbal reasoning). Institutional USMLE scores were matched to institutional GPA and MCAT averages from two or four years prior (for Step 1 or 2 CK, respectively) to account for the typical lag between matriculation and USMLE testing.

Statistical analysis

All analysis was at the institution level. We performed ordinary least squares linear regression analysis, with test of Pearson’s r for bivariate correlations. Conditions of linearity, nearly normal residuals, and homoscedasticity were checked [20]. Institutions with 3-year average USMLE performance outside a 95% prediction interval (regression residual ≥2 standard deviations, SD, from predicted) were identified [21]. Hypothesis tests were 2-sided with α = .05; ANOVA was used to confirm overall significance of multiple regressions. Statistical analysis was done using SPSS version 25.0 (SPSS Inc., Chicago, Illinois).

Results

In total, 89 (54 public and 35 private) of 131 US allopathic medical schools reported complete USMLE scores over the 3-year study period (68% reporting rate), representing 39,615 and 39,252 Step 1 and 2 CK examinees, respectively. Among reporting institutions, the institutional mean USMLE Step 1 score was 229.7 (SD 5.5) and Step 2 CK score was 238.3 (SD 4.7) (Table 1). GPA and MCAT scores showed minimal heterogeneity across the study years (data not shown). USMLE scores increased across the study years, which was also observed nationally. The average GPA, MCAT scores, and USMLE Step 1 scores for the 89 reporting institutions were slightly higher than national averages for all matriculants/examinees. Complete GPA, MCAT, and USMLE data for reporting institutions and nationally are provided in S1 Table.

Tab. 1. Average GPA, MCAT, and USMLE Step 1 and 2 CK score among 89 US allopathic medical schools and nationally.
Average GPA, MCAT, and USMLE Step 1 and 2 CK score among 89 US allopathic medical schools and nationally.

Predictors of institutional USMLE performance

The strongest predictor of institutional USMLE scores was prior student academic performance, including undergraduate GPA (Step 1, Pearson’s r = .64; Step 2 CK, r = .53; both P < .001) and MCAT score (Step 1, r = .84; Step 2 CK, r = .62; both P < .001). Numerous student body demographic and institutional factors had moderately strong correlations with institutional USMLE scores in bivariate regression; however, when controlling for GPA and MCAT, these correlations were weak and no longer significant (Table 2). For example, private institutions were correlated with higher USMLE Step 1 scores (r = .51, P < .001), but this correlation vanished after controlling for GPA and MCAT (r = .12, P = .42), as private institutions recruit students with higher MCAT scores compared to public institutions (mean 33.5 vs. 30.9, difference 2.7, 95% CI 1.9–3.5; P < .001).

Tab. 2. Linear regression between various institutional characteristics and institutional USMLE performance, without and with control for average institutional GPA and MCAT.
Linear regression between various institutional characteristics and institutional USMLE performance, without and with control for average institutional GPA and MCAT.

The final regression model utilizing GPA and MCAT achieved an adjusted R2 of 72% for Step 1 (standardized βMCAT 0.7, βGPA 0.2, model P < .001) and 41% for Step 2 CK (standardized βMCAT 0.5, βGPA 0.3, model P < .001). GPA added significant but incremental validity evidence over MCAT alone (Step 1, ΔR2 2%, P = .009; Step 2 CK, ΔR2 4%, P = .02); accordingly, for visualization, institutional USMLE was regressed on MCAT score alone (Fig 1).

Fig. 1. Regression analysis of institutional MCAT versus USMLE performance.
Regression analysis of institutional MCAT versus USMLE performance.
(A) Regression analysis of institutional average matriculant Medical College Admission Test (MCAT) score (2010–2012) versus institutional average US Medical Licensing Examination (USMLE) Step 1 score (2012–2014) across n = 89 US allopathic medical schools, representing 39,615 examinees. (B) Regression analysis of institutional average matriculant MCAT score (2008–2010) versus institutional average USMLE Step 2 Clinical Knowledge (CK) score (2012–2014) across n = 89 US allopathic medical schools, representing 39,252 examinees. For both plots, each bubble represents 3-year average at one institution, with bubble size reflecting number examined at each institution. Ordinary least squares best fit line (solid) and 95% prediction interval (dashed lines) are shown, with colored data points highlighting institutions outside the prediction interval.

Institutions with above- or below-predicted USMLE performance

Using the GPA and MCAT regression model, we identified a subset of institutions with 3-year average institutional USMLE scores statistically above or below predicted (Table 3).

Tab. 3. US allopathic medical schools with above- or below-predicteda institutional USMLE Step 1 or Step 2 CK performance, 2012–2014.
US allopathic medical schools with above- or below-predicted<em class="ref"><sup>a</sup></em> institutional USMLE Step 1 or Step 2 CK performance, 2012–2014.

Discussion

In this exploratory study of 89 US allopathic medical schools, we identified 5 institutions with above-predicted institutional USMLE performance based on the described model. The etiology of these institutions’ relative success (or the 3 unnamed institutions’ below-predicted performance) is unclear; we can only say that numerous demographic and institutional factors we assessed did not account for this variation. We hypothesize that unmeasured student factors that vary systematically between institutions (e.g., through admissions processes) or institution-specific factors (e.g., alignment of curricula with USMLE content) may explain these institutional differences. For example, medical schools that provide commercially available Step 1 question banks [22] or where students take Step 1 after the core clerkships [23] have reported improved institutional scores, demonstrating that unique institutional strategies can promote students’ USMLE success. Further study is needed to understand if the 5 institutions identified here have unique factors that promoted their students’ success on these exams.

We found that institutions’ average student GPA and MCAT accounted for substantial variation in institutional average USMLE Step 1 and Step 2 CK scores, which was expected based on prior studies at the individual [8,9] and institutional level [9,11,12]. Importantly, institutional demographic factors (such as percent minority students or biological sciences majors) were correlated with institutional USMLE performance in bivariate regression but were not significant after controlling for GPA and MCAT. National Institutes of Health research funding, which had been previously shown to correlate with institutional USMLE performance [13], was similarly not significant when controlling for GPA and MCAT.

Institutions with a propensity for matching students in the primary care specialties family medicine, pediatrics, and internal medicine–which have lower USMLE screening thresholds for residency interviews than other more “competitive” specialties [24]–tended to recruit students with lower GPA and MCAT scores, and thus lower institutional USMLE scores. As with other institutional factors, however, institutions’ primary care specialty rate was not associated with differential USMLE performance beyond its association with GPA and MCAT.

Such findings highlight the critical importance of controlling for prior academic performance when attempting to explain USMLE performance differences. However, we doubt that pre-medical students–a key consumer of the annual US News and World Report data–consider these covariates when interpreting institutional USMLE scores and identifying medical schools of interest. Indeed, undergraduates might conclude (erroneously) that private medical schools outperform public schools on the USMLE, when in fact students attending private schools have higher test scores at matriculation. There may be a role for better contextualizing this data so that pre-medical students can be informed consumers. The National Board of Medical Examiners (NBME), who produce the USMLE, are positioned to more rigorously explore the relationship between institutions and exam performance.

Limitations

This study relied on self-reported institutional data via a third-party publication, as the NBME does not publish institutional score performance. Misreporting is possible, although we validated the reported scores from several institutions. US News and World Report provides their methodology for data collection with each annual release [25], but do not state specifics related to how data is validated or standardized within- or between-schools. For example, it is unclear if institutions have discretion in how they formulate their institutional MCAT average, including how individuals with multiple test results are handled, which can introduce bias into the relationship between MCAT and USMLE performance [26].

Although we assessed numerous student and medical school factors, some potentially important covariates–such as percent of students with advanced degrees, curricular structure, timing of USMLE examinations, and school age–were not incorporated into this study but are important areas for future investigation. For example, some institutions have moved the USMLE Step 1 test window to after core clinical clerkships [23], with small benefits in scores and reduced failure rates [27].

Moreover, only 89 of 131 US allopathic medical schools (68%) reported complete data; non-reporters may differ in important ways. We found that reporting institutions, as compared to an average of all students nationally, had slightly higher average GPA and MCAT scores, with an associated 1.5-point higher average USMLE Step 1 score. Statistical comparisons of these differences are not advisable given the different units of reporting (institutions vs. individuals); yet the very small differences suggest that the reporting institutions were nationally representative. The relatively short 3-year study period does not preclude that the observed institutional outliers may represent random variation; replication with longer observation is needed. Finally, our study was ecological; no inference can be made that institution-level findings translate to individual students (i.e., the ecological fallacy), and indeed only institutional averages and counts were reported (without any measure of student-to-student variability). Nevertheless, the purpose of this study was only to compare institutions.

Conclusions

We found that institutional average GPA and MCAT scores correlate strongly with institutional USMLE performance. Numerous student demographic and institutional factors were insignificant when controlling for GPA and MCAT. We identified several institutions with significant above- or below-predicted USMLE performance, raising questions about inter-institutional USMLE performance parity. Methods to assess institutions’ overall performance on knowledge-based exams may offer a parameter to evaluate medical schools and their curricula, while providing prospective students with valuable data regarding these high-stakes exams. Additional study is needed to explore the etiology and durability of the observed performance differences, and to incorporate other student and institutional factors that may be important predictors of performance.

Supporting information

S1 Table [xlsx]
Average GPA, MCAT, and USMLE Step 1 and 2 CK for 89 US allopathic medical schools reporting complete data from 2012–2014, with associated national averages.


Zdroje

1. Haist SA, Katsufrakis PJ, Dillon GF. The evolution of the United States Medical Licensing Examination (USMLE): Enhancing assessment of practice-related competencies. JAMA. 2013;310(21): 2245–2246. doi: 10.1001/jama.2013.282328 24302081

2. US Medical Licensing Examination, 2019 Bulletin of Information. National Board of Medical Examiners. [cited 2018 Sep 15]. Available from: https://www.usmle.org/pdfs/bulletin/2019bulletin.pdf.

3. US Medical Licensing Examination, 2018 Performance Data. National Board of Medical Examiners. [cited 2019 Jul 17]. Available from: https://www.usmle.org/performance-data/default.aspx#2018_step-1

4. Andriole DA, Jeffe DB. Prematriculation variables associated with suboptimal outcomes for the 1994–1999 cohort of US medical school matriculants. JAMA. 2010;304(11): 1212–1219. doi: 10.1001/jama.2010.1321 20841535

5. National Resident Matching Program, Data Release and Research Committee: Results of the 2018 NRMP Program Director Survey. National Resident Matching Program. Washington, DC. [cited 2018 Sep 15]. Available from: http://www.nrmp.org/wp-content/uploads/2018/07/NRMP-2018-Program-Director-Survey-for-WWW.pdf

6. Jayakumar KL, Lipoff JB. US Medical Licensing Examination Step 2 Clinical Knowledge score in dermatology resident selection. JAMA Dermatol. 2017;153(11): 1193–1194. doi: 10.1001/jamadermatol.2017.2924 28832874

7. Jeffe DB, Andriole DA. Factors associated with American Board of Medical Specialties member board certification among US medical school graduates. JAMA. 2011;306(9): 961–970. doi: 10.1001/jama.2011.1099 21900136

8. Donnon T, Paolucci EO, Violato C. The predictive validity of the MCAT for medical school performance and medical board licensing examinations: A meta-analysis of the published research. Acad Med. 2007;82(1): 100–106. doi: 10.1097/01.ACM.0000249878.25186.b7 17198300

9. Basco WT Jr, Way DP, Gilbert GE, Hudson A. Undergraduate institutional MCAT scores as predictors of USMLE Step 1 performance. Acad Med. 2002;77(10S): 13–16.

10. Burk-Rafel J, Santen SA, Purkiss J. Study behaviors and USMLE Step 1 performance: Implications of a student self-directed parallel curriculum. Acad Med. 2017;92(11S): 67–74.

11. Hecker K, Violato C. How much do differences in medical schools influence student performance? A longitudinal study employing hierarchical linear modeling. Teach Learn Med. 2008;20(2): 104–113. doi: 10.1080/10401330801991915 18444195

12. Hecker K, Violato C. Medical school curricula: Do curricular approaches affect competence in medicine? Fam Med. 2010;41(6): 420–426.

13. Ghaffari-Rafi A, Lee RE, Fang R, Miles DJ. Multivariable analysis of factors associated with USMLE scores across US medical schools. BMC Med Ed. 2019;19: 154.

14. von Elm E, Altman DG, Egger M, Pocock SJ, Gotzsche PC, Vandenbroucke JP. The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) Statement: Guidelines for reporting observational studies. PLOS Med. 2007;4(10): e296. doi: 10.1371/journal.pmed.0040296 17941714

15. Best Medical Schools: Research. US News and World Report. 2008–2016 editions, online and print. [cited 2018 Sep 15]. Available from: http://grad-schools.usnews.rankingsandreviews.com/best-graduate-schools/top-medical-schools/research-rankings/ Subscription required.

16. Medical School Admission Requirements. Association of American Medical Colleges. Washington (DC): 2008–2012 editions.

17. FACTS Table 17. Association of American Medical Colleges. 2013. [cited 2016 Nov 6] Available by request from: https://www.aamc.org/data-reports/students-residents/report/facts

18. US Medical Licensing Examination, Performance Data. National Board of Medical Examiners. 2014–2016. Available from: https://www.usmle.org/performance-data/

19. USMLE Score Interpretation Guidelines. 2015 ed. Available from: https://www.usmle.org/pdfs/transcripts/USMLE_Step_Examination_Score_Interpretation_Guidelines.pdf

20. Altman N, Krzywinski. Regression diagnostics. Nat Meth. 2016;13: 385–386.

21. Werner J, Griebeler EM. New insights into non-avian dinosaur reproduction and their evolutionary and ecological implications: Linking fossil evidence to allometries of extant close relatives. PLOS ONE. 2013;8(8): e72862. doi: 10.1371/journal.pone.0072862 23991160

22. Banos JH, Pepin ME, Van Wagoner N. Class-wide access to a commercial Step 1 question bank during preclinical organ-based modules: a pilot project. Acad Med. 2018;93(3): 486–490. doi: 10.1097/ACM.0000000000001861 28817433

23. Daniel M, Fleming A, Grochowski C, Harnik V, Klimstra S, Morrison G, et al. Why not wait? Eight institutions share their experiences moving United States Medical Licensing Examination Step 1 after core clinical clerkships. Acad Med. 2017;92(11): 1515–1524. doi: 10.1097/ACM.0000000000001714 28422816

24. National Resident Matching Program, Data Release and Research Committee: Results of the 2018 NRMP Program Director Survey. National Resident Matching Program, Washington (DC): 2018. Available from: https://www.nrmp.org/wp-content/uploads/2018/07/NRMP-2018-Program-Director-Survey-for-WWW.pdf

25. Morse R, Martin E. Methodology: 2020 Best Medical Schools. Available from: https://www.usnews.com/education/best-graduate-schools/articles/medical-schools-methodology

26. Zhao X, Oppler S, Dunleavy D, Kroopnick M. Validity of four approaches of using repeaters' MCAT scores in medical school admissions to predict USMLE Step 1 total scores. Acad Med. 2010;85(10 Suppl): S64–67. doi: 10.1097/ACM.0b013e3181ed38fc 20881707

27. Jurich D, Daniel M, Paniagua M, Fleming A, Harnik V, Pock A, et al. Moving the United States Medical Licensing Examination Step 1 after core clerkships: An outcomes analysis. Acad Med. 2019;94(3): 371–377. doi: 10.1097/ACM.0000000000002458 30211755


Článek vyšel v časopise

PLOS One


2019 Číslo 11
Nejčtenější tento týden
Nejčtenější v tomto čísle
Kurzy

Zvyšte si kvalifikaci online z pohodlí domova

plice
INSIGHTS from European Respiratory Congress
nový kurz

Současné pohledy na riziko v parodontologii
Autoři: MUDr. Ladislav Korábek, CSc., MBA

Svět praktické medicíny 3/2024 (znalostní test z časopisu)

Kardiologické projevy hypereozinofilií
Autoři: prof. MUDr. Petr Němec, Ph.D.

Střevní příprava před kolonoskopií
Autoři: MUDr. Klára Kmochová, Ph.D.

Všechny kurzy
Kurzy Podcasty Doporučená témata Časopisy
Přihlášení
Zapomenuté heslo

Zadejte e-mailovou adresu, se kterou jste vytvářel(a) účet, budou Vám na ni zaslány informace k nastavení nového hesla.

Přihlášení

Nemáte účet?  Registrujte se

#ADS_BOTTOM_SCRIPTS#