Learn more: PMC Disclaimer | PMC Copyright Notice
Investigating the Impact of Preparation Strategies on USMLE Step 1 Performance
Abstract
Background
The USMLE Step 1 score helps differentiate applicants for competitive residency programs. Students frequently ask medical educators how to prepare for this high-stakes exam. Multiple resources exist such as books, training programs, and question banks. The purpose of this study is to provide medical educators with data on which resources correlate with higher exam scores.
Methods
164 medical students at West Virginia University School of Medicine were given a survey following the completion of Step 1. The survey contained questions about resources used and exam date. De-identified data was paired with class quartile rankings and Step 1 scores. Average Step 1 scores were calculated for each resource and a student's t-test was used to compare between groups. Test dates were also broken into quintiles and average Step 1 scores were compared between groups.
Results
Students who used the USMLE World question bank had higher Step 1 scores (M=229, SE=1.4) compared to students who did not use this resource (M=215, SE=4.2). Commercial preparation courses were not associated with improved scores. The timing of when the exam was taken was also not associated with significant differences between scores.
Conclusions
Interactive study options offer the best choice to improve USMLE Step 1 scores.
Introduction
The United States Medical Licensing Exam (USMLE) Step 1 is the first of a series of three board exams. It serves both as a licensing exam as well as a screening tool for residency directors. Scores have continued to rise on the USMLE Step 1 over the past decade (Sutton et al., 2014). The number of medical students applying to competitive residency programs is also on the rise (Rinard & Mahabir, 2010). Residency directors at competitive programs use the USMLE Step 1 score in conjunction with class ranking to screen potential applicants (Lotfipour et al., 2008). Those students meeting certain academic and professional criteria are offered an interview. USMLE Step 1 scores correlate with how well a student will do on a residency board exam (Kay, Jackson, & Frank, 2015). A primary goal of medical educators is to promote success for students and to help them fulfill their career ambitions. The key to this process is providing an excellent medical curriculum and offering informative guidance for USMLE Step 1 preparation (Lieberman et al., 2010). This study examined the extent to which test preparation courses help improve scores on the USMLE Step 1 as well as when to take the exam following the second year of medical school. Common courses and preparatory programs including Doctors in Training, First Aid and Board Review Series books, and Kaplan and USMLE World Question banks were evaluated in correlation to score differences on the USMLE Step 1. Testing time interval following second year medical school was broken into quintiles. Students preparing for this high-stakes examination used various strategies for preparation and outcomes varied accordingly for the different preparation strategies employed. Not surprisingly, students in the first quartile of the medical school class performed better on USMLE Step 1 than students in the fourth quartile. The objective of this study is to investigate which preparation strategies correlate with high scores on the USMLE Step 1 in order to improve academic guidance regarding the best methods to prepare for the exam.
Methods
The West Virginia University (WVU) Institutional Review Board committee approved this study. Students from the WVU School of Medicine graduating classes of 2013-2014 who chose to participate were given a cover letter that instructed him or her to the purpose of the research and to his or her role as a participant. A survey (Appendix 1) in the form of a handout was designed for current WVU medical students who completed the second year curriculum and passed the USMLE Step 1 exam. The survey included the date of the exam, date of expected graduation, and resources used for exam (First Aid, Board Review Series, USMLE World Question Bank, Kaplan Question Bank, and Doctors in Training). A numerical analysis template was created for each student with data from the survey. The number of students who used each resource was tabulated and the time period for taking the exam was broken up into quintiles. The data set was de-identified for personal information and subsequently paired with USMLE Step 1 exam scores and class quartile rank based on a numerical match system. The mean USMLE Step 1 score was calculated for students who used each reference material and was reported with standard error. A student's t-test was used to compare between groups. The effect size was calculated to determine practical significance. A Pearson's coefficient was determined for quartile ranking and USMLE Step 1 scores. The coefficient of determination was likewise calculated. P<0.05 was considered statistically significant for all values.
Results
From the graduating classes of 2013 and 2014, there were a total of 165 students. Of those, 164 students completed the survey (99.39%). The majority of students used multiple resources to prepare for the exam. The resources can be broadly grouped into books, training programs, and question banks. Of those students who completed the survey, 97.56% used the First Aid book and 42.07% used the Board Review Series books. 51.83% used the Doctors in Training program. 88.41% used the USMLE Question bank and 15.85% used the Kaplan Question Bank. Those students who used the Doctors in Training program inherently used the First Aid book because it was included as the course reference (Figure 1).
Students who used USMLE World Question bank had higher USMLE Step 1 scores (M=229, SE=1.4) than students who did not use this resource (M=215, SE=4.2). The difference was statistically significant t(162) = 3.3, p<0.05. The calculated effect size (d=.77) suggests that the difference in mean scores was also practically significant. No other statistically significant differences were found between groups (Figure 2).
The time since finishing second year medical school to taking the exam was broken into quintiles. The time interval until taking the exam did not have an effect on USMLE Step 1 scores, unlike for other medical exams (Malangoni et al., 2012). Those students who took the test closest to the end of second year medical school or closest to the beginning of third year medical school had slightly lower scores but these differences were not statistically significant (Figure 3).
Not surprisingly, a statistically significant difference in USMLE Step 1 scores was seen between class ranking quartiles. Those students who did well on the medical school curriculum performed well on USMLE Step 1. A strong negative correlation was seen for USMLE Step 1 scores and class ranking quartiles (r = −0.689). The 4th quartile students had the lowest USMLE Step 1 scores. The coefficient of determination is 0.47. Approximately 47% of the variance in USMLE Step 1 scores can be explained by knowing the class ranking quartile for the student. Our data showed no significant difference in USMLE Step 1 scores between students in the classes of 2013 and 2014 (2013: mean 226.4, N= 90 and 2014: mean 227.4, N=75) (Figure 4).
Discussion
Commercial test preparation courses, books, and question banks have become increasing popular due to the high-stake nature of the USMLE Step 1. Students often ask medical educators about the best methods to prepare for the exam. In response to these questions, some schools have even started a seminar series to discuss process-oriented exam preparation (Strowd & Lambros, 2010). The best advice to give students is mixed. Over the past several years, studies have reported conflicting results. The National Board of Medical Examiners Comprehensive Basic Science Self-Assessment, which most students take at the end of second year medical school, only accounts for 67% of the variance on USMLE Step 1 scores (Morrison et al., 2010). Therefore, test preparation does play some role in exam performance. Zhang and colleagues however showed that exam performance was closely related to class rank and was independent of preparation technique (Zhang, Rauchwarger, Toth, & O'Connell, 2004). Furthermore, scores on the Medical College Admissions Test (MCAT) correlate closely with performance on USMLE Step 1(Julian, 2005). Additionally, in a large group of students commercial test preparation courses provided no improvement on USMLE Step 1 scores (Werner & Bull, 2003). A novel approach, peer-designed review courses, on the other hand have been tied to better performance on the USMLE Step 1 (Alcamo, Davids, Way, Lynn, & Vandre, 2010). Therefore, traditional preparation techniques may need to be re-evaluated in favor of the more team based and problem solving peer review courses.
In this paper, we showed a unique series of results. Similar to other groups we found that commercial test courses had no statistical significance on exam score (Thadani, Swanson, & Galbraith, 2000). Not surprisingly, we also found that students in the top quartiles performed better on USMLE Step 1. These results were similar to what Zhang and colleagues found at University of Miami School of Medicine (UMSM). With the use of ANCOVA and the second year GPA as a covariate, Zhang's group found that participation in preparation courses had no significant effect on student USMLE Step 1 performance when second-year GPA was controlled (Zhang et al., 2004). The novel finding we did report is that the use of the USMLE World question bank provided a statistically significant improvement on USMLE Step 1 scores. The question bank is interactive and requires students to engage with the resource. Similar to peer-review courses it may be an improved alternative to traditional examination preparation.
Group study is a common method utilized by medical students in preparation for exams and recent evidence has supported its effectiveness. Medical education curriculums have started including problem based learning courses. The benefit of problem-based learning is the application of knowledge through discussion and team engagement.
Problem based learning has been shown to increase retention of material and enhance performance on exams (Imanieh, Dehghani, Sobhani, & Haghighat, 2014). A recent study employed the same techniques for USMLE Step 1 preparation. The Ohio State University College of Medicine (OSUCOM) conducted a study comparing the Step 1 scores of students who participated in a weekly peer designed and peer-led review course during the second year of medical school to other students at OSUCOM who did not participate in the preparation course. When the Step 1 scores were examined using ANCOVA, the participants of the peer-designed and peer-led review course scored significantly higher than the nonparticipants (Alcamo et al., 2010).
Another topic that students frequently consider is when to take the exam. Previously, when the USMLE Step 1 was administered as a paper-and-pencil test, it was administered only twice annually. The examination is now computerized and students have the option to take the USMLE Step 1 at anytime. Students often have anxiety about when to set their test date and frequently ask advice from medical educators. Students at WVU School of Medicine have approximately 7 weeks between the end of the second year curriculum and the start of third year clinical work. The school does not currently offer a sponsored review course, so most students either choose to study on their own for several weeks with several study resources or pay for a month long electronic study course such as Doctors in Training. The time until taking the exam was broken into quintiles. We reported no statistically significant difference in scores based on when the exam was taken. Pohl and colleagues also analyzed the relationship between students Step 1 scores and the amount of time that had passed since the completion of their second year medical course work. The study concluded that as long as the exam was scheduled with 2 months of finishing second year course work there was no statistically significant difference in scores based on when the exam was taken (Pohl, Robeson, Hojat, & Veloski, 2002).
Conclusion
We conducted this study in order to provide tangible evidence that medical educators may use in advising students on USMLE Step 1 preparation. Similar to other groups we have shown that commercial based preparation courses provided no statistically significant benefit on USMLE Step 1 scores. We also showed that the timing for when the exam is taken does not correlate with better or worse scores. What we did find is that study resources, which require student engagement and interaction, correlate with better Step 1 scores. Future work will examine the benefit of peer-review courses and increased problem based learning strategies. Problem based learning may be a way in which students can improve scores on this high-stakes exam. As the average score on USMLE Step 1 continues to increase, medical educators are charged with providing insightful advice to students on how best to succeed on this exam as well as successfully match into residency programs. Novel preparation approaches that engage students and help them participate in the learning process are sure to help students succeed.
Acknowledgements
Research reported in this publication was supported by the NIGMS of the National Institutes of Health under award number U54GM104942. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.
Glossary Terms
Problem based learning- team based learning strategy focusing on student participation, discussion, and application of knowledge. Doctors in Training- commercial USMLE Step 1 preparation course which uses a companion workbook and First Aid resource manual. USMLE World Question Bank- online question bank that organizes questions into 45-minute test blocks. The questions are traditionally three-part questions requiring knowledge integration from multiple disciplines.
Appendix 1
Footnotes
Contributors: PB organized the project, collected data, and helped write the manuscript. BL helped organize the project and write the manuscript. ZR and JB helped collect data through the surveys. SC was the faculty member in charge of the project. He helped provide guidance on statistics and manuscript assembly.
References
- Alcamo AM, Davids AR, Way DP, Lynn DJ, Vandre DD. The impact of a peer-designed and -led USMLE Step 1 review course: improvement in preparation and scores. Acad Med. 2010;85(10 Suppl):S45–48. doi: 10.1097/ACM.0b013e3181ed1cb9. [PubMed] [Google Scholar]
- Imanieh MH, Dehghani SM, Sobhani AR, Haghighat M. Evaluation of problem-based learning in medical students' education. J Adv Med Educ Prof. 2014;2(1):1–5. [PMC free article] [PubMed] [Google Scholar]
- Julian ER. Validity of the Medical College Admission Test for predicting medical school performance. Acad Med. 2005;80(10):910–917. [PubMed] [Google Scholar]
- Kay C, Jackson JL, Frank M. The Relationship Between Internal Medicine Residency Graduate Performance on the ABIM Certifying Examination, Yearly In-Service Training Examinations, and the USMLE Step 1 Examination. Acad Med. 2015;90(1):100–104. doi: 10.1097/ACM.0000000000000500. [PubMed] [Google Scholar]
- Lieberman SA, Ainsworth MA, Asimakis GK, Thomas L, Cain LD, Mancuso MG, Frye AW. Effects of comprehensive educational reforms on academic success in a diverse student body. Med Educ. 2010;44(12):1232–1240. doi: 10.1111/j.1365-2923.2010.03770.x. [PubMed] [Google Scholar]
- Lotfipour S, Luu R, Hayden SR, Vaca F, Hoonpongsimanont W, Langdorf M. Becoming an emergency medicine resident: a practical guide for medical students. J Emerg Med. 2008;35(3):339–344. doi: 10.1016/j.jemermed.2008.04.009. [PubMed] [Google Scholar]
- Malangoni MA, Jones AT, Rubright J, Biester TW, Buyske J, Lewis FR., Jr. Delay in taking the American Board of Surgery qualifying examination affects examination performance. Surgery. 2012;152(4):738–743. discussion 743-736. doi: 10.1016/j.surg.2012.07.001. [PubMed] [Google Scholar]
- Morrison CA, Ross LP, Fogle T, Butler A, Miller J, Dillon GF. Relationship between performance on the NBME Comprehensive Basic Sciences Self-Assessment and USMLE Step 1 for U.S. and Canadian medical school students. Acad Med. 2010;85(10 Suppl):S98–101. doi: 10.1097/ACM.0b013e3181ed3f5c. [PubMed] [Google Scholar]
- Pohl CA, Robeson MR, Hojat M, Veloski JJ. Sooner or later? USMLE step 1 performance and test administration date at the end of the second year. Acad Med. 2002;77(10 Suppl):S17–19. [PubMed] [Google Scholar]
- Rinard JR, Mahabir RC. Successfully matching into surgical specialties: an analysis of national resident matching program data. J Grad Med Educ. 2010;2(3):316–321. doi: 10.4300/JGME-D-09-00020.1. [PMC free article] [PubMed] [Google Scholar]
- Strowd RE, Lambros A. Impacting student anxiety for the USMLE Step 1 through process-oriented preparation. Med Educ Online. 2010;15 doi: 10.3402/meo.v15i0.4880. [PMC free article] [PubMed] [Google Scholar]
- Sutton E, Richardson JD, Ziegler C, Bond J, Burke-Poole M, McMasters KM. Is USMLE Step 1 score a valid predictor of success in surgical residency? Am J Surg. 2014;208(6):1029–1034. doi: 10.1016/j.amjsurg.2014.06.032. [PubMed] [Google Scholar]
- Thadani RA, Swanson DB, Galbraith RM. A preliminary analysis of different approaches to preparing for the USMLE step 1. Acad Med. 2000;75(10 Suppl):S40–42. [PubMed] [Google Scholar]
- Werner LS, Bull BS. The effect of three commercial coaching courses on Step One USMLE performance. Med Educ. 2003;37(6):527–531. [PubMed] [Google Scholar]
- Zhang C, Rauchwarger A, Toth C, O'Connell M. Student USMLE step 1 preparation and performance. Adv Health Sci Educ Theory Pract. 2004;9(4):291–297. doi: 10.1007/s10459-004-3925-x. [PubMed] [Google Scholar]





