Original Research Article - Journal of Contemporary Medical Education (2021)
CASE: BASED VIRTUAL PATIENT SIMULATION IMPACT on MEDICAL and GRADUATE MEDICAL EDUCATIONAL COMPETENCIES and MILESTONES: A META-ANALYSIS
Eric W Lam1, Ivan C Lam1, Miriam E Schwartz1, Deborah Woo2 and Douglas E Paull2*2Department of Medicine, Georgetown University School of Medicine, Washington DC, USA
Douglas E Paull, Department of Medicine, Georgetown University School of Medicine, USA, Email: dpaull78@comcast.net
Received: 26-Oct-2020 Published: 19-Nov-2020
Abstract
Objective: Virtual strategies have assumed an even greater role in medical education
and graduate medical education during the recent pandemic. The objective of this
study was to assess the impact of case-based, virtual patient simulation on medical
education and graduate medical education.
Methods: A literature search of years 2000 through 2019 discovered 2,285 potential
articles on virtual patient simulation. Fifty-four articles meeting the following criteria
were included in the meta-analysis: involved medical education or graduate medical
education participants; utilized case-based virtual patient simulations; and contained
enough data to calculate an effect size (number of participants in each study arm,
mean, and standard deviation of at least one measured learning outcome).
Results: Virtual patient simulation had a large overall impact on medical education
and graduate medical education learning outcomes, 0.88 (0.64-1.12), z=7.36, p<0.001. Effect sizes by competencies were patient care, 0.95 (0.62-1.28); medical knowledge,
0.69 (-0.06-1.44); interpersonal and communication skills, 0.52 (0.01-1.05);
professionalism, 1.32 (0.29-2.35); and systems-based practice, 0.71 (0.21-1.20). There
was, however, a high level of heterogeneity between studies (I2=92.6%) lessening the
certainty of the effect size summary.
Conclusions: The current study reinforces the results of previous meta-analyses
demonstrating the moderate to large effect of virtual patient simulation interventions
upon learning outcomes. Additionally, it highlights the effectiveness of virtual patient
simulation for medical education and graduate medical education competencies
beyond patient care, medical knowledge, and communication to include systems-
based practice and professionalism. Virtual patient simulation is well-suited to address
current challenges facing medical education and graduate medical education.
Keywords
Learning outcomes; Medical education; Graduate medical education; Virtual patient simulation
Introduction
Health care education is facing changes in recent times. With rapidly advancing technologies, increasing complexities of patient care, and numerous challenges to health care systems, a variety of education methods are needed to address the evolving needs of medical students, graduate medical trainees, and clinicians [1]. Simulationbased education, especially Virtual Patient Simulation (VPS), can help address emerging challenges and gaps by reaching a greater number of learners, supplementing the number and types of patients that learners experience (including rare clinical cases), addressing schedules and the lack of time for both learners and clinical educators, avoiding the medico-legal and ethical implications of skill acquisition from actual patients, and promoting patient safety concepts [2-4].
The current COVID-19 pandemic presents additional challenges to health care processes, team training, and education [5,6]. Given the limited number of persons allowed in gatherings, postponed or canceled surgeries, and limited rotations between training sites, Chick et al. proposed innovative methods of learning such as technology-enhanced modalities, including “the flipped classroom model, online practice questions, teleconferencing in place of in-person lectures, involving residents in telemedicine clinics, procedural simulation, and the facilitated use of surgical videos [5].” Kiely et al. recommended strategies for obstetrics team training and simulation-based education while maintaining social distancing; these recommended modalities include spatial, temporal, video-recording, video-conferencing, and virtual learning opportunities to effectively engage obstetrics team members during the COVID-19 pandemic [6]. Virtual strategies, though not a substitute for hands-on direct patient care and procedures, are an example of innovative modalities of healthcare education that can help address the need for rigorous educational learning during this time of the COVID-19 pandemic [5,6].
“Digital education (also known as electronic learning or digital learning) is the act of teaching and learning by means of digital technologies. It is an overarching term for an evolving multitude of educational approaches, concepts, methods, and technologies [7].” Examples of digital learning include mobile education, serious games, virtual patients, and virtual reality environments [2,7]. VPS or screen-based simulation is an example of digital education. “Virtual patient” simulation has been defined as “a specific type of computer program that simulates real-life clinical scenarios; learners emulate the roles of health care providers to obtain history, conduct a physical exam, and make diagnostic and therapeutic decisions [1,8].” A number of narrative reviews, systematic reviews, and meta-analyses regarding VPS have been published in the preceding years. Cook and Triola (2009) conducted a narrative literature review on VPS and proposed that VPS has the “unique and costeffective function to facilitate and assess the development of clinical reasoning [1].” They also recommended more research in instructional design, curricular integration, and using VPS to enhance clinical reasoning [1]. In a subsequent systematic analysis of 48 studies, Cook et al. (2010) concluded that virtual patients were “consistently associated with higher learning outcomes” when compared against studies with no intervention(s) [9].
Consorti et al. (2012) compared VPS to more traditional learning methods, either as an alternative learning method or as an addition to the usual curriculum [10]. They reported that VPS was not only effective as an educational tool for clinical reasoning and in clinical data gathering and interpretation, but also useful in preparing students for further education in the domains of communication skills and ethical reasoning [10]. Kononowicz et al. and the Digital Health Education Collaboration Group (2019) performed a systematic review and meta-analysis of 51 randomized controlled trials (RCT) of VPS in health professions education [2]. The participants in the studies were from a variety of disciplines. Skills that were improved in the studies included clinical reasoning, procedural, and team skills.
Objectives
The primary objective of this meta-analysis was to assess the impact of case-based, VPS on medical education and graduate medical education (GME). A secondary objective was to analyze those VPS design factors that may contribute to positive learning outcomes. We believe that the results of such a meta-analysis will be a timely and helpful tool for medical educators in developing, implementing, and evaluating digital education experiences, especially during current world challenges. The framework for the analysis was core competencies and milestones for medical education and GME described by the Accreditation Council for Graduate Medical Education (ACGME) and the American Board of Medical Specialties (ABMS) [11-13].
Methods
Data sources and searches
The Georgetown University Institutional Review Board (IRB) determined that the study was not considered research involving human subjects and therefore IRB approval was not required. A search strategy was formulated to collect articles pertaining to the effectiveness of screen-based VPS in medical education and GME. The primary search strategy employed Boolean search with a combination of the following terms: healthcare screen based simulation, high fidelity medical simulation, high fidelity simulation, patient modeling, patient simulation, patient simulation training, screen based simulation, screen based virtual patient education, screen based virtual patient simulation, simulation medical education, virtual patient, virtual patient education, virtual patient medical education, virtual patient safety, virtual patient simulation, virtual patient simulation education, virtual patient simulator, virtual patient surgery, virtual patient training, virtual patients, and virtual simulation. The following databases were searched: Cochrane, Google Scholar, JSTOR, Microsoft Academic, OvidMD, PubMed, and Science Direct (Figure 1). The search period was restricted to years 2000 through 2019 to maintain modern relevancy. Search results were compiled in Microsoft Excel. After the removal of duplicates, a total of 1,521 studies were found using this initial approach (Figure 2).
Figure 1. Literature search. Cumulative number of records retrieved from search of databases for years 2000-2019. Databases searched included Cochrane, Google Scholar, JSTOR, Microsoft Academic, OvidMD, PubMed, and Science Direct. Search terms consisted of: healthcare screen based simulation, high fidelity medical simulation, high fidelity simulation, patient modeling, patient simulation, patient simulation training, screen based simulation, screen based virtual patient education, screen based virtual patient simulation, simulation medical education, virtual patient, virtual patient education, virtual patient medical education, virtual patient safety, virtual patient simulation, virtual patient simulation education, virtual patient simulator, virtual patient surgery, virtual patient training, virtual patients, and virtual simulation.
Inclusion criteria and study selection
Titles, abstracts, and text of articles were screened and reviewed by three reviewers (DEP, EWL, MES). Discussion and consensus were used to resolve disagreements. Screen-based VPS design was part of the key inclusion criteria. Articles forwarded for study inclusion shared the following criteria derived from our PICO statement: 1) Population - include medical students and/or residents/fellows; 2) Intervention - case-based, screen-based VPS; 3) Comparison - RCT, pre/post, or experimental-control study design; and 4) Outcome - at least one learning outcome measured in two groups or two instances in time with sufficient data [i.e., mean (M), standard deviation (SD), number of participants (n)] to allow calculation of an effect size (ES). A total of 54 articles were ultimately selected for inclusion in the meta-analysis under these criteria and methods.
Data extraction and items
Full text of selected articles was collected, and data extraction occurred independently and in duplicate. Extracted data variables included study design, number of participants, learner level (medical students or residents/fellows), specialty (e.g., medicine, surgery), type of control group (e.g., didactic, standardized patient), VPS intervention features (e.g., feedback, practice), number of unique VPS scenarios, Kirkpatrick level (e.g., knowledge, performance) assessed, and data (i.e., M, SD, n) for each study and control group. The competencies assessed in each article were compared to the 18 milestones among five competencies outlined by ACGME including: patient care, medical knowledge, interpersonal and communication skills, professionalism, and systembased practice [14]. A sixth competency, practice-based learning and improvement was not included because all the articles and study interventions could be viewed as addressing this competency. The competency assigned to each article by three reviewers (DEP, EWL, MES) demonstrated moderate-good agreement, Fleiss’s Kappa 0.52. Disagreements between reviewers were resolved by consensus.
Data synthesis
ESs (Hedge’s g) were calculated using pooled SDs from comparison groups from each included article. A predetermined algorithm was used to determine which ES was entered for those studies reporting more than one summary, evaluable learning outcome: 1) the highest Kirkpatrick level learning outcome; and 2) if more than one method of assessment was present for the highest level of learning outcome, the ESs were averaged following previously published guidelines [15]. A random effects model was utilized to develop a forest plot and ES summary with 95% confidence intervals for the included VPS articles [16,17]. To assess for and address any publication bias, a funnel plot was utilized for the synthesis and an Egger’s regression applied to the dataset [18]. An adjusted random effects summary and confidence interval were then calculated and reported using an R0 trim and fill technique. Subgroup comparisons were pre-determined and related to the study questions regarding factors impacting learning outcomes for a VPS intervention [19]. Random effects models were used for subgroup analyses and statistical comparisons conducted with ANOVA.
Quality assessment
Heterogeneity between studies was evaluated by Q and I2 statistics [20,21]. The quality of studies was assessed using the 18-point Medical Education Research Study Quality Instrument (MERSQI) checklist tool [22]. There was substantial agreement between the two observers’ (DEP, EWL) who scored the MERSQI, Cohen’s Kappa of 0.56. Bias within individual studies was evaluated using the Cochrane handbook [23]. The risks for selection, attrition, detection, performance, and reporting biases were rated for each study as low (green), high (red), or unclear (yellow). Inter-rater reliability between the three reviewers (DEP, EWL, MES) rating bias for individual studies was substantial, Fleiss’s Kappa=0.62. Bias across all studies was calculated as percentages of low, high, and unclear ratings for each type of bias using a stacked bar approach.
Results
Study selection
Fifty-four studies were included in the final analysis, flow diagram, (Figure 2) [24-77]. Each study involved a VPS intervention involving medical students and/or residents/fellows that was based on case scenario(s) and included sufficient pre/post or study group/control data to calculate an ES.
Figure 2. Flowchart. Flowchart of record search strategy and selection of articles for full review leading to final 54 studies included in the meta-analysis based on inclusion criteria: 1) Population-include medical students and/or residents/fellow; 2) Intervention-casebased, screen-based VPS; 3) Comparison-RCT, pre-post, or experimental-control study design; and 4) Outcomeat least one learning outcome measured in two groups or two instances in time with sufficient data [i.e., mean (M), standard deviation (SD), number of participants (n)] to allow calculation of an effect size (ES). [Add note about Overlapping Publications.]
Study characteristics
Study characteristics are summarized in Table 1. Twentythree studies (23/54, 43%) employed randomization; 35 studies (35/54, 65%) had a control group; and 33 studies (33/54, 61%) included both pre and post data. The control group was another VPS group (13/35, 37%); didactic or traditional curriculum not involving simulation (10/35, 29%), standardized patient (7/35, 20%), high-fidelity simulation (2/35, 6%), and “other” (3/35, 8%). The studies included n=4,827 participants with M=89.4 participants per study; 37 studies (37/54, 68%) primarily involved medical students and 14 studies (14/54, 26%) residents/fellows. Three (3/54, 6%) studies included medical students or residents/ fellows combined or as part of an interprofessional team. Specialty designations included surgery (16/54, 29%), medicine (13/54, 24%), family medicine (13/54, 24%), psychiatry (6/54, 11%), pediatrics (3/54, 6%), obstetrics gynecology (2/54, 4%), and anesthesia (1/54, 2%).
Study (Year) | Study Design |
n | Learner Level |
Specialty | Control | VPS Features |
No. of Scenarios | GME Milestone |
MERSQI Score |
---|---|---|---|---|---|---|---|---|---|
Chon (2019) | PP | 140 | MS | SURG | NO | INT, FB, PRAC | 4 | MK-1 | 14.0 |
Sezer (2019) | R, PP, C | 88 | MS | FM | SP | INT, FB, PRAC | 1 | ICS-1 | 14.5 |
Fleiszer (2018) | C | 90 | MS | SURG | VPS | INT, FB | 1 | PC-3 | 14.5 |
Taekman (2017) | PP | 48 | OTHER | OB GYN | NO | INT, FB | 1 | SBP-1 | 7.5 |
Dankbaar (2017) | PP, C | 103 | MS | MED | DIDACT | INT, FB, PRAC | 1 | SBP-2 | 14.5 |
Tolsgaard (2016) | R, C | 45 | MS | MED | VPS | INT, PRAC | 4 | PC-1 | 15.0 |
McKendy (2016) | PP | 29 | RES | SURG | NO | FB, PRAC | 18 | MK-1 | 11.0 |
Sullivan (2016) | PP | 98 | MS | SURG | NO | INT, FB, PRAC | 2 | MK-1 | 13.0 |
Paull (2016) | C | 212 | RES | MED | HFS | INT, FB | 1 | SBP-2 | 11.0 |
Foster (2016) | R, C | 70 | MS | PSYCH | VPS | FB | 3 | PROF-1 | 16.0 |
Edelbring (2016) | C | 190 | MS | MED | VPS | INT, FB, PRAC | 1 | MK-1 | 12.0 |
Elledge (2016) | PP | 29 | MS | SURG | NO | INT, FB, PRAC | 5 | PC-2 | 10.5 |
Kleinert (a) (2015) | PP | 25 | MS | SURG | NO | INT, FB, PRAC | 4 | PC-1 | 12.0 |
Kleinert (b) (2015) | PP | 62 | MS | SURG | NO | INT, FB, PRAC | 3 | PC-2 | 9.0 |
Close (2015) | PP | 71 | RES | SURG | NO | INT, FB, PRAC | 20 | PC-2 | 12.0 |
Kleinsmith (2015) | C | 73 | MS | MED | SP | NO | 4 | PROF-1 | 12.5 |
Foster (2015) | R, PP, C | 67 | MS | PSYCH | OTHER | INT, FB | 1 | ICS-1 | 14.0 |
Johnson (2015) | PP | 52 | RES | PEDS | NO | NO | 2 | SBP-4 | 10.0 |
Woodham (2015) | C | 119 | MS | SURG | VPS | INT, FB | 1 | SBP-2 | 11.0 |
Pantziaras (a) (2015) | PP | 32 | RES | PSYCH | NO | NO | 1 | MK-1 | 12.0 |
Pantziaras (b) (2015) | PP | 32 | RES | PSYCH | NO | FB | 1 | PROF-3 | 9.5 |
Leung (2015) | PP, C | 130 | MS | ANESTH | VPS | INT, FB, PRAC | 6 | PC-2 | 13.5 |
Sperl-Hillen (2014) | R, PP, C | 341 | RES | FM | DIDACT | FB, PRAC | 18 | PC-2 | 16.0 |
Poulton (2014) | R, C | 81 | MS | MED | VPS | INT, FB | 5 | PC-2 | 12.5 |
Bediang (2013) | R, C | 20 | MS | FM | VPS | NO | 2 | PC-1 | 12.0 |
Harris (2013) | R, C | 120 | RES | FM | VPS | INT, FB, PRAC | 5 | PC-2 | 12.0 |
Funke (2013) | PP | 116 | MS | SURG | NO | INT, FB, PRAC | 6 | SBP-3 | 11.0 |
Tan (2013) | PP | 137 | MS | FM | NO | INT, PRAC | 1 | SBP-1 | 11.5 |
Courteille (2013) | R, C | 82 | RES | SURG | DIDACT | INT, FB | 1 | MK-1 | 11.5 |
Yang (2013) | PP, C | 31 | MS | SURG | OTHER | FB, PRAC | 3 | PC-3 | 11.5 |
Kononowicz (2012) | R, PP, C | 226 | MS | MED | DIDACT | INT, FB, PRAC | 6 | MK-2 | 13.0 |
Lin (2012) | C | 66 | MS | PSYCH | DIDACT | INT, FB | 1 | PC-1 | 10.0 |
Lunney (2012) | PP | 91 | RES | PEDS | NO | INT, PRAC | 10 | PROF-1 | 13.0 |
Persky (2011) | R, C | 76 | MS | FM | VPS | FB | 2 | PROF-3 | 13.0 |
Oliven (2011) | R, C | 262 | MS | FM | SP | FB, PRAC | 5 | PC-1 | 13.0 |
Botezatu (a) (2010) | R, C | 49 | MS | MED | DIDACT | INT, FB | 6 | PC-1 | 14.0 |
Botezatu (b) (2010) | R, C | 216 | MS | MED | DIDACT | INT, FB | 6 | PC-1 | 14.0 |
Andreatta (2010) | R, PP, C | 15 | RES | MED | SP | INT, FB, PRAC | 1 | SBP-4 | 15.0 |
Gucwa (2010) | PP | 16 | RES | FM | NO | NO | 1 | PROF-1 | 9.5 |
Kandasamy (2009) | R, PP, C | 62 | MS | SURG | OTHER | INT | 5 | PC-2 | 11.0 |
Deladisma (2009) | R, PP, C | 29 | MS | SURG | DIDACT | INT | 1 | PC-1 | 10.0 |
Vukanovic-Criley (2008) | PP, C | 82 | MS | MED | DIDACT | INT, FB, PRAC | 1 | PC-4 | 13.0 |
Youngblood (2008) | R, PP, C | 30 | OTHER | SURG | HFS | FB, PRAC | 6 | SBP-1 | 14.5 |
Boyd (2008) | PP | 101 | RES | OB GYN | NO | INT, FB | 1 | PROF-3 | 12.5 |
Deladisma (2007) | R, C | 84 | MS | FM | SP | NO | 1 | PROF-1 | 13.5 |
Sijistermans (2007) | PP | 134 | MS | PSYCH | NO | FB, PRAC | 10 | SBP-4 | 10.0 |
Vash (2007) | R, C | 48 | MS | SURG | DIDACT | INT, FB, PRAC | 14 | PC-1 | 12.5 |
Nendaz (2006) | C | 6 | MS | FM | SP | NO | 3 | MK-1 | 14.5 |
Stevens (2006) | C | 20 | MS | FM | VPS | NO | 1 | PROF-1 | 8.0 |
Ferguson (2006) | PP | 30 | RES | PEDS | NO | INT, FB, PRAC | 1 | PROF-1 | 12.5 |
Triola (2006) | R, PP, C | 55 | OTHER | MED | SP | INT, FB, PRAC | 4 | ICS-1 | 15.0 |
Dickerson (2006) | R, C | 17 | MS | MED | VPS | NO | 1 | PC-1 | 11.0 |
Bearman (a) (2001) | PP, C | 212 | MS | FM | VPS | INT, FB | 2 | ICS-1 | 14.0 |
Bearman (b) (2001) | R, C | 167 | MS | FM | VPS | INT, FB | 2 | ICS-1 | 10.0 |
Abbreviations: ANESTH=Anesthesia; C=Controls; DIDACT=Didactic; FB=Feedback to participants; FM=Family Medicine; GME=Graduate medical education; ICS=Interpersonal and communication skills; INT=Integrated into curriculum; MED=Medicine; MERSQI=Medical Education Research Study Quality Instrument; MK=Medical knowledge; MS=Medical students; n=Number of participants; NO=None; OB GYN=Obstetrics and Gynecology; OTHER=interprofessional participants that include residents or fellows; PC=Patient care; PEDS=Pediatrics;Â PP=Pre and post; PRAC=Practice; PROF=Professional; PSYCH=Psychiatry; R=Randomization; RES=Residents; SBP=Systems-based practice; SP=Standardized patient; SURG=Surgery; VPS=virtual patient simulation. |
The studies included n=217 virtual patient simulation scenarios with M=4.0 scenarios per study, (Table 1).
There were 90 evaluable learning outcomes in the 54 studies including learner satisfaction (22/90, 24%), knowledge (41/90, 46%), and performance (27/90, 30%), (Table 2). Eleven studies (11/54, 20%) included enough information to calculate an ES for more than one level of learning outcome, and 19 studies (19/54, 35%) recorded more than one evaluable metric within a single level of outcome. Among the 20 studies (20/54, 37%) assessing one or more measures of performance, assessment was by an embedded VPS scoring system (7/20, 35%); a standardized patient simulation (5/20, 25%); a combination of VPS and SP (5/20, 25%); or another method (3/20, 15%). In 26 studies (26/54, 48%) there was a reported time interval, M=6.2 weeks, between the intervention and the assessment. Fourteen studies (14/54, 26%) measured learning outcomes immediately after the VPS intervention. In 15 studies (15/54, 28%), the evaluation occurred after an unspecified period.
Study (Year) | Kirkpatrick Level | Assessment | VPS Study Group | M | n | SD | Comparison Group | M | n | SD | P-Value |
---|---|---|---|---|---|---|---|---|---|---|---|
Chon (2019) | K | TEST | VPS POST | 76 | 140 | 11.6 | PRE | 60.4 | 140 | 16.6 | <0.05 |
Sezer (2019) | K | TEST | VPS POST | 80.6 | 44 | 9.88 | PRE | 57.1 | 44 | 11.6 | <0.05 |
K | TEST | VPS POST | 80.6 | 44 | 9.88 | SP POST | 77.5 | 44 | 11.8 | NS | |
Fleiszer (2018) | P | VPS PROC | VPS SRS POST | 59 | 25 | 13 | VPS JRS POST | 51 | 57 | 12 | <0.05 |
P | VPS COMM | VPS SRS POST | 31 | 25 | 15 | VPS JRS POST | 28 | 57 | 14 | NS | |
Taekman (2017) | S | QUEST | VPS POST | 8.95 | 48 | 1.42 | PRE | 7.83 | 48 | 1.55 | <0.05 |
Dankbaar (2017) | K | TEST | VPS POST | 57.9 | 34 | 6.5 | DIDACT POST | 52.6 | 37 | 7.1 | <0.05 |
K | TEST | VPS GAME POST | 60.1 | 32 | 6.7 | DIDACT POST | 52.6 | 37 | 7.1 | <0.05 | |
Tolsgaard (2016) | K | TEST | VPS CONST POST | 61.4 | 20 | 5.2 | PRE | 58.9 | 20 | 7 | NS |
K | TEST | VPS CONST POST | 61.4 | 20 | 5.2 | VPS PR SLV POST | 62.6 | 19 | 5.7 | NS | |
P | SP | VPS CONST POST | 59.1 | 20 | 12.8 | VPS PR SLV POST | 60.8 | 19 | 11.5 | NS | |
McKendy (2016) | K | TEST | VPS POST | 55.4 | 26 | 6.6 | PRE | 59.6 | 29 | 8.1 | <0.05 |
Sullivan (2016) | P | VPS DIVERT | VPS POST | 67.9 | 76 | 29 | PRE | 35.1 | 98 | 34.8 | <0.05 |
P | VPS GI BLEED | VPS POST | 82.1 | 51 | 19.8 | PRE | 41.8 | 78 | 40.9 | <0.05 | |
Paull (2016) | S | QUEST | VPS POST | 4.6 | 108 | 0.8 | HFS POST | 4.6 | 104 | 0.7 | NS |
Foster (2016) | P | SP | VPS EMP FB POST | 2.91 | 35 | 0.16 | VPS NO FB POST | 2.27 | 17 | 0.21 | <0.05 |
Edelbring (2016) | S | QUEST | VPS STUD POST | 4.18 | 58 | 0.53 | VPS TEACH POST | 3.6 | 27 | 0.51 | <0.05 |
Elledge (2016) | S | VAS | VPS POST | 45.7 | 29 | 16.6 | PRE | 29.2 | 29 | 19.2 | <0.05 |
K | TEST | VPS POST | 13 | 29 | 3.56 | PRE | 10 | 29 | 2.52 | <0.05 | |
Kleinert (a) (2015) | K | TEST | VPS POST | 8.88 | 25 | 0.9 | PRE | 7.24 | 25 | 0.9 | <0.05 |
Kleinert (b) (2015) | K | TEST | VPS POST | 7 | 62 | 1 | PRE | 5 | 62 | 1 | <0.05 |
Close (2015) | P | VPS | VPS POST | 68 | 71 | 40 | PRE | 22 | 71 | 40 | <0.05 |
Kleinsmith (2015) | P | VPS + SP | VPS POST | 18.7 | 73 | 12.7 | SP POST | 14.5 | 73 | 9.1 | <0.05 |
Foster (2015) | P | SP | VPS POST | 22.4 | 34 | 3.7 | VIDEO POST | 21.6 | 33 | 4.2 | NS |
Johnson (2016) | S | QUEST | VPS POST | 3.72 | 219 | 0.76 | PRE | 3.64 | 221 | 0.84 | NS |
Woodham (2015) | S | QUEST | VPS VIDEO POST | 3.82 | 116 | 0.9 | VPS TEXT POST | 3.87 | 119 | 0.75 | NS |
Pantzarias (a) (2015) | K | TEST IMMED | VPS POST | 8.47 | 32 | 1.65 | PRE | 7.44 | 32 | 0.32 | <0.05 |
K | TEST 8 WKS | VPS POST | 8.38 | 26 | 2.02 | PRE | 7.44 | 32 | 0.32 | <0.05 | |
Pantzarias (b) (2015) | S | QUEST | VPS POST | 4.2 | 32 | 0.8 | PRE | 3.86 | 32 | 0.73 | <0.05 |
Leung (2015) | S | QUEST | VPS BRANCH POST | 5.19 | 130 | 2.1 | VPS LINEAR POST | 4.5 | 130 | 2.9 | <0.05 |
K | TEST MC | VPS BRANCH POST | 85 | 32 | 11.5 | PRE | 66 | 32 | 8.6 | <0.05 | |
K | TEST MC | VPS BRANCH POST | 85 | 32 | 11.5 | VPS LINEAR POST | 69 | 32 | 18.7 | <0.05 | |
K | TEST ESSAY | VPS BRANCH POST | 64 | 32 | 17.2 | PRE | 43 | 32 | 12.9 | <0.05 | |
K | TEST ESSAY | VPS BRANCH POST | 64 | 32 | 17.2 | VPS LINEAR POST | 49 | 32 | 17.2 | <0.05 | |
K | TEST-EOY | VPS BRANCH POST | 54 | 32 | 12.9 | PRE | 48 | 32 | 10 | <0.05 | |
K | TEST-EOY | VPS BRANCH POST | 54 | 32 | 12.9 | VPS LINEAR POST | 51 | 32 | 12.9 | NS | |
Sperl-Hillen (2014) | K | TEST | VPS POST | 5.3 | 92 | 1.8 | DIDACT POST | 4.1 | 128 | 1.6 | <0.05 |
Poulton (2014) | K | TEST | VPS BRANCH POST | 8.26 | 37 | 1.31 | VPS LINEAR POST | 6.94 | 43 | 1.62 | <0.05 |
Bediang (2013) | P | SP ARF | VPS ARF POST | 73.9 | 20 | 4.4 | VPS POST CONTROL | 63.8 | 20 | 11.7 | <0.05 |
P | SP CSH | VPS CSH POST | 66.1 | 20 | 6.2 | VPS POST CONTROL | 60.2 | 20 | 7.9 | <0.05 | |
Harris (2013) | S | QUEST | VPS POST | 211 | 32 | 28.1 | DIDACT POST | 211 | 50 | 28.4 | NS |
S | QUEST | VPS FAC POST | 222 | 30 | 21.6 | VPS RES POST | 204 | 90 | 28.4 | <0.05 | |
Funke (2013) | P | VPS | VPS CASE 6 | 62.3 | 116 | 5.6 | VPS CASE 2 | 53.9 | 116 | 5.6 | <0.05 |
Tan (2013) | S | QUEST | VPS POST | 3.14 | 127 | 0.76 | PRE | 2.17 | 127 | 0.81 | <0.05 |
K | TEST | VPS POST | 10 | 127 | 2.39 | PRE | 7.69 | 127 | 2.27 | <0.05 | |
Courteille (2013) | K | TEST RES | VPS POST | 10 | 20 | 1.1 | DIDACT POST | 9.9 | 21 | 1.1 | NS |
K | TEST MS | VPS POST | 9 | 18 | 1.3 | DIDACT POST | 9.4 | 23 | 1.4 | NS | |
Yang (2013) | K | TEST NBME | VPS POST | 86.5 | 33 | 7.4 | PRE | 83.5 | 36 | 9 | NS |
P | VPS | VPS POST | 70.8 | 31 | 25.9 | PRE | 56.7 | 27 | 34.8 | NS | |
Kononowicz (2012) | K | TEST | VPS POST | 48.3 | 47 | 3.2 | PRE | 36.9 | 47 | 3.4 | <0.05 |
K | TEST | VPS POST | 48.3 | 47 | 3.2 | DIDACT POST | 45.8 | 75 | 3.8 | <0.05 | |
Lin (2012) | P | CLIN EVAL | VPS POST | 88.2 | 32 | 3.1 | DIDACT POST | 85.2 | 34 | 3.9 | <0.05 |
Lunney (2012) | S | QUEST | VPS POST | 10.3 | 91 | 2.78 | PRE | 7.7 | 91 | 2.2 | <0.05 |
K | TEST | VPS POST | 12 | 91 | 1.55 | PRE | 8.19 | 91 | 1.79 | <0.05 | |
Persky (2011) | S | QUEST | VPS OBESE POST | 3.85 | 37 | 0.86 | VPS NON-OB POST | 2.6 | 39 | 0.72 | <0.05 |
Oliven (2011) | P | VPS + SP | VPS POST | 79.3 | 262 | 8.9 | SP POST | 82.3 | 262 | 7.9 | <0.05 |
Botezatu (a) (2010) | K | TEST HEMAT | VPS POST | 6.2 | 25 | 1.9 | PRE | 4.3 | 24 | 1.7 | <0.05 |
K | TEST CARDIOL | VPS POST | 7.9 | 25 | 1.2 | PRE | 6.1 | 24 | 1.7 | <0.05 | |
P | VPS HEMAT | VPS POST | 8 | 25 | 0.8 | PRE | 6.5 | 24 | 0.4 | <0.05 | |
P | VPS CARDIOL | VPS POST | 8.8 | 25 | 0.9 | PRE | 7.6 | 24 | 0.5 | <0.05 | |
Botezatu (b) (2010) | K | TEST HEMAT | VPS POST | 4.27 | 25 | 0.81 | PRE | 2.93 | 24 | 1.07 | <0.05 |
K | TEST CARDIOL | VPS POST | 4.6 | 25 | 0.69 | PRE | 3.7 | 24 | 0.9 | <0.05 | |
P | VPS HEMAT | VPS POST | 4.82 | 25 | 0.76 | PRE | 3.61 | 24 | 0.43 | <0.05 | |
P | VPS CARDIOL | VPS POST | 5.18 | 25 | 0.59 | PRE | 4.11 | 24 | 0.41 | <0.05 | |
Andreatta (2010) | K | TEST | VR POST | 16.7 | 7 | 3.04 | PRE | 17.1 | 7 | 3.63 | NS |
K | TEST | VR POST | 16.7 | 7 | 3.04 | SP POST | 18.5 | 8 | 2.62 | NS | |
P | VPS + SP | VR POST | 3.55 | 7 | 1.7 | SP POST | 3.47 | 8 | 0.41 | NS | |
Gucwa (2010) | S | QUEST | VPS POST | 3.88 | 16 | 0.44 | PRE | 3.71 | 16 | 0.72 | NS |
Kandasamy (2009) | K | TEST | VPS POST | 84.6 | 28 | 12.6 | PRE | 59.1 | 28 | 21.4 | <0.05 |
K | TEST | VPS POST | 84.6 | 28 | 12.6 | DIDACT POST | 74.3 | 27 | 15.3 | <0.05 | |
Deladisma (2009) | S | QUEST | VPS POST | 4.27 | 15 | 0.47 | DIDACT POST | 3.5 | 14 | 0.71 | <0.05 |
Vukanovic-Criley (2008) | K | TEST | VPS POST | 73.5 | 24 | 8.4 | PRE | 58.7 | 24 | 14 | <0.05 |
K | TEST | VPS POST | 73.5 | 24 | 8.4 | DIDACT POST | 59.5 | 42 | 15.4 | <0.05 | |
Youngblood (2008) | P | VPS + HFSÂ | VPS POST | 43.1 | 16 | 3.94 | PRE | 23.8 | 16 | 4.47 | <0.05 |
P | VPS + HFSÂ | VPS POST | 43.1 | 16 | 3.94 | HFS POST | 44.5 | 14 | 5.17 | NS | |
Boyd (2008) | S | QUEST * | VPS POST | 20.2 | 99 | 5.5 | PRE | 23.9 | 99 | 4.6 | <0.05 |
K | TEST | VPS POST | 12.4 | 99 | 2.4 | PRE | 10.4 | 99 | 2.3 | <0.05 | |
Deladisma (2007) | P | VPS + SP | VPS POST | 4.29 | 33 | 1.32 | SP POST | 3.24 | 51 | 1.06 | <0.05 |
Sijistermans (2007) | S | QUEST | VPS POST | 3.91 | 134 | 0.28 | PRE | 3.56 | 134 | 0.34 | <0.05 |
Vash (2007) | K | TEST | VPS POST | 18 | 23 | 2.9 | DIDACT POST | 13 | 22 | 3 | <0.05 |
Nendaz (2006) | P | VPS + SP | VPS POST | 61 | 6 | 24 | SP POST | 72 | 6 | 23 | NS |
Stevens (2006) | S | QUEST | VPS VERS 2 | 7.4 | 13 | 0.3 | VPS VERS 1 | 6.4 | 7 | 0.2 | <0.05 |
Ferguson (2006) | S | QUEST * | VPS POST | 37.5 | 29 | 9.17 | PRE | 51.9 | 29 | 10.8 | <0.05 |
K | TEST | VPS POST | 12.4 | 30 | 1.58 | PRE | 9.47 | 30 | 1.55 | <0.05 | |
Triola (2006) | S | QUEST | VPS POST | 2.97 | 23 | 0.6 | PRE | 2.14 | 23 | 0.64 | <0.05 |
S | QUEST | VPS POST | 2.97 | 23 | 0.6 | SP POST | 3.19 | 32 | 0.65 | NS | |
Dickerson (2006) | P | EXPERT EVAL | VPS SYN VOICE | 4.37 | 8 | 1.59 | VPS REC VOICE | 5 | 9 | 1.85 | NS |
Bearman (a) (2001) | P | SP | VPS NAR POST | 38.8 | 41 | 4.8 | PRE | 36.3 | 55 | 5.5 | <0.05 |
P | SP | VPS NAR POST | 38.8 | 41 | 4.8 | VPS PR SLV POST | 35.7 | 38 | 5.3 | <0.05 | |
Bearman (b) (2001) | S | QUEST | VPS NAR POST | 33.7 | 85 | 5.2 | VPS PR SLV POST | 32.7 | 82 | 5.2 | NS |
Abbreviations:Â | |||||||||||
ARF=Acute renal failure; BRANCH=Branching design; CARDIOL=Cardiology; CASE 2=Second case of a 6-case curriculum; CASE 6=Sixth case of a 6-case curriculum; CLIN EVAL=Clinical evaluation by instructor; COMM=Communication skills; CONST=Participants construct VPS scenario; CONTROL=Control group; CSH=Chronic subdural hematoma; DIDACT=Didactic or traditional curriculum; DIVERT=Diverticulitis case; EMP FB=Empathetic feedback; EXPERT EVAL=Expert Evaluation of videotapes; FAC=Faculty; GAME=Serious gaming; GI BLEED=Gastrointestinal bleeding case; HEMAT=Hematology; HFS=High-fidelity simulation; JRS=More junior participants; K=Knowledge; LIM=Limited time to complete virtual patient simulation; LINEAR=Linear design; M=Mean; n=Number of participants; NAR=Narrative design; NO FB=No feedback to participant; NON-OB=Non-obese virtual patient; NS=not statistically significant; OBESE=Obese virtual patient; P=Performance; POST=Post intervention whether VPS or control group;Â PR SLV=Problem-solving design; PRE=Pre intervention; PROC=Procedural knowledge or skills; QUEST=Question or survey; RES=Resident; S=Satisfaction or self-reported impact; SD=Standard deviation; SP=Standardized patient; SRS=More senior participants; STUD=Student regulated curriculum; TEACH=Teacher regulated curriculum; TEST=Knowledge-based test/examination; TEST 8 WKS=Test 8 weeks after intervention; TEST-EOY=End of year test; TEST ESSAY=Open-ended questions, essay style test; TEST IMMED=Test immediately following intervention; TEST MC=Multiple choice test; TEST MS=Test for medical students; TEST NBME=National Board of Medical Examiners scores; TEXT=Text replacing video in VPS; UNLIM=Unlimited time to complete virtual patient simulation; VAS=Visual analog scale; VERS 1=First version of VPS curriculum; VERS 2=Second version of VPS curriculum after improvements; VIDEO=VPS using videos in place of text; VPS=Virtual patient simulation; VPS REC VOICE=Virtual patient with a recorded voice;Â VPS SYN VOICE=Virtual patient with a recorded voice; VR=Virtual reality. | |||||||||||
*Lower scores are better | |||||||||||
GME core competencies addressed were: patient care (PC) (21/54, 39%), systems-based practice (SBP) (10/54, 18.5%), medical knowledge (MK) (8/54, 15%), professionalism (PROF) (10/54, 18.5%), and interpersonal and communication skills (ICS) (5/53, 9%), (Table 1). The most frequent milestones for PC, SBP, MK, PROF, and ICS, respectively, were: PC-1 - “Gathers and synthesizes essential and accurate information to define each patient’s clinical problem(s)” (10/21, 48%) and PC-2 - “Develops and achieves comprehensive management plan for each patient” (8/21, 38%); SBP- 1 - “Works effectively within an interprofessional team” (3/10, 30%), SBP-2 - “Recognizes system error and advocates for system improvement” (3/10, 30%), and SBP-4 - “Transitions patients effectively within and across health delivery systems” (3/10, 30%); MK-1 - “Clinical knowledge” (7/8, 88%); PROF-1 - “Has professional and respectful interactions with patients, caregivers and members of the interprofessional team” (7/10, 70%); and ICS-1 - “Communicates effectively with patients and caregivers” (5/5, 100%).
Study quality and risk of bias within studies
The MERSQI score (maximum score 18 points) for the 53 studies was 12.3 ± 2.0 with a median of 12.0. Possible bias within each of the 54 included studies is shown in Table 3. An aggregated assessment of bias across all included studies is shown in Figure 3. Challenging areas included selection bias (e.g., randomization) and performance bias (e.g., blinding of participants). On the other hand, the majority of included studies was rated as having a small risk of detection bias (e.g., blinding observers/raters) and attrition bias (e.g., high response rates). In 13% (35/270 evaluations) of instances, the level of possible bias could not be determined from the information included in the article.