Home Article
Medical students’ experiences and perspectives on simulation-based education
Medical students’ experiences and perspectives on simulation-based education

Article Type: Original Research Article History
Abstract

Purpose

We aimed to gain insight into student experiences of simulation-based education (SBE), particularly in regards to (1) current provision of SBE, (2) learning opportunities and (3) areas for improvements, innovation and focused initiatives. This was to create focussed initiatives that can assess and address specific needs to improve SBE for learners and educators alike.

Methods

Medical students were surveyed with 41 questions on their experiences of SBE using an 11-point Likert scale (0 – strongly disagree, 10 – strongly agree). Results were analysed by individual questions and presented as median (interquartile range) or percentage (n/N respondents). Data were analysed using Mann–Whitney U or Kruskal–Wallis tests (p < 0.05 for significance). Likert scales were analysed for internal consistency using Cronbach’s alpha.

Results

246 students participated, with 76.0% (187/246) completing all questions. 99.2% of students (235/237) had participated in SBE. The most valuable elements of SBE were learning a new skill under supervision (90.3%, 187/207), applying prior knowledge to a clinical scenario (73.4%, 152/187) and identifying gaps in knowledge/skill (73.4%, 152/187). Simulation was thought to improve medical knowledge (95.2%, 218/229) and technical skills (87.3%, 200/229). Twenty-one per cent (41/197) of students reported a negative experience and 23.5% (48/204) felt anxious. Students strongly agreed that simulation was beneficial to their training (9 [8–10]) and that there should be more SBE (8.5 [8–10]).

Conclusions

Medical students find SBE accessible and valuable to their education. By analysing student perspectives (such as self-reported negative experience), targeted areas for further research and focussed initiatives can be implemented.

Ensor, Sivasubramaniam, Laird, Roddis, Qin, Pacilli, Nestel, and Nataraja: Medical students’ experiences and perspectives on simulation-based education

Practice Points

  • 99.2% of the surveyed students had participated in simulation-based education (SBE) and 97% reported it to be a useful component of medical education.
  • The most valuable elements of SBE reported by students were learning a new skill under supervision, applying prior knowledge to a clinical scenario and identifying gaps in knowledge/skill.
  • One-fifth of students reported a negative experience during SBE; 24% felt anxious, 12% felt vulnerable and 1.5% felt unsafe.
  • Three-quarters of students felt that pre-briefing was beneficial for SBE and all participants were in favour of regular debriefing.
  • Ninety per cent of students think there should be more SBE in the medical curriculum.

Introduction

Simulation-based education (SBE) is a well-accepted component of medical education. SBE promotes skill acquisition and competence by practice in a safe learning environment [1]. This environment allows for provision of expert feedback which, combined with the principles of deliberate practice and mastery learning, facilitates patient safety by obtaining competency prior to patient contact [15]. SBE facilitates clinical experience in an era of decreasing exposure to meaningful clinical encounters [6,7].

Simulation encompasses a broad range of educational approaches and techniques. It may involve any activity that replaces or amplifies real experiences with guided experiences to evoke or replicate aspects of the real world in a fully interactive fashion [8]. SBE utilizes various modalities depending on the nature of the task, clinical scenario or objectives. Examples of these include role-play, simulated patients, part-task trainers (synthetic, animal or cadaveric), simulated scenarios, virtual reality simulation, laparoscopic bench trainers and combinations of these examples [1,9]. This allows application across a breadth of healthcare professions and levels of experience [1012].

Despite the ubiquity of SBE in medical education and assessment, the broader attitudes of learners towards SBE are not commonly assessed formally or systematically at an institutional level (versus for individual activities). Learner feedback is not routinely being sought in a field of education that is inherently learner centred. Whilst analysis of specific interventions on learning outcomes is of benefit, we propose that analysis of students’ attitudes towards SBE is also of benefit by means of identifying areas of improvement to enhance students’ experience and engagement in their education.

Current literature largely focuses on particular simulated tasks or scenarios, analysing students’ performance and/or feedback for a specific simulation [1119]. On the contrary, there is little research that assesses students’ attitudes towards SBE in its entirety. Joseph et al. [20] and Agha et al. [21] have published cross-sectional studies describing Indian and Saudi Arabian medical students’ (respectively) attitudes towards SBE. To our knowledge, there is no such study in the Australian context.

McGaghie et al. [22], in describing 12 components of SBE best practice, outlined that SBE should feature feedback, outcome measurement, instructor training and educational/professional context. Our study directly and indirectly assesses the application of these principles.

We aimed to gain insight into student experiences of SBE, particularly in regards to (1) current provision of SBE, (2) learning opportunities and (3) areas for improvements, innovation and focused initiatives. Consequently, we can create a platform by which focussed initiatives can assess and address specific needs to improve SBE for learners and educators alike.

Methods

We aimed to gain insight into students’ exposure to and attitudes towards SBE. Particular areas of interest were:

  • Gaining insight into the nature of the current provision of SBE.
  • Identifying student opportunities for learning through SBE.
  • Identifying student attitudes towards SBE (including what students value).
  • Identifying specific areas for improvement.
  • Identifying specific areas for innovation and focused initiatives.
  • Screening for negative emotional reactions to SBE.

Participants

Medical students from Monash University (Australia) were recruited between July 2018 and April 2020. Monash University’s medical school has approximately 2,000 students, offering undergraduate (Years 1–5) and graduate-entry (Years A–D) degrees.

Recruitment

We aimed to recruit 250 students: approximately 10% of the medical school cohort accounting for attrition. Convenience sampling was used to recruit participants on an opt-in basis, commonly after lectures and tutorials (with verbal reminders by investigators). Participants were presented with an explanatory statement at the beginning of the survey and the investigators’ contact details were provided if participants later chose to opt out. Recruiting investigators were not involved in medical school teaching to avoid perceived coercion. Inclusion criteria were any medical student at our institution, regardless of year level. There were no exclusion criteria to obtain a broad sample across the student cohort. To avoid selection bias and to obtain an accurate representation of the students’ population, domestic and international students were recruited from pre-clinical and clinical year levels, undergraduate and graduate-entry streams, and metropolitan and rural clinical schools. There is great diversity in medical student backgrounds and experiences, meaning that it was of importance to make sure all (as best as achievable) were represented. Regarding rurality, while there is a standardised curriculum, the way that this is taught can be variable according to the local clinical facilities and expertise.

Survey design

Current literature was reviewed in order to identify studies investigating students’ perspectives of SBE. Key themes were identified and questions were developed by a group of both educators and students in order to ensure survey relevancy to our cohort and effectiveness in achieving our aims. Relevant questions were adapted for our cohort from Agha et al. [21], Joseph et al. [20] and Dudas et al. [17]. Draft surveys were workshopped and trialled with student participants for feedback in regards to survey flow prior to formal distribution.

A scoping survey of 37 questions was provided to participants. Surveys were completed on paper or using QualtricsXM software (Qualtrics, London, United Kingdom) on a tablet device (iPad – Apple, Cupertino, CA).

Questions presented to participants were divided into the following themes:

  1. Demographic information.
  2. Familiarity with SBE and previous/current experience of SBE.
  3. Attitudes towards the use of SBE.
  4. Preferences regarding the utilisation of SBE in a medical school curriculum.

After recording demographic data and self-reported familiarity with SBE, participants were provided with a brief definition of SBE – ‘An educational technique that replaces or amplifies real experiences with guided experiences that evoke or replicate substantial aspects of the real world in a fully interactive manner’ [8] – as well as a list of examples from Nataraja et al. [9].

Survey questions (Appendix 1) were presented as multiple-option tick boxes and 5- or 11-point Likert scales. Eleven-point Likert scales assessed student agreement with a statement (0 – strongly disagree, 10 – strongly agree) and, for user simplicity, 5-point Likert scales assessed frequency of student experiences (0 – never, 4 – always). SBE-associated negative experiences were also assessed using a 5-point Likert scale which were categorized into ‘Yes’ (0–1), ‘Sometimes’ (2) and ‘No’ (3–4). Within these questions, anxiety was differentiated from negative feelings as some students may feel anxious with a positive effect (e.g. better performance), while others may experience negative feeling separate from anxiety (e.g. embarrassment). Multiple option questions assessed student SBE exposure or student preference in relation to the aspect of SBE being assessed. Some questions in the survey were adapted from Agha et al. [21], Joseph et al. [20] and Dudas et al. [17] in order to make them more suitable for our cohort.

Data analysis

Data were recorded in Microsoft® Excel® (Microsoft® Excel® for Mac, Version 16.42, USA; 2020) and analysed using GraphPad Prism® (PRISM 8 for macOS, Version 8.4.3; Graphpad Software Inc., USA; 2020) and IBM® SPSS® (Version 26.0, USA; 2020) software. Data were analysed and presented according to response rates for each question (represented as n/N). This was done to demonstrate how representative each question is of the total cohort. Continuous and ordinal categorical data were assessed for normality using a Shapiro–Wilk test. Data were assessed for normality. Parametric data were presented as mean ± standard deviation, and non-parametric data were presented as median and interquartile range (IQR). Parametric data were analysed for significance (p < 0.05) using a student’s t-test. Non-parametric data were analysed using the Mann–Whitney U -test or Kruskal–Wallis test. Cronbach’s alpha was used to assess internal consistency. Free text comments were analysed using qualitative content analysis [23].

Ethics committee

Ethics approval was granted from the Monash University Human Research Ethics Committee (HREC); Project ID 17383.

Results

Participant demographics

Two hundred and forty-six medical students participated, with 76.0% (187/246) completing all questions. Participants were predominantly female (66.8%, 163/244) and median age 22 [2123] years. Demographic characteristics are presented in Table 1.

Table 1:
Participant demographic characteristics
% (n/N *)
Gender
 Female 66.8% (163/244)
 Male 33.2% (81/244)
Year level
 Year 1 3.3% (8/245)
 Year 2A 5.7% (14/245)
 Year 3B 10.6% (26/245)
 Year 4C 59.6% (146/245)
 Year 5D 20.8% (51/245)
Undergraduate or graduate
 Undergraduate 82.9% (204/246)
 Graduate 17.1% (42/246)
Metropolitan or rural
 Metropolitan 76.6% (180/235)
 Rural 23.4% (55/235)
Domestic or international
 Domestic 88.5% (201/227)
 International 11.5% (26/227)

*N denotes the number of participants who responded to each question.

SBE previous experience

All but two students* reported previous participation in SBE (99.2%, 235/237), occurring on average every 4 [1–8] weeks. Both students who had not participated in SBE were in the preclinical phase of their medical education when there is a greater focus on the basic sciences. Most students (86.4% 203/235) had experienced SBE in an extracurricular environment, largely through university clubs (54.9%, 129/235), self-practice (48.9%, 115/235) and external skills-based courses (35.7%, 84/235). 43.2% (102/236) of students reported having paid for a simulation activity, with a median bracket cost of AUD$20–39 [$0–19 to $40–59].

Students participated primarily in simulated patients/role-play (97.9%, 230/235) or scenario-based simulations (92.3%, 217/235). Students also reported these methods as the most beneficial (Table 2). Synthetic part task trainers were beneficial (58.3%, 77/132), though few (6.1%, 8/132) students reported them as the most beneficial. There was perceived limited utility of augmented/virtual reality trainers (25.5%, 60/235) and online simulation (20.9%, 49/235) in comparison to more tactile forms of SBE.

Table 2:
Student exposure to various forms of SBE and debriefing
Modality of SBE Participant exposure, % (n/N *) SBE utility (rated first three preferences), % (n/N *) SBE utility (rated first preference), % (n/N *)
Simulated patient/role-play 97.9% (230/235) 80.3% (106/132) 39.4% (52/132)
Scenario-based simulation 92.3% (217/235) 87.1% (115/132) 40.2% (53/132)
Part task trainer – synthetic 83.8% (197/235) 58.3% (77/132) 6.1% (8/132)
Part task trainer – animal 59.2% (139/235) 31.8% (42/132) 4.6% (6/132)
Part task trainer – cadaveric 49.4% (116/235) 20.5% (27/132) 4.6% (6/132)
Augmented/virtual reality trainer 25.5% (60/235) 4.6% (6/132) 3.0% (4/132)
Online simulation 20.9% (49/235) 1.5% (2/132) 0.8% (1/132)
Laparoscopic bench trainer 43.4% (102/235) 12.9% (17/132) 1.5% (2/132)
No debriefing 8.9% (19/214) 0% (0/102) 0% (0/102)
Video-assisted debriefing 6.5% (14/214) 2.0% (2/102) 1.0% (1/102)
Peer feedback 67.8% (145/214) 77.5% (79/102) 7.8% (8/102)
Instructor feedback 89.3% (191/214) 98.0% (100/102) 82.4% (84/102)
Self-appraisal (to group) 44.9% (96/214) 29.4% (30/102) 2.0% (2/102)
Self-appraisal (individual) 36.4% (78/214) 13.7% (14/102) 0.0% (0/102)
Written feedback (including mark/grade) 23.4% (50/214) 23.5% (24/102) 1.0% (1/102)
Informal opportunistic feedback 37.9% (81/214) 30.4% (31/102) 5.9% (6/102)

Notes: Students were instructed to select all options that applied and then select the form(s) of SBE/debriefing which they found to be (A) beneficial and (B) single most beneficial.

*N denotes the number of participants who responded to each question.

Outcomes/benefits

The most common skills acquired from SBE were medical knowledge (95.2%, 218/229), technical skills, for example, suturing (87.3%, 200/229), and clinical judgement and skills by repeated practice (85.6%, 196/229). Students generally did not report that SBE contributed to leadership capability (37.6%, 86/229) or navigation of the hospital system (e.g. making referrals) (25.3%, 58/229).

Briefing/debriefing

Prior to SBE, students reported being briefed ‘most of the time’ (32.9%, 74/225), ‘about half of the time’ (31.1%, 70/225) or ‘rarely’ (24.9%, 56/225). This usually occurred as a verbal briefing before the simulation (88.9%, 200/225), pre-reading (61.3%, 138/225) or videos (34.2%, 7/225). Active preparations such as worksheets (18.2%, 41/225) and presentations/tasks (18.2%, 41/225) were less common. Three-quarters of students reported that pre-reading is beneficial in their preparation for SBE activities (76.9%, 160/208). Following SBE activities, a wide range of debriefing techniques were utilized (Table 2). However, 82.4% (84/102) of students reported instructor feedback as the most beneficial form of debriefing. Other favoured forms of debriefing were peer feedback (77.5%, 79/102) and self-appraisal (29.4%, 30/102). Even though some students reported debriefing to not occur (8.9%, 19/214), no students favoured SBE without debriefing.

Negative experiences in SBE

One-fifth (20.8%, 41/197) of students reported a previous negative experience during SBE. Almost a quarter (23.5%, 48/204) of students reported feeling anxious during SBE, while fewer reported feeling vulnerable (11.8%, 24/204) or unsafe (1.5%, 3/204). It was very rare to feel criticized by instructors (0.5%, 1/204) or peers (0.5%, 1/196), and the majority of students always (77.5%, 158/204) or sometimes (15.7%, 32/304) looked forward to SBE.

Student attitudes

Students rated their degree of agreement with 15 statements regarding the utility of SBE. A Cronbach’s alpha value of 0.89 demonstrated a high level of internal consistency among students’ responses.

Students were very much in favour of SBE, with 90.1% (191/210) of participants stating that SBE benefits their learning. When asked to rate aspects of SBE out of 10 (0 – not at all useful, 10 – very useful), SBE was viewed to benefit students’ learning (9 [8–10]), make their learning more engaging (9 [8–10]), and aide the transition from knowledge to application (9 [8–10]). While still in favour, students placed less value on aspects of SBE that involved interpersonal communication (score 7 [6–8]) and the development of leadership skills (score 7 [6–9]).

Students reported the most beneficial aspects of SBE to be learning a new skill under supervision (90.3%, 187/207), applying prior knowledge to a clinical scenario (73.4%, 152/187) and identifying gaps in knowledge/skill (73.4%, 152/187).

Qualitative analysis revealed areas of student concern regarding SBE, including ‘if the groups are too big it can waste time and become disengaging’, ‘self-appraisal with no guidance/lack of personalised feedback’, ‘when the scenario is not realistic or relevant’ and ‘anxiety and feelings of judgement’. Conversely, ‘working as a team in emergency scenarios’, ‘getting technique revised and critiqued by professionals’, and ‘practising while being able to make mistakes’ were all highly valued aspects of the SBE curriculum.

Students’ demographic characteristics and experiences during their medical degree did not affect the perceived utility of SBE (Table 2).

Overall, the majority of students (97.1%, 202/208) felt that SBE is a useful component of their education (9 [8–10]) and that it should be incorporated to a greater extent (8.5 [8–10]). There was no demographic group that found SBE to be more beneficial than others (Table 3).

Table 3:
Comparison between student perceived utility ratings of SBE according to demographic characteristics
Please rate how useful you find simulation on a scale from 0 to 10. Median score [IQR] p-Value
Gender
 Female 9 [8–10] 0.5*
 Male 9 [8–9]
Year Level
 Year 1 8 [7.5–9] 0.08
 Year 2A 9 [7.5–10]
 Year 3B 9 [8–10]
 Year 4C 9 [8–9]
 Year 5D 9 [8–10]
Undergraduate or graduate
 Undergraduate 9 [8–10] 0.6*
 Graduate 9 [8–10]
Metropolitan or rural
 Metropolitan 9 [8–10] 0.8*
 Rural 9 [8–10]

Notes: p-Value represents comparison of rating scores within each demographic category (e.g. median score of male versus female participants).

*Mann–Whitney U-test.

Kruskal–Wallis test

Discussion

SBE is an established and effective component of the modern medical curriculum. When utilized as a supplement to patient interactions, SBE provides a safe environment by which learners can acquire new skills or achieve greater competency [9,24]. While participant feedback regarding particular SBE interventions is published, it is uncommon for cohort-wide perspectives to be sought [1118,25]. This is to the disadvantage of learners and educators, alike, as informed student feedback provides valuable information in order to implement specific initiatives and prioritize learner-centred needs.

This study aimed to gain insight into medical students’ exposure to and attitudes towards SBE as a means of identifying potential areas of improvement, and to facilitate further assessment and intervention. Two previous studies, Joseph et al. [20] and Agha et al. [21], have surveyed medical students’ perspectives on SBE, though only one recruited a similar sample size to our study. Our study is unique because it is the first to investigate broader student perspectives in a Western educational setting and also in that it assesses negative emotional reactions to SBE. Although our study was conducted at a single university, unlike the previous studies, our cohort spans multiple year levels and includes domestic and international undergraduate and graduate-entry students from metropolitan and rural clinical schools. This is also unique in comparison to previous literature.

Agha et al. [21] performed a cross-sectional survey of 115 Saudi Arabian medical students, using a 5-point Likert scale to investigate their satisfaction with SBE. They found that despite challenges such as lack of staff/facility availability, instructor knowledge, and student cooperation the majority of students were satisfied with SBE (85%, 98/115) and felt that it improved their knowledge and skills. Similarly, Joseph et al.’s [20] cross-sectional survey of 247 Indian medical students reflected a general (though to a lesser degree) favour toward SBE (72.5%, 179/247), whereby SBE was viewed as a positive means of creating a varied learning experience and addressing the perceived problem of limited student access to patient encounters.

Students’ responses to the statement ‘simulation has helped me to improve my…’. Students were asked to select all options that applied
Figure 1:

Students’ responses to the statement ‘simulation has helped me to improve my…’. Students were asked to select all options that applied

Students’ responses to how often they have had negative experiences during SBE using a 5-point Likert scale (0 – Never, 4 – Always). Results are presented as: ‘Yes’ – Likert scale 0–1; ‘Sometimes’ – Likert scale 2; ‘No’ – Likert scale 3–4
Figure 2:

Students’ responses to how often they have had negative experiences during SBE using a 5-point Likert scale (0 – Never, 4 – Always). Results are presented as: ‘Yes’ – Likert scale 0–1; ‘Sometimes’ – Likert scale 2; ‘No’ – Likert scale 3–4

Our study demonstrated a high proportion of participants who favoured SBE as a useful component of medical education (97.1%), with 90% believing that SBE directly improved their learning. One potential factor reflecting the high satisfaction rate in our survey is that the majority of participants (99.2%) had previously participated in SBE, compared to 67% of Agha et al.’s [21] cohort.

An indirect measure of students finding benefit in SBE is that 86.4% of students experienced SBE in an extra-curricular environment, 43.2% having paid to do so. This highlights the fact that students desire and actively seek out opportunities for SBE as a means of further education and advancement. It should be noted, however, that we do not advocate for students being unduly charged for SBE experiences.

Students’ responses assessing medical student perception and attitudes towards SBE (11-point Likert scale)
Figure 3:

Students’ responses assessing medical student perception and attitudes towards SBE (11-point Likert scale)

Although debriefing is common and a well-recognised component of SBE [22], (pre-)briefing is also essential for reflective learning. In our study, approximately one-third of participants respectively experienced pre-briefing ‘most of the time,’ ‘about half of the time,’ or ‘rarely.’ The reason for this is beyond the scope of our findings the reason for this and will require further investigation with qualitative analysis. We argue that pre-briefing should occur more often, as even a four-minute preparatory briefing before a simulation can statistically improve clinical performance and cognitive appraisal ratios (perceived resources versus perceived demands) for learners [26]. Briefing is also reported to improve the learning experience for even the observer, promoting them from a passive to an active participant (and thereby learner) in the simulation [27]. Notably, students perceive the importance of preparation for SBE, with three-quarters of our sample reporting that pre-reading is a beneficial component of their preparation for SBE. It is not clear from our study whether pre-reading achieves objective improvements in performance, or rather a greater sense of comfort via an improved cognitive appraisal ratio; this should be investigated further.

In a workplace culture that places value on confidence and technical ability, the role of SBE as a means to encourage learning through mistakes and reflection has the potential to create negative experiences if not implemented appropriately [28]. It is unsurprising that SBE sometimes triggers an emotional response from the learner and although this can often be beneficial, this is only to a certain point where negative emotion can reduce the capacity of one’s working memory [29]. A quarter of participants in our study (23.5%) reported feelings of anxiety associated with SBE and one-fifth (20.8%) had reported a previous negative experience during SBE. Despite this, only 1.5% ever reported having felt unsafe (albeit this number should be reduced further). It must be noted that the exact definition of ‘unsafe’ may be interpreted differently by students and educators, thus requiring further qualitative analysis to in order to better understand the intricacies of this statement. While it is clear that there is room for improvement, reassuringly over three-quarters of students reported looking forward to participating in SBE. Our findings are in keeping with qualitative analyses regarding students’ anxiety to perform in SBE [30]. In SBE environments, students may find learning enjoyable even if they experience a degree of negative emotion associated with it, described by Keskitalo as ‘hard fun’ [31]. Ng et al. [28] state that faculties are essential in creating a safe environment where mistakes can be made by learners without having to ‘constantly save face and appear certain’ – reassuringly <1% of our participants ever felt criticized by instructors or peers. While the Australian Medical Council (AMC) offers guidance regarding support services for students in need, they do not explicitly state that a student should feel safe in their learning environment [32]. Although most students in our survey reported feeling safe, perhaps the proportion of students reporting negative experiences or anxiety could be better addressed with specific initiatives if an element of safety were to be included in AMC Standards.

This single-centre study provides a broad insight regarding the current medical student experiences of SBE. Whilst we recognize that the goal of improving learner outcomes should be improved by implementation and evaluation of specific interventions, the benefit of our study is that by understanding and appealing to the experiences of the general student cohort, new interventions can be better tailored to address gaps within the current student experience and, thus, improve learner outcomes. For example, our findings suggest that while students see the benefit of SBE and wish to engage with it further, new interventions could include an element of formal (pre-)briefing given the disproportionate perception of high value versus limited participation. Given our broad sample and high rates of survey completion, and high Cronbach’s alpha values for internal consistency, we believe our results to be a valid representation of our population. However serial assessments over time and among different cohorts are needed to assess the reliability of our findings.

One key finding of our study to highlight is regarding the negative experiences of SBE. While we recognize that the majority of students enjoy SBE, no single experience will appeal to 100% of a cohort, and that student ‘enjoyment’ is not the ultimate goal of SBE, our findings suggest that there is a cohort of students who feel anxious and vulnerable while participating in SBE – this must be addressed. Given that these experiences were present in a population that reported higher levels of satisfaction than previous studies, one could infer that such negative experiences are not limited to our cohort, though being a single-institutional study, this could not be generalized. We, therefore, recommend that further analysis across multiple centres be conducted in order to better understand the root causes of these negative experiences, whether negative experiences have objective impacts on SBE learning outcomes, and whether focussed initiatives can be implemented to address this.

Our study is limited in generalizability given that it is a single centre (despite capturing varied educational programs within this). In addition, given that the survey was conducted using a convenience sampling method over a 21-month period, it is likely that students’ exposure to SBE could have changed. However, recruiting students at various stages across the academic year would likely reduce the effect of certain periods being associated with higher/lower likelihoods of SBE exposure.

Given that this study has been developed by our research team in order to answer specific questions, it has not yet been validated across a greater cohort.

Conclusions

SBE is a highly valued and beneficial aspect of medical education across many settings and all year levels. In addition, students have ample access, experience and provision of SBE. Students stated a preference for SBE to support the development of ‘technical’ skills. While the majority of students look forward to SBE and wish for it to be incorporated to a greater degree in the medical school curriculum, a proportion of students have negative experiences which should be further assessed and addressed. We encourage medical education providers to actively seek feedback regarding student experiences of SBE. By analysis of students’ feedback regarding experiences, opportunities and areas for improvement, targeted areas for further research and focussed initiatives (from policy to individual practice) can be implemented to further improve the efficacy, participation and enjoyment of learner-centred SBE.

Supplementary material

Supplementary data are available at The International Journal of Healthcare Simulation online.

Declarations

Acknowledgements

None declared.

Authors’ contributions

None declared.

Funding

None declared.

Availability of data and materials

None declared.

Ethics approval and consent to participate

None declared.

Competing interests

The authors report no conflict of interest.

References

Weller JM, Nestel D, Marshall SD, et al. Simulation in clinical teaching and learning. Medical Journal of Australia. 2012;196:594594.

Issenberg SB, Mcgaghie WC, Petrusa ER, et al. Features and uses of high-fidelity medical simulations that lead to effective learning: a BEME systematic review. Medical Teacher. 2005;27:1028.

McGaghie WC, Siddall VJ, Mazmanian PE, et al. Lessons for continuing medical education from simulation research in undergraduate and graduate medical education: effectiveness of continuing medical education: American College of Chest Physicians Evidence-Based Educational Guidelines. Chest. 2009;135:62S68S.

Kunkler K. The role of medical simulation: an overview. International Journal of Medical Robotics. 2006;2:203210.

Ryu WHA, Dharampal N, Mostafa AE, et al. Systematic review of patient-specific surgical simulation: toward advancing medical education. Journal of Surgical Education. 2017;74:10281038.

Laschinger S, Medves J, Pulling C, et al. Effectiveness of simulation on health profession students’ knowledge, skills, confidence and satisfaction. International Journal of Evidence-Based Healthcare. 2008;6:278302.

Olson LG, Hill SR, Newby DA. Barriers to student access to patients in a group of teaching hospitals. Medical Journal of Australia. 2005;183:461463.

Gaba DM. The future vision of simulation in health care. BMJ Quality & Safety. 2004;13:i2i10.

Nataraja RM, Webb N, Lopez PJ. Simulation in paediatric urology and surgery, Part 2: an overview of simulation modalities and their applications. Journal of Pediatric Urology. 2018;14:125131.

10 

Elliott S, Murrell K, Harper P, et al A comprehensive systematic review of the use of simulation in the continuing education and training of qualified medical, nursing and midwifery staff: JBI Library of Systematic Reviews. 2011;9:538587.

11 

Turatsinze S, Willson A, Sessions H, et al. Medical student satisfaction and confidence in simulation-based learning in Rwanda – pre and post-simulation survey research. African Journal of Emergency Medicine. 2020;10:8489.

12 

Zendejas BM MSc, Brydges RP, Wang ATM, et al. Patient outcomes in simulation-based medical education: a systematic review. Journal of General Internal Medicine. 2013;28:10781089.

13 

McIvor WR. Experience with medical student simulation education. Critical Care Medicine. 2004;32:S66S69.

14 

Shah A, Carter T, Kuwani T, et al. Simulation to develop tomorrow’s medical registrar. Clinical Teaching. 2013;10:4246.

15 

Brandão C, Collares C, Marin H. Student perception on high-fidelity simulation during the medical clerkship. Studies in Health Technology and Informatics. 2013;192:960.

16 

Maurice J, Wood E, Jack K. Endoscopy simulation for medical students. Clinical Teaching. 2014;11:416420.

17 

Dudas RA, Colbert-Getz JM, Balighian E, et al Evaluation of a simulation-based pediatric clinical skills curriculum for medical students. Simulation in Healthcare. 2014; 9:2132.

18 

Morrissey B, Jacob H, Harnik E, et al. Simulation in undergraduate paediatrics: a cluster-randomised trial. Clinical Teaching. 2015;13:337342.

19 

Everitt A, Seale J, Butchers C, et al Mixing pre-clinical and clinical medical students in full patient simulation: impacts and outcomes. MedEdPublish; 6. Epub ahead of print 22 November 2017. doi: 10.15694/mep.2017.000210.

20 

Joseph N, Nelliyanil M, Jindal S, et al. Perception of simulation-based learning among medical students in South India. Annals of Medical and Health Science Research. 2015;5:247252.

21 

Agha S, Alhamrani AY, Khan MA. Satisfaction of medical students with simulation based learning. Saudi Medical Journal. 2015;36:731736.

22 

McGaghie WC, Issenberg SB, Petrusa ER, et al. A critical review of simulation-based medical education research: 2003–2009. Medical Education. 2010;44:5063.

23 

Hsieh H-F, Shannon SE. Three approaches to qualitative content analysis. Qualitative Health Research. 2005;15:12771288.

24 

Nataraja RM, Webb N, Lopez P-J. Simulation in paediatric urology and surgery. Part 1: an overview of educational theory. Journal of Pediatric Urology. 2018;14: 120124.

25 

Everitt A, Seale J, Butchers C, et al Mixing pre-clinical and clinical medical students in full patient simulation: impacts and outcomes. MedEdPublish; 6. Epub ahead of print 2017. doi: 10.15694/mep.2017.000210.

26 

Evain JN, Perrot A, Vincent A, et al. Team planning discussion and clinical performance: a prospective, randomised, controlled simulation trial. Anaesthesia 2019;74:488496.

27 

O’Regan S, Molloy E, Watterson L, et al Observer roles that optimise learning in healthcare simulation education: a systematic review. Advances in Simulation. 2016;1:4.

28 

Ng SL, Kangasjarvi E, Lorello GR, et al. ‘There shouldn’t be anything wrong with not knowing’: epistemologies in simulation. Medical Education. 2019;53:10491059.

29 

Fraser K, Ma I, Teteris E, et al. Emotion, cognitive load and learning outcomes during simulation training. Medical Education. 2012;46:10551062.

30 

Paskins Z, Peile E. Final year medical students’ views on simulation-based teaching: a comparison with the Best Evidence Medical Education Systematic Review. Medical Teacher. 2010;32:569577.

31 

Keskitalo T, Ruokamo H. Students’ emotions in simulation-based medical education. Journal of Interactive Learning Research. 2017;28:149159.

32 

Australian Medical Council. Standards for Assessment and Accreditation of Primary Medical Programs by the Australian Medical Council. 2012., https://www.amc.org.au/wp-content/uploads/2019/10/Standards-for-Assessment-and-Accreditation-of-Primary-Medical-Programs-by-the-Australian-Medical-Council-2012.pdf [Accessed 9 March 2021].
* Both undergraduate students in preclinical Years 1 and 2A. Undergraduate pre-clinical Years 1 and 2 are equivalent to Graduate-entry pre-clinical Year A before the medical school curriculum converges in the clinical Year 3B.