Virtual simulation is an innovative knowledge translation strategy that can be as effective as in-person simulation, while being more accessible and cost-efficient, particularly in situations where in-person interaction is difficult (e.g. during online or remote courses) [1]. According to the International Nursing Association for Clinical Simulation and Learning (INACSL), there are four main phases in the design and implementation of simulation-based education: presimulation activities, the simulation experience, debriefing and evaluation [2]. However, robust debriefing following virtual simulations is often neglected which can negatively impact student learning [3].
Debriefing is essential for learning and includes a guided reflection to provide feedback, identifying existing gaps in knowledge or competencies, and discussion to improve future performance [4]. For effective implementation, the facilitator needs to have adequate knowledge and skills related to debriefing frameworks [4]. We use the term virtual debriefing to refer to synchronous online debriefing. Although similar to an in-person debrief, the facilitator must also deal with technical challenges related to engaging learners in an online platform. Virtual debriefing of virtual simulations poses additional challenges as instructors may not observe learner performance and must support self-assessment of performance and identification of learning needs. Additionally, nurse educators in our networks reported gaps in experience or training in virtual debriefing and/or debriefing virtual simulations.
To support educators in using virtual simulations to augment or replace clinical experiences, the Canadian Alliance of Nurse Educators using Simulation (CAN-Sim) developed an online module to address identified professional development needs. The online debriefing module consists of two different components: (1) a Facilitator Guide to Virtual Debriefing (https://can-sim.ca/); and (2) a Synchronous Virtual Debriefing Virtual Simulation Game (VSG) (https://can-sim.ca/). The debriefing guide is a 2-hour interactive presentation that focuses on Debriefing for Meaningful Learning [5]. The guide includes a video demonstration of a virtual debriefing using the Zoom videoconferencing platform. The VSG provides the opportunity for participants to apply the knowledge learned from the debriefing guide as they progress through a series of decision points. To better understand the impact of that online debriefing module, this study aims to evaluate the efficacy of the virtual debriefing professional development module to improve competency of nurses’ educators.
A pretest–post-test survey design was used to evaluate the effect of the debriefing module. To measure self-perceived competence, the professional development for virtual debriefing learning outcomes assessment rubric was used at three time points: (1) baseline pretest; (2) post-test 1 after the facilitator guide; (3) post-test 2 after the synchronous virtual debriefing VSG. Participants were nurse educators who were faculty members, clinical instructors and/or lab instructors of a college in Texas, USA. Ethical approval was obtained from the Austin Community College Institutional Research Review Committee and the Queen’s University Health Sciences & Affiliated Teaching Hospitals Research Ethics Board. After data collection, the survey data were analysed using SPSS Software Version 25.0 through descriptive statistics (frequencies, means, and standard deviations) and repeated measures ANOVA and post hoc analysis.
In the development of our e-learning module, we emphasized the creation of interactive and engaging content to enhance the learning experience and potential impact. Through measuring the outcomes and impact of such interventions we sought to assess the effectiveness and practical relevance of the module.
The total number of participants varied across different measurement time points: pretest (n = 34); post-test 1 (n = 27); post-test 2 (n = 30). The mean pretest knowledge of participants for each competency is displayed in Table 1. Using a repeated measures ANOVA with a Greenhouse–Geisser correction (see Table 2), the mean total learning outcome scores were statistically significantly different [F(1.665,34.972) = 22.259, p < .001]. In addition, we report the partial eta-squared (η p2) which reflects the effect size, with a value greater than 0.14 indicating a large effect size. Similarly, the differences between the pretest and the two post-test scores for each competency were also statistically significant, with large effect sizes.
Measure | Mean score (SD) | ||
---|---|---|---|
Pretest | Post-test 1 | Post-test 2 | |
1. Applying debriefing | 3.22 (1.45) | 3.83 (0.98) | 4.43 (0.66) |
2. Comparing facilitation methods | 2.64 (1.29) | 3.73 (1.03) | 4.45 (0.74) |
3. Establishing a climate of respect and confidentiality | 3.52 (1.34) | 4.30 (0.88) | 4.70 (0.47) |
4. Facilitating engagement in reflection | 3.39 (1.27) | 3.96 (0.88) | 4.48 (0.67) |
5. Providing formative feedback | 3.30 (1.11) | 3.83 (0.98) | 4.35 (0.78) |
Total score | 15.77 (5.42) | 19.45 (4.07) | 22.32 (2.92) |
Note: Scores for outcomes 1 through 5 were measured out of five.
Measure | Pretest to post-test 1 | Post-test 1 to post-test 2 | Repeated measures ANOVA | ||||||
---|---|---|---|---|---|---|---|---|---|
Mean difference | Std. error | Sig | Mean difference | Std. error | Sig | F | p | ηp2 | |
Competency 1 | −.609 | .241 | .058 | −.609 | .151 | .002 | 14.63 | <0.001 | .399 |
Competency 2 | −1.091 | .236 | <.001 | −.727 | .210 | .007 | 29.97 | <0.001 | .588 |
Competency 3 | −.783 | .235 | .009 | −.391 | .186 | .141 | 12.51 | <0.001 | .363 |
Competency 4 | −565 | .216 | .048 | −.522 | .207 | .059 | 10.78 | <0.001 | .329 |
Competency 5 | −.522 | .176 | .022 | −.522 | .165 | .013 | 14.14 | <0.001 | .391 |
Total Score | −3.682 | .977 | .003 | −2.864 | .774 | .004 | 22.56 | <0.001 | .515 |
Competency 1 – Debriefing into a synchronous virtual debriefing session to support learners to gain a deeper understanding of their actions and thought processes.
Competency 2 – Compare various methods of facilitating a virtual debriefing for nursing students to identify the debriefing method that best aligns with the learning experience.
Competency 3 – Establish a climate of respect and confidentiality related to the content of the debriefing discussion to support psychological safety.
Competency 4 – Facilitate participant’s engagement in the reflective process of virtual simulated experience to assist with knowledge translation and application.
Competency 5 – Provide formative feedback following a virtual experience to ensure participants have met the learning objectives.
Post hoc analysis of the learning outcomes assessment rubric total scores demonstrated a significant effect from pretest to post-test 1 (p = .003) which reflects the impact of the facilitator guide, and a significant effect from post-test 1 to post-test 2 (p = .004) which reflects the impact of the VSG on nurse educators’ perceived competency related to virtual debriefing. Pairwise comparisons for the individual competencies were significant for 4 out of 5 of the facilitator guide scores, and 3 out of 5 of the virtual simulation game scores.
Our findings demonstrate that an online professional development module that includes a virtual simulation game can significantly improve nurse educators’ self-perceived competency related to facilitating virtual debriefing. Using pre-packaged professional development modules is a cost-effective way to provide flexible and easily accessible platform education and allows for greater scalability as it can be utilized by a large number of nurse educators globally and help overcome geographical barriers. Also, modules may be used as an introduction to or a refresher on debriefing in a virtual environment. Collecting data from a diverse group of nurse educators, strengthened the potential for validity, reliability and generalizability of our findings and supports the module’s potential utility for other end users.
Further modules could be developed and evaluated that focus on different debriefing frameworks for virtual debriefing to provide nurse educators with debriefing options that align best with their context and learner needs. Lastly, further investigation using a multi-site setting is needed to further assess the effectiveness of this model; thus, a larger evaluation study is currently being conducted by our research team.
MLF: conceptualization and design of study, major contributor to manuscript. AS: major contributor to manuscript. LK: major contributor to manuscript. JB: contributor to manuscript. EZ: conceptualization and design of study, contributor to manuscript. JT: conceptualization and design of study, contributor to manuscript.
None declared.
None declared.
Ethics approval was received from the Austin Community College IRB Committee and the Queen’s University Health Sciences and Affiliated Teaching Hospitals Research Ethics Board.
None declared.
1.
2.
3.
4.
5.