- Research
- Open access
- Published:
Observation of behavioural skills by medical simulation facilitators: a cross-sectional analysis of self-reported importance, difficulties, observation strategies and expertise development
Advances in Simulation volumeĀ 8, ArticleĀ number:Ā 28 (2023)
Abstract
Background
The association between team performance and patient care was an immense boost for team-based education in health care. Behavioural skills are an important focus in these sessions, often provided via a mannikin-based immersive simulation experience in a (near) authentic setting. Observation of these skills by the facilitator(s) is paramount for facilitated feedback with the team. Despite the acknowledgement that trained facilitators are important for optimal learning, insight into this observation process by facilitators is limited.
Objectives
What are the self-reported current practices and difficulties regarding the observation of behavioural skills amongst facilitators during team training and how have they been trained to observe behavioural skills?
Methods
This cross-sectional study used a pilot-tested, content-validated, multi-linguistic online survey within Europe, distributed through a non-discriminative snowball sampling method. Inclusion was limited to facilitators observing behavioural skills within a medical team setting.
Results
A total of 175 persons filled in the questionnaire. All aspects of behavioural skill were perceived as very important to observe. The self-perceived difficulty of the behavioural skill aspects ranged from slightly to moderately difficult. Qualitative analysis revealed three major themes elaborating on this perceived difficulty: (1) not everything can be observed, (2) not everything is observed and (3) interpretation of observed behavioural skills is difficult. Additionally, the number of team members health care facilitators have to observe, outnumbers their self-reported maximum. Strategies and tools used to facilitate their observation were a blank notepad, co-observers and predefined learning goals. The majority of facilitators acquired observational skills through self-study and personal experience and/or observing peers. Co-observation with either peers or experts was regarded as most learn some for their expertise development. Overall, participants perceived themselves as moderately competent in the observation of behavioural skills during team training.
Conclusions
Observation of behavioural skills by facilitators in health care remains a complex and challenging task. Facilitatorsā limitations with respect to attention, focus and (in)ability to perform concomitant tasks, need to be acknowledged. Although strategies and tools can help to facilitate the observation process, they all have their limitations and are used in different ways.
Background
The global interest in so-called behavioural skills increased significantly over the past decades. This was especially due to the association between behavioural skills and preventable medical errors, the latter resulting in unnecessary patient morbidity and mortality [1,2,3,4,5].
Behavioural skills or non-technical skills (NTS) can be defined as the cognitive, social and personal skills that complement ātechnical skillsā and contribute to safe and efficient task performance [6]. Cognitive NTS is often subdivided into situation awareness, decision-making, coping with stress and managing fatigue, whereas social NTS consists of task management, leadership, communication and teamwork [7, 8]. Related terms for NTS commonly found in literature are ācrewā or ācrisis resource managementā (CRM), risk management, āsoft skillsā and ācrisis avoidance resource managementā. Murphy et al. [9] proposed the use of ābehavioural skillsā as a new lexicon over the term NTS as more accurately describing the range and complexity of interpersonal skills.
With the acknowledgement of behavioural skillsā importance in the prevention of medical errors, many team training programs focussing on these aspects were developed. Currently, well-established literature is available which underscores the importance of team-based behavioural skills training [10, 11] and evidence correlates these interventions to improved team performance and even improved patient outcomes and reduced mortality [12,13,14,15,16,17].
Commonly, these team-based trainings are offered through āimmersive simulationā, during which teams are confronted with a scenario-based mannikin in a setting as authentic as possible. Feedback is usually facilitated in the debriefing phase by the facilitator. This debriefing phase is paramount in simulation-based training as this is where most learning occurs [18,19,20]. Savoldelli et al. [21] found that simulation training without feedback did not lead to improvement of behavioural skills in anaesthesia trainees. Furthermore, a meta-analysis showed properly facilitated debriefings to be three times more effective compared to non-facilitated debriefings [22]. Additionally, self-assessment of behavioural skills by participants is reported to be difficult, underscoring the importance of trained facilitators to guide this paramount process of debriefing [22,23,24].
Roussin and Weinstock [25] described a classification for the broad range of simulation-based training programmes currently available. The so-called zone 3 and 4 training entities encompasses ānative teamā training (i.e. without role-playing), with a focus on behavioural skills and a facilitator (rather than an instructor) to guide the (post-event) debriefing. This is the type of training meant wherever this article refers to as āteam trainingā.
Problem definition
The process of āfacilitated feedbackā during debriefing encompasses different stages: first of all, facilitators have to make observations of the behavioural skills. Then, they have to interpret and evaluate these observed aspects before they can facilitate a feedback conversation with the team. Despite the acknowledgement that trained faculty is paramount for optimal learning, insight into how facilitators observe behavioural skills and how their training can impact behavioural skill observation, and therefore, participantsā learning is still limited [26]. The available literature has mainly focused on āstandardisingā this observation process at the end, by providing facilitators with tools, e.g. ābehaviourally anchored rating scalesā, to assess team performance [27]. This however does not describe what facilitators find important to observe, nor their reasoning and interpretation. More so, observation and provision of feedback in behavioural skills can be regarded as deceptively āsimpleā due to the availability of many easy-to-use frameworks [8]. The contrary is the case, as feedback on behavioural skills remains challenging and requires trained faculty, even when applying robust frameworks [10, 28,29,30,31]. The process of becoming trained and experienced faculty in observing behavioural skills can be envisaged to span many years. However, such āexpertise developmentā studies regarding the process of behavioural skills feedback (i.e. observation, interpretation, evaluation and feedback) are scarce [32].
To summarise, experienced faculty is essential for the observation of behavioural skills and feedback and therefore team learning. However, sufficient knowledge of contemporary practice on the observation of behavioural skills by simulation facilitators, as well as potential areas for improvement is still limited.
Objectives
This study aims to answer the following overall research question:
-
āWhat are the self-reported current practices and difficulties regarding observation of NTS amongst facilitators during team training and how have they been trained to observe these skills?ā
The study addresses the following three sub-questions:
-
(1)
Which aspects of NTS find health care facilitators most important and/or most difficult to observe and why?
-
(2)
What strategies and tools (e.g. checklist) do health care facilitators utilise in their daily practice?
-
(3)
How are health care facilitators trained for observing NTS and what kind of future training needs do they have?
Methods
Study design
This study used a cross-sectional methodology using a multi-linguistic online survey within Europe. The survey and all corresponding documents (i.e. participantsā information) were made available in four languages (Dutch, English, German and French).
Participants
Facilitators supervising NTS team training sessionsāpreviously mentioned SimZones 3 or 4āin healthcare were allowed to enrol in the study. Since the SimZones classification model might not be widely known, inclusion criteria were described as being a āteacherā, āinstructorā or āfacilitatorā, who provides feedback on non-technical skills during medical team training, using a scenario-based āimmersive simulationā in a (near) authentic setting.
Development and testing
The survey was constructed based on available literature regarding faculty development and observation of non-technical skills as well as interviews with five experts: an experienced psychologist in human factors and behavioural skills (WvL), two experienced simulation facilitators (AB, WvM) and two educational specialists (DV, DD). The survey was pilot-tested by seven participants who were excluded from participation in the final survey. Two double questions were split into separate questions, several small textual changes were made and one question was removed due to inconsistent interpretation.
The final survey consisted of five clusters: (1) demographic data, (2) level of experience, (3) observational aspects and difficulty, (4) observation strategy and (5) level of education and training. The survey consisted a total of 42 questions (including the consent questions), with the following distribution: fourteen multiple choice questions, nine open-text questions, twelve Likert-scale questions, three slider-scale questions and four matrix table questions. The exact number of questions presented to the participant was dependent on the provided answers as āadaptive questioningā would only present relevant questions based on previous given input to minimise drop-out.
Translations were done by three different native speakers who all had academic language proficiency (i.e. Dutch-French, Dutch-German and Dutch-English).
Ethical review board and informed consent
The study was approved by the independent Dutch Association for Medical Education Ethical Review Committee (ID 2019.1.5).
The participantās information letter was included in the distributive e-mails sent out to the participants. This letter included the aims and goals of the study, participantsā anticipated efforts (anticipated survey completion of 15Ā min), participantsā rights for withdrawal, data handling, ethical approval, research team information and contact details. At the start of the survey, participants were presented with an option to review this information letter, followed by an active opt-in informed consent question. Participation was anonymous; all demographic data used for descriptive analyses regarding age group, gender, background, country of employment and type of their organisation were voluntary questions.
Data collection
The survey was distributed as an āopen surveyā via e-mail through local, regional and national simulation centres and organisations using a non-discriminative snowball sampling method. The survey link was open for responses between May and November 2019. Participants had 1Ā week to complete their survey. Afterwards, the system would close their record and mark their attempt as āincompleteā. Partially completed surveys were included for analysis to limit non-response bias. Raw data were stored in Qualtrics online platform (Qualtricsā¢, Provo, UT, USA). Respondents were prevented from taking the survey more than once using Qualtrics build-in functionality, which used cookies to prevent users from accessing the survey link more than once.
Data analysis
Quantitative data were analysed using descriptive statistical analysis using IBM Statistic Package for the Social Sciences (IBM Corp.Ā© SPSS Statistics for Windows, Version 28.0.0.0). A data entry record with no questions answered past the informed consent and demographic section was regarded a dropout. Level of experience was calculated as the amount of days a facilitator had observed non-technical skills. Exploration of the association between level of experience and Likert-scale results were analysed using Spearmanās Rho test. Qualitative data on open-ended questions were analysed by applying open, axial and selective coding by two researchers (GB, LM) independently using an iterative, constant comparison approach. Any discrepancies in coding were discussed until a consensus was reached.
Results
After a 7-month period, 175 people had participated in the survey. Twenty-seven participants were excluded either because of dropout (Nā=ā23, 13%) or non-observation of NTS (Nā=ā4, 2%), leaving a total of 148 responses for analysis (Fig.Ā 1).
Table 1 provides an overview of participantsā demographic data. The sample included a variety of facilitators, both men (51.3%) and women (48.0%) aged between 20 and 69Ā years. Medical doctors (48.6%) and (registered) nurses (41.2%) made up the vast majority of respondents. Participants were employed in over seventeen countries, with the majority working in the Netherlands (51.4%). Most respondents were employed (only) in university medical centres (44.6%), teaching hospitals (15.5%) and teaching facilities (23.6%) such as simulation centres or universities of applied sciences. The participantsā experience, calculated as (median, IQR) number of NTS observation days was 216Ā days [80ā520].
Self-reported importance and difficulty of NTS
Table 2 provides the results of the respondentsā perceived importance and difficulty of various NTS items. Perceived importance (scales 1ā5; mean [SD]) ranged from very important for ācoping with stress and fatigueā (score 3,7 [0.8]) to extremely important for ācommunicationā (score 4,8 [0.4]). Perceived difficulty on the other hand (scale 1ā5; mean [SD]) ranged from slightly difficult for āleadershipā (score 1,8 [0.8]) to moderately difficult for ācoping with stress and fatigueā (score 3,1 [1.1]). Spearmanās rho analysis showed no association between the level of experience and any of the NTS items.
Participants were asked to provide an explanation on why certain non-technical skills were perceivedāslightly to moderatelyādifficult to observe. Qualitative analysis of these responses (Nā=ā93) revealed three major themes as shown in Table 3.
Theme 1: Not everything can be observed
Not all aspects of non-technical skills are expressed to the facilitator, and therefore, some cannot be observed directly. As an example, āmental processesā or āmental framesā of the participants are of utmost importance but cannot be observed directly. Also, the scripted training session itself might limit the expression of certain non-technical skills, e.g. due to time constraints or a particular focus of the training (i.e. scenario script).
Theme 2: Not everything is observed
Even if non-technical skills are expressed (either verbally or nonverbally), they can be missed easily by the facilitator, e.g. because the facilitator is focusing on a (pre)specified other aspect. It might also occur due to the subtlety of the expression, multiple expressions at the same time, hearing difficulties (e.g. audio difficulties of the simulation setting) and/or fatigue of the facilitator.
Theme 3: Interpretation of the observed NTS is difficult
Even when expressions or behaviours are observed by the facilitator, subsequent interpretation and analysis are difficult. Reported reasons for this are as follows: (1) interpretation is a āsubjective processā, influenced by the facilitatorsā own frame of reference and/or experience; (2) not all NTS have clear standards and/or observable criteria, e.g. āattitudesā are reported to be (more) difficult to observe due to this phenomenon; (3) interpretation might be complex as performance fluctuates over time during the training session; (4) non-technical and technical skills might be intertwined hampering cause-effect interpretation and (5) participants might express different behaviour compared to their authentic workplace setting.
When participants were asked to indicate the percentage of observed NTS (in relation to the amount of NTS being expressed) in their average training session on a sliding scale (0ā100%), and the mean [SD] result is 57% [17.6]. Furthermore, the mean [SD] number of team members facilitators observed, outnumbered their personal reported maximum 5.5 [2.0] and 3.9 [1.8], respectively; t(122)ā=ā8323; pā<ā0.001). This finding was universal amongst all facilitators and not associated with the level of experience within the sample population (Ļā=ā0.084, nā=ā122, pā=ā0.360).
Facilitators were asked to elaborate on the barriers and enablers regarding the observation of NTS. Qualitative analysis (Nā=ā106) revealed some overlap with the observational difficulties previously described. In addition, āother concomitant tasksā such as role play in the scenario (i.e. patient, senior staff member) or observation of technical skills were most often cited as important barriers. Quote: [Translated] āYou have to perform multiple tasks as a facilitator: provide information [to the participants], [scenario] role play, control the technology [i.e. mannikin] and make observations, so you cannot pay full attention to the observation of NTS.ā [R43] Reported enablers on the other hand included the use of multiple (co-)observers, having predetermined learning goals to focus on during the observation and making use of video recordings. Quote: [Translated] āThere is a lot happening in a very short time span and I also need to focus on the content. I am always glad there are two of us.ā [R21] Furthermore, personal knowledge and experience were mentioned as well for identifying relevant behavioural skills: e.g. āItās very hard to pick up on everything, training [and] practicing help[ed] to improve skills at recognition.ā [R115].
Observation strategies and tools used for observing NTS
Several strategies (e.g. co-observer, predefined learning goals) and tools (e.g. checklists/frameworks) can be utilised while observing NTS. The frequency by which health care facilitators used these is visualised in Fig.Ā 2 (see also Supplementary Table S1).
A blank notepad, co-observers and predefined NTS learning goals were used most frequently. The use of a checklist/framework varied widely: 18% always, 29% most of the time, 14% about half the time, 20% sometimes and 19% never. All participants stating they used a checklist/framework at least sometimes were asked to specify this, including whether they had received formal training in its application (see Supplementary Table S2). The percentage of trained faculty was under 50% (range 0ā45%) for each framework or checklist, with the exception of the 15-item CRM framework by Rall et al. [33] (12/14, 86%).
Participants were asked about the perceived usefulness of frameworks for their observation practice. Answers varied from: ānot at all usefulā 5%, āslightly usefulā 23%, āmoderately usefulā 26% and āvery usefulā 31% to āextremely usefulā 15%. An additional open-ended question was asked to explore this further (see Table 4). Reported benefits of using a checklist/framework were the provision of a structure, a more objective observation and beneficial for research purposes. Reported disadvantages on the other hand were hindering of real-time observation, not suitable for every observation, less useful when more experienced, subjectivity of framework interpretation and resulting in a shift from learner-to-instructor-centred debrief.
Expertise development in NTS observation
Participants were asked to specify their attended training modalities with regard to their expertise development in NTS observation. Observing peers, self-study/own experience and direct supervision by expert(s) were selected most often by 64%, 62% and 45% of participants, respectively. The usefulness score (mean, [SD]) on a 5-point Likert scale was highest for direct supervision by expert(s), 4.21 [0.67], as shown in Table 5.
Participants were given the option to comment on what aspects were most contributing to their learning curve with regard to the observation of NTS. Qualitative analysis (Nā=ā53) is in line with the quantitative data as āreflection on oneās individual observation, either by peers or expertsā is mentioned by 49% (26/53). Quote: [Translated] āObserving the same scenario together and discussing what you have seen [is most useful].ā [R25].
Overall, given their faculty development, participants perceived themselves as āmoderately competentā for observing NTS (mean [SD] 3.42 [0.71]). There was no association between the level of experience and self-perceived overall competence (Ļā=ā0.122, Nā=ā120, pā=ā0.185).
Finally, participants were asked whether they had future training needs (Question: āAre there any aspects you would like (additional) training for regarding the observation of NTSā). A total of 48% (57/119) answered positively whereas 52% (62/119) declined on having future training goals. There was neither an association between future training needs and self-perceived overall competence (rpbāāā0.125, Nā=ā119, pā=ā0.099), nor with the level of experience (rpb 0.128, Nā=ā119, pā=ā0.167). Respondents who answered positively on having future training aspects were asked to specify these in an open-ended question. The majority of these answers related to either receiving direct feedback from other peers/experts, e.g. [Translated] ā⦠More direct supervision and observation from trainersā [R18] and [Translated] āI would like to see a training with side-by-side NTS analysis by participants and experts of some pre-recorded video fragments.ā [R25] or related to ātraining benefitsā in general, e.g. [Translated] āExtra training is always beneficialā [R20]. Two other interesting quotes stood out: āI would like to have [had] formal training during my early practice as [an] observer.ā [R120] and [Translated] āā¦repetitive NTS observation training with some certification as many people are currently observing NTS without any proof of education/trainingā. [R36].
Discussion
This study endorses the opinion that observation of non-technical/behavioural skills is a complex task also within the medical community. The main results of the three explored research questions will be discussed consecutively in the sections below.
Medical facilitators regard all of the behavioural skills at least āvery importantā to observe. Given the corresponding (only) slight to moderate difficulty with these observations, one can think that observation of behavioural skills is at first sight a simple task. However, when questioned more deeply, three themes were distilled, explaining why observation of behavioural skills is a rather challenging task. First, not everything can be observed; mostly mental frames of the participants cannot be observed directly and need to be explored based on the observed subsequent behaviour which is in line with publications recommending facilitators to question participantsā rationale for their actions in the debrief (i.e. ādouble-loop learningā) [18, 24, 34,35,36,37]. Second, not everything is observed; despite the expression of a behavioural skill, it might not have been observed by the facilitator. This might be intentionally by a distinct focus of the facilitator or unintentionally due to the subtlety or simultaneously of other behavioural skills or from distractions of concomitant tasks. Interestingly, facilitators report that the amount of team members they observe on their average training outnumbers their personal maximum. Salas et al. [27] advocated already that one facilitator should observe only two team members in complex teams so as not to overlook any interactions. With the average amount of team members being almost twice as high, it might not be a surprise that facilitators self-report they capture only half of the expressed behavioural skills [38].
Third, even when behavioural skills are observed by medical facilitators, interpretation is difficult, e.g. due to the personal frames of reference from the facilitators. This difficulty is in line with previous publications on limited inter-rater agreement, especially in larger, more complex, multidisciplinary teams [39, 40]. Gingerich et al. [41] however described nicely how to look at this facilitatorsā cognition from different research perspectives, including the meaningful idiosyncratic stance, stressing that different interpretations between facilitators are not necessarily a bad thing and combining information from multiple observers and perhaps even participants itself can offer advantages for the debriefing phase [27, 41,42,43,44]. The technique of having multiple observers for āco-debriefingā is nicely described by Cheng et al. [45] and appears to be used often by facilitators in our study.
The majority of facilitators had predefined behavioural skills learning goals at the start of their training. This allows for a tailored scenario selection and provides the facilitator with a clear aspect to focus on, also recommended by Salas et al. [27]. Although many checklists/frameworks are available for observation of behavioural skills, its use varied particularly within our population as did their reasoning for (not) using them. We hypothesised that novice facilitators would use some checklist/framework to structure their observation; however, we could not find an association between checklist/framework use and level of experience. We might have missed ātrue noviceā facilitators and/or the perceived disadvantages of checklist/framework use outweigh their perceived benefits in everyday practice. Interestingly, of all the participants that used a checklist/framework, only a minority used well-validated frameworks published in the literature, a phenomenon also addressed by Jenkins [46]. Frameworks can indeed have conflicting properties as reported in our study, i.e. sub-team specific frameworks (e.g. non-technical skills for surgeons [47], scrub practitioners [48] or anaesthetists [49]) have more clearly described behaviours, helping facilitatorsā interpretation whereas āwhole teamā frameworks (e.g. clinical teamwork scale [50], coordination behaviour in acute care teams [51]) are wider applicable possibly benefiting facilitators by gaining experience quicker. Jepsen et al. [8] provided a nice overview of 23 frameworks applicable to the observation of behavioural skills. Either way, facilitators need to provide their learners with detailed feedback to enhance meaningful learning [38] and behavioural frameworks might serve as a ācognitive aidā and prevent facilitatorsā fixation. On the other hand, if the observational task is already perceived as too challenging, having to complete yet another (concomitant) task might feel burdensome rather than helpful. Although the impact of checklists/frameworks on feedback and student learning is beyond our scope, using identical frameworks with multiple observers, it could facilitate an inter-rater discussion, thereby enhancing facilitatorsā insight into their personal frame of reference.
When it comes to expertise development in behavioural skill observation, the majority of the facilitators were auto-didacts, who acquired their observation skills through either self-study/experience or from observing peers. Given the facilitatorsā overall self-perceived moderate competence score coupled with the likely selection bias, it is interesting that only half of the facilitators reported future training desires. A study exploring simulation instructorsā self-reported needs, revealed that a āsense of competenceā, i.e. achieving their highest level of professional functioning through self-reflection, was indeed present amongst many [52]. Since the vast majority of our sample population was made up of doctors and nurses being actively employed in health care facilities, time constraints and/or workload might be a relevant hypothesis to examine further, especially when striving for (regularly) trained faculty [53, 54]. Deployment of nonclinical faculty has been proposed to overcome time constraints for expertise development and re-certification within the medical domain [53, 55].
Strengths and limitations
To the authorsā knowledge, this study is the first to address self-perceived current practices and expertise development of behavioural skills observation amongst health care facilitators. The study provides relevant insights for different behavioural (research) domains such as expertise development, implementation of observational frameworks and organisation of (medical) team training. In order to focus on observation and interpretation of behavioural skills, we purposefully did not include questions on team feedback (i.e. ādebriefingā). One of the important disadvantages of this study methodology is the inability to validate self-reported data. In addition, the cross-sectional sample strategy most likely skewed our sample population towards the more intrinsically motivated facilitators. Quantification of a response rate to analyse this further is unfortunately not possible since there is no formal registration for health care facilitators observing behavioural skills within Europe. In addition, half of the sample population was employed in the Netherlands, although we could not find this to be a confounding variable in the results.
Future research directions
More insight into what determines qualitative behavioural skill observation would allow tailoring of facilitatorsā expertise development and guidelines for everyday practice. For example: in what way does the facilitator-team member ratio influence the quality of the observed behavioural skills? Is the amount of self-reported observations by the facilitators correct, and if so, what behavioural aspects are observed and why? In addition, medical facilitators tend to be mostly auto-didacts who appreciate co-observation with peers or experts. Research analysing how facilitators learn from each other and experts while observing medical teams can further benefit expertise development.
Future practice directions
Meanwhile, the three themes distilled in this study hold important implications for everyday practice. First, since not everything can be observed directly, facilitators need to reserve sufficient time for the (learner-centred) debriefing to unravel theāunobservableāparticipantsā thoughts. Second, the shortcomings of the human facilitator should be acknowledged. Facilitators cannot fulfil concomitant tasks in addition to observing behavioural skills and might even limit their attention further to (only) a (pre)specified aspect. This also necessitates multiple faculty members to run a (complex) team training, such as an āoperatorā for controlling the mannikin and multiple co-observers. Despite this, facilitators need to be aware of their personal frame of reference compared to that of others.
Conclusions
Observation of behavioural skills by facilitators in health care remains a complex and challenging task. Facilitatorsā limitations with respect to attention, focus and (in)ability to do concomitant tasks need to be acknowledged. Although strategies and tools can help to facilitate the observation process, they all have their limitations and are used in different ways.
Availability of data and materials
The dataset of the current study is available from the corresponding author upon reasonable request.
Abbreviations
- NTS:
-
Non-technical skills
References
Mazzocco K, Petitti DB, Fong KT, Bonacum D, Brookey J, Graham S, et al. Surgical team behaviors and patient outcomes. Am J Surg. 2009;197(5):678ā85.
Siu J, Maran N, Paterson-Brown S. Observation of behavioural markers of non-technical skills in the operating room and their relationship to intra-operative incidents. Surgeon. 2016;14(3):119ā28.
Gawande AA, Zinner MJ, Studdert DM, Brennan TA. Analysis of errors reported by surgeons at three teaching hospitals. Surgery. 2003;133(6):614ā21.
Institute of Medicine Committee on Quality of Health Care in A. To err is human: building a safer health system. In: Kohn LT, Corrigan JM, Donaldson MS, editors. To err is human: building a safer health system. Washington (DC): National Academies Press (US) Copyright 2000 by the National Academy of Sciences. All rights reserved; 2000.
Dedy NJ, Bonrath EM, Zevin B, Grantcharov TP. Teaching nontechnical skills in surgical residency: a systematic review of current approaches and outcomes. Surgery. 2013;154(5):1000ā8.
Flin R, OāConnor P, Crichton M. Safety at the sharp end. A guide to non-technical skills. Surrey: Ashgate; 2008. p. 336.
Flin R, Maran N. Basic concepts for crew resource management and non-technical skills. Best Pract Res Clin Anaesthesiol. 2015;29(1):27ā39.
Jepsen RMHG, Ćstergaard D, Dieckmann P. Development of instruments for assessment of individualsā and teamsā non-technical skills in healthcare: a critical review. Cogn Technol Work. 2015;17(1):63ā77.
Murphy P, Nestel D, Gormley GJ. Words matter: towards a new lexicon for 'nontechnical skillsā training. Adv Simul (Lond). 2019;4:8.
Sevdalis N, Hull L, Birnbach DJ. Improving patient safety in the operating theatre and perioperative care: obstacles, interventions, and priorities for accelerating progress. Br J Anaesth. 2012;109(Suppl 1):i3ā16.
Schmitt M, Blue A, Aschenbrener CA, Viggiano TR. Core competencies for interprofessional collaborative practice: reforming health care by transforming health professionalsā education. Acad Med. 2011;86(11):1351.
Manser T. Teamwork and patient safety in dynamic domains of healthcare: a review of the literature. Acta Anaesthesiol Scand. 2009;53(2):143ā51.
Julia Neily J, Mills PD, Young-Xu Y, Carney BT, West P, Berger DH, et al. Association between implementation of a medical team training program and surgical mortality. JAMA. 2010;304(15):1693ā700.
Catchpole KR, Dale TJ, Hirst DG, Smith JP, Giddings TA. A multicenter trial of aviation-style training for surgical teams. J Patient Saf. 2010;6(3):180ā6.
Marshall DA, Manus DA. A team training program using human factors to enhance patient safety. AORN J. 2007;86(6):994ā1011.
Awad SS, Fagan SP, Bellows C, Albo D, Green-Rashad B, De la Garza M, et al. Bridging the communication gap in the operating room with medical team training. Am J Surg. 2005;190(5):770ā4.
Carpenter JE, Bagian JP, Snider RG, Jeray KJ. Medical team training improves team performance: AOA critical issues. J Bone Joint Surg Am. 2017;99(18):1604ā10.
Kolbe M, Grande B, Spahn DR. Briefing and debriefing during simulation-based training and beyond: content, structure, attitude and setting. Best Pract Res Clin Anaesthesiol. 2015;29(1):87ā96.
Issenberg SB, McGaghie WC, Petrusa ER, Lee Gordon D, Scalese RJ. Features and uses of high-fidelity medical simulations that lead to effective learning: a BEME systematic review. Med Teach. 2005;27(1):10ā28.
Lederman LC. Debriefing: toward a systematic assessment of theory and practice. Simul Gaming. 1992;23(2):145ā60.
Savoldelli GL, Naik VN, Park J, Joo HS, Chow R, Hamstra SJ. Value of debriefing during simulated crisis management: oral versus video-assisted oral feedback. Anesthesiology. 2006;105(2):279ā85.
Tannenbaum SI, Cerasoli CP. Do team and individual debriefs enhance performance? A meta-analysis. Hum Factors. 2013;55(1):231ā45.
Arora S, Miskovic D, Hull L, Moorthy K, Aggarwal R, Johannsson H, et al. Self vs expert assessment of technical and non-technical skills in high fidelity simulation. Am J Surg. 2011;202(4):500ā6.
Burns CL. Using debriefing and feedback in simulation to improve participant performance: an educatorās perspective. Int J Med Educ. 2015;6:118ā20.
Roussin CJ, Weinstock P. SimZones: an organizational innovation for simulation programs and centers. Acad Med. 2017;92(8):1114ā20.
Gordon M, Farnan J, Grafton-Clarke C, Ahmed R, Gurbutt D, McLachlan J, et al. Non-technical skills assessments in undergraduate medical education: a focused BEME systematic review: BEME Guide no. 54. Med Teach. 2019:41(7)732-45.
Salas E, Reyes DL, Woods AL. The assessment of team performance: observations and needs. Innovative Assessment of Collaboration. Methodology of Educational Measurement and Assessment; 2017. p. 21ā36.
Hull L, Arora S, Stefanidis D, Sevdalis N. Facilitating the implementation of the American College of Surgeons/Association of Program Directors in Surgery phase III skills curriculum: training faculty in the assessment of team skills. Am J Surg. 2015;210(5):933-41 e2.
Baker DP, Dismukes RK. A framework for understanding crew performance assessment issues. Int J Aviat Psychol. 2002;12(3):205ā22.
Boet S, Bould MD, Layat Burn C, Reeves S. Twelve tips for a successful interprofessional team-based high-fidelity simulation education session. Med Teach. 2014;36(10):853ā7.
Lindqvist SM, Reeves S. Facilitatorsā perceptions of delivering interprofessional education: a qualitative study. Med Teach. 2007;29(4):403ā5.
Russ S, Hull L, Rout S, Vincent C, Darzi A, Sevdalis N. Observational teamwork assessment for surgery: feasibility of clinical and nonclinical assessor calibration with short-term training. Ann Surg. 2012;255(4):804ā9.
Rall M, Gaba D, Howard SK, Dieckmann P. Human performance and patient safety. In: Gropper M, Eriksson L, Fleisher L, Wiener-Kronish J, Cohen N, Leslie K, editors. Millerās Anesthesia; 2010.
Rudolph JW, Simon R, Raemer DB, Eppich WJ. Debriefing as formative assessment: closing performance gaps in medical education. Acad Emerg Med. 2008;15(11):1010ā6.
Rudolph JW, Simon R, Rivard P, Dufresne RL, Raemer DB. Debriefing with good judgment: combining rigorous feedback with genuine inquiry. Anesthesiol Clin. 2007;25(2):361ā76.
Cheng A, Morse KJ, Rudolph J, Arab AA, Runnacles J, Eppich W. Learner-centered debriefing for health care simulation education: lessons for faculty development. Simul Healthc. 2016;11(1):32ā40.
Fey MK, Roussin CJ, Rudolph JW, Morse KJ, Palaganas JC, Szyld D. Teaching, coaching, or debriefing with good judgment: a roadmap for implementing āwith good judgmentā across the SimZones. Adv Simul (Lond). 2022;7(1):39.
Rosen MA, Weaver SJ, Lazzara EH, Salas E, Wu T, Silvestri S, et al. Tools for evaluating team performance in simulation-based training. J Emerg Trauma Shock. 2010;3(4):353ā9.
Eppich W, Nannicelli AP, Seivert NP, Sohn MW, Rozenfeld R, Woods DM, et al. A rater training protocol to assess team performance. J Contin Educ Health Prof. 2015;35(2):83ā90.
Feldman M, Lazzara EH, Vanderbilt AA, DiazGranados D. Rater training to support high-stakes simulation-based assessments. J Contin Educ Health Prof. 2012;32(4):279ā86.
Gingerich A, Kogan J, Yeates P, Govaerts M, Holmboe E. Seeing the āblack boxā differently: assessor cognition from three research perspectives. Med Educ. 2014;48(11):1055ā68.
van der Vleuten CP, Schuwirth LW. Assessing professional competence: from methods to programmes. Med Educ. 2005;39(3):309ā17.
Govaerts MJ, Schuwirth LW, Van der Vleuten CP, Muijtjens AM. Workplace-based assessment: effects of rater expertise. Adv Health Sci Educ Theory Pract. 2011;16(2):151ā65.
Weller J, Shulruf B, Torrie J, Frengley R, Boyd M, Paul A, et al. Validation of a measurement tool for self-assessment of teamwork in intensive care. Br J Anaesth. 2013;111(3):460ā7.
Cheng A, Palaganas J, Eppich W, Rudolph J, Robinson T, Grant V. Co-debriefing for simulation-based education: a primer for facilitators. Simul Healthc. 2015;10(2):69ā75.
Jenkins B. Training and assessment of non-technical skills in the operating theatre: where next? Anaesthesia. 2015;70:893ā906.
Geraghty A, Paterson-Brown S. Non-technical skills for surgeons (NOTSS). Surg Infect (Larchmt). 2020;38(10):612ā6.
Flin R, Mitchell L, McLeod B. Non-technical skills of the scrub practitioner: the SPLINTS system. ORNAC J. 2014;32(3):33ā8.
Flin R, Patey R. Non-technical skills for anaesthetists: developing and applying ANTS. Best Pract Res Clin Anaesthesiol. 2011;25(2):215ā27.
Guise JM, Deering SH, Kanki BG, Osterweil P, Li H, Mori M, et al. Validation of a tool to measure and promote clinical teamwork. Simul Healthc. 2008;3(4):217ā23.
Kolbe M, Burtscher MJ, Manser T. Co-ACTāa framework for observing coordination behaviour in acute care teams. BMJ Qual Saf. 2013;22(7):596ā605.
Frei-Landau R, Levin O. Simulation-based learning in teacher education: using Maslowās Hierarchy of needs to conceptualize instructorsā needs. Front Psychol. 2023;14:1149576.
Hull L, Arora S, Symons NR, Jalil R, Darzi A, Vincent C, et al. Training faculty in nontechnical skill assessment: national guidelines on program requirements. Ann Surg. 2013;258(2):370ā5.
Eva KW, Bordage G, Campbell C, Galbraith R, Ginsburg S, Holmboe E, et al. Towards a program of assessment for health professionals: from training into practice. Adv Health Sci Educ Theory Pract. 2016;21(4):897ā913.
Hull L, Sevdalis N. Advances in teaching and assessing nontechnical skills. Surg Clin North Am. 2015;95(4):869ā84.
Acknowledgements
The authors like to thank Anique Baten and Wouter van Leeuwen for their feedback on survey construction; Eva Schaffrath and Célia Dejond for translation of the survey, Gijs Bongaarts for his help in data analysis and all the colleagues who have helped in survey distribution, including but not limited to Ulrich Strauch (Dutch Society of Simulation in Healthcare, Dutch Society of Intensive Care Medicine), Marc Lazarovici (Deutsche Gesellschaft zur Förderung der Simulation in der Medizin e.V.), Lukas Drabauer (Alpha Medical Concepts), Isabelle Bragard (Centre de Simulation Médicale Université de Liège), Ingrid Hoekstra (Regional Ambulance Service Utrecht), Ron Brendel (METS Centre Bilthoven), Marijn Boer (Martini Hospital Groningen) and Hans van Schuppen (Amsterdam University Medical Centre) as well as Armand ten Haaf, Jeroen Hoebers, Henk Metz and Ger Dautzenberg (Department of Simulation in Healthcare, Maastricht University Medical Centre).
Funding
Not applicable.
Author information
Authors and Affiliations
Contributions
All authors were involved in the study design and manuscript preparation. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisherās Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Additional file 1: Table S1.
Frequencies of strategies and tools used by facilitators. Table S2. Frequencies of different checklists/frameworks used by facilitators.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Mommers, L., Verstegen, D., Dolmans, D. et al. Observation of behavioural skills by medical simulation facilitators: a cross-sectional analysis of self-reported importance, difficulties, observation strategies and expertise development. Adv Simul 8, 28 (2023). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s41077-023-00268-x
Received:
Accepted:
Published:
DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s41077-023-00268-x