Learning outcomes for communication skills across the health professions: a systematic literature review and qualitative synthesis

Affiliations.

  • 1 Faculty of Medicine, Nursing and Health Sciences, Monash University, Melbourne, Victoria, Australia.
  • 2 Department of Medical Education, University of Melbourne, Melbourne, Victoria, Australia.
  • 3 Department of Surgery (Austin), University of Melbourne, Melbourne, Victoria, Australia.
  • 4 Department of Physiotherapy, Monash University, Melbourne, Victoria, Australia.
  • PMID: 28389493
  • PMCID: PMC5558817
  • DOI: 10.1136/bmjopen-2016-014570

Objective: The aim of this study was to identify and analyse communication skills learning outcomes via a systematic review and present results in a synthesised list. Summarised results inform educators and researchers in communication skills teaching and learning across health professions.

Design: Systematic review and qualitative synthesis.

Methods: A systematic search of five databases (MEDLINE, PsycINFO, ERIC, CINAHL plus and Scopus), from first records until August 2016, identified published learning outcomes for communication skills in health professions education. Extracted data were analysed through an iterative process of qualitative synthesis. This process was guided by principles of person centredness and an a priori decision guide.

Results: 168 papers met the eligibility criteria; 1669 individual learning outcomes were extracted and refined using qualitative synthesis. A final refined set of 205 learning outcomes were constructed and are presented in 4 domains that include: (1) knowledge (eg, describe the importance of communication in healthcare), (2) content skills (eg, explore a healthcare seeker's motivation for seeking healthcare),( 3) process skills (eg, respond promptly to a communication partner's questions) and (4) perceptual skills (eg, reflect on own ways of expressing emotion).

Conclusions: This study provides a list of 205 communication skills learning outcomes that provide a foundation for further research and educational design in communication education across the health professions. Areas for future investigation include greater patient involvement in communication skills education design and further identification of learning outcomes that target knowledge and perceptual skills. This work may also prompt educators to be cognisant of the quality and scope of the learning outcomes they design and their application as goals for learning.

Keywords: Communication Skills; Health Professions Education; Learning Outcome; Qualitative synthesis; Systematic Review.

Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.

Publication types

  • Systematic Review
  • Clinical Competence*
  • Communication*
  • Educational Measurement
  • Health Occupations / education*
  • Health Personnel / education*
  • Professional Competence
  • Research article
  • Open access
  • Published: 29 April 2021

A qualitative systematic review and thematic synthesis exploring the impacts of clinical academic activity by healthcare professionals outside medicine

  • Lisa Newington   ORCID: orcid.org/0000-0001-6954-2981 1 , 2 ,
  • Mary Wells   ORCID: orcid.org/0000-0001-5789-2773 1 , 2 ,
  • Adine Adonis   ORCID: orcid.org/0000-0002-5372-8025 1 ,
  • Lee Bolton   ORCID: orcid.org/0000-0002-7104-9376 1 , 2 ,
  • Layla Bolton Saghdaoui   ORCID: orcid.org/0000-0003-1229-773X 1 , 2 ,
  • Margaret Coffey   ORCID: orcid.org/0000-0001-8898-3432 1 , 2 ,
  • Jennifer Crow   ORCID: orcid.org/0000-0003-4692-1799 1 ,
  • Olga Fadeeva Costa   ORCID: orcid.org/0000-0003-0319-5045 1 , 2 ,
  • Catherine Hughes 1 ,
  • Matthew Savage 1 ,
  • Lillie Shahabi   ORCID: orcid.org/0000-0001-5268-0771 1 , 3 &
  • Caroline M. Alexander   ORCID: orcid.org/0000-0003-1816-8939 1 , 2  

BMC Health Services Research volume  21 , Article number:  400 ( 2021 ) Cite this article

7269 Accesses

29 Citations

106 Altmetric

Metrics details

There are increasing opportunities for healthcare professionals outside medicine to be involved in and lead clinical research. However, there are few roles within these professions that include time for research. In order to develop such roles, and evaluate effective use of this time, the range of impacts of this clinical academic activity need to be valued and understood by healthcare leaders and managers. To date, these impacts have not been comprehensively explored, but are suggested to extend beyond traditional quantitative impact metrics, such as publications, citations and funding awards.

Ten databases, four grey literature repositories and a naïve web search engine were systematically searched for articles reporting impacts of clinical academic activity by healthcare professionals outside medicine. Specifically, this did not include the direct impacts of the research findings, rather the impacts of the research activity. All stages of the review were performed by a minimum of two reviewers and reported impacts were categorised qualitatively according to a modified VICTOR (making Visible the ImpaCT Of Research) framework.

Of the initial 2704 identified articles, 20 were eligible for inclusion. Identified impacts were mapped to seven themes: impacts for patients; impacts for the service provision and workforce; impacts to research profile, culture and capacity; economic impacts; impacts on staff recruitment and retention; impacts to knowledge exchange; and impacts to the clinical academic.

Conclusions

Several overlapping sub-themes were identified across the main themes. These included the challenges and benefits of balancing clinical and academic roles, the creation and implementation of new evidence, and the development of collaborations and networks. These may be key areas for organisations to explore when looking to support and increase academic activity among healthcare professionals outside medicine. The modified VICTOR tool is a useful starting point for individuals and organisations to record the impact of their research activity. Further work is needed to explore standardised methods of capturing research impact that address the full range of impacts identified in this systematic review and are specific to the context of clinical academics outside medicine.

Peer Review reports

There is compelling evidence that research active healthcare organisations have improved care performance compared to their non-research active counterparts [ 1 ]. Examples include patients feeling better informed about their condition and medication, having greater confidence in their healthcare staff [ 2 ], greater staff adherence to treatment guidelines [ 3 ] and lower mortality rates [ 3 , 4 ]. In the UK, this has resulted in correlation between research activity and the national healthcare inspection rating [ 4 ].

Traditionally, healthcare research has been associated with medical professionals (doctors), with approximately 5% of UK medical consultants working in clinical academic roles [ 5 , 6 ]. Clinical academics engage in clinical practice and also conduct and lead programmes of applied health and/or social care research, often directly aimed at improving patient care and care pathways [ 7 ].

Healthcare professionals outside medicine are increasingly developing the expertise to lead clinically relevant research, with the aim of 1% of this workforce being employed in clinical academic roles by 2030 [ 8 ]. Healthcare professions outside medicine include: nursing, midwifery, the allied health professions (art therapists, dietitians, drama therapists, music therapists, occupational therapists, orthoptists, operating department practitioners, osteopaths, podiatrists, prosthetists/orthotists, paramedics, physiotherapists, radiographers, and speech and language therapists), clinical psychologists, healthcare scientists and pharmacists. Within the UK, the drive to increase the clinical academic workforce is supported by a targeted Health Education England/National Institute for Health Research funding stream specifically for these professions [ 9 ], and through fellowship funding from a number of national health charities. Similar schemes exist elsewhere [ 10 , 11 , 12 , 13 ].

As clinical academic activity increases, there is a need to evaluate its impact at both individual and organisational levels, and across the short to longer term. Several frameworks have been designed to guide impact assessments for healthcare research and these have recently been systematically reviewed to create a summary framework [ 14 ]. However, the focus is on evaluating individual programmes of research, rather than the impact of collective research activity within an organisation. This is also the case with other research impact assessment tools [ 15 ].

Outside the medical professions, the impacts of dedicated allied health professional (AHP) research roles have been systematically reviewed to explore their outcomes in terms of building research capacity and culture. Wenke and Mickan described varied roles, but most often these centred on the development of researchers’ own research projects and their dissemination [ 10 ]. Additional responsibilities included supervising others and developing strategies to promote research activity. These roles were found to have positive impacts on individual research skills, research outputs and research culture, however other areas of impact were not assessed, such as patient outcomes, changes to clinical training or practice guidelines, or increased investment. Importantly, only one study described practising clinicians with dual clinical and research roles, and other non-medical professionals, such as nurses and midwives, were excluded from the review.

Existing reviews have only included published research studies, thus overlooking impact reports that have been compiled by individual healthcare organisations or collaborations [ 16 , 17 ]. Such documents contain valuable insights and reflections on clinical academic programmes, often in the form of case studies, and are useful for other healthcare providers supporting or developing their own clinical academic strategies.

The current systematic review was developed in order to understand the full range of impacts of non-medical clinical academic roles. A cross-disciplinary approach was taken to include clinical academic activity among nurses, midwives, AHPs and other non-medical healthcare professionals.

The review protocol was pre-registered with the Open Science Foundation [ 18 ] and followed the PRISMA protocol reporting guidelines [ 19 ]. The primary review question was: what are the reported impacts of clinical academic activity among practising healthcare professionals outside medicine?

Selection criteria

Impact was not pre-defined for the purposes of this review, and eligible articles were those reporting any form of impact that was attributed to clinical academic activity carried out by non-medical healthcare professionals. This did not include the reported outcomes of clinical research studies, rather the impact of these individuals being involved in research activity. Clinical academic activity was defined as the review intervention, and was the involvement of practising clinicians in research. This included specific research roles, such as research fellowships or combined clinical academic positions, in addition to other protected research time or opportunities to be involved in research. The review population was defined as healthcare professionals outside medicine. Full eligibility criteria, including the list of eligible healthcare professions, are provided in Table  1 .

Search strategy

Ten healthcare databases and four grey literature repositories were individually searched by the lead author between December 2019 and January 2020. Search locations are shown in Fig.  1 . The example search strategy for Medline is provided in Additional File  1 , which included i) terms for the different non-medical clinical disciplines, and ii) terms for clinical academic activity, combined with iii) terms for impact. The search strategy was developed and piloted by LN, CMA and MW, with additional assistance from a healthcare librarian (LG). There were no restrictions for country of publication, but due to time and resource limitations, articles were restricted to those available in the English language. An additional Google search for ‘impact of clinical academic nursing, midwifery and allied health’ was conducted in a naïve browser and the first 50 hits recorded.

figure 1

PRISMA flowchart

Eligibility assessment

Identified references were exported into Covidence ( Covidence.org ) and duplicates removed. Title and abstract screening were performed in two stages. Firstly, the lead author and two members of the review team (JC and OFC) independently screened out those articles which clearly met the exclusion criteria. Any disagreements were resolved by discussion. Secondly, the lead author and two members of the review team (LS and LB) independently screened the remaining articles against both the inclusion and exclusion criteria. Again, any disagreements were resolved by discussion. Full text screening was independently performed by LN and one of five members of the review team (AA, MC, LS, OFC and LBS); disagreements were resolved by MW and CMA.

Quality assessment

The mixed methods appraisal tool (MMAT) was used to evaluate the quality of the included studies [ 20 ]. The assessment form was piloted prior to use and modified to include key components of the qualitative checklist proposed by Walsh et al. [ 21 ]. The quality assessment form template is provided in Additional File  2 . Quality assessment was conducted independently by LN and one of eight members of the review team (AA, LBS, CL, LS, MC, JJ, LB and JC). Disagreements were resolved by discussion. All relevant studies were included regardless of their MMAT score.

Data extraction

Data extraction was completed independently by LN and one of the eight members of the review team, listed above, and was conducted in parallel to the quality assessment. Data items were extracted using a pre-piloted electronic form under the headings listed in Table  2 . The two independent data extraction forms for each article were compared and harmonised by MS and CH, who referred back to the original articles where disagreement occurred. The impacts of the clinic academic activity were broadly characterised using the section headings from the VICTOR (making Visible the ImpaCT Of Research) framework [ 22 ]. The VICTOR framework was developed for individual research teams to record the impact of their study and is endorsed by the UK National Institute for Health Research [ 23 ]. It comprises a series of open ended questions categorised under seven sections: health benefits, safety and quality improvements for research participants and carers during the study; service and workforce impacts; research profile of the organisation and research capacity; economic impacts; organisation’s influence and reputation; knowledge generation and exchange. Following piloting for this systematic review, the headings were modified to include a section for impacts to the individual, and research profile and research capacity and the organisation’s influence and reputation were merged.

Data synthesis

Extracted data for each of the pre-identified categories of impact (Table 2 ) were independently analysed by two members of the review team (LN and one of CH, AA, MS, JC, LSB, MC and LB) to create a thematic synthesis. This involved independently coding the data to identify recurring, unique and contradictory content and using the codes to independently summarise the content of the theme in a series of sub-themes [ 24 ]. The findings were discussed and agreed together by the two independent reviewers for each category of impact. The final analysis for all categories of impact was discussed and refined by CMA, MW and LN.

Study characteristics

A total of 2704 articles were identified after removal of duplicates, of which 20 met the review eligibility criteria (Fig. 1 ) [ 13 , 16 , 17 , 25 , 26 , 27 , 28 , 29 , 30 , 31 , 32 , 33 , 34 , 35 , 36 , 37 , 38 , 39 , 40 , 41 ]. The most common reasons for exclusion were that the study population did not involve clinical academics, that is clinicians who were also involved in research activity; or the study assessed the amount of research interest/activity, rather than the impacts of this activity (Additional File  3 ). Of the included articles, nine reported qualitative data [ 17 , 25 , 29 , 30 , 31 , 33 , 36 , 38 , 41 ], three reported quantitative data [ 28 , 35 , 37 ] and eight reported a mixture of both [ 13 , 16 , 26 , 27 , 32 , 34 , 39 , 40 ]. Sixteen were peer reviewed journal articles and four were organisational reports. Publication dates ranged from 2003 to 2019 and the geographical distribution was: Europe (including the UK) 10; North America 5; Australasia 4; Middle East 1 (Table  3 ).

Participants

A variety of healthcare professions outside medicine were included. Eight articles involved mixed professional groups, most commonly nursing, midwifery and one or more of the allied health professions. Nine articles were specific to nursing, two to allied health professions and one to pharmacy. The nature of the clinical academic activity was not consistent. All articles discussed clinicians conducting research in clinical practice, however some also incorporated formal educational components at masters or doctoral level and others involved short programmes of research training and/or mentorship (Table 3 ).

Methodological assessment

Outcomes of the MMAT assessment [ 20 ] are shown in Fig.  2 and are available in full via the Open Science Framework [ 18 ]. No articles met all quality assessment criteria, although three qualitative studies were rated as having a single area of concern [ 31 , 38 , 41 ]. Common issues with study quality and risk of bias for qualitative and mixed methods studies were a lack of clarity in how the findings were derived from the data and a lack of coherence between data, analysis and interpretation. Common issues with quantitative and mixed methods studies were a lack of information about the measurement tools/methods and a lack of consideration of response bias. Common issues across all study types were inadequate sampling methods and a lack of reporting of ethics/other approvals.

figure 2

Quality assessment scores for included articles using the Mixed Methods Assessment Tool

Reported impacts of clinical academic activity

Reported impacts were categorised into seven main themes based on the modified section headings for the VICTOR framework (Table 2 ). The distribution of the reported impact themes across the articles included in this review are presented in Table  4 . Extracted and coded data for each theme was used to generate the sub-themes, which are described below and presented with illustrative excerpts from the included articles. The full framework of themes, sub-themes and additional quotes is provided in Additional File  4 . Most of the extracted impacts were positive in nature, reflecting the aims of the included papers (Table 3 ), however challenges for the individual clinical academics and their healthcare teams were also described.

Impacts for patients

The reported impacts for patients focused on beneficial changes to service provision that arose as a result of local clinical academic activity, and wider access to evidence-based healthcare as a result of the promotion of evidence-based practice across research-active teams and departments:

“This project demonstrated significant improvements in the neutropenic patient pathway, enhancing experience and outcomes for patients and a reduction in unnecessary admissions.” [ 27 ]
“Behaviours learned during the programmes provided benefits for improving the quality of care delivered within services … Respondents reported they discussed evidence with colleagues, searched the literature for evidence updates, questioned and used evidence to inform their practice following completion of the programme.” [ 16 ]

It was suggested that these practice-changes were associated with improved patient/carer experiences, particularly because clinical academic research was directed at issues that were meaningful to patients and focused on patient experience in addition to clinical outcomes:

“My research has demonstrated the benefits of these kinds of approaches both to patients’ quality of life, and patients, carers, friends and families’ experience of palliative care services.” [ 36 ]
"[It is a] synergistic relationship, as research knowledge improves the care I provide, but the close patient contact allows me to identify areas that require further research." [ 29 ]

Furthermore, clinical academics reported a drive to challenge and improve their own clinical practice and that of their team, with the goal of improving all aspects of patient care:

“Participants expressed how their involvement in the innovation increased their observation of their own clinical practice, brought the research back to practice, enhanced practice development and the clinical role overall while contributing to improved patient care.” [ 34 ]

Impacts on service provision and workforce

I. clinical service provision.

In addition to the identified impacts to patients (theme 1), changes in practice as a result of the clinical academic activity were also regarded as beneficial to the clinical service through improved care delivery and pathways . This included the introduction of new equipment, better integration of clinical teams, efficiencies, cost-savings and securing new clinical funding:

“Research positions supported projects that led to changes in service delivery models, with [one] manager commenting, ‘ … it’s amazed me that through the research grant that she got for that project, she has now generated for the Health Service recurrent money for the full time [implementation of the] … rural allied health model’.” [ 41 ]
“One participant’s intervention removes the need for GPs’ referral for physiotherapy, potentially saving ‘multimillion pounds’ across the NHS [National Health Service], and has subsequently been recognised in the NHS long-term plan.” [ 39 ]

It was reported that clinical academics remained up to date with the relevant literature and clinical guidelines and were able to translate research into practice and implement the evidence . They were also able to share these skills with their clinical team to support their colleagues in the delivery of high-quality care:

“As a clinical academic midwife my aim is to bring more evidence into practice and assist other midwives in doing the same.” [ 29 ]
“Most of this reported activity focused on reviewing published evidence in relation to clinical practice but participants also reported involvement in facilitating/enhancing research skills in other clinicians.” [ 37 ]

The need to find backfill was reported as a negative consequence of releasing clinical staff for research . In some instances, it was not possible to find backfill at the same clinical grade as the clinical academic, potentially leaving a deficit in the clinical service. However, backfill posts were also seen as opportunities for other clinicians to gain experience by acting-up into the role:

“Release from the workplace, despite the employer grant, was in some cases problematic. Finding appropriately skilled staff to cover services particularly in highly specialised areas and, or recruiting to short term, often part time, vacancies were challenges. In contrast some managers saw this as an opportunity to give other staff the chance to act up, for succession planning, or worked creatively to make release possible.” [ 16 ]
“Protected time for the APNs [Advanced Practice Nurses] (i.e., at least one day a week) to engage in research activities was crucial to the program and, at times, difficult to achieve.” [ 32 ]

The majority of the described clinical academic activity centred on short-term secondments, which elicited issues with the return to clinical practice from a research role . These included difficulties in maintaining evidence-based practice due to time constraints and a lack of opportunity to use the research skills that had been developed:

“On my return to work I was unable to continue to facilitate evidence-based practice as much as I would have liked due to time constraints. Within my working role there was no dedicated time to devote to evidence-based activity.” [ 37 ]

ii. Clinical academic workforce

Clinical academic infrastructure was described in terms of fellowship and career pathways. Steps were taken to ensure visibility of these opportunities to facilitate the development of research capacity. However, it was also noted that the absence of established clinical academic career structures resulted in a perceived lack of value of these skills and caused difficulties for clinical managers when trying to plan their service:

“Clinical and academic mentorship exposed the Chief Nurse Fellows to clinical academic career role models, which in turn raised the profile of this alternative career route.” [ 27 ]
“There’s a huge untapped workforce … with the right support and time we could be doing things more effectively and more efficiently, but that isn’t necessarily valued in organisations. We’ve got to see this many patients, (we’re) not using our skills of criticality, reflectivity; we’re not going to innovate and change practice.” [ 39 ]
"Imagine being able to continue my research and tie the results directly to clinical practice. I would like there to be an opportunity of this kind. However, there is a lack of services for nurses with higher academic qualifications who want to develop clinical practice.” [ 25 ]

Challenges encountered in balancing the clinical and research components of the role were widely reported. Clinical duties were given priority and some articles reported unclear expectations for the research roles, in contrast to established clinical job descriptions:

“A culture that prioritises practice in the current context means that the doing of nursing work only is seen as core business. This, together with the need for managers and clinicians to make quick decisions in order to achieve short term goals, operates as a disincentive to rigorous research activity at ward level. Within this context, the expectations for those with a research component in their role, is at times unclear.” [ 30 ]

In addition, some articles recounted strategies that aided the development of research skills and clinical academic roles:

“Also, I think the option of having 0.5FTE [full-time equivalent) backfill was good, as it allowed greater flexibility for staff who have roles that are difficult to backfill full-time, also I thought it was useful to have more thinking time, and time to access support, get feedback etc. Full-time research can be very intense especially when you are not conditioned for it.” [ 13 ]

Where clinical academics identified access to resources and support , this was identified as a positive asset. However, a number of articles recalled issues with a lack of manager awareness and support, and insufficient funding for research activities and computer software:

“I have a very supportive divisional head nurse and have been appointed into a trailblazer post; we haven’t got anything similar within the organisation. So there’s real potential to forge out innovative ways in which clinical academics can fulfil that remit of working in clinical practice and undertaking research, but also pave the way for others that want to come up.” [ 39 ]
“Biggest challenge: Getting managers on board, in particular releasing staff to take advantage of internship opportunities offered by HEE [Health Education England] Wessex, and recognising that research is essential to the core business of the Trust.” [ 17 ]
“It’s not the scheme, but greater staff access to relevant software (such as SPSS) would be useful.” [ 13 ]

Impacts to the research profile, culture and capacity

Research profile, culture and capacity were interlinked and several of the reported impacts spanned all aspects of this theme. However, winning research funding and other awards , publication of journal articles and clinical guidelines, and conference presentations were primarily considered as contributing to the organisation’s research profile:

“Since completing the programme - One Chief Nurse Fellow was the first UK nurse to be recognised by the Daisy Foundation and has received a Daisy Award for Extraordinary Nurses. Others have also received nominations and were shortlisted for national nursing awards. In addition, two of the projects are featured on The Academy of Fabulous NHS Stuff.” [ 27 ]
“As a result of the research activities, seven manuscripts were submitted and accepted for peer-reviewed publications.” [ 28 ]

The provision of research training and support , and organisational-wide research engagement and participation were commonly reported as beneficial impacts to the organisations’ research culture and capacity:

“Improved attitudes towards research were noted by a clinician, ‘ … research isn’t this incredibly difficult thing that only very special people can do. Actually, it’s attainable by many and it was quite inspiring actually … I don’t know that that would have been their view prior to this position developing that profile.” [ 41 ]
“It’s about allowing people to engage with research and become enthused by it. It’s also about having the right leaders who are able to take the step back and say: ‘This is a good use of your time’. You can find better ways of giving care if you have a culture that values research.” [ 36 ]

There were also aspects where the lack of research culture within an organisation was seen as a challenge for research involvement:

“There is the perception that doing research is an ‘imposition, on clinical nursing staff, that it is not ‘real nursing work’; rather, it is a ‘luxury’. When invited by nurse researchers to participate in research activities, clinical nurses often say ‘I don’t have time for this’ and the general attitude is ‘we’ll think about it’, ‘if I have time’, or ‘tell us about the result’. Indeed, in some instances there is a perception amongst discussion group members that projects are undermined through gate-keeping behaviour and lack of support by the ‘sceptics’.” [ 30 ]

Clinical academics were able to promote evidence-based practice among their peers, for example by sharing resources and setting up journal clubs or other special interest groups. This facilitated a shift towards research becoming embedded in practice :

“I now exhort other colleagues to question day-to-day practice and we have introduced a journal club.” [ 29 ]
“The group process allowed scholars to participate in joint problem solving and enhanced their ability to apply current research to questions arising from clinical practice. The scholars were expected to serve as clinical resources to others in the healthcare system.” [ 33 ]

Three additional elements were identified that encompassed both research profile and research culture/capacity. These were: building local and external collaborations , creating visible clinical academic opportunities and being seen as an attractive place to work :

“We have increasing numbers of staff involved in research activity, studying for MRes and PhDs, and their research is closely related to their professional practice and aims to improve care. We have five research themes with [the healthcare organisation’s] nurses/midwives leading these and staff linking into these themes for their masters or doctoral study and we are starting to build groups of staff at different points on a clinical academic career pathway. Many staff present their research nationally and internationally and publish widely and some are part of national expert groups, linked to their research.” [ 17 ]
“People can choose where they want to work. They’ll be looking for organisations that are aspirational. So actually offering innovative career pathways that can intellectually challenge, but also have that direct patient care element, is going to be attractive to a lot of people.” [ 39 ]

Economic impacts

The funding required to support the clinical academic activity was generally sourced from outside the clinical organisation. Reported benefits of receiving research funding included dedicated time for research training and activity and bringing in additional money to the clinical service:

“Increased grant income – the value of successful non-medic research grants in 2014–2015 (the last financial year the outcome of all grant applications is known) was £923,495. Appointments of clinical academic posts were achieved by securing external grant funding, use of research capability funding to pump prime, and commitment to 50:50 funding from academic partners.” [ 17 ]

However, there were also issues where funding for research was not available, or was repurposed from clinical budgets :

“Management are more than happy to support research initiatives in principle, however, [they are] usually unable to provide [this] support as they have extremely tight budgets and other clinical management demands.” [ 30 ]
“The initiative was resourced by the reallocation of nursing/ODP vacancy funding within each clinical division.” [ 27 ]

As discussed in theme 2 (provision of clinical service), it was proposed that clinical academic activity was associated with financial gains in terms of cost-savings and efficiencies , although the difficulty of capturing this data was recognised:

“The amount of money saved by using the scanner and avoiding catheterisation was estimated to be around £1.2m per year. This did not include the cost of bacteraemia attributed to urinary tract infections. Savings associated with using a scanner, such as fewer treatment delays and overnight stays in hospital, were recognised as additional savings. The set up and running costs of a scanner were estimated to be met within six months to two years, after which significant ongoing cost efficiencies would be realised over its eight to ten year lifespan.” [ 17 ]
“The work had led to a decrease in patients’ clinical stay following surgery from six to four days, resulting in savings that allowed additional needs and demands to be met. We were very open and transparent with the data, and clinical practice changed. We calculated that there was a total of 28,000 bed-days saved per year as a result of this work.” [ 36 ]

Financial implications for the clinical academics were also reported. These were largely negative and included reductions in salary and pension contributions, and the need to self-fund conference attendance:

“Well one of the big decisions I had to make about whether or not to accept the role (associated with a research training award) was the hours and the money because it’s moving to full time, which is fine... Therefore, by moving to full time but losing my enhancements I’ll be on around the same as I get on a good month when I have done lots of nights and weekends. But by working as I’m doing there won’t be any opportunity to do extra shifts, any overtime. So a lot of it was money.” [ 16 ]
“The presenters had to either self-fund their travel and conference registration, or apply for travel scholarships through internal or external opportunities.” [ 28 ]

Impacts on staff recruitment and retention

The lack of clinical academic career opportunities was noted as a challenge that individuals wishing to maintain a dual role needed to negotiate. This related to the sub-theme ‘balancing clinical and academic components of the role’ discussed in theme 2 (impacts for the clinical academic workforce), and was identified as a potential driver for individuals to return to full-time clinical work, or move into purely academic roles after completion of their clinical academic activity:

“The organisational system is perceived as unfamiliar with, and unsupportive of, non-physician clinician scientist positions, and, in consequence, active job crafting is necessary to obtain positions in which such individuals can exert both roles. Dual positions are often a personal combination of jobs instead of being offered from within one institution.” [ 31 ]
"Unfortunately, I wasn’t offered the chance to implement the results of my study in my organisation, due to the lack of development positions. So, as a result, I have applied for, and been given, a position as a teacher at the university college." [ 25 ]

A particular challenge for healthcare managers was the need to provide backfill or make other arrangements to enable the release of clinical staff for research activities. The need to recruit to backfill posts was discussed in theme 2 (impact to clinical service provision), specifically the sub-themes: release of clinical staff for research and return to clinical practice from a research role.

Several articles reported strategies to support clinical academics and increase awareness and access to clinical research opportunities. Where successful, it was suggested that these aided the retention and career progress of staff who were involved in clinical academic activity. Such strategies also contributed to job satisfaction and recruitment more generally and were closely linked with the impacts to the organisation’s research profile, research culture and capacity (theme 3):

“Since completing their year as a fellow, the entire pilot cohort still works within the organisation, with five of them having moved into junior leadership positions. Although we cannot assume that this would not have been their career trajectory had they not undertaken this fellowship, the skills developed and demonstrated through the initiative are essential for the job specifications of more senior posts.” [ 27 ]
“Interviewees reported that their department was seen as a more ‘attractive employer’ and was ‘attracting higher calibre staff’. Clinicians described staying in the health service to undertake research, ‘because these opportunities do exist, these really fabulous clinicians that we have just might stay’.” [ 41 ]

Impacts to knowledge exchange

Contributions to knowledge transfer were reported in all articles, and there was a large overlap with theme 3 (research profile, culture and capacity). Knowledge exchange activities included formal dissemination , such as conference presentations and posters, publications, being an invited speaker and winning prizes and awards that further highlighted the value of the work. It was recognised that there would be a delay between completion of the research activity and delivery of these research outputs:

“Interviewees provided numerous and diverse examples of presentations at grand rounds, poster presentations and oral presentations at both local and international conferences, including one interviewee who noted her team had presented their research project findings at three international and three local conferences. Another interviewee stated, ‘We have presented at a couple of conferences and we presented at a … convention or meeting and we actually got an award for our poster’.” [ 26 ]
“Individuals need time in the role as well to get some momentum, get the relationships in the department, get the research programs going and there's usually a delay until you start to see the pure research outputs.” [ 41 ]

Clinical academics also played a role in developing networks and collaborations aiding the transfer of knowledge among clinical and academic communities and patient populations. This included sharing their expertise with clinical colleagues and other practice improvement strategies aimed at the implementation of clinical guidelines and delivery of evidence-based practice. Again there was a large overlap with theme 2 (impacts to clinical service provision), specifically the sub-theme translation of research into practice and evidence implementation:

“What I notice clearly is that I’m very well informed about scientific evidence and sharing this information with my colleagues. [I ask them] did you read this? And [I] pass on knowledge in that way.” [ 31 ]
“After being able to demonstrate the success of the approach locally, David was asked to help with a national roll out, organising with colleagues an audit of all 22 orthopaedic units across Scotland over 12 weeks.” [ 36 ]

Impacts to the clinical academic

Many of the themes and sub-themes of impact discussed above had also had a direct influence on the individuals involved in clinical academic activities. Developing networks and collaborations, discussed in theme 6 (knowledge exchange), and building local and external collaborations, discussed in theme 3 (research profile, culture and capacity), were similarly interpreted as individual clinical academics developing their networks and influence :

“I think the impact of the role on me has been quite incredible. … how much you learn about the different disciplines and then develop those networks … it’s been a huge learning curve.” [ 41 ]

Furthermore, clinical academics reported a change in their attitude to clinical practice, with greater reflection and questioning of established practice, which was also reflected in theme 1 (impacts for patients), particularly in terms of improved clinical practice and access to evidence-based healthcare:

“They felt that they had developed from “doers” to “thinkers”, in that they felt more aware of and reflective in relation to their colleagues … The nurses perceived progress in acquiring new knowledge, in spite of language barriers, and recognised the value of scientific knowledge for clinical practice. They experienced healthcare in a ‘new light’ through their knowledge development.” [ 25 ]

The development of research and leadership skills was identified as a beneficial effect of being involved in clinical academic activity. In some articles, it was suggested that this unlocked new career opportunities for the individuals involved, however in many instances there were no existing roles within the organisation for the research-active clinicians to aspire to, as discussed in theme 5 (recruitment and retention):

“The majority agreed or strongly agreed that they felt more confident developing a research question (94%/49), searching (87%/45) and appraising (90%/47) literature, challenging practice using evidence (85%/44), assisting others to use critical appraisal skills (79%/41) and engage in the clinical academic training pathway (87%/45).” [ 16 ]
“Despite their achievements during the PhD, many participants expressed anxieties about their future careers, having been made to move aside clinically in order to progress their academic ambitions, rather than being able to develop their academic and clinical skills in tandem. For example a dietician said: Recently I’ve had to step out of my area of expertise … I’m just doing general, allergies, weight management, which is not my area, but I need to pay the mortgage.” [ 39 ]

Clinical academics appeared to find self-fulfilment with their roles, and described a passion for their contributions to clinical research:

“It is actually exciting to learn that the world does not work the way you thought it did”. 70 [ 38 ]. “I want to do this for me, but I also want to do it for my daughters to show that women can be in science and can lead in these fields and yes we might have to juggle family things and children, but you can do it.” [ 39 ]
“What accomplishments are you most proud of? Knowing that I am now a subject expert – I get phone calls asking, `How would you handle this?’.” [ 40 ]

The challenges and sacrifices of clinical academic roles were widely recounted, examples of which have been included in the previous themes. Financial implications were discussed in theme 4 (economic impacts, sub-theme financial implications for the clinical academic) and the challenge of combining clinical and academic work and identities were included in theme 2 (impacts for the clinical academic workforce, sub-theme balancing clinical and academic components of the role) and theme 5 (impacts on staff recruitment and retention, sub-theme maintaining a dual role). A lack of time for clinical academic activity was another widely reported issue:

“When you are really interested in something or passionate about it, you use whatever time you have, even if it means writing your proposal after your regular hours at home.” [ 38 ]
“The time taken to do research is often underestimated and considerable time and effort is often put into preparing a grant application which ultimately may not be successful. Focusing on meeting deadlines and the progress of a project means that less attention can be given to other aspects of work roles. Ultimately, doing research without adequate support or funding becomes a constant juggle.” [ 30 ]

Finally, a few articles reported reflections on what it takes to be a clinical academic . These included the qualities of determination, tenacity and resilience, and serendipity:

“The demanding expectations surrounding a clinical academic role were described by interviewees (participants and managers) and the characteristics and behaviours that were perceived as required for success. These included confidence, doggedness and resilience, reflective skills, criticality, and growing political know-how to better navigate organisations.” [ 16 ]
“Mentorship (from nursing, midwifery and medical colleagues), determination, tenacity, resilience and serendipity have been key factors in achieving success.” [ 17 ]

This systematic review identified 20 articles that discussed elements of the impact of clinical academic activity among healthcare professionals outside medicine. With the addition of a theme for the impacts to the clinical academic, all reported types of impact could be mapped to the VICTOR framework creating the following themes: impacts for patients; impacts for service provision and workforce; impacts to research profile, culture and capacity; economic impacts; impacts on staff recruitment and retention; impacts to knowledge exchange; and impacts to the clinical academic. In order to develop and evaluate clinical academic roles for healthcare professionals outside medicine, the range of impacts of this clinical academic activity need to be understood and valued by healthcare leaders and managers. This review has systematically identified and mapped the nature of the impacts reported in the literature, and forms a valuable resource for healthcare services looking to develop and evaluate these roles at local and national levels.

Within the main headings of impact described above, we identified several similar sub-themes that cut across the different categories of impact. Sub-themes described the content of each of the categories of impact and included perceived enablers of creating the desired impact and associated detrimental features. Notably, the sub-themes that reflected the challenge of maintaining or balancing the clinical and academic components of the role contributed to four main themes. Within clinical service provision (theme 2i), this related to the need for clinical services to manage both the release of clinical staff for research, and their return to clinical practice after research secondments. For the clinical academic workforce (theme 2ii), this led to individuals and team members being required to adapt to the different pace and duties associated with research and clinical work. The process of showcasing a visible clinical academic pathway (that incorporated both research and clinical activities) was identified as key feature of building research profile, culture and capacity (theme 3), and being able to offer suitable clinical academic posts was important for staff recruitment and retention (theme 5). Finally, being able to work and develop in both clinical and academic roles offered self-fulfilment for the individual clinical academic (theme 7). Similarly, the creation and implementation of new evidence was also a component of several themes, as was the development of collaborations and networks.

The multifaceted nature of research impact identified in this review illustrates that different aspects of clinical academic activity may be perceived as having both positive and negative impacts. Furthermore, these conflicting impacts may apply to the same individual or across different stakeholders. Different aspects of research impact may be more or less important in different contexts and the relative value of these different impacts will need to be considered to enable meaningful evaluation [ 42 , 43 , 44 ].

This systematic review was deliberately broad in scope to allow the identification of the whole range of impacts associated with clinical academic activity. The lack of an agreed and consistently used definition of clinical academic proves problematic and has been discussed elsewhere [ 45 , 46 ]. Clinical academic activity (defined here as the involvement of practising clinicians in research) in the included articles incorporated a range of research fellowships and research training programmes, and/or in-practice mentorship and research support. The aims of the articles varied. Many described and evaluated specific interventions that were aimed at increasing research activity among healthcare professional group(s), while others provided summary case studies of individuals who had been successful in a clinical academic role.

The inclusion of grey literature increased the breadth of the review, particularly given the finding that the impacts of non-medical healthcare research are underrepresented in the academic literature [ 47 ]. However, it is acknowledged that the methodological quality of the included institutional reports was lower than the standards for peer-reviewed publication. Data obtained from the institutional reports were largely positive reflections of strategies that had been put into place to encourage and support research activity among healthcare professionals outside medicine. Expressions of the less positive aspects of these strategies may therefore have been excluded or not collected by the authors. Importantly, the distribution of the seven identified themes of impacts did not differ between organisational reports and peer review journals, illustrating that the types of impact that were considered important by study participants and organisations were similar.

Existing reviews on the impact of clinical academic activity have focused on individual clinical groups within the non-medical workforce [ 10 ] or at the level of the healthcare institution [ 1 ]. Boaz et al. found that healthcare organisations which deliberately integrated research into their practice and fostered research engagement reported improved healthcare performance including clinical outcomes and processes of care, and our review also identified similar improvements. However, Boaz et al.’s review included papers focused primarily on research led by clinical academic doctors, and the impacts of the research processes on the clinical teams and the individuals involved was not reported [ 1 ]. Wenke and Mickan identified four themes of impact associated with allied health research positions based in clinical settings: increased individual research skills and participation; increased research activity; improved research culture and attitudes; and increased team and organisational level skills [ 10 ]. These features were also described within the current review under the themes: impact for patients; impacts to the clinical academic; impacts to research profile, culture and capacity; and impacts for service provision and workforce. In addition, we also identified impacts on staff recruitment and retention, knowledge exchange and economic impacts.

The distribution of impacts reported in the current review did not differ in relation to the clinical groups involved in each of the included articles, indicating that similar methods of capturing the impact of clinical academic research activity could be applied across the professions outside medicine, rather than being discipline-specific. A similar systematic review search strategy could also be applied to assess the reported impact of clinical academic activity by alternative and complementary therapy practitioners.

The impacts reported in the included articles were largely qualitative reports. Some studies incorporated quantitative data capture tools, such as the Research Capacity and Culture tool [ 48 , 49 ] and the WReN (Wessex Research Network) spider [ 50 , 51 ], or counts of publications, presentations and awards. The quantitative measures appeared to explore a discrete component of research impact, whereas the qualitative data provided a broad picture of the impacts in different contexts and uncovered both intended and unintended consequences of the research activity. Reed et al. proposed five impact evaluation typologies (experimental and statistical methods, systems analysis methods, textual, oral and arts-based methods, indicator-based approaches, and evidence synthesis approaches); our findings illustrate the first and third of these categories [ 44 ]. Future work should explore standardised methods of capturing the research impact that address the full range of impacts identified in this systematic review and are specific to the context of clinical academics outside medicine. With the desire for at least 1% of the UK NMAHP workforce to be clinical academics by 2030 [ 8 ], policy makers will need to consider, promote or potentially mitigate the different types of impacts that this systematic review identified in connection with these roles.

The VICTOR framework [ 22 ] was used to guide data extraction in the current review, with a priori modifications to include the impacts to the individual clinical academic and to merge the category relating to the organisation’s influence and reputation with the category for research profile, culture and capacity. No further refinements were made during the analysis process as all reported impacts were able to be mapped. While other research impact frameworks exist [ 14 , 44 ], our findings suggest that the VICTOR tool may be a good starting point for capturing the nature of research impact that is important for clinical academic healthcare research outside the medical professions, and it is already endorsed for use in the UK [ 23 ]. The identification of sub-themes that crossed one or more of the main impact themes indicate that these may be key areas to explore, particularly for organisations looking to support and increase academic activity among these clinical groups.

Limitations

The systematic review team comprised research-active clinicians from professions outside medicine, and therefore the review took place through this lens. Steps were taken to facilitate objectivity, including: a clearly defined protocol [ 18 ]; two or more reviewers independently conducting each stage of the review; inclusion of reviewers with different levels of clinical and research experience; and the provision of oversight by senior (clinically-active) academics. We acknowledge that the thematic analysis and coding of the extracted data may have been interpreted differently by reviewers from different backgrounds. Furthermore, the definition of clinical academic activity used in our review differs from that used elsewhere [ 45 , 52 ]. However, the absence of an agreed definition has been recognised [ 45 ] and the overlap of our findings with the existing literature support our review processes and findings [ 1 , 10 ].

As with other reviews of research impact, there is a risk that relevant studies were excluded due to poor indexing in the medical databases [ 14 ]. We took the additional step of including grey literature searches in both established repositories and through a naïve web search engine in an attempt to maximise the identification of eligible articles, but accept that articles may still have been missed if they were not identified through these mechanisms. We did not formally screen the reference lists of the included articles, and accept that this may have yielded additional studies.

The MMAT tool used for methodological assessment of the included articles was designed for the appraisal of mixed methods studies [ 20 ], although our review included mixed methods, qualitative, and quantitative articles. The MMAT was chosen to allow the same criteria to be applied across all included articles using the relevant sub-sections as appropriate. Quality assessment scores were not used to determine how the extracted data was incorporated into the thematic synthesis. We acknowledge that the presented synthesis therefore includes the findings from studies and organisational reports across the spectrum of methodological quality. However, no theme or sub-theme solely comprised data from articles that were assessed to be of lower quality.

The application of alternative impact frameworks would have yielded different theme headings, as these were taken directly from the VICTOR terminology. However, the use of an established impact assessment tool aided transparency and consistency of data extraction and categorisation. The coding and resulting theme descriptions were created through an inductive process that explored the meaning of the extracted data, rather than looking to specifically fit it to the VICTOR headings. The descriptions of the content of each theme illustrate the available data.

Finally, the aim of our review was to capture the range of impacts associated with clinical academic activity outside medicine. We have highlighted the key themes of impact and described the characteristic content of these themes. While this work contributes to the existing discussion around research impact, it does not explore the utility of capturing and comparing the reported impacts using a standardised method in a clinical research setting.

Twenty articles were identified that reported the impact of clinical academic activity among the healthcare professions outside medicine. These impacts could be mapped using a modified VICTOR framework and were classified as: impacts for patients; impacts for service provision and workforce; impacts to research profile, culture and capacity; economic impacts; impacts on staff recruitment and retention; impacts to knowledge exchange; and impacts to the clinical academic. With our addition of impacts to clinical academics, the VICTOR tool may be a useful starting point for individuals and organisations to record the impact of their research activity, although further work is needed to establish its utility. This review identified several sub-themes of impact that crossed one or more of the main themes: the challenges and benefits of balancing clinical and academic roles; the creation and implementation of new research evidence; and the development of collaborations and networks. These are likely to be key areas for organisations to explore when looking to support and increase academic activity among healthcare professionals outside medicine.

Availability of data and materials

The datasets generated and analysed during the current review are available in the Open Science Framework (OSF) repository ( https://osf.io/gj7se ). All other relevant data are included as supplementary files.

Abbreviations

allied health professional (includes: art therapists; drama therapists; music therapists; chiropodists/podiatrists; dietitians; occupational therapists; operating department practitioners; orthoptists; osteopaths; paramedics; physiotherapists; prosthetists and orthotists; radiographers; and speech and language therapists)

Association of UK University Hospitals

Department of Health and Social Care (UK)

Mixed Methods Appraisal Tool

nurses, midwives and allied health professionals

Nursing, Midwifery and Allied Health Professions Research Unit (Scotland)

preferred reporting items for systematic reviews and meta-analyses

making Visible the ImpaCT Of Research framework, a research impact capture tool

Boaz A, Hanney S, Jones T, Soper B. Does the engagement of clinicians and organisations in research improve healthcare performance: a three-stage review. BMJ Open. 2015;5(12):e009415. https://doi.org/10.1136/bmjopen-2015-009415 .

Article   PubMed   PubMed Central   Google Scholar  

Jonker L, Fisher SJ, Dagnan D. Patients admitted to more research-active hospitals have more confidence in staff and are better informed about their condition and medication: results from a retrospective cross-sectional study. J Eval Clin Pract. 2020;26(1):203–8. https://doi.org/10.1111/jep.13118 .

Article   PubMed   Google Scholar  

Majumdar SR, Roe MT, Peterson ED, Chen AY, Gibler WB, Armstrong PW. Better outcomes for patients treated at hospitals that participate in clinical trials. Arch Intern Med. 2008;168(6):657–62. https://doi.org/10.1001/archinternmed.2007.124 .

Jonker L, Fisher SJ. The correlation between National Health Service trusts’ clinical trial activity and both mortality rates and care quality commission ratings: a retrospective cross-sectional study. Public Health. 2018;157:1–6. https://doi.org/10.1016/j.puhe.2017.12.022 .

Article   CAS   PubMed   Google Scholar  

Medical Schools Council. Clinical academia. 2018. https://www.medschools.ac.uk/our-work/clinical-academia . Accessed 4 Mar 2020.

Healthcareers NHS. Clinical academic medicine. 2017. https://www.healthcareers.nhs.uk/explore-roles/doctors/career-opportunities-doctors/clinical-academic-medicine . Accessed 1 Oct 2020.

National Institute for Health Research. Building a research career: A guide for aspiring clinical academics (excluding doctors and dentists) and their managers. 2016. https://www.clahrc-eoe.nihr.ac.uk/2016/02/building-a-research-career-a-guide-for-aspiring-clinical-academics-and-their-managers-from-the-nihr/ . Accessed 16 Oct 2020.

Westwood G, Richardson A, Latter S, Macleod Clark J, Fader M. Building clinical academic leadership capacity: sustainability through partnership. J Res Nurs. 2018;23(4):346–57. https://doi.org/10.1177/1744987117748348 .

National Institute for Health Research. HEE-NIHR Integrated Clinical Academic Programme. 2019. https://www.nihr.ac.uk/explore-nihr/academy-programmes/hee-nihr-integrated-clinical-academic-programme.htm . Accessed 4 Mar 2020.

Wenke R, Mickan S. The role and impact of research positions within health care settings in allied health: a systematic review. BMC Health Serv Res. 2016;16(1):355. https://doi.org/10.1186/s12913-016-1606-0 .

Hill NL, Yevchak A, Kolanowski AM, Penrod FJ, Milone-Nuzzo PF, Sawyer AM, et al. What it takes: perspectives from developing nurse scientists. J Nurs Educ. 2014;53(7):403–9. https://doi.org/10.3928/01484834-20140617-01 .

Munro S, Hendrix CC, Cowan LJ, Battaglia C, Wilder VD, Bormann JE, et al. Research productivity following nursing research initiative grants. Nurs Outlook. 2019;67(1):6–12. https://doi.org/10.1016/j.outlook.2018.06.011 .

Wenke R, Weir KA, Noble C, Mahoney J, Mickan S. Not enough time for research? Use of supported funding to promote allied health research activity. J Multidiscip Healthc. 2018;11:269–77. https://doi.org/10.2147/JMDH.S157034 .

Cruz Rivera S, Kyte DG, Aiyegbusi OL, Keeley TJ, Calvert MJ. Assessing the impact of healthcare research: a systematic review of methodological frameworks. PLoS Med. 2017;14(8):e1002370. https://doi.org/10.1371/journal.pmed.1002370 .

Adam P, Ovseiko PV, Grant J, Graham KEA, Boukhris OF, Dowd AM, et al. ISRIA statement: ten-point guidelines for an effective process of research impact assessment. Heal Res Policy Syst. 2018;16(1):8. https://doi.org/10.1186/s12961-018-0281-5 .

Article   Google Scholar  

Hiley J, Begg C, Swift A, Topping A. West midlands clinical academic careers Programmes for nurses, midwives, allied health professions, pharmacists and healthcare scientists (NMAHPPS). Evaluation Report 2018. https://www.birminghamhealthpartners.co.uk/education/clinical-academic-careers-programmes/ . Accessed 6 Mar 2020.

Association of UK University Hospitals. Transforming healthcare through clinical academic roles in nursing, midwifery and allied health professions: A practical resource for healthcare provider organisations. 2016. https://councilofdeans.org.uk/2016/11/transforming-healthcare-through-clinical-academic-roles-in-nursing-midwifery-and-allied-health-professions/ . Accessed 18 Oct 2020.

Newington L, Wells M, Alexander CM. What are the impacts of clinical academic activity among practising healthcare professionals outside medicine? A systematic review (protocol). OSF Registry. 2019; https://osf.io/gj7se .

Moher D, Shamseer L, Clarke M, Ghersi D, Liberati A, Petticrew M, et al. Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015 statement. Syst Rev. 2015;4. https://doi.org/10.1186/2046-4053-4-1 .

Hong QN, Pluye P, Fàbregues S, Bartlett G, Boardman F, Cargo M, et al. Improving the content validity of the mixed methods appraisal tool: a modified e-Delphi study. J Clin Epidemiol. 2019;111:49–59. https://doi.org/10.1016/j.jclinepi.2019.03.008 .

Walsh D, Downe S. Appraising the quality of qualitative research. Midwifery. 2006;22(2):108–19. https://doi.org/10.1016/j.midw.2005.05.004 .

Cooke J, Jones N, Holliday J. VICTOR (making visible the ImpaCT of research) pack. A resource for healthcare organisation 2019. https://www.e-repository.clahrc-yh.nihr.ac.uk/visible-impact-of-research/ . Accessed 11 Dec 2019.

National Institute for Health Research. A VICTOR-y for measuring research impact in the NHS. 2019. https://www.nihr.ac.uk/blog/a-victor-y-for-measuring-research-impact-in-the-nhs/12081 . Accessed 6 Nov 2020.

Thomas J, Harden A. Methods for the thematic synthesis of qualitative research in systematic reviews. BMC Med Res Methodol. 2008;8(1):45. https://doi.org/10.1186/1471-2288-8-45 .

Bäck-Pettersson S, Jensen KP, Kylén S, Sernert N, Hermansson E. Nurses’ experiences of participation in a research and development programme. J Clin Nurs. 2013;22(7-8):1103–11. https://doi.org/10.1111/j.1365-2702.2012.04297.x .

Black AT, Ali S, Baumbusch J, McNamee K, Mackay M. Practice-based nursing research: evaluation of clinical and professional impacts from a research training programme. J Clin Nurs. 2019;28(13-14):2681–7. https://doi.org/10.1111/jocn.14861 .

Bramley L, Manning JC, Cooper J. Engaging and developing front-line clinical nurses to drive care excellence: evaluating the chief nurse excellence in care junior fellowship initiative. J Res Nurs. 2018;23(8):678–89. https://doi.org/10.1177/1744987118808843 .

Chan R, Gardner G, Webster J, Geary A. Building research capacity in the nursing workforce: the design and evaluation of the nurse researcher role. Aust J Adv Nurs. 2010;27:62–9.

Google Scholar  

Department of Health and Social Care. Developing the Role of the Clinical Academic Researcher in the Nursing, Midwifery and Allied Health Professions. 2012. https://www.gov.uk/government/publications/developing-the-role-of-the-clinical-academic-researcher-in-the-nursing-midwifery-and-allied-health-professions . Accessed 27 Sep 2020.

Higgins I, Parker V, Keatinge D, Giles M, Winskill R, Guest E, et al. Doing clinical research: the challenges and benefits. Contemp Nurse. 2010;35(2):171–81. https://doi.org/10.5172/conu.2010.35.2.171 .

Kluijtmans M, de Haan E, Akkerman S, van Tartwijk J. Professional identity in clinician-scientists: brokers between care and science. Med Educ. 2017;51(6):645–55. https://doi.org/10.1111/medu.13241 .

Leung D, Widger K, Howell D, Nelson S, Molassiotis A. Mentoring advanced practice nurses in research: recommendations from a pilot program. Can Oncol Nurs J. 2012;22(1):31–41. https://doi.org/10.5737/1181912x2213135 .

Brooks Carthon J, Holland S, Gamble K, Rothwell H, Pancir D, Ballinghoff J, et al. Increasing research capacity in a safety net setting through an academic clinical partnership. J Nurs Adm. 2017;47(6):350–5. https://doi.org/10.1097/NNA.0000000000000491 .

McKee G, Codd M, Dempsey O, Gallagher P, Comiskey C. Describing the implementation of an innovative intervention and evaluating its effectiveness in increasing research capacity of advanced clinical nurses: using the consolidated framework for implementation research. BMC Nurs. 2017;16(1):21. https://doi.org/10.1186/s12912-017-0214-6 .

Nazer LH, Tuffaha H, Jaddoua S. A program to increase research productivity among hospital pharmacists. J Pharm Pract. 2017;30(3):336–41. https://doi.org/10.1177/0897190016645432 .

Nursing Midwifery and Allied Health Professsions Research Unit. A clinical academic approach for nurses, midwives and allied health professionals - it’s a no-brainer! 2017. https://www.nmahp-ru.ac.uk/media/microsites/nmahp-ru/documents/Clinical-academic-approach-for-nurses-midwives%2D%2DAHPs.pdf . Accessed 27 Apr 2020.

Pomeroy VM, Tallis RC, Stitt E. Dismantling some barriers to evidenced-based rehabilitation with “hands-on” clinical research secondments. Physiotherapy. 2003;89(5):266–75. https://doi.org/10.1016/S0031-9406(05)60038-7 .

Siedlecki SL, Albert NM. Research-active clinical nurses: against all odds. J Clin Nurs. 2017;26(5-6):766–73. https://doi.org/10.1111/jocn.13523 .

Trusson D, Rowley E, Bramley L. A mixed-methods study of challenges and benefits of clinical academic careers for nurses, midwives and allied health professionals. BMJ Open. 2019;9(10):e030595. https://doi.org/10.1136/bmjopen-2019-030595 .

Turkel MC, Ferket K, Reidinger G, Beatty DE. Building a nursing research fellowship in a community hospital. Nurs Econ. 2008;26(1):26–34.

PubMed   Google Scholar  

Wenke RJ, Ward EC, Hickman I, Hulcombe J, Phillips R, Mickan S. Allied health research positions: a qualitative evaluation of their impact. Heal Res Policy Syst. 2017;15(1):6. https://doi.org/10.1186/s12961-016-0166-4 .

Greenhalgh T, Raftery J, Hanney S, Glover M. Research impact: a narrative review. BMC Med. 2016;14(1):78. https://doi.org/10.1186/s12916-016-0620-8 .

Bayley JE, Phipps D. Building the concept of research impact literacy. Evid Policy. 2019;15(4):597–606. https://doi.org/10.1332/174426417X15034894876108 .

Reed MS, Ferré M, Martin-Ortega J, Blanche R, Lawford-Rolfe R, Dallimer M, et al. Evaluating impact from research: a methodological framework. Res Policy. 2021;50(4):104147. https://doi.org/10.1016/j.respol.2020.104147 .

Carrick-Sen D, Moore A, Davidson P, Gendong H, Jackson D. International perspectives of nurses, midwives and allied health professionals clinical academic roles: are we at tipping point? Int J Pract Learn Heal Soc Care. 2019;7(2):1–15. https://doi.org/10.18552/ijpblhsc.v7i2.639 .

Newington L, Alexander CM, Wells M. What is a clinical academic? Qualitative interviews with healthcare managers, research-active nurses and other research-active healthcare professionals outside medicine. J Clin Nurs. 2021;00. https://doi.org/10.1111/jocn.15624 .

Kelly D, Kent B, McMahon A, Taylor J, Traynor M. Impact case studies submitted to REF 2014: the hidden impact of nursing research. J Res Nurs. 2016;21(4):256–68. https://doi.org/10.1177/1744987116646153 .

Holden L, Pager S, Golenko X, Ware RS. Validation of the research capacity and culture (RCC) tool: measuring RCC at individual, team and organisation levels. Aust J Prim Health. 2012;18(1):62–7. https://doi.org/10.1071/PY10081 .

Matus J, Wenke R, Hughes I, Mickan S. Evaluation of the research capacity and culture of allied health professionals in a large regional public health service. J Multidiscip Healthc. 2019;12:83–96. https://doi.org/10.2147/JMDH.S178696 .

Luna Puerta L, Apfelbacher C, Smith H. Proliferation of the WReN spider, an instrument to measure health professionals’ experience of research: a bibliographic study. BMC Med Educ. 2019;19(1):255. https://doi.org/10.1186/s12909-019-1693-9 .

Smith H, Wright D, Morgan S, Dunleavey J, Moore M. The ‘research spider’: a simple method of assessing research experience. Prim Heal Care Res Dev. 2002;3(3):139–40. https://doi.org/10.1191/1463423602pc102xx .

Council of Deans of Health. Nursing, midwifery and allied health clinical academic careers in the UK. 2018. http://councilofdeans.org.uk/wp-content/uploads/2018/08/Nursing-midwifery-and-allied-health-clinical-academic-research-careers-in-the-UK.pdf . Accessed 18 Feb 2020.

Download references

Acknowledgements

Thank you to Clare Leon-Villapalos, Melanie Almonte, Jinju James and Gemma Clunie for their support with screening, data extraction and quality assessment; and to Lisa Gardner for her assistance in refining the search strategy.

This research was funded by the NIHR Imperial Biomedical Research Centre (BRC). The views expressed are those of the authors and not necessarily those of the NIHR or the Department of Health and Social Care.

Author information

Authors and affiliations.

Imperial College Healthcare NHS Trust, Education Centre, Charing Cross Hospital, Fulham Palace Road, London, W6 8RF, UK

Lisa Newington, Mary Wells, Adine Adonis, Lee Bolton, Layla Bolton Saghdaoui, Margaret Coffey, Jennifer Crow, Olga Fadeeva Costa, Catherine Hughes, Matthew Savage, Lillie Shahabi & Caroline M. Alexander

Department of Surgery and Cancer, Faculty of Medicine, Imperial College London, London, UK

Lisa Newington, Mary Wells, Lee Bolton, Layla Bolton Saghdaoui, Margaret Coffey, Olga Fadeeva Costa & Caroline M. Alexander

Department of Brain Sciences, Faculty of Medicine, Imperial College London, London, UK

Lillie Shahabi

You can also search for this author in PubMed   Google Scholar

Contributions

LN, MW and CMA designed the review. LN, JC, OFC, LS and LB completed the title and abstract screening. LN, AA, MC, LS, OFC, MA and LBS completed the full text screening. LN, AA, LBS, CL, LS, MC, JJ, LB and JC completed quality assessment and data extraction. Disagreements during screening were resolved by MW and CMA. Differences in data extraction were reviewed by MS and CH. LN, AA, LB, LBS, MC, JC, MS and CH thematically analysed the extracted data, and LN, MW and CMA completed the thematic synthesis. LN drafted the manuscript with major contributions from MW and CMA. All authors reviewed and approved the final manuscript.

Corresponding author

Correspondence to Lisa Newington .

Ethics declarations

Ethics approval and consent to participate.

Ethics approval was not required for this systematic review.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s note.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1..

Example search strategy for Medline. Search strategy terms.

Additional file 2.

Mixed Methods Appraisal Tool. Quality appraisal form and instruction for use.

Additional file 3.

Articles excluded during full text screen and reasons for exclusion. References of the excluded articles and reason for exclusion.

Additional file 4.

Coding framework for the types of impact identified and exemplar quotes. Full listing of the theme headings with descriptors and supplementary quotes.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ . The Creative Commons Public Domain Dedication waiver ( http://creativecommons.org/publicdomain/zero/1.0/ ) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Cite this article.

Newington, L., Wells, M., Adonis, A. et al. A qualitative systematic review and thematic synthesis exploring the impacts of clinical academic activity by healthcare professionals outside medicine. BMC Health Serv Res 21 , 400 (2021). https://doi.org/10.1186/s12913-021-06354-y

Download citation

Received : 24 November 2020

Accepted : 31 March 2021

Published : 29 April 2021

DOI : https://doi.org/10.1186/s12913-021-06354-y

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Research impact
  • Clinical academics
  • Allied health professions
  • Systematic review
  • Thematic synthesis

BMC Health Services Research

ISSN: 1472-6963

qualitative synthesis and systematic review in health professions education

Europe PMC requires Javascript to function effectively.

Either your web browser doesn't support Javascript or it is currently turned off. In the latter case, please turn on Javascript support in your web browser and reload this page.

Search life-sciences literature (44,058,344 articles, preprints and more)

  • Free full text
  • Citations & impact
  • Similar Articles

Analysing synthesis of evidence in a systematic review in health professions education: observations on struggling beyond Kirkpatrick.

Author information, affiliations.

  • Maudsley G 1

ORCIDs linked to this article

  • Taylor D | 0000-0002-3296-2963

Medical Education Online , 01 Dec 2020 , 25(1): 1731278 https://doi.org/10.1080/10872981.2020.1731278   PMID: 32228373  PMCID: PMC7170338

Abstract 

Free full text , analysing synthesis of evidence in a systematic review in health professions education: observations on struggling beyond kirkpatrick, gillian maudsley.

a Department of Public Health & Policy, The University of Liverpool, Liverpool, UK

David Taylor

b Medical Education & Physiology, College of Medicine, Gulf Medical University, Ajman, United Arab Emirates

Background : Systematic reviews in health professions education may well under-report struggles to synthesize disparate evidence that defies standard quantitative approaches. This paper reports further process analysis in a previously reported systematic review about mobile devices on clinical placements.

Objective : For a troublesome systematic review: (1) Analyse further the distribution and reliability of classifying the evidence to Maxwell quality dimensions (beyond ‘Does it work?’ ) and their overlap with Kirkpatrick K-levels. (2) Analyse how the abstracts represented those dimensions of the evidence-base. (3) Reflect on difficulties in synthesis and merits of Maxwell dimensions.

Design : Following integrative synthesis of 45 K2–K4 primary studies (by combined content–thematic analysis in the pragmatism paradigm): (1) Hierarchical cluster analysis explored overlap between Maxwell dimensions and K-levels. Independent and consensus-coding to Maxwell dimensions compared (using: percentages; kappa; McNemar hypothesis-testing) pre- vs post-discussion and (2) article abstract vs main body. (3) Narrative summary captured process difficulties and merits.

Results : (1) The largest cluster (five-cluster dendrogram) was acceptability–accessibility–K1–appropriateness–K3, with K1 and K4 widely separated. For article main bodies, independent coding agreed most for appropriateness (good; adjusted kappa = 0.78). Evidence increased significantly pre–post-discussion about acceptability (p = 0.008; 31/45→39/45), accessibility, and equity-ethics-professionalism. (2) Abstracts suggested efficiency significantly less than main bodies evidenced: 31.1% vs 44.4%, p = 0.031. 3) Challenges and merits emerged for before, during, and after the review.

Conclusions : There should be more systematic reporting of process analysis about difficulties synthesizing suboptimal evidence-bases. In this example, Maxwell dimensions were a useful framework beyond K-levels for classifying and synthesizing the evidence-base.

  • Introduction

Synthesizing messy evidence in health professions education can be more art than science, more pragmatism than finesse, and more trouble apparently than it's worth. Nevertheless, exploring the evidence enough to make useful recommendations and to critique and improve the process must be better than investigatory inanition waiting for the perfect evidence-base.

Whether viewed as what is probably, plausibly, or often generally true, the nature of the ‘evidence’ [ 1 ] and the nature of its ‘good quality’ are debatable concepts. Approaches to synthesis are eclectic in Best Evidence Medical Education (BEME) systematic reviews [ 2 ]. In-depth consideration is limited though [ 3 ] about how to classify and analyse diverse evidence that thwarts standard quantitative approaches to the systematic review, and qualitative research synthesis is contested [ 4 , 5 ]. In making sense of a difficult synthesis, a post hoc standard veneer of slick reporting might hide how researchers cycle through confusion and clarity. While Kirkpatrick’s four-level outcomes-based model or taxonomy has been popular for imposing some order in systematic reviews in health professions education, it is useful to analyse how other frameworks might enhance the processes of organizing and synthesis.

The popularity of the Kirkpatrick model

The Kirkpatrick model has been popular for straightforward, practical evaluation of training interventions by K1 = reaction, K2 = learning, K3 = behaviour, and K4 = results [ 6 ], especially in medical education [ 7 , 8 ]. Many BEME systematic reviews have used the model to filter and summarize [ 1 ], and its use to organize and synthesize systematic reviews and ‘quality-score’ quantitative evidence extends across diverse health-care education [ 9 – 11 ]. The common inferences that these ‘levels’ are a causal sequence (K1→K2→K3→K4) or hierarchical in value (K4>K3>K2>K1) have attracted criticism though [ 7 , 8 ].

Whether or not Kirkpatrick meant a causal hierarchy, he did view K1→K4 as increasing in complexity and meaning – thorough evaluation might use all levels [ 6 ]. Moreau [ 12 ] noted that the New World Kirkpatrick Model [ 13 ] improved on three criticisms: – Difficulties evaluating K3/K4 : Evaluation now also focused on ways of promoting the application of learning (K3) and contributions to organizational goals (K4). – Ignoring confounding and intervening variables : Each K-level now included various personal and organizational influences. – Inferring an unproven causal chain (K1 → K2 → K3 → K4 ): Considering ‘chains of evidence’ rather than implying a causal hierarchy of levels allowed non-sequential use of the levels. Such advances complemented re-thinking evidence synthesis beyond ‘effectiveness = Does it work?’ .

Building beyond Kirkpatrick and ‘what works?’, using an example

The ‘What works?’ question could be dismissed as a narrow interpretation of effectiveness that uses only quantitative evidence for justification, or it could be interpreted more widely from whatever evidence supports ‘whether (it works)’ plus: ‘how, why, and in what circumstances (the context)?’. ‘The difficulty of evaluating any educational philosophy in a scientific manner’ occupied the early days of seeking best evidence (available) in medical (and other health professions) education [ 14 , p.1]. The push to widen the horizons of BEME reviews has continued [ 15 ]. Systematic reviews in health services research and public health have long since been dealing with condensing complexity into concise counsel.

Petticrew [ 16 , p.2] argued that the question for complex interventions should be:

“What has happened previously when this intervention [has] been implemented across a range of contexts, populations and subpopulations, and how have those effects come about?”

‘Does it work?’ becomes ‘meaningless and usually unanswerable’ (p.2) for complex interventions, and narrative reviews seek evidence to reduce uncertainty rather than to derive a precise effect-size. Even weak studies provide illumination when a field is still in development and:

“evidence synthesis often is, and should be, an exercise in Bayesian decisionmaking, and reducing uncertainty, and not hypothesis testing” (p.5).

BEME review 52 [ 17 ] investigated: ‘What works best for health professions students using mobile (hand-held) devices for educational support on clinical placements?’ in an underdeveloped evidence-base. This was about a complex intervention and required versatile interpretation and classification of a mash-up of ‘whether’ (justification), ‘how/why’ (clarification), and ‘what’ (description) evidence [ 2 , 18 ]. Of the K2–K4 primary empirical studies included (K1-only studies were excluded), 46.7% (21/45) were mixed methods, 33.3% quantitative, and 20.0% qualitative research. K3 (86.7%) and S3-strength evidence ( Conclusions can probably be based on the results) (55.6%) [ 19 , 20 ] predominated. There were only five L6 (randomized controlled trials) and two L5 (longitudinal) designs [ 21 ]. About three-quarters had supplementary K1 evidence and 53.3% had K4 evidence, mostly K4b. Inter-observer agreement on filtering abstracts was good (e.g. 92.1% in the final 2016 update, kappa = 0.64, p < 0.0001).

BEME review 52 concluded about mobile devices as educational support that [ 17 ]:

They supported students’ learning on clinical placement via: assessment; communication; clinical decision-making; logbook or notetaking; and most often accessing information.

In the hidden and informal curricula, ‘what happened’ was that students were:

bothered about : actual and perceived disapproval of peers, clinicians or educators, and patients; confidentiality and privacy; and security aspects,

side-tracked by : social connectivity (or other private use) and hectic clinical settings,

confused by : policy ambiguity.

That review moved beyond ‘Does it work?’ and beyond K-levels. Much of the synthesis of the required ‘What works best …?’ question did implicitly answer ‘What has happened previously with use of mobile devices on clinical placements?’ . Maxwell’s dimensions from health services research [ 22 , 23 ] then helped to widen horizons. Used in evaluating quality of care, this 3As & 3Es framework considers:

Acceptability (What do users prefer and how satisfied are they?). Accessibility (How reachable is the service? What are the barriers?). Appropriateness (How relevant is the service to needs?). Effectiveness (Does it work? What are the outcomes?). Efficiency (How are outputs to inputs? What are costs?). Equity (How fair is it?).

Adapting these dimensions ( Table 1 ) helped to organize evidence and deliberate about synthesis. Previous BEME reviews did not feature this framework, warranting further analysis.

Distribution and reliability: Maxwell dimensions (and final clustering with Kirkpatrick K-levels noted).A systematic review (Maudsley et al. 2019): What works best for health professions students using mobile (hand-held) devices for educational support on clinical placements? Evidence from n = 45 studies

In a preliminary analysis, BEME review 52 reported that the commonest Maxwell evidence-profiles (just under one-half) supported accessibility, appropriateness, acceptability, and effectiveness, or those plus efficiency, with little about equity-ethics-professionalism. Further analysis would gainfully explore the usefulness of that additional framework beyond Kirkpatrick, how both frameworks overlapped, and insights about presenting that body of evidence. As Regehr [ 24 , p.34] argued:

“We are bound to learn more from our own work and that of others if we systematically examine and document our struggles than if we loudly proclaim our successes.”

In extracting transferable messages about the process of a systematic review, the aims here were to (1) Analyse further the distribution and reliability of classifying the evidence to Maxwell quality dimensions (beyond ‘Does it work?’ ) and their overlap with K-levels. (2) Analyse how the abstracts represented those dimensions of the evidence-base. (3) Reflect on difficulties in the synthesis and merits of Maxwell dimensions.

  • Materials & methods

Supplementary to BEME systematic review 52, the co-authors of this paper analysed their independent and consensus classification of evidence from the 45 primary studies [ 17 : 3,228 ‘initial hits’ on 1988–2016 bibliographic database search]. This involved much immersion and deliberation on alternative interpretations, within the pragmatism paradigm [ 25 ]. These articles presented K2, K3, or K4 +/–K1 evidence, as K1-only articles had been excluded. ‘Self-reported’ evidence was allowed. For that review, a combined deductive content analysis and thematic analysis [ 3 , 26 , 27 ] focused on integrative synthesis to summarize the evidence systematically, quantifying as appropriate [ 28 ]. QSR NVivo 10 assisted data handling.

After calibrating 10 articles together, each reviewer also coded evidence in each article to one or more:

Maxwell dimensions of quality adapted to the mobile device (left column, Table 1 ): acceptability, accessibility, appropriateness, effectiveness, efficiency, and equity (expanded to equity-ethics-professionalism).

Disagreements were resolved by discussion. Discussion of these classifications also informed the overall integrative synthesis.

For each Maxwell dimension, analysis in IBM SPSS Statistics 24 calculated the percentage of the articles where, respectively, the main body presented or the abstract suggested empirical evidence. Cohen kappa measured the reliability of the independent Yes/No classifications of the main body (Reviewer 1 vs Reviewer 2). Imbalance in average prevalence of Yes vs No (near the extremes rather than 50%) [ 29 , 30 , p.260] prompted use of prevalence-adjusted, bias-adjusted kappa (PABA-K) with 95% confidence interval.

Hypothesis-testing compared Yes-No for each Maxwell dimension from:

combined pre-agreement (i.e. Yes = both independently coded to Yes; No = one or both coded to No) vs final agreed classification (post-discussion consensus) of main body of article.

abstract vs main body.

McNemar hypothesis-testing treated both these comparisons as paired, under the null hypothesis of no difference in Yes-No.

SPSS 24 hierarchical cluster analysis explored how Maxwell dimensions and K-levels overlapped. An initial basic, simple-linkage, nearest-neighbour cluster analysis suggested the order for variables to enter a between-groups linkage analysis, clustering by variable (measuring Squared Euclidean distance; binary variable: 1 = Yes, 2 = No, rescaled to 0–1). A dendrogram summarized, from the left, stronger relationships with shorter horizontal fork-prongs, with vertical lines ‘joining’ variables (x-axis distance) ( Figure 1 ) [ 31 ]. Coherence of findings and a ‘scree plot’ (agglomeration coefficient vs stage) suggested how many clusters to declare.

qualitative synthesis and systematic review in health professions education

SPSS hierarchical five-cluster dendrogram: Maxwell dimensions and K-levels

Narrative summary captured reflection on the difficulties in synthesis [ 28 ]. Both reviewers independently outlined four main difficulties in synthesis and four main merits in using Maxwell dimensions in synthesis and analysis, then agreed a final summary in reflective discussion.

Distribution and reliability of using Maxwell dimensions and overlap with K-levels

All bar one study provided evidence for appropriateness of mobile device use to learning needs (including caregiving and patient safety aspects) (44/45, 97.8%), 86.7% each for acceptability and accessibility, with 73.3% and 44.4% for effectiveness and efficiency, respectively, but only just over one-quarter for equity-ethics-professionalism ( Table 1 ). For the main body of articles, independent observations of Maxwell dimensions agreed best for appropriateness (good, PABA-K = 0.78). Classifying to acceptability, accessibility, effectiveness, and equity-ethics-professionalism (mostly about digital professionalism) showed moderate agreement (PABA-K = 0.47–0.51). Classifying to efficiency reached only fair agreement (PABA-K = 0.33), broadly interpreted as consideration of outputs to inputs, such as costs, saving or making the most of time or effort in learning or providing care, or allowing timely feedback.

The largest cluster in the five-cluster dendrogram of K-levels and Maxwell dimensions was acceptability–accessibility–K1–appropriateness–K3, with effectiveness nearby, quite dissimilar from the efficiency–K4b cluster ( Figure 1 ). K2a–K2b clustered stronger than K4a–equity-ethics-professionalism.

How abstracts represented Maxwell quality dimensions of evidence

Acceptability was classified as ‘present’ significantly more post-discussion (same in 82.2%, p = 0.008; 31/45→39/45) and likewise for accessibility (same in 77.8%, p = 0.002; 29/45→39/45) and equity-ethics-professionalism (same in 75.6%; p = 0.001; 1/45→12/45).

If the abstract suggested evidence of a Maxwell dimension, the main body included that evidence. The main body presented evidence for acceptability and efficiency significantly more than the abstract: 86.7% vs 66.7%, p = 0.004; 44.4% vs 31.1%, p = 0.031, respectively. For accessibility, appropriateness, effectiveness (p = 0.063, respectively) and equity-ethics-professionalism (p = 0.250), the excess was not statistically significant.

Reflecting on difficulties in synthesis and analysis and merits of Maxwell dimensions

Combined observations about the main struggles in synthesis and analysis aggregated around the time ( Figure 2 ):

before: insufficient guidance about such mixed evidence and concerns about transgressing qualitative research ‘rules’,

during: ill-defined outcomes and methods, requiring much translation, particularly for equity-ethics-professionalism evidence,

after: much effort in reporting analysis of the process and much potential to be misconstrued .

qualitative synthesis and systematic review in health professions education

Likewise, for the list of Maxwell dimensions, main merits aggregated around the time ( Figure 2 ):

before: a simplified starting-point, adaptable to the intervention,

during: a lens for better and wider understanding of implementation and impact, intended or otherwise,

after: a structured framework for reflection, prompting deliberation to consensus about tricky evidence and its wider worth (and possible gaps), and being consistent with application of Maxwell dimensions by systematic reviews elsewhere in health services research .

Systematic reviews in medical education require more open discussion of difficulties in synthesizing suboptimal evidence-bases and more systematic reporting of process analysis. In the example explored here, Maxwell dimensions [ 22 , 23 ] helped to classify and synthesize the variegated quantitative, qualitative, and mixed methods evidence-base of a systematic review [ 17 ] meaningfully and moderately reliably, when only one-third of articles reported ‘quantitative-only’ research. The dimensions also helped to illuminate a coherent relationship with K-levels. Health professions education systematic reviews have not used this framework previously and very few health services research systematic reviews have reported using it [ 32 – 34 ]. As with the Kirkpatrick model, Maxwell dimensions were valuable in simplifying the approach [ 35 ] plus giving a deliberative framework for reviewers to share understanding in a tricky integrative synthesis and consider possible gaps. While the Maxwell list of ‘characteristics’ might not amount to a ‘framework’ for health-care performance measurement and improvement [ 36 ], here Maxwell dimensions provided a ‘framework for concepts’ of educational support. This built beyond ‘Does it work?’ .

When BEME review 52 [ 17 ] reported much potential for mobile devices to support health professions students on clinical placement (via their transitions, meta-learning, and care contribution, but requiring policy to address negative informal and hidden curricula), this explored beyond ‘Does it work?’ . Besides the usual recommendation for improved reporting of primary research, BEME review 52 recommended that ‘effectiveness’-reviews extend beyond a simplistic approach of just ‘What works?’ , echoing Eva [ 37 ] and Regehr [ 24 ]. This justified further scrutiny. This also reflected the broader horizons of health services evaluation [ 38 ] and challenges in systematic reviews of complex interventions in public health and social sciences [ 39 , 40 ], including the need to synthesize haphazard evidence. The ‘stainless steel’ law of such systematic reviews remained that ‘the more rigorous the review, the less evidence there will be to suggest that the intervention is effective’ [ 39 , p.758]. In qualitative synthesis:

“study findings are systematically interpreted through a series of expert judgements to represent the meaning of the collected work. … the findings of qualitative studies – and sometimes mixed-methods and quantitative research – are pooled.” [ 3 , p.253]

Better narrative synthesis is required [ 39 ] plus better blending with quantitative observations, as appropriate. Here, further analysis confirmed Maxwell dimensions to be a useful extra framework to prompt much deliberation on evidence, improve understanding, and represent complexity.

Maxwell dimensions typified health-care evaluation in the UK National Health Service (NHS) during its 1990s quality management and ‘internal’ market phase [ 41 ], providing a ‘characteristics model’ of quality [ 42 ]. In health services research, these dimensions have guided the integrative systematic review of ‘What is the effect of non-medical prescribing in primary care and community settings on patient outcomes?’ , exploring beyond ‘Does it work?’ [ 32 ]. A similar US Institute of Medicine [ 43 ] list of ‘characteristics’ (explicitly mentioning patient safety) was also popular, guiding classification of evidence in a systematic review of pay-for-performance in UK general practice [ 33 ]. While Donabedian’s much revered structure-process-outcome framework for measuring health-care quality might have been an alternative [ 44 ], it did not intuitively have as much potential for exploring the K-levels. Berwick and Fox [ 45 ] considered that the Donabedian framework was not necessarily patient-centred or viewing health care as a system. The Maxwell dimensions allowed the evidence in the educational context to be student-centred and to be viewed holistically as if part of a learning system. Several major NHS reforms and many alternative quality indicators later, more complex representations of quality have superseded Maxwell dimensions, e.g. to analyse patient-professional co-production of knowledge and health [ 46 ].

Nevertheless, Maxwell dimensions remain a basic, durable, practical starting-point for evaluating services [e.g. 47 , 48 ], notably dental in recent years [e.g. 49 , 50 ]. Relevant to workforce development, Halter et al.’s [ 34 ] systematic review of the impact of physician associates on secondary care used Maxwell dimensions as ‘outcome’ search-terms and then reportedly to organize the main messages (albeit without presenting or discussing the latter). Here, the dimensions adapted well to making sense of a research evidence mash-up about mobile devices in clinical placements for the future health professions workforce.

While the evidence is not obliged to represent all dimensions, the extra Maxwell lens highlighted potentially underrepresented aspects such as efficiency and equity-ethics-professionalism and an evidence distribution more towards the 3As. There was a gap in the evidence about quantifying efficiencies in learning or care provided. There was also a gap in the evidence about ‘fairness’ of the use of mobile devices.

Abstracts tended to omit much supplementary evidence about acceptability understandably (given exclusion of K1-only papers) but omitting evidence about efficiency suggested lower sensitivity of title-abstract filtering on this dimension. Despite moderately reliable independent observations, inter-observer discussion significantly increased the proportion classified to acceptability, accessibility, and especially equity-ethics-professionalism, probably reflecting refinement of definitions but also the subtlety of some evidence. A coherent clustering with K-levels confirmed that broadly interpreting ‘effectiveness’ across research types reached well beyond K2a-K2b randomized controlled trial–type evidence. Widening impact to ‘making a difference’ for the organization or for other people involved K4a and K4b clustering with, respectively, equity-ethics-professionalism and efficiency – and not with K1, so even if K-levels were non-hierarchical, this suggested that K1 and K4a/b differ substantively.

Yardley and Dornan [ 8 ] found ‘K2 and below’ to show suboptimal sensitivity as a BEME exclusion-filter for their review-question (about early workplace experience in undergraduate medical education). For BEME review 52 though, excluded K1-only papers did not illuminate its review-question further. Suitability of K-levels to filter and K-levels and Maxwell dimensions to organize and summarize trustworthy evidence depends on the review-question.

Strengths here were that both Kirkpatrick model and Maxwell dimensions were applied with critical scepticism and within the pragmatism paradigm, which enhanced: deliberation and synthesis; quantification of key aspects ( Table 1 , Figure 1 ); and mixing of qualitative and quantitative analytical approaches [ 51 ]. Furthermore, despite problematic calibration of Maxwell coding for equity-ethics-professionalism, the final coding appeared robust.

The evidence-base on which this ‘process analysis’ focused was relatively small, yet its eclecticism was both strength and weakness. Calibration about efficiency needed more attention. It was also unsurprising if evidence about acceptability and accessibility was uncommon, given exclusion of K1-only articles, but useful K1 supplementary evidence still featured. While using cluster analysis on 12 variables for n = 45 ignored a 2 m ‘rule-of-thumb’ sample-size (where m = number of variables) [ 31 ] and might be seen as overkill (or overreliant on hypothesis-testing [ 40 ]), the five-cluster dendrogram illuminated the Kirkpatrick–Maxwell relationship. Exploration of Maxwell coding against strength of evidence and study design may well also be merited but would require a larger evidence-base.

Beyond the convenience of Kirkpatrick outcome-levels for filtering abstracts and summarizing outcome-evidence, Maxwell dimensions helped to promote Regehr’s [ 24 ] preferred imperatives for medical education evidence: gaining a rich understanding and representing complexity . This contrasted with what he called the dominant imperatives: seeking proof  [ ‘that something works’ , 37 , p. 295] and generalizable simplicity .

Reviewer-pairs or teams must calibrate and critique such classification tools with care to widen analytical horizons robustly. Here, deliberative synthesis of ‘whether’, ‘how/why’, and ‘what’ concepts [ 2 , 18 , 52 , 53 ] applied Maxwell dimensions to illuminate the process of implementing the educational intervention (using mobile devices on clinical placements) as well as broadly interpreting outcomes. To improve systematicity [ 54 , 55 ] and thoroughness for tricky integrative synthesis in systematic reviews, reflective deliberation and supplementary analyses about such tools are required, particularly their conceptual integrity and trustworthiness. Maxwell dimensions at least give a practical framework for organizing, deliberating about, and synthesizing key concepts when struggling beyond Kirkpatrick in a messy evidence-base. Waiting for the perfect evidence-base to synthesize would be unhelpful for the topic. As Glass [ 56 , p.4] highlighted:

“A common method of integrating several studies with inconsistent findings is to carp on the design or analysis deficiencies of all but a few studies – those remaining frequently being one’s own work or that of one’s students or friends – and then advance the one or two ‘acceptable’ studies as the truth of the matter. This approach takes design and analysis too seriously, in my opinion. I don’t condone a poor job of either; but I also recognize that a study with a half dozen design and analysis flaws may still be valid. […] … I believe the difference [in results between poorly-designed and the best-designed studies] to be so small that to integrate research results by eliminating the ‘poorly done’ studies is to discard a vast amount of important data.”

Wilson and Lipsey [ 57 , p.420] confirmed this in analysing 250 meta-analyses:

“It appears that low method quality functions more as error than as bias, reducing the confidence that can be placed in the findings but neither consistently over- nor underestimating program effects.”

In health professions education, it is challenging to undertake systematic multi-component mixed methods reviews that attempt to arrange, interpret, and summarize evidence about ‘What is the effect of this complex intervention?’ [ 58 , p.2]. Such reviews may well attempt to ‘configure’ and (less so) ‘aggregate’ [ 58 ] an evidence-base that is quite a mess, epistemologically or otherwise. More tools and analysis of their use are required for the synthesis of jumbles of evidence.

  • Practice points

There should be more systematic reporting of process analysis about difficulties synthesizing suboptimal evidence-bases in health professions education.

Maxwell dimensions are potentially useful for evaluating educational interventions and synthesizing messy, variegated evidence-bases.

Kirkpatrick model and Maxwell dimensions applied to such an evidence-base clustered coherently together and should be applied with critical scepticism to improve understanding and represent complexity.

To improve systematicity and thoroughness of systematic reviews, especially when synthesis is tricky, reflective deliberation and supplementary analyses about the conceptual integrity and trustworthiness of ‘filtering and classification’ tools are warranted.

Be aware that abstracts may well omit certain types of substantive evidence reported in the main body, thus reducing the sensitivity of title-abstract filtering.

  • Acknowledgments

We would like to thank our other BEME review 52 co-authors for their contribution to that systematic review on which this process analysis built.

Small extracts of an early version of this work were presented at the Association for Medical Education in Europe (AMEE) conference in Helsinki, Aug-2017 (poster).

  • Disclosure statement

No potential conflict of interest was reported by the authors.

Full text links 

Read article at publisher's site: https://doi.org/10.1080/10872981.2020.1731278

Citations & impact 

This article has not been cited yet.

Impact metrics

Alternative metrics.

Altmetric item for https://www.altmetric.com/details/78837488

Similar Articles 

To arrive at the top five similar articles we use a word-weighted algorithm to compare words from the Title and Abstract of each citation.

A Best Evidence Medical Education (BEME) systematic review of: What works best for health professions students using mobile (hand-held) devices for educational support on clinical placements? BEME Guide No. 52.

Maudsley G , Taylor D , Allam O , Garner J , Calinici T , Linkman K

Med Teach , 41(2):125-140, 28 Nov 2018

Cited by: 11 articles | PMID: 30484351

The educational effects of portfolios on undergraduate student learning: a Best Evidence Medical Education (BEME) systematic review. BEME Guide No. 11.

Buckley S , Coleman J , Davison I , Khan KS , Zamora J , Malick S , Morley D , Pollard D , Ashcroft T , Popovic C , Sayers J

Med Teach , 31(4):282-298, 01 Apr 2009

Cited by: 172 articles | PMID: 19404891

Scoping reviews in health professions education: challenges, considerations and lessons learned about epistemology and methodology.

Thomas A , Lubarsky S , Varpio L , Durning SJ , Young ME

Adv Health Sci Educ Theory Pract , 25(4):989-1002, 25 Nov 2019

Cited by: 20 articles | PMID: 31768787

Qualitative synthesis and systematic review in health professions education.

Bearman M , Dawson P

Med Educ , 47(3):252-260, 01 Mar 2013

Cited by: 97 articles | PMID: 23398011

Building capacity for education research among clinical educators in the health professions: A BEME (Best Evidence Medical Education) Systematic Review of the outcomes of interventions: BEME Guide No. 34.

Ahmed R , Farooq A , Storie D , Hartling L , Oswald A

Med Teach , 38(2):123-136, 26 Nov 2015

Cited by: 11 articles | PMID: 26610023

Europe PMC is part of the ELIXIR infrastructure

Log in using your username and password

  • Search More Search for this keyword Advanced search
  • Latest content
  • For authors
  • Browse by collection
  • BMJ Journals More You are viewing from: Google Indexer

You are here

  • Volume 7, Issue 4
  • Learning outcomes for communication skills across the health professions: a systematic literature review and qualitative synthesis
  • Article Text
  • Article info
  • Citation Tools
  • Rapid Responses
  • Article metrics

Download PDF

  • Charlotte Denniston 1 ,
  • Elizabeth Molloy 1 , 2 ,
  • http://orcid.org/0000-0003-2941-2298 Debra Nestel 1 , 3 ,
  • Robyn Woodward-Kron 2 ,
  • Jennifer L Keating 4
  • 1 Faculty of Medicine, Nursing and Health Sciences , Monash University , Melbourne, Victoria , Australia
  • 2 Department of Medical Education , University of Melbourne , Melbourne, Victoria , Australia
  • 3 Department of Surgery (Austin) , University of Melbourne , Melbourne, Victoria , Australia
  • 4 Department of Physiotherapy , Monash University , Melbourne, Victoria , Australia
  • Correspondence to Charlotte Denniston; charlotte.denniston{at}monash.edu

Objective The aim of this study was to identify and analyse communication skills learning outcomes via a systematic review and present results in a synthesised list. Summarised results inform educators and researchers in communication skills teaching and learning across health professions.

Design Systematic review and qualitative synthesis.

Methods A systematic search of five databases (MEDLINE, PsycINFO, ERIC, CINAHL plus and Scopus), from first records until August 2016, identified published learning outcomes for communication skills in health professions education. Extracted data were analysed through an iterative process of qualitative synthesis. This process was guided by principles of person centredness and an a priori decision guide.

Results 168 papers met the eligibility criteria; 1669 individual learning outcomes were extracted and refined using qualitative synthesis. A final refined set of 205 learning outcomes were constructed and are presented in 4 domains that include: (1) knowledge (eg, describe the importance of communication in healthcare), (2) content skills (eg, explore a healthcare seeker's motivation for seeking healthcare),( 3) process skills (eg, respond promptly to a communication partner's questions) and (4) perceptual skills (eg, reflect on own ways of expressing emotion).

Conclusions This study provides a list of 205 communication skills learning outcomes that provide a foundation for further research and educational design in communication education across the health professions. Areas for future investigation include greater patient involvement in communication skills education design and further identification of learning outcomes that target knowledge and perceptual skills. This work may also prompt educators to be cognisant of the quality and scope of the learning outcomes they design and their application as goals for learning.

  • Communication Skills
  • Learning Outcome
  • Qualitative synthesis
  • Systematic Review
  • Health Professions Education

This is an Open Access article distributed in accordance with the Creative Commons Attribution Non Commercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited and the use is non-commercial. See: http://creativecommons.org/licenses/by-nc/4.0/

https://doi.org/10.1136/bmjopen-2016-014570

Statistics from Altmetric.com

Request permissions.

If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.

Strengths and limitations of this study

As far we are aware, no previous paper has reported a comprehensive systematic literature review and qualitative synthesis of learning outcomes in communication skills across the health professions.

Learning outcomes were restructured using qualitative synthesis to remove duplication and ambiguity; and to be concise, defined and accessible to educators developing individualised education programmes.

The final list provides a comprehensive summary of published literature for consideration by patients, educators, learners and other stakeholders.

The search strategy was limited to the English language and the specific education of communication skills. By restricting the search strategy, we may have excluded papers that contained relevant learning outcomes.

Introduction

It is widely acknowledged that skilled communication is essential for all health professionals. For the purposes of this paper, communication is defined as a two-way process involving speech, writing or non-verbal means that aim to create shared interpretation for those involved. 1 As such, effective communication skills are emphasised as a target for health professional education programmes locally 2 and internationally (including medicine, 3 physiotherapy 4 and across interprofessional groups 5 ). Over the past 20 years, there has been a growth in communication skills research, and the implementation of communication skills programmes in entry-level medical programmes has become commonplace worldwide. 6 , 7 Despite its recognition as a key competency for practice, communication skills teaching has not been routinely adopted in all health professional programmes, and discourse related to communication skills pedagogy remains less common outside medicine. 8

The literature on learning and teaching communication skills includes guides and consensus statements for doctor–patient communication (such as Calgary Cambridge Referenced Observation Guides, the UK Consensus Statement, the Kalamazoo Consensus Statement and the German Basel Consensus Statement). 9–12 The content of these documents could be presented as intended learning outcomes (ie, what learners are expected to be able to know and/or do after participation in an education programme). However, many of the individual items in these documents target multiple constructs and do not directly translate into specific learning goals.

Learning outcomes in education design

For many years, researchers in education have proposed alternate ways of structuring goals in educational design. From the 1960s, the term ‘instructional objective’ was used to define a statement expressing what learners should be able to do at the end of the learning period. 13 These instructional objectives were designed to be constructively aligned with methods for teaching and assessment. 14 Learning objectives have been described as focusing more discretely on observable knowledge, attitudes and skills, whereas the more contemporary term ‘outcome’ is defined as a broader statement of what is achieved and assessed at the end of a course of study. 15 , 16 More recently, competency-based medical education (CBME) argues for education organised around a defined set of competencies towards which learning is targeted to achieve proficiency. 17 Competencies have been defined as the observable capabilities of a health professional (ie, knowledge, skills, values and attitudes). 18 Grant 19 and Prideaux 20–22 argue that the differentiation of terms is not greatly important, as long as the goals for learning are clear. Regardless of semantics, it is acknowledged that some aspects of healthcare practice are difficult to define in these terms, and the identification of an objective, outcome or a competency may not specify exactly what is to be achieved. 20 , 23 This argument suggests that the ‘sum’ of healthcare practice is far greater than the ‘parts’ and that any reduction of complex human behaviour into objectives or competencies may be seen as unhelpful. 24 This stance resonates with reports within the communication skills literature, which argue that communication also cannot be separated into its parts. 25

Communication as a skill

Although learners may develop many aspects of communication through experience prior to university, 26 , 27 effective communication in the context of healthcare practice is highly technical and is likely to require training, deliberate targeted practice and feedback to develop skilled performance. 28 , 29

One school of thought in the communication literature is that positioning communication as a set of behavioural skills is reductionist and mistaken. 25 , 30 , 31 It has been argued that the ‘atomisation’ of communication into discrete observable skills may not take into consideration the complexity of a communication interaction, and nor does it consider the authenticity and creativity required in complex practice. 25 , 31 These authors also argue that the deployment of a communication skill does not necessarily equate to skilled communication. 25 , 30 It is important here therefore to make a distinction between learning and assessing skilled communication. Salmon and Young 25 may be correct when they contend that by assessing communication skills using a reductionist checklist, there is a risk of trivialising the communication interaction. This is particularly the case if assessment procedures require the demonstration of the many discrete communication skills proposed (ie, the ‘parts’) regardless of the context. Such rating requirements may inaccurately judge the ‘whole’. 22 , 25 , 32 However, breaking a complex phenomenon, like communication, into discrete ‘parts’ may allow a novice learner to appreciate what may be required for their own skill development. 33 Targeted performance development and self-regulated learning require targeted learning goals: therefore, the clarification of these skills has the potential to be useful for learners and educators. 34–36 Learning the component ‘parts’ of a communication interaction may allow for skill acquisition and improved performance of the ‘whole’. 37 With increased proficiency, learners would then be able to use these skills in a flexible, personal and creative way depending on the context. 33 , 38

Communication skills learning outcomes

To the best of our knowledge, there is only one published review identifying communication skills learning outcomes relevant to multiple health professional groups. In 2013, Bachmann et al 8 produced the European consensus on learning objectives (HPCCC: Health Professions Core Communication Curriculum). Bachmann et al use the term learning ‘objective’ in their report to which we attribute the same meaning as the term learning ‘outcome’. The HPCCC presents learning objectives that are based on the literature and a medical communication consensus statement and were developed using an extensive Delphi process. In total, 121 communication experts from 16 countries and 15 professions reviewed learning objectives in four stages. Each stage included a review of drafts, ranking of statements and comments regarding acceptability. 8 This process drew on a large community of practice within Europe that has a significant focus on communication in healthcare. The HPCCC presents learning objectives that go beyond ‘health professional-patient’ communication and target interprofessional and intraprofessional communication skills. However, the report omits the key quality indicator of detail about the literature review processes, subsequently limiting potential for replication. 39 While the Delphi process and the quantitative and qualitative analysis methods were clearly described, the development of the initial draft list of statements is less clear.

Research aim

The work presented here builds on the contributions of previous studies and aims to identify and analyse communication skills learning outcomes via a systematic review and to present the results in a synthesised list.

Methodology

Research design.

A systematic review of the literature (stage 1) was completed to assemble published learning outcomes relevant to health professional communication skills. A parsimonious set of learning outcomes was then developed through an explicit and iterative process of qualitative synthesis (stage 2).

Research team

The research team for stage 1 was the core research team (CD, EM, JLK). Researchers in the field (RW-K, DN, FK: see acknowledgements) joined the core researchers, expanding the research team to six for stage 2 of this study. The stage 2 team consisted of educational researchers, social scientists and an educational linguist based in medical education. Four of the members had clinical backgrounds (physiotherapy). The members had published extensively across health professions education and communication skills research in areas that included scale development, healthcare simulation education, communication skills teaching, curriculum and resource development, feedback and assessment, clinical education and interprofessional education. All members were actively involved in the education of prequalification and postqualification health professional students and medical education research.

Theoretical perspective and guiding principles

This work forms part of a broader research programme exploring the social construction of skilled communication in the health professions. According to social constructionism, knowledge and meaning are constructed through the interaction of a learner and the surrounding environment; therefore, multiple realities exist and there is no ‘true’ interpretation of a phenomenon. 40 In this work, we did not set out to define a single truth in relation to learning targets for communication skills. Instead, we investigated published interpretations of this phenomenon and synthesised the findings for application in education. The research team considered that the results of this review would provide a foundation for others to interpret rather than a prescriptive list.

As a research team, we also ascribed to the notions of person-centredness during the qualitative synthesis phase of this study. Aligning with social constructionism, patient-centredness appreciates the individual and social dimensions of a phenomenon. Values of person-centredness including acknowledging patient-as-person and the multiple other persons involved in healthcare practice (eg, colleagues, students and the health professional themselves), and sensitivity to another's perspective and preferences 41 informed the choice of language in the results of this work.

Stage 1: systematic literature review

The first stage of this work was a systematic literature review to identify published learning outcomes of health professional communication skills programmes.

Inclusion and exclusion criteria

For inclusion, papers must have described learning outcomes within an education programme targeting the development of communication skills. Participating learners had to be health professionals and could be of any health profession and level of education. Any statement describing what learners were expected to know and/or do after participation in a programme was included, irrespective of the terminology used (eg, learning outcomes, objectives, targets and goals). Knowledge, behavioural and attitudinal learning outcomes were included. Papers were excluded if not available in English, if the education targeted English language fluency or the learning outcomes related to improvement of communication disorders (eg, those related to deafness or aphasia). Any study design was eligible for review.

Search strategy

A comprehensive search was conducted in August 2016, on the full holdings of MEDLINE, PsycINFO, ERIC, CINAHL plus and Scopus databases. Search terms included health professional, communication, training and their synonyms. The strategy used to search OVID databases is presented in online supplementary appendix A . The full yield from each database was exported to a bibliographic management system and duplicates deleted. On the basis of the title and abstract, papers that were ineligible were deleted. The remaining papers were read in full, and their eligibility was assessed against the inclusion and exclusion criteria. The primary author (CD) independently reviewed all papers and consulted with authors JK and EM on papers when eligibility was difficult to determine.

supplementary appendix

Data extraction

Table 1 describes the data that were extracted from each paper. All data were descriptively analysed and synthesised for discussion. Learning outcomes described in each paper were extracted and referenced to the source using an alphanumeric identifier. They were pooled and common learning outcomes were collated under preliminary themes, and subthemes, in a process of thematic analysis for commonality. 42 Preliminary thematic groups rendered the data set manageable for qualitative synthesis. The core research team (CD, EM, JLK) collaborated to delete duplicated learning outcomes (eg, those with exact wording) that were revealed as the learning outcomes were collated.

  • View inline

Items for data extraction

Quality analysis

This work does not resemble a traditional systematic review, in that quality assessment was not carried out on the included papers as we were not attempting to establish the validity of the recommended learning objectives. A systematic review approach, however, was used to identify eligible papers to enable transparency and replication and to address the research aim. Aligning with this aim, our focus was to extract the learning outcomes used in education design and review the content and quality of these statements. In other words, we applied a quality assessment filter to the learning outcomes (see decision guide, figure 1 ), not to the papers themselves.

  • Download figure
  • Open in new tab
  • Download powerpoint

Decision guide.

Stage 2: qualitative synthesis

In stage 2, learning outcomes identified in stage 1 were synthesised by the research team through an iterative process of individual, paired and group work. An initial meeting of the expanded research team was convened, and a random subset of learning outcomes was taken from the data set to rehearse the process of learning outcome synthesis. Synthesis was guided by the notions of person-centredness, and the decision guide is shown in figure 1 . The decision guide was adapted from that used by Dalton 43 in the refinement of assessment targets for competency to practice. This set of criteria was deemed the most fitting to guide the construction of learning outcomes. The decision guide was used to judge the quality of the learning outcomes and to address semantics in order to emphasise the important characteristics when constructing statements to be used for goal setting and practice. 28 , 36 , 43

Following this group meeting, the remaining learning outcomes were divided into five equivalent sections (eg, equivalent work per group member) and each member completed an individual review of one section. The primary author (CD) reviewed all sections.

Following this individual work, the primary author met with each member to compare work. After consensus was reached on all five sections, the core research team (CD, EM, JLK) reviewed themes and subthemes to reflect the synthesised data set. The refined learning outcomes were returned to the full group for consideration with a request for individual review prior to finalisation. A final group meeting enabled consensus regarding content, definition of terms and structure and presentation of the document.

Throughout the synthesis process, the number of learning outcomes was reduced as the research group collapsed multiple learning outcomes into one representative learning outcome, as further duplicates were identified, and as the group modified learning outcomes to align with the decision guide (see table 2 for an example of the qualitative synthesis process). At each step of this iterative process, the research group engaged in discussion to resolve different interpretations of the data, to highlight concepts that were absent from the data set, to resolve inconsistencies in language/terms and to ensure the learning outcomes continued to reflect the source literature. The primary author (CD) made an audit trail of decisions throughout the analysis. The audit trail included clear documentation of the process to make transparent the methods used in learning outcome refinement. 44

Example of qualitative synthesis

Stage 1: literature review

A total of 168 papers were included for review ( figure 2 ). Full reference details of eligible papers are presented in online supplementary appendix B ; further details are available on request from the primary author. A summary is presented below and in table 3 .

Flow of papers into the review.

The highest proportion of included papers was primary research (64%). There was a variety of study designs including randomised controlled trials investigating change in behaviour and posteducation learner evaluations. The majority of papers were published between 2010 and August 2016 (46%), originated from the USA (56%), described communication teaching in medicine (55%) and in qualified health professionals (56%). Learners in oncology (10%) outnumbered learners in other fields of practice. Learners in second year were most common in the prequalification population (8%). Many papers did not report a theoretical perspective of their work (58%); of those that did, the most common perspective reported was patient-centred care principles (7%).

Many included papers (40%) did not report any specific communication models that informed teaching. If cited, the most frequent models were the Kalamazoo Consensus Statement (7%), Calgary-Cambridge Referenced Observation Guides (5%) and the Three Functions of the Medical Interview (5%); a variety of other communication models were reported <4% of the time. Other factors that informed educational design were stakeholder engagement, with 51% of reports documenting turning to clinicians, researchers and the literature to design the learning outcomes of their programmes. Only 5% reported engaging patients as stakeholders in the development of programmes.

A total of 1669 learning outcomes for communication skills education were extracted from eligible papers; duplicated learning outcomes were deleted, reducing the list to 1073.

The 1073 learning outcomes identified in stage 1 were condensed to a final set of 205 learning outcomes ( figure 3 ). The learning outcomes are presented in themes of ‘knowledge’, ‘content’, ‘process’ and ‘perceptual’ skills. Three of these themes are based on the seminal work of Kurtz, Silverman and Draper. 45 The research group added a fourth theme, knowledge (see table 4 for detailed definitions and other terms used in the final learning outcomes document). ‘Knowledge’ learning outcomes made up 20% of the final list of learning outcomes. They relate to knowledge of the characteristics and modes of communication in healthcare and how emotions and relationships affect communication in healthcare. Learning outcomes within the ‘content’ and ‘process’ themes dominate the literature and therefore are represented heavily in the final set (36% and 35%, respectively). These learning outcomes include the important ‘what’ and ‘how’ of communication skills. 46 The learning outcomes within the content domain have been presented in the chronological sequence of a healthcare interaction for ease of representation. 47 , 48 Learning outcomes under the ‘perceptual’ skill theme made up 9% of learning outcomes in the final results. This final list of learning outcomes covers a wide range of skills proposed for communication teaching and is presented in online supplementary appendix C .

Final learning outcomes: themes and definitions

Flow of learning outcome (LO) refinement.

The aim was to identify and analyse communication skills learning outcomes via a systematic review and present results in a synthesised list for consideration and use by educators and researchers across the health professions. From 168 included papers, 205 learning outcomes for communication skills evolved through qualitative synthesis. As far as we are aware, no previous review has reported on a comprehensive systematic literature review to identify learning outcomes for communication skills and presented them for uptake across the health professions.

What did we learn about communication skills learning outcomes?

Patients are less likely to be involved in education design.

Patient-centred communication is considered fundamental to effective healthcare delivery, but these principles are not consistently used as a basis for communication skills teaching, with only 7% of papers reporting patient-centred care as an underpinning theoretical perspective. In addition, only 5% of papers reported engaging patients in any aspect of educational design. Although valued in health professions education, patient engagement is not well defined. 49 Regan de Bere and Nunn 50 have proposed ideas for the future pedagogy of patient (and public) involvement in health profession education. It seems appropriate that future communication skills educational design follows this lead and engages patients (and public) in the design of teaching and learning to explore what patients expect from health professionals, in terms of communication, as well as to acknowledge patients and patient-centred care as a unifying focus in health professions education. 50

Knowledge and perceptual skills are less common

Unlike the learning outcomes reported in the HPCCC, this paper includes learning outcomes relating to behaviour and knowledge. Knowledge forms part of the initial phases of scaffolding for learning and assessment, 51 and in learning theories describing skill development, a learner usually needs to know the terms and context of behaviour before they learn the behaviour itself. 52 In the past, knowledge-based learning outcomes have been criticised for being overemphasised in approaches to educational design, at the expense of behavioural learning outcomes. 18 In the communication skills literature, perhaps the emphasis has swung too far away from the knowledge agenda with most publications only reporting behavioural learning outcomes. The presented list brings communication skills in line with other skills teaching, by acknowledging the integration of knowledge and behaviour in ‘technical’ skill development. 28

Learning outcomes under the perceptual skill theme made up only 9% of learning outcomes in the final results. These learning outcomes relate to awareness and evaluation of self, others and context, which are considered important in reflective, self-regulating health professionals. 34 , 53 Despite ample literature promoting reflective practice in the health professions education field in general, 54 published programmes on communication skill development seem less focused on incorporating learning objectives to target these evaluative skills. The low level of reporting of perceptual skill objectives may reflect the challenge in defining these skills so that they are objective and measurable. Healthcare communication is unique, complex and nuanced, and therefore objectivity when defining these desired communication skills can be elusive. 26 However, clearly articulating these important skills is key to defining learning goals for performance development.

Language can be modified for greater application

During the qualitative synthesis process, the research group identified many of the learning outcomes that were specific to communication with a patient but that could equally apply to other communication interactions. The term ‘patient’ appeared in many included learning outcomes and was replaced in the refined list with the term ‘healthcare seeker’. The term healthcare seeker was more broadly applicable across healthcare contexts and acknowledged the active role that an individual can have in regard to his/her health. While the original learning outcome may have referred to communication between a health professional and care seeker, in some instances we found that learning outcomes could be easily reshaped to apply to communication with a different ‘communication partner’ entirely (eg, colleague or student). This parallel between skills in a therapeutic interaction and a collegial or educational interaction has been identified by others. 55 , 56

The research team also replaced all profession-specific terms identifying the target learner (eg, dentist) with the term ‘health professional’ as most learning outcomes had potential application to a variety of professions. The identification of these common learning outcomes may provide a platform for interprofessional education and facilitate a shift from ‘siloed’ health professional education to ‘collaborative’ practice, and education, within which communication is considered an essential common skill. 57

Learning objectives are often unclear or absent

Previous reviews reporting communication skills in medical education have criticised the quality of published learning outcomes, reporting that they were unclear or absent in many papers on communication education. 47 , 58 This was reflected in the current review with only 208 papers, of the 945 full text screened, reporting specific learning outcomes for the education described. Like Cegala and Lenzmeier Broz, 47 the authors acknowledge that the word count of many journals limits the inclusion of exhaustive lists of learning outcomes. However, even when learning outcomes were reported, they often targeted two to three constructs per statement, included value-laden words such as ‘good’ (eg, use good non-verbals) and many were ambiguous in their intention (eg, be able to apply the necessary communication skills adequately in a simulation). During stage 2 qualitative synthesis, the use of the decision guide focused attention on the design features of useful learning outcomes and guided the synthesis process. For example, many learning outcomes such as ‘show empathy’ were not accompanied by a descriptor or defined example; this rendered them neither useful, nor measurable or transparent (see figure 1 , decision guide). For learning outcomes such as this, we made the decision to combine all learning outcomes describing ‘empathic’ behaviour under a single learning outcome ‘demonstrate empathy in the following ways’: after which a list of observable behaviours were assembled (see item no. 136 in online supplementary appendix C ). This method of illustrating these ‘slippery’ learning outcomes, with observable behaviours, aimed to shape the learning outcome into a useful, measurable and transparent structure that would guide educators and learners.

Study strengths and limitations

This paper has a number of limitations. By casting a wide net in this literature review, we drew from the worldwide interest in this topic. In targeting a breadth of perspectives, we identified diverse studies, from educators across professions, which informed the results. A side effect of the broad search was the challenge of dealing with a large amount of data. Multiple stages of thematic analysis and refinement were required to create a manageable data set. The process of refinement necessarily included removal or rewording of items. We cannot be sure that this has not led to omissions in the final set of learning outcomes. However, the iterative nature of the qualitative synthesis process meant that all attempts were made to produce a representative list. Although these results represent a wide range of literature, they do not include input from other key stakeholders (ie, care seekers, representatives from different healthcare professions, learners and educators). Therefore, we do not claim that this is a complete list of learning outcomes illustrative of all stakeholders' needs. By focusing on specific communication skills education, we may have excluded communication learning outcomes in patient safety or clinical reasoning education programmes. In addition, we excluded papers reporting on communication disorders or English language fluency which may have also provided relevant learning outcomes.

Implications for practice and future research

This work provides educators and learners with a comprehensive set of learning outcomes for educational design and goal setting in healthcare communication. Aligning with a social constructionist perspective, we invite stakeholders to take these learning objectives and reinterpret them based on their context. This work may prompt educators to reflect on how knowledge and perceptual skills are taught in domains such as communication and to encourage educators to be cognisant of the quality of the learning outcomes they design.

Further research is required to interrogate these learning outcomes for accuracy and completeness by engaging with stakeholders (including care seekers) to align targets for learning with targets for practice. Future research might also explore the parallels between communication skills needed in care seeker interactions and those required for other healthcare conversations (eg, between healthcare professionals, trainees or policymakers). Perhaps by positioning communication skills as skills that are required with any ‘communication partner’, we can equalise the gravitas generally afforded to ‘doctor–patient’ communication in this field of literature.

Conclusions

This paper presents a synthesis of the vast literature in communication skills teaching and a list of 205 learning outcomes. These learning outcomes have been categorised into four domains and provide educators from across the healthcare professions with a basis from which to develop learning goals and programmes relevant to their setting.

Acknowledgments

The authors acknowledge Dr Fiona Kent for her contribution to the qualitative synthesis phase of this review.

  • McAllister L ,
  • O'Keefe M ,
  • Henderson A ,
  • ↵ Royal College of Physicians and Surgeons of Canada . CanMEDS Physician Competency Framework—Communicator. Secondary CanMEDS Physician Competency Framework—Communicator . 2015 . http://www.royalcollege.ca/portal/page/portal/rc/canmeds/framework/communicator
  • ↵ Chartered Society of Physiotherapy . Q.A. standards. Section 7: communication. Secondary Q.A. standards. Section 7: communication . 2015 . http://www.csp.org.uk/professional-union/professionalism/csp-expectations-members/quality-assurance-standards/section-7-co
  • ↵ Interprofessional Education Collaborative Expert Panel . Core competencies for interprofessional collaborative practice: report of an expert panel . Washington : Interprofessional Education Collaborative , 2011 .
  • Silverman J
  • Bachmann C ,
  • Abramovitch H ,
  • Barbu CG , et al
  • Silverman JD
  • von Fragstein M ,
  • Silverman J ,
  • Cushing A , et al
  • Kiessling C ,
  • Dieterich A ,
  • Fabry G , et al
  • Rodgers AT ,
  • Touchie C ,
  • Cate OT , et al
  • van Dalen J
  • Simon R , et al
  • van de Wiel HB
  • van den Eertwegh V ,
  • van Dulmen S ,
  • van Dalen J , et al
  • Dreyfus H ,
  • Berkhout JJ ,
  • Helmich E ,
  • Teunissen PW , et al
  • O'Neill G ,
  • Donnelly R ,
  • Fitzmaurice M
  • Deveugele M
  • Liberati A ,
  • Tetzlaff J , et al
  • Gibson WJ ,
  • Rodgers BL ,
  • Benson J , et al
  • Cegala DJ ,
  • Lenzmeier Broz S
  • Veldhuijzen W ,
  • van der Weijden T , et al
  • Knight LV ,
  • Wilkinson CE
  • Regan de Bere S ,
  • Perron NJ , et al
  • Bhutta ZA , et al
  • Hulsman RL ,
  • Winnubst JAM , et al

Contributors CD, EM and JLK designed the work. CD conducted the systematic review search and data extraction in stage 1. All authors (CD, EM, JLK, DN and RW-K) were involved in qualitative synthesis and analysis in stage 2. CD drafted the first version of the manuscript. All authors reviewed and revised the manuscript and approved the final version.

Funding This research received no specific grant from any funding agency in the public, commercial or not-for-profit sectors.

Competing interests None declared.

Provenance and peer review Not commissioned; externally peer reviewed.

Data sharing statement No additional data are available.

Read the full text or download the PDF:

IMAGES

  1. systematic review and thematic synthesis of qualitative literature

    qualitative synthesis and systematic review in health professions education

  2. (PDF) A qualitative systematic review and thematic synthesis exploring

    qualitative synthesis and systematic review in health professions education

  3. How to do a systematic review

    qualitative synthesis and systematic review in health professions education

  4. PPT

    qualitative synthesis and systematic review in health professions education

  5. (PDF) A qualitative review of the design thinking framework in health

    qualitative synthesis and systematic review in health professions education

  6. (PDF) Digital storytelling in health professions education: a

    qualitative synthesis and systematic review in health professions education

VIDEO

  1. What is Qualitative Evidence Synthesis?

  2. Data Analysis

  3. QUALITATIVE SYNTHESIS OF SYSTEMATIC REVIEWS

  4. There Are Four Versions of the PRISMA Flow Diagram

  5. THE LITERATURE REVIEW FAMILY

  6. Rehabilitation Measures Database

COMMENTS

  1. Qualitative synthesis and systematic review in health professions education

    Abstract. Context: Formal qualitative synthesis is the process of pooling qualitative and mixed-method research data, and then drawing conclusions regarding the collective meaning of the research. Qualitative synthesis is regularly used within systematic reviews in the health professions literature, although such use has been heavily debated in ...

  2. Qualitative synthesis and systematic review in health professions education

    2. CONTEXT Formal qualitative synthesis is the. process of pooling qualitative and mixed-. method research data, and then drawing con-. clusions regarding the collective meaning of. the research ...

  3. Qualitative synthesis and systematic review in health professions

    Qualitative synthesis is regularly used within systematic reviews in the health professions literature, although such use has been heavily debated in the general literature. This controversy arises in part from the inherent tensions found when generalisations are derived from in‐depth studies that are heavily context‐dependent.

  4. Qualitative synthesis and systematic review in health professions

    Formal qualitative synthesis is the process of pooling qualitative and mixed-method research data, and then drawing conclusions regarding the collective meaning of the research. Qualitative synthesis is regularly used within systematic reviews in the health professions literature, although such use has been heavily debated in the general ...

  5. REV ISS WEB medu 12092 47-3 252.

    REV ISS WEB medu 12092 47-3 252. Qualitative synthesis and systematic review in health professions education. Margaret Bearman1& Phillip Dawson2. CONTEXT Formal qualitative synthesis is the process of pooling qualitative and mixed- method research data, and then drawing con- clusions regarding the collective meaning of the research.

  6. Qualitative synthesis and systematic review in health professions

    We explore three representative qualitative synthesis methodologies: thematic analysis; meta-ethnography, and realist synthesis. These can be understood across two dimensions: integrative to interpretative, and idealist to realist. Three examples are used to illustrate the relative strengths and limitations of these approaches. Original language.

  7. In this issue

    They argue that authors should be transparent about their methodologies to enable better critical analysis of review findings. Bearman M, Dawson P. Qualitative synthesis and systematic review in health professions education. Med Educ; 47: 252-60. Improving the quality of qualitative research. Stenfors-Hayes and colleagues introduce ...

  8. Series: Practical guidance to qualitative research. Part 7: Qualitative

    To date, many qualitative synthesis methods exist. Qualitative synthesis includes review types that utilise qualitative ... Dawson P. Qualitative synthesis and systematic review in health professions education. ... Peters MDJ, Stern C, et al. Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping ...

  9. Qualitative synthesis and systematic review in health professions education

    Context Formal qualitative synthesis is the process of pooling qualitative and mixed-method research data, and then drawing conclusions regarding the collective meaning of the research. Qualitative synthesis is regularly used within systematic reviews in the health professions literature, although such use has been heavily debated in the general literature.

  10. Qualitative synthesis and systematic review in health professions

    Qualitative synthesis and systematic review in health professions education - Margaret Bearman interview Item Preview podcast_medical-education-podcasts-201_qualitative-synthesis-syst_1000404395979_itemimage.png . remove-circle Share or Embed This Item. Share to Twitter. Share to Facebook. Share to Reddit.

  11. Qualitative synthesis and systematic review in health professions education

    Context Formal qualitative synthesis is the process of pooling qualitative and mixed‐method research data, and then drawing conclusions regarding the collective meaning of the research. Qualitative synthesis is regularly used within systematic reviews in the health professions literature, although such use has been heavily debated in the general literature. This controversy arises in part ...

  12. Qualitative synthesis and systematic review in health professions education

    Context Formal qualitative synthesis is the process of pooling qualitative and mixed-method research data, and then drawing conclusions regarding the collective meaning of the research. Qualitative synthesis is regularly used within systematic reviews in the health professions literature, although such use has been heavily debated in the general literature.

  13. Learning outcomes for communication skills across the health

    Design: Systematic review and qualitative synthesis. Methods: A systematic search of five databases (MEDLINE, PsycINFO, ERIC, CINAHL plus and Scopus), from first records until August 2016, identified published learning outcomes for communication skills in health professions education. Extracted data were analysed through an iterative process of ...

  14. A qualitative systematic review and thematic synthesis exploring the

    Background There are increasing opportunities for healthcare professionals outside medicine to be involved in and lead clinical research. However, there are few roles within these professions that include time for research. In order to develop such roles, and evaluate effective use of this time, the range of impacts of this clinical academic activity need to be valued and understood by ...

  15. Analysing synthesis of evidence in a systematic review in health

    Background: Systematic reviews in health professions education may well under-report struggles to synthesize disparate evidence that defies standard quantitative approaches.This paper reports further process analysis in a previously reported systematic review about mobile devices on clinical placements.Objective: For a troublesome systematic review: (1) Analyse further the distribution and ...

  16. Learning outcomes for communication skills across the health

    Objective The aim of this study was to identify and analyse communication skills learning outcomes via a systematic review and present results in a synthesised list. Summarised results inform educators and researchers in communication skills teaching and learning across health professions. Design Systematic review and qualitative synthesis. Methods A systematic search of five databases ...

  17. Qualitative systematic reviews: their importance for our understanding

    This article outlines what a qualitative systematic review is and explores what it can contribute to our understanding of pain. ... Eldridge S, et al. A systematic review and meta-synthesis of the impact of low back pain on people's lives. BMC ... Dawson P. Qualitative synthesis and systematic review in health professions education. Med Educ ...

  18. Health Care Simulation in Person and at a Distance

    Abstract. Distance simulation is a method of health care training in which the learners and facilitators are in different physical locations. Although methods of distance simulation have existed in health care for decades, this approach to education became much more prevalent during the COVID-19 pandemic. This systematic review studies a subset ...

  19. Analysing synthesis of evidence in a systematic review in health

    In-depth consideration is limited though [Citation 3] about how to classify and analyse diverse evidence that thwarts standard quantitative approaches to the systematic review, and qualitative research synthesis is contested [Citation 4, Citation 5]. In making sense of a difficult synthesis, a post hoc standard veneer of slick reporting might ...