Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Evaluation of the clinical effectiveness of a work-based mentoring programme to develop clinical reasoning on patient outcome: A stepped wedge cluster randomised controlled trial

  • Aled Williams ,

    Contributed equally to this work with: Aled Williams, Alison Rushton, James J. Lewis, Ceri Phillips

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Writing – original draft, Writing – review & editing

    aled.williams10@wales.nhs.uk

    Affiliation University Hospital of Wales Physiotherapy Department, Cardiff and Vale University Health Board, Cardiff, Wales, United Kingdom

  • Alison Rushton ,

    Contributed equally to this work with: Aled Williams, Alison Rushton, James J. Lewis, Ceri Phillips

    Roles Conceptualization, Methodology, Project administration, Supervision, Writing – review & editing

    Affiliation School of Sport, Exercise and Rehabilitation Sciences, University of Birmingham, Birmingham, England, United Kingdom

  • James J. Lewis ,

    Contributed equally to this work with: Aled Williams, Alison Rushton, James J. Lewis, Ceri Phillips

    Roles Data curation, Formal analysis, Supervision, Validation, Writing – review & editing

    Affiliation Y Lab (Public Services Innovation Lab for Wales), School of Social Sciences, Cardiff University, Cardiff, Wales, United Kingdom

  • Ceri Phillips

    Contributed equally to this work with: Aled Williams, Alison Rushton, James J. Lewis, Ceri Phillips

    Roles Conceptualization, Methodology, Project administration, Supervision, Writing – review & editing

    Affiliation College of Human and Health Sciences, Swansea University, Swansea, Wales, United Kingdom

Abstract

Background

Despite persistent calls to measure the effectiveness of educational interventions on patient outcomes, few studies have been conducted. Within musculoskeletal physiotherapy, the effects of clinical mentoring on postgraduate physiotherapists have been explored, but its impact on patient outcomes is unknown. The objective of this trial was to assess the effectiveness of a work-based mentoring programme to facilitate physiotherapist clinical reasoning on patient outcomes.

Methods

In a stepped-wedge cluster RCT in the musculoskeletal physiotherapy outpatient departments of a large NHS organisation, 16 physiotherapists were randomised by cluster to receive the intervention—150 hours of mentored clinical practice—at one of 3 time periods; control was usual training. 441 patients submitted outcome measures: Patient-Specific Functional Scale (PSFS) (primary outcome measure), EQ-5D-5L, patient activation and patient satisfaction (secondary outcome measures). A further secondary outcome measure of physiotherapist performance was collected by an independent assessor observing the physiotherapists practice.

Results

80.0% of intervention patients achieved clinically significant PSFS scores compared with 63.8% of control patients. Binary logistic regression analysis modelling for time, cluster and patient characteristics showed strong statistical evidence for this difference (p = 0.023; odds ratio 4.24, 95%CI 1.22, 14.79). Physiotherapist performance scores improved from a mean of 47.8% (SD 3.60) pre-intervention to a mean of 56.0% (SD 4.24) (p<0.001). There was no statistical evidence for differences between groups on other secondary outcomes.

Conclusion

This is the first study that we aware of that provides patient outcomes measurement of an established educational intervention in physiotherapy, providing evidence that this type of intervention positively impacts patient outcomes and physiotherapist performance. This provides a basis for further research in education across other healthcare disciplines and outcome measures.

Introduction

Minimal research has investigated the effect of healthcare education and training on quality of care and patient outcome. The paucity of evidence has led the UK Department of Health to call for research investigating healthcare education and patient outcomes [1], specifically measuring staff and patient experience, clinical effectiveness and safety [2]. This call to measure the association between healthcare intervention and patient outcome is not new in the medical education literature, where it has long been argued that the education and development of clinicians should be evaluated on the basis of whether they achieve better health outcomes [38]. Different levels of evidence have been outlined, advocated and widely used for healthcare education, as illustrated by Kirkpatrick’s 4-level method for training evaluation [9] and Moore’s 7-level [10] outcomes model of continuing medical education. Table 1 below illustrates the longer established Kirkpatrick’s model along with examples of typical research measurements:

thumbnail
Table 1. Kirkpatrick’s 4 levels of training evaluation [11].

https://doi.org/10.1371/journal.pone.0220110.t001

It has been generally acknowledged that “Level 4 evidence” of patient outcomes on Kirkpatrick’s scale is difficult to obtain [1317] commonly leading to lower level “surrogate outcomes” such as competency and clinician performance being used to evaluate educational interventions [3, 7]. Where level 4 outcomes have been obtained in medicine [18, 19] and nursing [2022] the most frequently used outcomes to evaluate training and education are mortality rates, length of time in theatre, length of stay in hospital, complication rates and patient satisfaction. While this represents progress within the published research by exploring experience and safety (adverse events and patient mortality rates], measuring the clinical effectiveness of clinician education on patient outcomes continues to lack published evidence. Multiple authors across specialties in healthcare continue to call for the “gold standard” of high-level patient outcomes in the recent literature [12, 15, 16, 2327].

This is illustrated within the field of musculoskeletal physiotherapy, where there are only two published studies to the best of our knowledge which have explored the clinical effectiveness of an educational intervention for physiotherapists on their patient outcomes. The first [28, 29] showed no effect on patient outcomes of an 8-day psychosocial course when compared with the outcomes of a waiting list control group of clinicians. The fact that the educational intervention was delivered away from the clinical context of the clinicians was a weakness of this study and conceded by the authors to be a likely factor in the lack of effect on patient outcomes. The second [30] demonstrated a positive effect on patient outcomes by utilising an educational approach within the usual clinical context of the clinicians for patients with neck pain and disability. However, this trial has a narrow clinical focus on the cervical spine, as well as some methodological issues (retrospective data collection, limited confounding variables factored into analysis, individuals in different arms of the trial working in the same clinic introducing the risk of contamination) which caveat these positive findings. Apart from studies using patient outcome to measure effectiveness, multiple studies have explored education for advanced practice at postgraduate level, where mentoring in the clinical environment is used nationally and internationally to develop clinical reasoning and other features of expert clinical practice [3136]. Qualitative studies investigating the clinical mentoring component of such postgraduate programmes have demonstrated positive effects on physiotherapist performance and career, and the mentored practice component of postgraduate programmes is consistently evaluated as an effective component of Masters level education on transforming practice [3139]. While the effects of this educational intervention have been explored on the physiotherapist, to the best of our knowledge, its clinical effectiveness on patient outcomes has not been investigated to date.

In summary, there is a gap in the current evidence of the effectiveness of educational interventions as measured by patient outcome; the objective of this trial was to evaluate the clinical effectiveness of a work-based mentoring programme that aims to develop clinical reasoning for physiotherapists, on patient outcome. This objective was achieved as follows.

Methods

Design

A stepped wedge cluster randomised controlled trial (SWT) was designed and registration was assigned on 31/07/2012 (ISRCTN79599220); there are no ongoing or related trials for this intervention and therefore none are registered. The protocol was published [40] in line with the SPIRIT 2013 statement [41]. There were no deviations from the protocol. Clusters were defined as groups of physiotherapists (departments) within the outpatient musculoskeletal physiotherapy service of a large NHS organisation. This organisation had 6 sites for service delivery each based in a separate hospital with specific staffing localised to each site. The 6 sites were organised into pairs for training purposes, and so to reduce the risk of any potential contamination across sites, those pairs formed the 3 clusters (North, South and West) which was the unit of randomisation. Fig 1 illustrates the flow of participants through the trial (using the CONSORT 2010 extension [42] incorporating a flow diagram structure used in reporting of recent SWTs [43, 44]) as well as the patient outcome data collected at each point in each arm of the trial.

The rationale for a clustered design was the nature of the intervention and outcome. The training of clinicians was a cluster-level intervention as such education programmes are frequently targeted at departments or practices, and it is assumed that physiotherapists within a department might share their experiences [4547]. The effect of the intervention on patient outcome also required measurement of clusters of patients treated by the physiotherapists before and after the intervention.

In the SWT design, the intervention is rolled out to all clusters over time, with the order in which they receive the intervention being randomised [48]. This design was selected primarily for logistical and pragmatic reasons as this research was interested in the effects of this intervention being rolled out using outcomes for which it has not previously been considered, and the logistical and practical infeasibility of delivering this intervention (which is time and specialist ‘heavy’) in a parallel or crossover design [4953].

Intervention: The intervention was rolled out as illustrated in Fig 2 below, with white cells representing control periods, grey cells representing intervention periods; each cell also represents a data collection point.

The first step (time point) corresponds to the baseline measurement where none of the clusters had received the intervention, which is usual practice for this design [53]. At each subsequent step a cluster of participating physiotherapists crossed over from control to receive the intervention. The intervention was delivered at the start of the 6-month time-period, to allow for consolidation and application of the programme, before data collection occurred at the end of the time-period.

The intervention was a 150-hour clinical mentorship programme aimed at facilitating clinical reasoning. The rationale for the intervention was drawn from the educational standards document of the International Federation of Orthopaedic Manipulative Physical Therapists (IFOMPT) [54], a non-governmental federation promoting international excellence and unity in clinical and academic standards in the field of musculoskeletal physiotherapy; it is a subgroup of the World Confederation for Physical Therapy, which is a recognised partner of the World Health Organization. A minimum of 150 hours of mentored clinical practice is recommended for students, where the clinical mentor is a member of the member organization of IFOMPT. Furthermore, the clinical mentoring component of masters programmes that has been explored in terms of its impact on physiotherapist performance and career (as highlighted in the introduction) utilises this same educational approach. The intervention was delivered by mentors who were members of the Musculoskeletal Association of Chartered Physiotherapists (MACP), the UK member organisation of IFOMPT, having qualified at MSc / PgD level from a higher education establishment and who also had previous experience in delivering such mentorship at post-graduate level. The intervention took place in the usual clinical context of the participating physiotherapists, consisting of the mentors observing the participating physiotherapists assessing and treating new and follow-up patients, discussing and facilitating clinical reasoning processes immediately after the patients left the clinic. The model of clinical reasoning utilised for discussion was the dialectical model of Edwards et al [55] derived from qualitative studies of expert physiotherapists incorporating diagnostic and management aspects of clinical reasoning (such as narrative, collaborative and hypothetico-deductive reasoning) [5658].

Control: During the control steps of the trial, participants received their usual training allocation which involved in-service training on current evidence applied to practice (4 hours per month), technique sessions on practical skills (30 minutes per week), as well as mentoring sessions observing clinical reasoning (1.5 hours per month). This choice of control was on the basis that the comparison group needed to be plausible and fair [59] and that comparisons with non- intervention controls fail to inform on the issue of selecting the most effective methods from multiple available options [60].

Participants

All qualified physiotherapists working in the selected NHS organisation whose majority of time practising was in the musculoskeletal outpatient context were approached to participate. Consecutive consenting patients attending the outpatient musculoskeletal physiotherapy service for treatment by the participating physiotherapists during data collection periods who were eligible to participate were also approached. Exclusion criteria are outlined in Table 2 below:

Ethical considerations.

Ethical approval for the trial was sought and obtained from the regional Research Ethics Committee (ref: 12/WA/0078) on 20/04/2012 and informed written consent was gained from the participating physiotherapists and their patients. The physiotherapist recruitment period was from 01/09/2012 to 28/02/2013 and patient recruitment was from 01/04/2013 to 31/05/2014 (as seen in Fig 3). Follow up was at discharge with a cut-off at 12 months post-recruitment; no patients exceeded this cut-off. Participation in the trial was on a voluntary basis. Physiotherapists were reassured that refusal to consent to be part of this trial would not influence their receipt of training or have any other impact on their professional development and participating patients were assured that refusal to consent to complete outcome data for this trial would have no impact on their care.

thumbnail
Fig 3. Patients recruited, and data collected in each cluster period n (%).

https://doi.org/10.1371/journal.pone.0220110.g003

Outcome measures

Patient level outcomes.

The primary outcome measure was the proportion of patients achieving clinically significant improvements on the Patient Specific Functional Scale (PSFS), a patient-specific tool for measuring health-related quality of life (HRQL). The tool asks that patients select their most relevant activities or functions and then rate their ability to complete this activity on a 11-point scale at a level experienced prior to injury or change in functional status, where "0" represents “unable to perform” and "10" represents “able to perform at prior level”. Patients were asked to select a value that best described their current level of ability on each activity. These outcomes were collected in collaboration with the patient’s physiotherapist (which is the advised method for use of the tool [73]) and entered on to the patient’s questionnaire containing the other outcome measures. Clinical significance was set at a level using established data for the minimal clinically important difference (MCID) being reported to be 3 points for each single identified activity and 2 points for aggregate activities. The proportion of patients achieving clinically significant outcomes (i.e. achieving the level for MCID) was calculated.

The secondary outcome measures at patient-level were EQ-5D-5L, the Patient Activation Measure and the MedRisk instrument. The EQ-5D-5L is a two-section generic measure of HRQL [74], the descriptive section yielding a 5 digit health state profile which was then converted into a single index value and the visual analogue scale giving respondent’s self-rated health on a scale of 0–100; higher scores in both sections indicate higher levels indices of health. The Patient Activation Measure (PAM) is a measurement of the knowledge, skills and confidence a person has in managing their own health and healthcare [68], delivering a patient activation score of between 0 and 100, higher scores indicating greater levels of activation. The MedRisk instrument for Measuring Patient Satisfaction with Physical Therapy Care (MRPS) is a measure of patient satisfaction in the musculoskeletal physiotherapy context [72], mean scores being used to assess satisfaction from 1.0 (highly dissatisfied) to 5.0 (highly satisfied). These secondary measures all have established measurement properties [6769, 7480]. The patients completed self-assessment questionnaires which contained a unique identifier for the treating physiotherapist on each questionnaire to ensure blinding and sent to the lead researcher. Patient outcomes were completed at baseline (after the first appointment with the physiotherapist) and endpoint (discharge); timescales between these points varied between patients depending on the length of the course of treatment from 3 weeks to 13 weeks. No patients crossed over from control to intervention periods.

Physiotherapist level outcomes.

Physiotherapist performance was assessed by each physiotherapist being observed with a new patient and a follow-up patient and then discussing their practice. This was performed by an assessor who was blinded (to whether the participant had received the intervention), independent (from outside the organisation) and experienced (familiar with using the performance criteria). Two physiotherapist-level outcomes were used; both derived from the assessment of Master’s level post-graduate student performance in the musculoskeletal context [32]. One outcome was the score and the other outcome was the score divided into 10% bands (e.g. 40–49%; 50–59%). The specific tool selected for measurement of performance was selected on the basis of its supporting empirical data where the construct of the assessment tool was supported by triangulation of data from multiple methods and studies [31, 33, 81, 82].

Sample size

Analysis was to be performed using 2 different estimates of intergroup differences; clinically, the more relevant use of the primary outcome measure was estimated to be 10% more patients achieving clinically significant outcomes on PSFS but as there was limited data on which to accurately base a sample size calculation, a second estimate was used for the power calculation. Sample size calculations followed the approach outlined for stepped-wedge designs in Hussey and Hughes [53] and used conventional values of 0.80 for statistical power and 0.05 for statistical significance, and were performed for a realistic sample of 12 participating physiotherapists organised into 3 clusters. A mean improvement of 1 point on the PSFS with a standard deviation of 2.0 (assuming a PSFS difference of 0.5 across the control arm and a 1.5 difference across the intervention arm) with an ICC of 0.05, a sample of 432 patients was required. An increase in the assumed ICC to 0.5 would result in a small drop in power to 77%.

Randomisation and blinding

The unit of randomisation was the cluster, defined as physiotherapy departments. The participating physiotherapists and clinical mentors implementing the intervention were aware of which cluster was receiving the intervention [49, 52]. The clusters of participating physiotherapists were randomly allocated to the sequence of intervention (to receive the intervention in time period 2, 3 or 4), by computer programme [83] by one of the clinical mentors. To ensure allocation concealment and the blinding of the lead researcher to the sequence generation and intervention allocation, this mentor also allocated a unique identification code to each physiotherapist that was used on all outcome measure questionnaires and kept a sealed copy of the key code linking codes to physiotherapists. This mentor was involved in delivering mentoring in all 3 clusters and was not involved in any data analysis. The key code was only opened by the lead researcher once all outcome data had been collected and analysed from each of the 4 time points. The independent assessor of physiotherapist performance was blinded as to whether the physiotherapist has received the intervention during their assessment. Patients were also blinded to whether their treating physiotherapist had received the intervention. In summary, patient outcome assessment was standardized but not blinded to the data collector, physiotherapist outcome assessment was blinded to the data collector and data analysis was blinded.

Statistical methods

The characteristics of the patient populations exposed and unexposed to the intervention in the trial are summarised using frequencies and percentages, means and standard deviations, as appropriate. The null hypothesis, that there was no difference in the proportion of eligible patients achieving clinically significant outcomes before and after their physiotherapists had exposure to the intervention, was tested using a logistic regression model with a binary outcome—whether or not each patient achieved the MCID on the PSFS.

As this was a SWT, it was essential to adjust outcomes for cluster [84] and for step (time) [85]. This was achieved by fitting simple logistic/linear regression with linear trend for step and fixed effect for cluster. The analysis took account of the within cluster correlation by adjusting for each cluster with a fixed effect. This is not possible in a parallel cluster randomised trial, due to the multicollinearity between cluster and trial arm, which would over-parameterise the model. However, because this is a stepped wedge trial, where each cluster is in both the control arm and then the intervention arm, this is a valid approach. A recent article has highlighted different extensions where time can be fitted to the model as a random or linear variable [84]. In our trial with only 3 clusters it was not possible to estimate the variance parameter of a random effect. In addition, estimating cluster, step and arm all as fixed effects would over-parameterise the model; as a compromise step was fitted as a linear trend. This adjustment for cluster and step is reported as the minimally adjusted model. In addition, a second model additionally adjusted independent variables that were deemed important to consider including: patients’ duration of symptoms (chronicity being a likely indicator of poor outcome [86]); exercise history (previously shown to be an indicator of outcome [87]); OrebroMPQ (predictive of poor outcome [88, 89]); initial PAM (patients’ ability to self-manage being likely to influence outcome [69, 90, 91]); and other patient factors (initial PSFS, EQ5D index, sex, age, ethnicity) [87]. This is referred to as the fully-adjusted model. Significance was set at the 5% level and 95% confidence intervals were reported. Null hypotheses and analyses for secondary outcomes took a similar form to that for the primary outcome. Two sensitivity analyses were also performed to address missing data for best (all missing data achieving MCID) and worst case (all missing data not achieving MCID) scenarios. All analysis was by intention to treat.

Results

The baseline characteristics of the patients recruited to the trial are outlined in Table 3. The control arm contained more female patients, and more patients who had less of a background in exercise. The intervention arm contained more patients who the Örebro questionnaire classifies as high risk i.e. at higher likelihood of a poor outcome. It was important, therefore, to model for these factors in the final analysis. Other between group differences appear small.

thumbnail
Table 3. Baseline characteristics of exposed and unexposed patients (N = 441).

https://doi.org/10.1371/journal.pone.0220110.t003

Recruitment and follow-up to the trial

The stepped wedge CRT design was carried out as represented in Fig 3. 441 patients were recruited by 16 physiotherapists (power calculations required 432 patients). The clusters were different sizes–North being the biggest and West the smallest.

Primary outcome measure

Using the established data for the minimal clinically important difference (MCID) of 3 points for each single identified activity and 2 points for aggregate activities, the proportion of patients achieving the level for MDC and MCID was 80.0% of the patients in the intervention arm compared with 63.8% of patients in the control arm (Fig 4).

thumbnail
Fig 4. Distribution of clinically significant Improvements in PSFS by trial arm excluding (a) and including (b) missing values.

https://doi.org/10.1371/journal.pone.0220110.g004

Using logistic regression analysis, in the minimally adjusted model the difference in the proportion of patients achieving the MCID between the intervention and control arm gives an odds ratio of 4.14 and a p value of 0.009. In the fully adjusted model, the odds ratio strengthens to 4.24 with a p value of 0.023 (Table 4).

thumbnail
Table 4. Summary of primary outcome measure with adjusted and unadjusted intervention effects.

https://doi.org/10.1371/journal.pone.0220110.t004

Sensitivity analyses.

To explore the potential effect of missing data, two sensitivity analyses were performed, assuming all patients with missing outcome data to have made a good clinical outcome (for a best-case scenario), and then assuming all patients with missing outcome data to have made a poor clinical outcome (for a worst-case scenario). In the first sensitivity analysis, all missing outcomes were recoded as clinically significant. This adjustment still yielded a statistically significant value for the intervention in the minimally adjusted model (p = 0.018) with an odds ratio of 3.38 (95%CI 1.24, 9.27) as well as in the fully adjusted model (p = 0.020) with an odds ratio of 4.23 (95%CI 1.25, 14.31). In the second sensitivity analysis, recoding all missing outcomes as clinically non-significant, the effect of the intervention was again statistically significant in the minimally adjusted model (p = 0.012) with an odds ratio of 3.19 (95%CI 1.29, 7.89) but this statistical significance is lost in the fully adjusted model (p = 0.074, odds ratio 2.74 95%CI 0.91, 8.27).

A further analysis was performed by aggregating the PSFS data and comparing mean changes as a second view of the primary outcome, as pre-planned in the protocol. This was performed to explore whether similar effects were seen using scale data and linear regression modelling as were seen using the binary data with logistic regression. During the power calculation for the study, a difference of 1.0 points was used as the estimate for what could be perceived as clinically significant. The minimally adjusted model (using the same factors as used for the logistic regression modelling for the binary values) yielded a difference of +1.35 points for the intervention over control and is statistically significant (95%CI 0.14 to 2.55; p = 0.028). However, this loses statistical significance in the fully adjusted model to +0.42 points (95%CI -0.61 to 1.45; p = 0.42).

Secondary outcome measures

Physiotherapist performance.

The physiotherapists’ performances improved as can be seen in Fig 5. Fig 5A shows the mean score of the physiotherapists’ performance as 47.8 prior to the intervention, and 55.1 post intervention (post I). A subsequent measure of performance post-intervention shows a further mean increase to 56.9 (post II). Fig 5B captures the clinical and educational relevance of these score changes. Displaying performance by banding, 75% of physiotherapists prior to intensive mentoring would have been below the mark of 50, which is the pass mark for Masters level performance on the University assessment criteria. This reduces to only 6.7% of physiotherapists post-intervention, with 53.3% of physiotherapists achieving a score in a higher band of between 60 and 70, which none achieved prior to the intervention.

Patient outcomes.

The effect on patients measured using the primary outcome measure (selected amongst other reasons due to its responsiveness) was not reflected in the secondary outcome measures (Table 5). There was no statistical evidence for any differences between groups on generic HRQL (measured using EQ-5D-5L index and VAS), patient satisfaction (measured using the MRPS), or patient activation (measured with the PAM). It should be noted that there were more missing data in this section, with approximately a third of secondary outcome measures incomplete with little difference by trial arm. For the MRPS there were 77 incomplete outcomes in the unexposed arm (31.2%) and 71 in the exposed arm (36.6%); for the PAM and EQ5D there were 78 incomplete outcomes in the unexposed arm (31.2%) and 71 in the exposed arm (36.6%).

thumbnail
Table 5. Summary of secondary outcome measure by exposure to the intervention, with adjusted and unadjusted intervention effects.

https://doi.org/10.1371/journal.pone.0220110.t005

Discussion

Primary outcome

Facilitating physiotherapists’ clinical reasoning through work-based mentoring significantly increases the proportion of their patients achieving clinically significant patient-specific HRQL outcomes. While the statistical evidence demonstrated is important, the increase from just under 2 in 3 patients attending the physiotherapy service achieving the MCID on PSFS to 4 in 5 patients achieving this is a welcome increase for physiotherapists, referring clinicians, managers and not least of all, patients. This is especially important given that this outcome measure is set for the functional activities that patients themselves have identified as being most affected by their musculoskeletal condition and therefore the most relevant to address in the rehabilitation process. The greater effect of this educational intervention over that of usual training could be attributed to its nature: the facilitation of clinical reasoning, with a focus on narrative reasoning (exploring and understanding the patient’s lived experience), diagnostic reasoning (incorporating analytic and non-analytic reasoning), management reasoning (both instrumental and communicative strategies) as well as developing collaborative reasoning and consensual approaches to arriving at a diagnosis and selecting management goals and options are likely to foster a more patient-centred and efficient approach [55, 57]. It could also be attributed to its context; being conducted in the workplace allowed real time feedback specific to the clinician in the context of their usual work. A third factor could be its delivery being regular over several weeks, making reinforcement and practice change more likely [30].

Secondary outcomes

The clinical effectiveness of the intervention on patient-specific outcomes is also reflected in improvements in physiotherapist performance, as seen in the blinded measures pre- and post-intervention. The finding of 75% of the participants being below the pass mark for a masters clinical module is unsurprising as the participants were without the input of a taught masters curriculum leading up to their assessment, making the improvements in performance in response to the intervention impressive in that over 90% of the participants would have ‘passed’ on a tool that measures performance at masters level. In addition, over half of the participants were assessed to be at a higher banding than the pass level. The validity of these findings are supported by the assessments being performed by a blinded assessor who was familiar in using the tool as a clinical mentor for the University where the tool is normally used. One may question whether these improvements were not due to the intervention but due to improvements over time with the physiotherapists learning through experience. The changes in physiotherapist performance (Fig 3A) show measures taken at approximately 12-month intervals, with 1 set of pre-intervention and 2 sets of post-intervention measures. The biggest increase is the 7.3-point difference between pre-intervention and post intervention compared with the 1.8-point difference between the 2 post-intervention measures, pointing to the biggest gains being made due to the intervention rather than just over time.

The findings of the primary outcome measure in this trial are not reflected in the secondary outcome measures of generic HRQL, satisfaction or activation. Possible reasons for this lack of effectiveness may be due to the trial being underpowered for these measures, as the power calculation for this trial was performed on the primary outcome measure of the PSFS. Furthermore, this may have been compounded by missing data being greater for these secondary measures. This was owing to the PSFS data being able to be obtained by the physiotherapist during a telephone call to notify patients who were not present to complete the questionnaire at discharge that a postal questionnaire was being sent. The secondary measures being self-report questionnaires that required the patient to complete and return them meant that they could not be captured by telephone. A second reason may be the responsiveness of the instruments. Studies exploring the comparative responsiveness of the PSFS have demonstrated it to be more responsive than condition-specific outcomes used in the MSK context [66, 92, 93]. While the EQ-5D-5L was selected due to its extensive evidence base and validity as well as being accessible and its broad currency across multiple disciplines, it was not selected as the primary outcome measure on the basis that, as a generic measure of HRQL, it was likely to be less responsive than the patient-specific measure of HRQL, the PSFS [94, 95]. The lack of significant difference in patient satisfaction may well be attributable to a ceiling effect. International data as presented in a meta-analysis of use of the MRPS across outpatient settings in northern Europe, North America, the United Kingdom and Ireland shows patient satisfaction with MSK physiotherapy to be consistently high, with a pooled estimate of 4.44 (95% CI 4.41, 4.46) out of 5.0 [77]. In this trial, the patients in the intervention group reported a mean satisfaction of 4.6 which is higher than the data from the international studies. Ceiling effects are considered to be present if more than 15% of respondents achieve the highest possible total score [96] and indeed, the most common response to the questionnaire was 5.0 out of 5.0, with 14% of the sample submitting this response. While this figure is just outside the cut-off for ceiling effects, with the level of satisfaction so high for the control group, achieving statistically evidence for improvements in satisfaction for the intervention group was extremely challenging. The lack of difference between groups in patient activation may be attributable to the nature of the training, requiring a more specific educational approach on developing patients’ skills in self-management. The intervention in this trial was to enhance the physiotherapists’ clinical reasoning skills within a framework that was explicitly patient-centred. Conceivably, this should encompass empowering the patient’s self-management skills, and equipping them to manage their own health effectively. However, the results of our trial show that while the physiotherapy service improves patients’ levels of activation, in that both arms of the trial showed improvements in PAM scores, the intervention added no value in this area of patient care. The developers of the PAM suggest approaches to training of clinicians specifically tailored to improve their ability to improve patient activation [97].

Comparison with existing literature

The results of this trial contrast with those of a trial by Overmeer and colleagues [28, 29] investigating the effect of an 8-day university course for physiotherapists on patient outcome. While the education programme did produce changes in clinicians’ attitudes, beliefs and knowledge, there was no effect on patient outcomes. The authors propose that this lack of effect may be due to the delivery of the education being away from the clinical context, and that the educational strategy most likely to change practice behaviour (and therefore clinical outcomes) is to educate within the clinical context and provide direct feedback on the clinician’s encounter with actual patients [28]. This is precisely the strategy that the intervention in the current trial employed and may explain the difference in results.

One trial by Cleland and colleagues [30] has previously demonstrated a positive effect on patient outcomes of an educational intervention to the treating physiotherapists, in the private sector in the United States. Employing a parallel design, both groups of clinicians received education via a 2-day course on the management of neck pain. The intervention group received ongoing education (part of which was delivered in the clinical setting); the control group received no further education. While the changes in pain scores were not significantly different for patients treated by the two groups of physiotherapists, reductions in disability scores were significantly greater (p = 0.019) in patients of the intervention group, with a mean difference of 4.2 points (95% CI 0.69 to 7.7) on the Neck Disability Index. Our trial has similar results to this trial but for a much broader caseload of patients (the focus of Cleland’s trial is only on patients with neck pain and uses a condition-specific measure). Also, the Cleland trial only controlled for the patient variables of age, gender and initial pain and disability scores; with prognostic factors not explored. There was no exploration of therapist factors or performance change and the intervention being delivered to individuals rather than clusters gave the potential for contamination bias (clinicians from different groups worked in the same clinics). It is worth noting that while the educational interventions were quite different between our trial and that of Cleland’s, both were ongoing and involved actual patients within the physiotherapist’s own clinical practice setting, supporting the assertion that ongoing education that occurs within the clinicians’ context is likely to affect patient outcome [28, 29].

Looking outside of the physiotherapy literature at other disciplines shows other examples of where level 4 evidence has been delivered to measure clinical effectiveness. This trial contributes new evidence to this broader literature in terms of the type of level 4 evidence provided, with the PSFS being aligned with the World Health Organisation’s International Classification of Functioning, Disability and Health (ICF) framework [98100] and measuring functional activities selected by—and therefore highly relevant to—the patient. Analysing the level 4 outcomes used in medicine shows that those most frequently used are the absence of harm, or procedural success measured by time taken. Reviews of simulation-based medical education [18] and residency training on patient outcomes [19] and studies on the effect of baccalaureate degrees on Nursing [2022] use mortality rates, length of time in theatre, length of stay in hospital, complication rates and patient satisfaction as the patient outcomes to evaluate training and education. This trial provides an alternative pattern for selecting outcomes for educational interventions. Accepting that different disciplines will have different norms and expectations of what to measure, and that surgical approaches carry high risk, the focus on health and function, activation and participation of our trial provides a positive outcomes approach by measuring functional activities identified as meaningful by the patient. Other disciplines could utilise such outcomes alongside their traditional measures to give a more holistic appraisal of the educational interventions under exploration. Furthermore, this trial provides evidence that is robust in that it compares the intervention with an educational control and demonstrates statistical and clinically significant improvements over this control. The reviews of the medical education disciplines cited above found that some of the studies showed no effects, or small effects when compared with other training approaches, or positive effects that were not statistically significant.

Our trial also builds on findings of qualitative studies exploring masters courses where this type of intervention is used, adding quantitative data to the qualitative themes identified. It is important to clarify that this is not a claim for the superiority of quantitative data over the established qualitative findings; rather, it provides new Level 3 outcomes evidence (outcomes which focus on the change in clinical behaviour in routine practice resulting from training) through direct observation of the participants in the clinical setting which is the ideal means of assessment [12]. By utilising a blinded assessor and a validated tool in a randomised controlled trial format we were able to demonstrate statistically significant change across the intervention, thus substantiating the claims of practice change in response to this type of intervention. Furthermore, these statistically significant changes in performance are reflected in statistically significant changes in the primary outcome measure of the patients treated by the participants.

Clinical Implications

This trial answers the calls from the medical education literature to evidence educational interventions for clinicians with patient outcomes. The clinical effectiveness of the intervention may be due to its nature, evidencing claims and convictions that have been asserted in the medical education literature for some time that training applied in context and incorporating real time feedback for clinicians is most likely to affect patient outcome [63, 101105]. Clinicians who are seeking to undertake external courses or higher education degrees to develop their level of expertise, may wish to select those that offer modules involving clinical placement or mentoring in the workplace. Hospital managers, rather than reducing training budgets and opportunities [106, 107] might wish to develop and fund training programmes for clinical staff with the goal of achieving improved patient outcomes. Training co-ordinators may wish to review the context within which their training occurs; could it be delivered over a sustained period, more specifically to the learning needs of the clinician, with real time feedback?

Strengths and weaknesses

This is the first rigorous trial that has measured the clinical effectiveness of an educational intervention not just on clinician performance but also on patient outcomes, using a responsive patient-centred primary outcome measure. The diverse patient population represented here provides generalisability and the trial approach comes from a “realistic evaluation” standpoint, being delivered and measured in an NHS organisation which could be reasonably replicated.

A limitation of this trial is that the patients providing outcome data were required to be recruited by the physiotherapists participating in the trial. It has been highlighted elsewhere that in SWTs where recruitment is extended over time or there is no blinding to the intervention there is a risk of selection bias [108]. Even though clusters were allocated to receive the intervention in a random order and patients were blinded to their therapists’ intervention exposure, it was not possible to blind the participating therapists; introducing the possibility that physiotherapists could have recruited patients more likely to have a positive outcome after their intervention while neglecting to recruit patients more likely to have a negative outcome. The results show that demographics between the intervention and control patients were similar on all factors other than gender (and these factors were included in the regression model for analysis) and so there is no evidence of this occurring; nonetheless, the recruitment of patients by participating physiotherapists is a methodological weakness.

Unanswered questions and future research

While this trial has shown a positive effect on patient-specific outcomes, it did not have an effect on all outcomes. In particular, the lack of effect on patient activation was surprising, raising the question ‘how do we train our clinicians in order to improve patients’ abilities to manage their own health?’ Future research might also explore the reproduction of similar interventions in broader professional contexts and the subsequent effect on patient outcome.

Conclusion

This is the first trial to measure the clinical effectiveness of an educational intervention with level 4 outcomes in the field of physiotherapy. It is also the first trial to provide level 4 evidence for the clinical effectiveness of an educational intervention that is utilised in higher education modules in masters degrees aimed at developing expertise in physiotherapy. As such this trial adds to the rich qualitative data supporting such an intervention by providing a quantitative viewpoint and data to show that such an intervention ‘works’ in terms of delivering improved patient outcomes, augmented by the secondary outcome of level 3 evidence demonstrating performance change in physiotherapist clinical reasoning alongside the improvements in patient outcome.

Acknowledgments

We thank the patients who consented for their outcomes data to be used in this trial and the physiotherapists and their mentors who committed to this trial. We thank Sean Grove (Clinical Specialist, Bristol Community Health) for his independent assessment of physiotherapy performance. We also thank Professor Kerenza Hood (Director of the Centre for Trials Research at Cardiff University) for technical advice on trial design and methodology and Doctor Alan Watkins, Swansea University Medical School for the sample size calculations. We thank the Cardiff and Vale University Health Board for their support in allowing the study to run in their physiotherapy service.

References

  1. 1. Nolte E, Fry CV, Winpenny E, Brereton L. Use of outcome metrics to measure quality in education and training of healthcare professionals—A scoping review of international experiences.; 2011.
  2. 2. DoH. Developing the Healthcare workforce: From Design to Delivery. In: Health Do, editor. 2012.
  3. 3. Prystowsky JB, Bordage G. An outcomes research perspective on medical education: the predominance of trainee assessment and satisfaction. Med Educ. 2001;35(4):331–6. pmid:11318995
  4. 4. Chen FM, Bauchner H, Burstin H. A call for outcomes research in medical education. Acad Med. 2004;79(10):955–60. pmid:15383351
  5. 5. Fineout-Overholt E, Johnston L. Evaluation: an essential step to the EBP process. Worldviews Evid Based Nurs. 2007;4(1):54–9. pmid:17355411
  6. 6. Fincher RM, White CB, Huang G, Schwartzstein R. Toward hypothesis-driven medical education research: task force report from the Millennium Conference 2007 on educational research. Acad Med. 2010;85(5):821–8. pmid:20520035
  7. 7. Kalet AL, Gillespie CC, Schwartz MD, Holmboe ES, Ark TK, Jay M, et al. New measures to establish the evidence base for medical education: identifying educationally sensitive patient outcomes. Academic Medicine: Journal Of The Association Of American Medical Colleges. 2010;85(5):844–51.
  8. 8. Whitcomb ME. Research in medical education: what do we know about the link between what doctors are taught and what they do? Acad Med. 2002;77(11):1067–8. pmid:12431913
  9. 9. Kirkpatrick DL, Kirkpatrick JD. Evaluating training programs: The four levels. 3rd ed. San Francisco: Berrett-Koehler.; 2006.
  10. 10. Moore DJ, Green J, Gallis H. Achieving desired results and improved outcomes: integrating planning and assessment throughout learning activities. Journal Of Continuing Education In The Health Professions. 2009;29(1):1–15. pmid:19288562
  11. 11. Kirkpatrick D. Measuring and Evaluating Impact. In: Biech E, editor. The ASTD Handbook for Workplace Learning Professionals. Danvers, MA: ASTD Press; 2008.
  12. 12. Dewan M, Walia K, Meszaros ZS, Manring J, Satish U. Using Meaningful Outcomes to Differentiate Change from Innovation in Medical Education. Academic Psychiatry. 2017;41(1):100–5. pmid:26976400
  13. 13. Schneider SM, Chisholm CD. ACGME Outcome Project: Phase 3 in Emergency Medicine Education. Academic Emergency Medicine. 2009;16(7):661–4. pmid:19500076
  14. 14. Gonnella JS, Hojat M. Medical education, social accountability and patient outcomes. Medical Education. 2012;46(1):3–4. pmid:22150187
  15. 15. O’Malley PG, Pangaro LN. Research in medical education and patient-centered outcomes: Shall ever the twain meet? JAMA Internal Medicine. 2016;176(2):167–8. pmid:26641544
  16. 16. Van Melle E, Gruppen L, Holmboe ES, Flynn L, Oandasan I, Frank JR, et al. Using Contribution Analysis to Evaluate Competency-Based Medical Education Programs: It’s All About Rigor in Thinking. Academic Medicine. 2017;92(6):752–8. pmid:28557934
  17. 17. Pritchard L. Getting the right outcome. Medical Education. 2003;37(11):952–3. pmid:14629406
  18. 18. Zendejas B, Brydges R, Wang AT, Cook DA. Patient Outcomes in Simulation-Based Medical Education: A Systematic Review. Journal of General Internal Medicine. 2013;28(8):1078–89. pmid:23595919
  19. 19. van der Leeuw RM, Lombarts KM, Arah OA, Heineman MJ. A systematic review of the effects of residency training on patient outcomes. BMC Medicine. 2012;10(1):65.
  20. 20. Aiken LH, Clarke SP, Cheung RB, Sloane DM, Silber JH. Educational Levels of Hospital Nurses and Surgical Patient Mortality. JAMA: the journal of the American Medical Association. 2003;290(12):1617–23. pmid:14506121
  21. 21. Friese CR, Lake ET, Aiken LH, Silber JH, Sochalski J. Hospital Nurse Practice Environments and Outcomes for Surgical Oncology Patients. Health services research. 2008;43(4):1145–63. pmid:18248404
  22. 22. McHugh MD, Kelly LA, Smith HL, Wu ES, Vanak JM, Aiken LH. Lower Mortality in Magnet Hospitals. Medical care. 2013;51(5):382–8. pmid:23047129
  23. 23. Delisle DR, Nash DB. Evaluating the Effect of Physician Fellowship Programs on Surgical Outcomes for Coronary Artery Bypass Grafting Procedures. American Journal of Medical Quality. 2017;32(3):322–9. pmid:27259870
  24. 24. Kanter SL. Proposals to Strengthen the Link Between Medical Education and Better Health for Individuals and Populations. Academic Medicine. 2011;86(11):1329. pmid:22030626
  25. 25. Dauphinee WD. Educators must consider patient outcomes when assessing the impact of clinical training. Med Educ. 2012;46.
  26. 26. Golub RM. Medical education theme issue 2016: Call for papers. JAMA. 2016;315(10):989-.
  27. 27. Norcini J. Understanding learning in the workplace for those who practise: we can’t wait another 50 years. Medical Education. 2016;50(1):18–20. pmid:26695462
  28. 28. Overmeer T, Boersma K, Denison E, Linton SJ. Does teaching physical therapists to deliver a biopsychosocial treatment program result in better patient outcomes? A randomized controlled trial. Phys Ther. 2011;91(5):804–19. pmid:21451098
  29. 29. Overmeer T, Boersma K, Main CJ, Linton SJ. Do physical therapists change their beliefs, attitudes, knowledge, skills and behaviour after a biopsychosocially orientated university course? J Eval Clin Pract. 2009;15(4):724–32. pmid:19674225
  30. 30. Cleland JA, Fritz JM, Brennan GP, Magel J. Does continuing education improve physical therapists’ effectiveness in treating neck pain? A randomized clinical trial. Phys Ther. 2009;89(1):38–47. pmid:18988663
  31. 31. Rushton A, Lindsay G. Developing clinical expertise for healthcare professionals through masters courses. International Journal of Therapy & Rehabilitation. 2007;14(4):156–61.
  32. 32. Rushton A, Lindsay G. Defining the construct of masters level clinical practice in manipulative physiotherapy. Manual Therapy. 2010;15(1):93–9. pmid:19748815
  33. 33. Rushton A, Lindsay G. Defining the construct of Masters level clinical practice in healthcare based on the UK experience. Med Teach. 2008;30(4):e100–7. pmid:18569652
  34. 34. Rushton A, Lindsay G. Developing clinical expertise through clinical placement at masters level. International Journal of Therapy & Rehabilitation. 2007;14(6):252–8.
  35. 35. Petty NJ, Scholes J, Ellis L. The impact of a musculoskeletal masters course: Developing clinical expertise. Man Ther. 2011;16(6):590–5. pmid:21708476
  36. 36. Petty NJ, Scholes J, Ellis L. Master’s level study: learning transitions towards clinical expertise in physiotherapy. Physiotherapy. 2011;97(3):218–25. pmid:21820540
  37. 37. Rushton A, Moore A. International identification of research priorities for postgraduate theses in musculoskeletal physiotherapy using a modified Delphi technique. Manual Therapy. 2009;In Press, Corrected Proof.
  38. 38. Green A, Perry J, Harrison K. The influence of a postgraduate clinical master’s qualification in manual therapy on the careers of physiotherapists in the United Kingdom. Manual Therapy. 2008;13(2):139–47. pmid:17307015
  39. 39. Stathopoulos I, Harrison K. Study at Master’s Level by Practising Physiotherapists. Physiotherapy. 2003;89(3):158–69.
  40. 40. Williams AL, Phillips CJ, Watkins A, Rushton AB. The effect of work-based mentoring on patient outcome in musculoskeletal physiotherapy: study protocol for a randomised controlled trial. Trials. 2014;15:409. pmid:25344736
  41. 41. Chan A- W, Tetzlaff JM, Altman DG, Laupacis A, Gøtzsche PC, Krleža-Jerić K, et al. SPIRIT 2013 Statement: Defining Standard Protocol Items for Clinical Trials. Annals of Internal Medicine. 2013;158(3):200–7. pmid:23295957
  42. 42. Campbell MK, Piaggio G, Elbourne DR, Altman DG. Consort 2010 statement: extension to cluster randomised trials. BMJ: British Medical Journal. 2012;345.
  43. 43. Gruber JS, Reygadas F, Arnold BF, Ray I, Nelson K, Colford JM. A Stepped Wedge, Cluster-Randomized Trial of a Household UV-Disinfection and Safe Storage Drinking Water Intervention in Rural Baja California Sur, Mexico. The American Journal of Tropical Medicine and Hygiene. 2013;89(2):238–45. pmid:23732255
  44. 44. Hemming K, Ryan R, Gill P, Westerby P, Jolly K, Marshall T. Targeted case finding in the prevention of cardiovascular disease: a stepped wedge cluster randomised controlled trial. The British Journal of General Practice. 2016;66(651):e758–e67. pmid:27528707
  45. 45. DiGuiseppi C, Coupland C. The design and use of cluster randomised controlled trials in evaluating injury prevention interventions: part 1. Rationale, design and informed consent. Injury Prevention. 2010;16(1):61–7. pmid:20179039
  46. 46. Gielen AC, Wilson MEH, McDonald EM, Serwint JR, Andrews JS, Hwang W, et al. Randomized trial of enhanced anticipatory guidance for injury prevention. Archives of Pediatrics & Adolescent Medicine. 2001;155(1):42–9.
  47. 47. Edwards SJ, Braunholtz DA, Lilford RJ, Stevens AJ. Ethical issues in the design and conduct of cluster randomised controlled trials. BMJ. 1999;318(7195):1407–9. pmid:10334756
  48. 48. Hemming K, Haines TP, Chilton PJ, Girling AJ, Lilford RJ. The stepped wedge cluster randomised trial: rationale, design, analysis and reporting. BMJ. 2015;350.
  49. 49. Mdege N, Man M, Brown C, Torgersen D. Systematic review of stepped wedge cluster randomised trials shows that design is particularly used to evaluate interventions during routine implementation. J Clin Epidemiol. 2011;64.
  50. 50. Gracey CF, Haidet P, Branch WT, Weissmann P, Kern DE, Mitchell G, et al. Precepting humanism: strategies for fostering the human dimensions of care in ambulatory settings. Acad Med. 2005;80.
  51. 51. Brown C, Lilford R. The stepped wedge trial design: a systematic review. BMC Med Res Methodol. 2006;6.
  52. 52. Brown C, Hofer T, Johal A, Thomson R, Nicholl J, Franklin BD, et al. An epistemology of patient safety research: a framework for study design and interpretation. Part 2. Study design. Qual Saf Health Care. 2008;17.
  53. 53. Hussey M, Hughes J. Design and analysis of stepped wedge cluster randomised trials. Contemporary Clin Trials. 2007;28.
  54. 54. IFOMPT. Educational Standards In Orthopaedic Manipulative Therapy. Part A: Educational Standards 2016. IFOMPT; 2016.
  55. 55. Edwards I, Jones MA. Clinical Reasoning and Expert Practice. Expertise in Physical Therapy Practice (Second Edition). Saint Louis: W.B. Saunders; 2007. p. 192–213.
  56. 56. Christensen N, Jones MA, Higgs J, Edwards I. Dimensions of Clinical Reasoning Capability. In: Higgs J, Jones MA, Loftus S, Christensen N, editors. Clinical Reasoning in the Health Professions, Volume 1. Amsterdam: Butterworth-Heinemann; 2008.
  57. 57. Edwards I, Jones M, Carr J, Braunack-Mayer A, Jensen GM. Clinical reasoning strategies in physical therapy. Phys Ther. 2004;84(4):312–30; discussion 31–5. pmid:15049726
  58. 58. Edwards I, Jones M, Hillier S. The interpretation of experience and its relationship to body movement: a clinical reasoning perspective. Man Ther. 2006;11(1):2–10. pmid:16359908
  59. 59. Murray E, Todd C, Modell M. Can general internal medicine be taught in general practice? An evaluation of the University College London model. Medical Education. 1997;31(5):369–74. pmid:9488860
  60. 60. Cook DA, Thompson WG, Thomas KG, Thomas MR, Pankratz VS. Impact of Self-Assessment Questions and Learning Styles in Web-Based Learning: A Randomized, Controlled, Crossover Trial. Academic Medicine. 2006;81(3):231–8. pmid:16501263
  61. 61. Alderson D. Developing expertise in surgery. Medical Teacher. 2010;32(10):830–6. pmid:20854159
  62. 62. Benner P, Tanner C, Chesla C. Expertise in nursing practice. New York: Springer; 1996.
  63. 63. Eraut M. Learning contexts. Learning in Health & Social Care. 2006;5(1):1–8.
  64. 64. Jensen GM, Gwyer J, Hack LM, Shepard KF. Expert Practice in Physical Therapy. Expertise in Physical Therapy Practice (Second Edition). Saint Louis: W.B. Saunders; 2007. p. 145–73.
  65. 65. Hefford C, Abbott JH, Arnold R, Baxter GD. The Patient-Specific Functional Scale: Validity, Reliability, and Responsiveness in Patients With Upper Extremity Musculoskeletal Problems. Journal of Orthopaedic & Sports Physical Therapy. 2012;42(2):56–65.
  66. 66. Horn KK, Jennings S, Richardson G, Vliet DV, Hefford C, Abbott JH. The patient-specific functional scale: psychometrics, clinimetrics, and application as a clinical outcome measure. J Orthop Sports Phys Ther. (2011 Oct 25):2012 Jan;42(1):30–42.
  67. 67. Fowles JB, Terry P, Xi M, Hibbard J, Bloom CT, Harvey L. Measuring self-management of patients’ and employees’ health: Further validation of the Patient Activation Measure (PAM) based on its relation to employee characteristics. Patient Education and Counseling. 2009;77(1):116–22. pmid:19356881
  68. 68. Hibbard JH, Mahoney ER, Stockard J, Tusler M. Development and testing of a short form of the patient activation measure. Health Serv Res. 2005;40(6 Pt 1):1918–30. pmid:16336556
  69. 69. Hibbard JH, Stockard J, Mahoney ER, Tusler M. Development of the Patient Activation Measure (PAM): conceptualizing and measuring activation in patients and consumers. Health Serv Res. 2004;39(4 Pt 1):1005–26. pmid:15230939
  70. 70. Hefford C, Kemp L, Abbot JH, Arnold R, Baxter GD, Taylor W. The Patient Specific Functional Scale: responsiveness and validity in upper or lower limb musculoskeletal disorders … Otago Rehabilitation and Disability Research Theme Meeting, 4–5 December 2008, School of Physiotherapy, University of Otago, Dunedin, New Zealand. Physical Therapy Reviews. 2009;14(1):3–4.
  71. 71. Beattie PF, Nelson R, Murphy DR. Development and preliminary validation of the MedRisk instrument to measure patient satisfaction with chiropractic care. J Manipulative Physiol Ther. 2011;34(1):23–9. pmid:21237404
  72. 72. Beattie P, Turner C, Dowda M, Michener L, Nelson R. The MedRisk Instrument for Measuring Patient Satisfaction With Physical Therapy Care: a psychometric analysis. J Orthop Sports Phys Ther. 2005;35(1):24–32. pmid:15754601
  73. 73. Chatman AB, Hyams SP, Neel JM, Binkley JM, Stratford PW, Schomberg A, et al. The Patient-Specific Functional Scale: measurement properties in patients with knee dysfunction. Phys Ther. 1997;77(8):820–9. pmid:9256870
  74. 74. Herdman M, Gudex C, Lloyd A, Janssen MF, Kind P, Parkin D, et al. Development and preliminary testing of the new five-level version of EQ-5D (EQ-5D-5L). Quality of Life Research. 2011;20(10):1727–36. pmid:21479777
  75. 75. Beattie P, Turner C, Dowda M, Michener L, Nelson R. The MedRisk instrument for measuring patient satisfaction with physical therapy care: a psychometric analysis. J Orthop Sports Phys Ther. 2005;35.
  76. 76. Beattie PF, Pinto MB, Nelson MK, Nelson R. Patient satisfaction with outpatient physical therapy: instrument validation [corrected] [published erratum appears in PHYS THER 2002 Aug;82(8):827]. Phys Ther. 2002;82.
  77. 77. Hush JM, Cameron K, Mackey M. Patient Satisfaction With Musculoskeletal Physical Therapy Care: A Systematic Review. Physical Therapy. 2011;91(1):25–36. pmid:21071504
  78. 78. Green CA, Perrin NA, Polen MR, Leo MC, Hibbard JH, Tusler M. Development of the Patient Activation Measure for mental health. Adm Policy Ment Health. 2010;37(4):327–33. pmid:19728074
  79. 79. Janssen MF, Pickard AS, Golicki D, Gudex C, Niewada M, Scalone L, et al. Measurement properties of the EQ-5D-5L compared to the EQ-5D-3L across eight patient groups: a multi-country study. Quality of Life Research. 2013;22(7):1717–27. pmid:23184421
  80. 80. EuroQol—a new facility for the measurement of health-related quality of life. The EuroQol Group. Health Policy (Amsterdam, Netherlands). 1990;16(3):199–208.
  81. 81. Rushton A, Lindsay G. Developing clinical expertise for healthcare professionals through masters courses. Int J Ther Rehab. 2007;14.
  82. 82. Rushton A, Lindsay G. Defining the construct of masters level clinical practice in manipulative physiotherapy. Man Ther. 2010;15.
  83. 83. Agar M. Speaking of Ethnography. London: Sage; 1986.
  84. 84. Hemming K, Taljaard M, Forbes A. Analysis of cluster randomised stepped wedge trials with repeated cross-sectional samples. Trials. 2017;18(1):101. pmid:28259174
  85. 85. Rothman KJ, Greenland S, Lash TL. Modern epidemiology. Philadelphia: Wilkins; 2008.
  86. 86. Mehta N, Inturrisi CE, Horn SD, Witkin LR. Using Chronic Pain Outcomes Data to Improve Outcomes. Anesthesiology Clinics. 2016;34(2):395–408. pmid:27208717
  87. 87. Resnik L, Hart DL. Using Clinical Outcomes to Identify Expert Physical Therapists. Physical Therapy. 2003;83(11):990–1002. pmid:14577826
  88. 88. Linton SJ, Boersma K. Early Identification of Patients at Risk of Developing a Persistent Back Problem: The Predictive Validity of The Örebro Musculoskeletal Pain Questionnaire. The Clinical Journal of Pain. 2003;19(2):80–6. pmid:12616177
  89. 89. Westman A, Linton SJ, Öhrvik J, Wahlén P, Leppert J. Do psychosocial factors predict disability and health at a 3‐year follow‐up for patients with non‐acute musculoskeletal pain? A validation of the Örebro Musculoskeletal Pain Screening Questionnaire. European Journal of Pain. 2008;12(5):641–9. pmid:18086541
  90. 90. Hibbard JH, Greene J, Tusler M. Improving the outcomes of disease management by tailoring care to the patient’s level of activation. Am J Manag Care. 2009;15(6):353–60. pmid:19514801
  91. 91. Hibbard JH, Tusler M. Assessing activation stage and employing a "next steps" approach to supporting patient self-management. J Ambul Care Manage. 2007;30(1):2–8. pmid:17170632
  92. 92. Abbott JH, Schmitt JS. The Patient-Specific Functional Scale was valid for group-level change comparisons and between-group discrimination. Journal of Clinical Epidemiology. 2014b;67(6):681–8.
  93. 93. Pengel LHM, Refshauge KM, Maher CG. Responsiveness of Pain, Disability, and Physical Impairment Outcomes in Patients With Low Back Pain. Spine. 2004;29(8):879–83. pmid:15082988
  94. 94. Abbott JH, Schmitt J. Minimum Important Differences for the Patient-Specific Functional Scale, 4 Region-Specific Outcome Measures, and the Numeric Pain Rating Scale. Journal of Orthopaedic & Sports Physical Therapy. 2014a;44(8):560–4.
  95. 95. Resnik L, Dobrzykowski E. Guide to outcomes measurement for patients with low back pain syndromes. J Orthop Sports Phys Ther. 2003;33(6):307–16; discussion 17–8. pmid:12839205
  96. 96. Terwee CB, Bot SDM, de Boer MR, van der Windt DAWM, Knol DL, Dekker J, et al. Quality criteria were proposed for measurement properties of health status questionnaires. Journal of Clinical Epidemiology. 2007;60(1):34–42. pmid:17161752
  97. 97. Insignia Health. Coaching for Activation 2017 [https://www.insigniahealth.com/products/cfa.
  98. 98. Fairbairn K, May K, Yang Y, Balasundar S, Hefford C, Abbott JH. Does the Patient-Specific Functional Scale (PSFS) reflect the International Classification of Functioning, Disability and Health (ICF)? New Zealand Journal of Physiotherapy. 2010;38(2):69-.
  99. 99. Fairbairn K, May K, Yang Y, Balasundar S, Hefford C, Abbott JH. Mapping Patient-Specific Functional Scale (PSFS) Items to the International Classification of Functioning, Disability and Health (ICF). Physical Therapy. 2012;92(2):310–7. pmid:22074939
  100. 100. Jette AM, Norweg A, Haley SM. Achieving meaningful measurements of ICF concepts. Disabil Rehabil. 2008;30(12–13):963–9. pmid:18484391
  101. 101. Eraut M. Feedback. Learning in Health & Social Care. 2006;5(3):111–8.
  102. 102. Eraut M. Sharing practice: problems and possibilities. Learning in Health & Social Care. 2004;3(4):171–8.
  103. 103. Dall’Alba G, Sandberg J. Unveiling Professional Development: A Critical Review of Stage Models. Review of Educational Research. 2006;76(3):383–412.
  104. 104. Petty NJ, Morley M. Clinical expertise: Learning together through observed practice. Man Ther. 2009;14(5):461–2. pmid:19564128
  105. 105. Johns C. Guided Reflection: A Narrative Approach to Advancing Professional Practice [Book]. John Wiley & Sons, Ltd. (UK); 2010 [318p.]. http://search.ebscohost.com/login.aspx?direct=true&db=nlebk&AN=330502&site=ehost-live.
  106. 106. Kanneganti C. Training key to healthcare delivery BMA.org: BMA; 2015 [https://www.bma.org.uk/connecting-doctors/policy_debate/b/weblog/posts/training-key-to-healthcare-delivery.
  107. 107. Merrifield N. Nurse training at risk from CPD budget cuts: Nursing Times; 2016 [https://www.nursingtimes.net/news/education/exclusive-nurse-training-at-risk-from-cpd-budget-cuts/7005710.article.
  108. 108. Hemming K, Haines TP, Chilton PJ, Girling AJ, Lilford RJ. The stepped wedge cluster randomised trial: rationale, design, analysis, and reporting. BMJ: British Medical Journal. 2015;350.