Abstract
Background Doctor–patient communication is an essential component of general practice. Improvement of GPs' communication patterns is an important target of training programmes. Available studies have so far failed to provide conclusive evidence of the effectiveness of educational interventions to improve doctor–patient communication.
Aim To examine the effectiveness of a learner-centred approach that focuses on actual needs, to improve GPs' communication with patients.
Design of study Randomised controlled trial.
Setting One hundred volunteer GPs in the Netherlands.
Method The intervention identified individual GPs' deficiencies in communication skills by observing authentic consultations in their own surgery. This performance assessment was followed by structured activities in small group meetings, aimed at remedying the identified shortcomings. Outcomes were measured using videotaped consultations in the GPs' own surgery before and after the intervention. Communication skills were rated using the MAAS-Global, a validated checklist.
Results The scores in the intervention group demonstrated a significant improvement compared with those of the control group (95% confidence interval = 0.04 to 0.75). The effect size was moderate to large (d-value = 0.66). The level of participation significantly contributed to the effectiveness. Largest improvement was found on patient-centred communication skills.
Conclusion The approach of structured individual improvement activities based on performance assessment is more effective in improving communication skills than current educational activities.
INTRODUCTION
Continuing Medical Education (CME) has changed considerably in the last decade as a result of theories and evidence of the effectiveness of educational activities.1–4 In the current view, education of practicing physicians is learner-centred, focuses on day-to-day clinical practice and ideally integrates several approaches.5,6 This approach was adopted rapidly in many countries, even though there is conflicting evidence on its effectiveness.6–8
We were particularly interested in the effectiveness of this approach to enhance doctors' communication skills. Doctor–patient communication is a core component of health care and has a strong impact on patient satisfaction and adherence to advice and treatment.9 A patient-centred approach in particular appears to favour these outcomes positively.10 Although both doctors and patients put a high value on good communication, performance in day-to-day practice was found wanting.11 Well-designed studies of the effectiveness of educational interventions to improve communication skills are scarce and demonstrated little benefit.12–15 We hypothesised that a learner-centred approach, comprising individual performance assessment followed by small group meetings, tailored to doctors' individual needs may be more successful in improving doctor–patient communication than traditional CME approaches. A randomised controlled trial was conducted to investigate the effects of such a programme for GPs. We also examined whether the level of participation in the intervention had an effect on the outcome, for example, whether the more active participants achieved better results. The feasibility of the intervention has been described previously.16
METHOD
Subjects
All GPs (n = 1066) in three regions in the south of the Netherlands (Zuidoost Brabant, Noord-Brabant Noordoost, Limburg) received a letter informing them about the study and asking whether they would be interested in participating (November 2001). Subsequently, they were invited to enrol (May 2002). We aimed at a maximum of 100 and a minimum of 80 participants. Power calculations using historical data showed these were sufficient to detect a 10% difference in the MAAS-global score with 80% power and a 5% risk of a type 1 error.17 The first 100 subscribing GPs enrolled in the study and a questionnaire was used to obtain participants' personal and professional characteristics (for example experience with modern educational methods).
Randomisation
Randomisation was performed after sampling professional characteristics and before baseline measurement. GPs participated individually, but randomisation was done at practice level. Practices were allocated as a whole to either the intervention or the control arm of the study. This was done to avoid the possibility of closely collaborating colleagues, randomised to different arms of the study, benefiting from mutual educational experiences. Practices were stratified by region and experience in working with modern educational methods; these were identified in an expert panel prior to the study as possible effect modifiers. Subsequently balanced randomisation of practices was done. Within strata, practices were allocated to either the intervention or the control condition by the principal researcher and the projects research assistant using the ‘allocate random data’ option in SPSS.
Educational interventions
Intervention group. The intervention consisted of assessment, selection of global topics for improvement, feedback and revalidation activities. Assessment was the first step to identify aspects of GPs' performances where improvement was likely to be beneficial.18,19 Three topics were assessed. One of these topics, for example, doctor–patient communication, is the focus of this paper. For each participant two topics for further improvement were selected. This was done by relating assessment results to standards for desired care.20
How this fits in
Doctor–patient communication is an essential component of general practice, and improvement of GPs' communication skills is an important target of training programmes for GPs. Available studies have so far failed to provide conclusive evidence of the effectiveness of educational interventions to improve doctor–patient communication. This study demonstrates that a learner-centred approach focusing on individual GPs' deficiencies in communication skills is more effective than the traditional approach.
GPs with scores above the standard focused on other topics than communication, but did not drop out of the study. The performance of GPs was assessed by video observation in daily practice.17 Participants received individual feedback on their assessment results and the selected topics from a GP. These were regional GPs who were invited to participate only for this specific task in the project, and who were trained by the authors. Feedback was given on each participant's mean communication score in relation to the criterion-referenced standard and mean scores of 10–15 colleagues. Also, detailed feedback was on the scores on items and the standards for these items, thus enabling GPs to identify aspects requiring improvement. Within 2 weeks after feedback, a series of three to four small group meetings started. The meetings lasted 2 hours and the groups consisted of four to six participating GPs and a tutor. The tutors were practising regional GPs, without specific expertise in doctor–patient communication, and were prepared with an 8-hour training course in supporting their colleagues' learning process. The training consisted of an introduction in the background and rationale of the intervention and of educational exercises that focused especially on the first two meetings. Participants also received a manual with information about quality improvement, examples, checklists and diagrams to fill out. The manual contained no information about doctor–patient communication. The content of the meetings was the participants' own responsibility. GPs started by defining personal goals for improvement, then analysed barriers to change, developed practice development plans including time schedules, and finally evaluated their results.14,21,22 Meetings were held within a fixed time frame of 7 months; participants were free to choose the time and place of the meetings. GPs received 2 hours of credits for performance assessment and feedback and 2 hours for each attended meeting.
Control group
The control group was offered existing written CME material about doctor–patient communication developed by the Dutch College of General Practitioners. It consists of paper cases, questionnaires for self-assessment and information about appropriate performance. No feedback of assessments or small group meetings were involved in this study arm. GPs received 15 hours of credits when they returned the self-assessment questionnaires.
Variables and instruments
Doctor–patient communication
Communication skills were evaluated using videotaped consultations in GPs' own surgeries; the psychometric properties of this procedure have been described earlier.17 Observation cameras were installed in the consultation and examination room. All consultations were videotaped provided the patient consented. GP observers, who received a 6-hour training, selected eight consultations for each observation period using validated selection criteria.17 Observers rated communication skills using a validated instrument (MAAS-Global).23 This instrument includes 12 items on doctor–patient communication in initial consultations, rated on a scale ranging from 0 to 6: items are broadly defined but anchored in detailed rating criteria. A short description of the items is given in Box 1.
Box 1. MAAS-Global rating list for consultation skills of doctors addressing doctor–patient communication. Each item is scored on a scale ranging from 0 to 6. For item 2 and 4 the rating ‘not applicable’ is an additional option.
Communication skills for each separate phase
Introduction
Giving the patient room to tell his story
General orientation on the reason for visit
Asking about other reasons for visit
Follow-up consultation
Naming previous complaints, requests for help and management plan
Asking about adherence to management plan
Asking about the course of the complaint
Request for help
Naming requests for help, wishes or expectations
Naming reasons that prompted the patient to come now
Completing exploring request for help
Physical examination
Instructions to the patient
Explanation of what is being done
Treating the patient with care and respect
Diagnosis
Naming findings and diagnosis/hypothesis
Naming causes or the relation between findings and diagnosis
Naming prognosis or expected course
Asking for patient's response
Management
Shared decision-making, discussing alternatives, risks and benefits
Discussing feasibility and adherence
Determining who will do what and when
Asking for patient's response
Evaluation of consultation
General communication skills
Exploration
Exploring requests for help, wishes or expectations
Exploring patient's response to information given within patient's frame of reference
Responding to non-verbal behaviour and cues
Emotions
Information giving
Summarisations
Structuring
Logical sequence of phases
Balanced division of time
Announcing (history taking, examination other phases)
Empathy
Concerned, inviting and sincerely empathetic
In intonation, gesture and eye contact expressing empathy in brief verbal responses
GPs were videotaped before and after the intervention period (May 2002 to May 2003). The intervention group received their scores on the first observation as feedback. The control group received feedback after the study was completed. Because of a possible decay effect, the second observation was videotaped 2–6 months after the intervention. For the second observation the observers were blinded regarding the group that participants belonged to, that is, intervention or control group. For the first observation, blinding was not feasible, because limited time was scheduled between assessment and feedback in the intervention group.
Educational activities
The educational activities of participants were monitored using checklists and questionnaires. GPs providing feedback filled out a checklist to ensure that all relevant aspects were discussed. Tutors registered participants' attendance at group meetings. Potential bias by CME activities on doctor–patient communication besides those included in the study was identified by questionnaire.
Statistical analysis
Analyses were performed for item scores and total score. For each participant the pre and post MAAS-Global scores were calculated as the mean of all items over all observations. The item scores were calculated as the means of items over all consultations. The effect of the intervention was analysed by linear regression, using the post score as dependent variable and the pre score and treatment (intervention or control condition) as independent variables. Possible outliers affecting the analyses were identified by means of scatter plots and excluded from analyses. To identify the contribution of the educational intervention in the post score, we calculated the 95% confidence interval (CI) of the treatment effect. Analyses were done as ‘intention-to-treat’ analyses if data were available: all participants, whether they participated in educational activities on doctor–patient communication or not, were included in the analysis.
The effect size (d-value, a commonly used term to compare the effects of educational interventions) was defined on the basis of the post versus pre scores (post-pre-score) for the participants in the intervention and control group.24 For the two groups the average and standard deviation of the post-pre-score were calculated. The effect size was obtained by subtracting the average scores for the two groups, and dividing it by the pooled standard deviation.24
To check whether participation in the small group meetings influenced the effectiveness, a second regression analysis was performed. As participants with high scores were allocated to other topics in the study, this analysis was performed only for the participants in intervention group that were allocated to the small group meetings on communication. In this analysis, the post score was used as dependent variable, and the pre score and the percentage of meetings attended were used as independent variables.
RESULTS
The enrolment questionnaire was returned by 670 GPs (63%); 174 (26%) indicated that they were interested to participate in the study and 100 GPs, representing 87 practices, enrolled. Personal and professional characteristics were obtained; 49 participants were randomised to the intervention arm and 51 to the control arm. Flow of participants in the study and reasons for drop out are given in the Figure 1.
Figure 1 Flow chart of participants in the study and reasons for dropout.
Of the 44 GPs in the intervention arm, 22 were assigned to a programme of four small group meetings about communication skills and 15 to three meetings. The remaining seven GPs did not focus on doctor–patient communication but on other topics. They did not drop out of the study. All participants received feedback according to schedule, except one (reason = no time). Of the 37 GPs, 22 attended all meetings and four GPs did not attend any meeting. The mean number of meetings attended was 2.75. Reasons for non-attendance were: working in solo practice and being unable to attend meetings during office hours (n = 3); and ‘not believing that communication training can be effective’(n = 1). Of the GPs in the intervention group, 14 reported CME activities (mean = 4.6 hours; standard deviation [SD] = 2.7) on communication not related to the study.
In the control arm, 16 out of 43 GPs ordered the CME materials and 16 GPs in the control group reported 6.6 hours (SD = 5.4) of CME on doctor–patient communication not related to the study; in total 24 GPs in the control arm of the study were involved in educational activities on doctor–patient communication.
The second observation was completed successfully for 38 GPs in the intervention group and 38 GPs in the control group, and used for analysis. Reasons for not participating in the second observation are given in the flow chart. GPs who did and those who did not complete the study showed no difference in scores on the first series of observed consultations (mean score = 2.35 ± 0.66 versus 2.34 ± 0.69; not significant).
Over 10 000 consultations were recorded, out of which two samples of eight consultations per participant were selected. The selection criteria met in the samples were comparable in both observations and for both groups. The mean total scores and scores on the items of the MAAS-Global are presented in Table 1. No outliers in scores were identified. Regression analysis showed a significant effect of both the treatment and the pre scores. The explained variance (R2) was 0.10. Participants reported eight items as personal improvement goals, and improvement was seen on five of these (Table 1). The corresponding effect size (d-value) was 0.66, indicating a moderate to large effect.24
Table 1 Item and total scores of GPs in intervention (n = 38) and control group (n = 38), and the differences in pre and post scores between intervention and control group, the corresponding P-value and the 95% CI of the differences in improvement between both groups.
The effect of participation was analysed by a regression analysis for the participants in the intervention arm allocated to the small group meetings, adding the percentage of meetings attended (attendance) as an independent variable. In this analysis (R2 = 0.39) the effect of the pre scores (B = 0.73; 95% CI = 0.26 to 1.20) and the effect of the attendance (B = 1.30; 95% CI = 0.57 to 2.04) were found to be significant.
DISCUSSION
Summary of main findings
The educational approach involving assessment of communication in daily practice and personalised learning activities guided by structured small group work proved to be an effective way of improving doctor–patient communication. The size of the effect was moderate to large. A dose-response effect was seen as the attendance to the intervention significantly contributed to the improvement of the scores. The participants showed most improvement on patient-centred items. This study showed assessment of GPs performance in the real process of care delivery to be an effective first step of an educational programme. GPs considered the assessment outcomes as an acceptable basis for educational activities, and selected aspects of their performance with low scores for improvement. GPs were also found to be capable of drawing up effective personal development plans with support only for the process of change, given the improvement that was seen on the reported improvement goals. The participating GPs had to get used to the approach investigated in this study, as the majority reported having no experience with the methods used in the intervention. We expect that repeated use of the programme may yet enhance its effectiveness.
Strengths and limitations of the study
Participants in the intervention group undertook more study-related activities to improve doctor–patient communication than their colleagues in the control group. This may be interpreted as bias, but in our view it should be regarded as an effect of the intervention, that is, the assessments to drive learning activities. Educational activities not related to the study may have influenced the results, but, as GPs in the control arm attended more of those activities than those in the intervention arm, this only emphasises the effectiveness of the intervention. Selection of randomised groups by attrition may cause bias: although we cannot rule out this effect, no differences in available pre scores were found between those who did and did not drop out of the study.
An unexpected finding is the deterioration of the scores of the control group. As the baseline scores of the control group are very near the post-intervention scores of the intervention group, it may appear that the effects must be attributed to chance, instead of the intervention. Although we have no satisfying explanation for the scores in the control group, we have arguments to attribute the effect to the intervention itself. First, the effects in the intervention group are seen in the participants that actively participated in the small groups. Second, the effects are seen in those items that were reported by participant as goals for improvement. And third, the regression analysis used is the preferred method in case of baseline differences.25
The effects of the intervention may be temporary and fade over time. Therefore, we planned the second measurement 2–6 months after the intervention. Whether improvements are permanently embedded in GPs' communication behaviour is still not clear.
Comparison with existing literature
This study provides evidence that supports the current view on continuing professional development, which stresses the importance of learning activities being learner-centred, focused on day-to-day clinical practice and integrating several approaches.5–8 Convincing evidence on educational interventions that improve the performance on doctor–patient communication towards desired standards is lacking.12–15 Therefore, the study also provides an approach for learning and improving in practice for those interested in the specific domain of doctor–patient communication.
Implications for future research
The findings of this study cannot easily be generalised to other topics or to the general population of GPs. The participants in our study were not representative as they were motivated to experience this new approach. Effectiveness may be different when GPs are not really interested in this educational method. Also, the assessment method of video observation in daily practice is ideal for this purpose, as it is personalised and performancebased, and yields videotapes that can be used in the small group meetings. Further research will have to shed light on the effectiveness of this approach for other topics, assessment formats and participating doctors. This approach may be implemented as an evidence-based educational approach to support continuing professional in the domain of doctor–patient communication.
Notes
Funding body
The study was funded by ZonMw, the Netherlands Organisation for Health Research and Development (project number 1615.0005)
Ethics committee
Approved by the Medical Ethical Committee of the Maastricht Academic Medical Centre (00-172)
Competing interests
The authors have stated that there are none.
- Received July 13, 2004.
- Revision received January 26, 2005.
- Accepted January 31, 2006.
- © British Journal of General Practice, 2006.