|Home | About | Journals | Submit | Contact Us | Français|
This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Many health professionals lack the skills to find and appraise published research. This lack of skills and associated knowledge needs to be addressed, and practice habits need to change, for evidence-based practice to occur. The aim of this before and after study was to evaluate the effect of a multifaceted intervention on the knowledge, skills, attitudes and behaviour of allied health professionals.
114 self-selected occupational therapists were recruited. The intervention included a 2-day workshop combined with outreach support for eight months. Support involved email and telephone contact and a workplace visit. Measures were collected at baseline, post-workshop, and eight months later. The primary outcome was knowledge, measured using the Adapted Fresno Test of Evidence-Based Practice (total score 0 to 156). Secondary outcomes were attitude to evidence-based practice (% reporting improved skills and confidence; % reporting barriers), and behaviour measured using an activity diary (% engaging/not engaging in search and appraisal activities), and assignment completion.
Post-workshop, there were significant gains in knowledge which were maintained at follow-up. The mean difference in the Adapted Fresno Test total score was 20.6 points (95% CI, 15.6 to 25.5). The change from post-workshop to follow-up was small and non-significant (mean difference 1.2 points, 95% CI, -6.0 to 8.5). Fewer participants reported lack of searching and appraisal skills as barriers to evidence-based practice over time (searching = 61%, 53%, 24%; appraisal 60%, 65%, 41%). These differences were statistically significant (p = 0.0001 and 0.010 respectively). Behaviour changed little. Pre-workshop, 6% engaged in critical appraisal increasing to 18% post-workshop and 18% at follow-up. Nearly two thirds (60%) were not reading any research literature at follow-up. Twenty-three participants (20.2%) completed their assignment.
Evidence-based practice skills and knowledge improved markedly with a targetted education intervention and outreach support. However, changes in behaviour were small, based on the frequency of searching and appraisal activities. Allied health educators should focus more on post-workshop skill development, particularly appraisal, and help learners to establish new routines and priorities around evidence-based practice. Learners also need to know that behaviour change of this nature may take months, even years.
Regardless of the desire to keep up to date, behaviour change is difficult. Even the most motivated of health professionals face barriers when attempting to stay up to date by finding, reading and using research. The primary barriers reported by allied health professionals for not using research in practice are a perceived lack of time, and lack of skills and knowledge when searching for and appraising research [1-5]. Yet little has been done in allied health to address these barriers.
Continuing professional education, particularly workshops can improve knowledge but the impact on behaviour is less impressive. In a recent randomised controlled trial, Taylor and colleagues found that after six months, a half-day workshop on critical appraisal had improved knowledge but not the evidence-seeking behaviour of health professionals, including physiotherapists . These findings are consistent with systematic reviews on the effectiveness of teaching evidence-based medicine and critical appraisal skills [7-9]. Therefore, a major challenge for educators is effecting knowledge transfer and behaviour change.
What appears to help promote behaviour change is partly the way in which education is delivered, and partly the provision of follow-up . In their Cochrane review, Thompson O'Brien and colleagues  recommended that researchers investigate components of workshops that contribute to effectiveness, such as practicing skills during and after a workshop, providing follow-up outreach support, and providing feedback on behaviour change. These collective strategies are often referred to as a 'multifaceted' intervention, where 'champions' of evidence-based practice market concepts to the profession, links are maintained with learners after training, and reminders and feedback are used to encourage behaviour change [10-14].
The effectiveness of a multifaceted intervention, aimed at improving evidence-based practice behaviours has received limited attention by allied health researchers. In addition to Taylor and colleagues , Stevenson, Lewis and Hay completed another randomised controlled trial involving education and training of British physiotherapists . While their one-day workshop on searching and appraisal skills was interactive, the overall intervention was not multifaceted, and no follow-up support was provided. Furthermore, the authors focussed on measuring (and changing) attitudes using self-report measures. No objective tests of skill or knowledge, nor prospective measures of behaviour were used. Another randomised controlled trial  has reported the use of a multi-faceted intervention program, designed to increase the use of evidence-based practice. However, this study was conducted with public health physicians, not allied health professionals, and was published in 2003 after the current study had finished. Therefore, this important study by Forsetlund and colleagues will be discussed later.
In summary, the current study adopted a multi-faceted approach involving an interactive 2-day workshop for occupational therapists, written materials, use of opinion leaders and champions, a discussion list and information service, and outreach follow-up support. At the time of planning this study (2001), skills-based workshops were commonly used to introduce allied health professionals to evidence-based practice, but no efficacy studies had been reported.
The aim of this before and after study was to measure the effect of a multifaceted intervention on evidence-based practice, on the knowledge and skills, behaviour and attitudes of occupational therapists. We hypothesised that post-intervention, participants would demonstrate improved skills and knowledge, report fewer barriers to adopting evidence-based practice, particularly lack of skills and knowledge, use their skills more frequently at work, and that these changes would be maintained over time.
The study commenced in November 2001 and concluded in March 2003, with approval from the University of Western Sydney ethics committee. A before and after study design was used. There was no control group. The intervention is described in detail below.
To be included, participants had to be a qualified, employed occupational therapist, working in the state of New South Wales, Australia during the period of the study. An advertisement was distributed by email and post, to a wide range of private and public employers. Therapists were invited to participate, and to encourage junior and senior colleagues, as well as friends to enroll. This method of 'snowball' sampling aimed to recruit at least 100 occupational therapists. No eligibility restrictions were placed on health sector, work location or clinical specialty. A total of 232 therapists expressed interest, with 114 being recruited.
The primary intervention was a 2-day workshop on evidence-based practice, held in February 2002. Of the 114 recruited, 106 attended the workshop. To accommodate the large number of participants, three 2-day workshops were conducted over a month. Each workshop provided the same content but on different weekends over a month. The authors conducted the workshops in a metropolitan city, with the assistance of a health librarian. The authors were considered 'expert' clinicians, each having over 15 years of occupational therapy experience in public and private health sectors and had attended short skills-based courses on evidence-based practice. This experience and knowledge helped us to choose important occupational therapy questions and studies for appraisal during the workshop.
The workshop included lectures, practical sessions and small group discussion focussed around six topics: the process of evidence-based practice; writing focussed clinical questions; searching electronic databases; critical appraisal of qualitative and quantitative research; interpreting statistics in randomised controlled trials; and overcoming barriers/making the change to evidence-based practice.
The workshop used principles of andragogic or adult learning theory,  and social cognitive theory  to help participants engage with the new 'innovation', evidence-based practice. Social cognitive theory aims to promote learning and behaviour change by increasing the self-efficacy of learners (see Bradley and colleagues for an excellent review) . The format and content of sessions, and clear learning objectives were developed with a steering committee comprising five occupational therapy clinicians, two allied health experts in evidence-based practice, a health librarian and two service users. During the workshop, new skills and knowledge such as writing clinical questions, searching databases, and interpreting statistics were modelled by presenters, using worked examples. Time was set aside after each session for individuals to reflect, consider how they would apply the new skills and knowledge, and write personal learning goals to be achieved post-workshop. During the final workshop session, the change process and potential barriers to adopting evidence-based practice were actively discussed. A short presentation describing Rogers five-stage model of innovation diffusion  and Prochaska and DiClemente's transtheoretical model of change [21,22] provided the stimulus for discussion.
To promote post-workshop skill development participants were invited to develop a learning contract which included a critically appraised topic (CAT). These CATs could be completed individually or in pairs as an optional 'assignment'. Participants wrote a clinical question about the effectiveness of an occupational therapy intervention, to focus their CAT. By writing a CAT, it was hoped that participants would develop improved search and appraisal skills, and document the practice implications of research appraised. Although the use of CATs as a learning assignment had not previously been reported in this context, hypothetical assignments have been used .
The follow-up outreach support consisted of regular email and telephone contact, and an optional workplace visit. Support was provided by an expert occupational therapy practitioner (ML), employed as a research assistant/project manager. The visits (n = 82) provided help with searching and appraisal, and monitored progress with the assignment. An email list was set up to facilitate communication. Information distributed via this list included resources and websites and answers to frequently asked questions. Reminders and individual feedback were provided about the assignment. Between March 2002 and February 2003, approximately 180 email messages were sent to this list, most by the project manager (ML). Approximately 225 email messages, mostly questions, were received from participants. The number of phone calls made and received was not logged. Completed CATs were presented at a one-day conference in February 2003, and uploaded to a new website.
A written questionnaire and the Adapted Fresno Test of competence in evidence-based medicine  were used to measure knowledge, the primary outcome. The questionnaire also captured data on attitudes to evidence-based practice (secondary outcome). A written activity diary and assignment completion were used to measure behaviour (secondary outcome). Copies of the measures are available upon request from the first author (AM).
All measures except the activity diary were collected before, immediately after (i.e. at the end of day two) and 8-months post-workshop. The activity diary was collected prospectively on five occasions: for 3-weeks pre-workshop, then for 8-months during 2002 with bi-monthly collection (March-April; May-June; July-August; September-October).
The 8-page questionnaire was developed from existing instruments [3,23-25]. Pilot-testing was conducted with eight occupational therapists from other states and territories in Australia. Minor changes were made to the content, layout and formatting in response to feedback. Participants completed the questionnaire on three occasions. The first questionnaire was mailed out, and returned on the first morning of the workshop (baseline). The second questionnaire was completed at the end of the workshop (post-workshop). The third and final questionnaire was distributed and returned by mail, approximately eight months after the workshop (follow-up).
The questionnaire contained three sections, and took approximately 20 minutes to complete. Section 1 recorded demographic data, perceived barriers to adopting evidence-based practice, and strategies used to overcome barriers. Section 2 required participants to rank (from one to five) how frequently they relied on different sources of information when making treatment decisions. Section 3 evaluated attitudes, knowledge and skill with regard to evidence-based practice, by asking each participant whether they 'agreed', 'disagreed' or were 'unsure' about specific statements. Examples of statements used were: 'An electronic database such as PubMed can only be accessed from hospital and university libraries' and 'The p value is a measure of reliability'. These questions objectively tested participants' knowledge, and had correct/incorrect answers. Other questions asked about self-reported skills, ability and knowledge. For example, 'I am aware of and have used a range of electronic databases' and 'I feel confident that I can critically appraise research evidence'.
An adapted version of the Fresno test of competence in evidence-based medicine  was used to objectively measures skills and knowledge. The original Fresno test was designed to evaluate the effectiveness of a university curriculum on evidence-based medicine, and includes 12 short-answer questions, focussed around clinical scenarios relevant to family practice residents. Respondents are asked to write a focussed clinical question based on a scenario, list sources of information that could answer the question (for example, books and electronic databases), then comment on study designs and statistics reported in published papers. Internal consistency for the original Fresno test items, using Cronbach's alpha is 0.88, indicating a satisfactory level of agreement. Inter rater reliability is good to excellent, with correlations ranging from 0.72 to 0.96 for individual test questions, and 0.97 for total test scores .
For the current study, the original Fresno Test was adapted by AM to include clinical scenarios relevant to occupational therapists. Five of the 12 more advanced statistical questions were removed (for example, those about sensitivity, specificity, numbers needed to treat), since these would not be taught in the workshop curriculum. The new test, referred to hereafter as the 'Adapted Fresno Test', asked participants to choose one of two new scenarios and answer seven related questions (see Table Table1).1). The Adapted Fresno Test took about 20 minutes to complete.
Three sets of different clinical scenarios were written for each test administration (i.e baseline, post-workshop and follow-up), to avoid a practice effect. Diagnoses included in the clinical scenarios were low back pain, traumatic brain injury, occupational overuse syndrome, depression, osteoarthritis and carpal tunnel syndrome. Interventions included transcutaneous electrical nerve stimulation, group education programs, workplace exercises, cognitive behaviour therapy, exercise programs and splinting (see example in Table Table1).1). Analysis of internal consistency for the three versions of the Adapted Fresno Test yielded a Cronbach's alpha score ranging from 0.72 to 0.84, indicating an acceptable level of consistency for the adapted instrument .
The seven questions in the Adapted Fresno Test were scored using standardised grading criteria, similar to those reported by Ramos and colleagues . The minimum test score was zero, and the maximum 156 for the seven questions (marking criteria available upon request). Each completed test took about 20 minutes to score. Scoring of the Adapted Fresno Test was evaluated for interrater reliability, and involved two raters independently scoring 20 completed tests (10 from pre-workshop and 10 from post-workshop) after receiving a two-hour training session . Interrater reliability results ranged from poor to excellent depending on the test question being scored. Intraclass correlation coefficients (ICCs) ranged from 0.20 to 0.96 for the pre-workshop test (0.88 for the total score), and 0.41 to 0.92 for the post-workshop test (0.87 for the total score). In the pre-workshop survey, Questions 1, 4 and 5 had an ICC below 0.80 (0.20, 0.23 and 0.53 respectively). In the post-workshop survey, Questions 1 and 3 had an ICC below 0.70 (0.41 and 0.57 respectively). Further refinement of the Adapted Fresno Test scoring system is therefore indicated; implications for study results will be addressed later under 'limitations'.
Behaviour change was measured using a concurrent activity diary, designed by AM for use in the study, and provided in paper or electronic format. Participants were asked to record only those activities that related to evidence-based practice, such as searching, reading research-related articles, critical appraisal and teaching others about evidence-based practice.
The following information was recorded in columns in participants' diaries, then subsequently analysed: date and nature of activity; what prompted the activity; start and finish times; whether the activity was conducted alone or not; if and how practice changed as a result of engaging in the activity. To improve accuracy, participants were asked to complete the activity diary contemporaneously for three weeks before and eight months after the workshop. Diaries were returned by fax, email or post on a bi-monthly basis. A research assistant contacted participants if their bi-monthly activity diary had not been returned.
Completion of the assignment or CAT at the end of 2002 reflected engagement in the first three steps of the process of evidence-based practice (writing a focussed question, searching for evidence, and critically appraising the evidence). This outcome was recorded as completed/not completed.
Descriptive statistics, including means and percentages, were used to compare outcomes.
For the Adapted Fresno Test (the primary outcome measure), and based on statistical advice, paired t-tests were used to evaluate change in objective knowledge. Differences in mean total scores and confidence intervals (95% CI's) were calculated. A power calculation could not conducted in advance because the Adapted Fresno Test had not previously been used as an outcome measure, and expected means and standard deviations were unknown. Rather than performing a post-hoc power analysis, we examined the width of the confidence intervals (95%) for the estimated effects (pre-workshop to post workshop difference). When confidence intervals are reported in this way, a post-hoc power analysis is redundant . Improvements of 10% (15.6 points) in the mean total score post-workshop, and 15% (23.4 points) at follow-up were considered clinically/educationally important, compared to baseline.
Non-parametric statistics were used for all other measures. Repeated measures were computed for the three occasions, rather than comparing paired samples, to reduce the risk of type 1 errors. Friedman's tests were used to test the hypotheses that over time: (a) more participants would be able to correctly answer the knowledge test questions, (b) more participants would feel confident and able to use published research in their work, (c) fewer participants would report barriers to evidence-based practice, and (d) more participants would conduct searches, read and appraise research than had done so before.
For knowledge test questions, the responses 'agree/disagree/unsure' were recoded as 'correct/incorrect'. For questions related to self-reported skills, confidence and ability, the responses 'agree/disagree/unsure' were recoded as 'yes/no'.
Non-responders were not included in analyses, post-intervention or at follow-up.
Demographic data for the 114 participants are presented in Table Table2.2. The majority held an undergraduate degree in occupational therapy, worked full-time as a clinician in a metropolitan city, and had been graduated for 10 years or more. One third held a postgraduate qualification. One third worked in a regional or rural area. Just over 50% of the sample worked in the private sector.
Immediately post-workshop, 106 questionnaires and Adapted Fresno Tests were returned for analysis (92.9%). At follow-up, 51 (44.7%) of the possible 114 questionnaires and Adapted Fresno Tests were returned. For the activity diaries, 79 (69.3%) were returned for analysis in the first 8-week time-period post-workshop. At follow-up, only 40 of the possible 114 diaries (35.1%) were returned.
At baseline, the mean total score for the Adapted Fresno Test was 57/156 (range 0 to 126). Only 19% of participants achieved the 50% 'pass mark' of 78/156. Post-workshop, the mean total test score was 78/156 (range 32 to 124), with 77% achieving the pass mark. At follow-up, the mean total score was 82/156 (range 36 to 155), with 61% achieving the pass mark (see Table Table33).
There were statistically significant and educationally important differences in knowledge when pre-workshop and post-workshop total scores were compared (mean difference 20.6 points, 95% CI, 15.6 to 25.5), and when pre-workshop and follow-up total scores were compared (mean difference 23.1 points, 95% CI, 14.7 to 31.6, see Table Table4).4). Although differences between post-workshop and follow-up were small and non-significant, the knowledge gains were maintained after eight months.
There was no important difference in the baseline scores of recent graduates compared to more experienced therapists. At baseline, the mean total score on the Adapted Fresno Test for recent graduates (previous five years) was 59.8, compared to 56.5 for all others.
The three questions testing critical appraisal skills and knowledge were analysed separately for the Adapted Fresno Test (Questions 5, 6 and 7 = sub-total of 72). There were statistically significant differences in appraisal knowledge when pre-workshop and post-workshop sub-totals were compared (mean difference 6.4 points, 95% CI, 3.0 to 9.8), and when pre-workshop and follow-up sub-totals were compared (mean difference 9.9 points, 95% CI, 4.9 to 15.0, see Table Table4).4). Differences between post-workshop and follow-up were small and non-significant. The mean sub-total for these three questions did not, however, reach the 50% pass mark of 36/72 even at follow-up.
As hypothesised there were statistically significant differences in the proportion of therapists over time who could correctly answer questions about a good clinical question, PubMed and Medline, Cochrane reviews, confidence intervals, p-values and single case design research (see Table Table5).5). Knowledge gains were maintained at follow-up.
Based on the self-report questionnaire, there was an immediate increase in the proportion of therapists who felt their skills, knowledge and confidence had improved post-workshop (see Table Table6).6). These changes and proportions were maintained at follow-up. As hypothesised, there were statistically significant differences in the proportion of therapists over time who felt confident: generating a clinical question, using electronic databases, conducting databases searches alone, using a computer and the internet, and appraising research.
Critical appraisal was challenging for many participants. Less than one quarter of participants felt confident with their appraisal skills pre-workshop (18%) and post-workshop (24%). However, the proportion rose a little at follow-up to 36%. These differences over time were statistically significant (see Table Table66)
Immediately after the workshop, a higher percentage of participants (94%) reported lack of time as a barrier than had done so before (75%), as they became aware of what evidence-based practice involved. Lack of time remained an ongoing concern for 88% at follow-up. These differences were statistically significant (see Table Table77).
As hypothesised, there was a significant difference (a decrease) in the proportion of therapists who felt their searching and appraisal skills were a barrier to evidence-based practice (see Table Table77).
Searching was perceived to be less of a problem than appraisal. The proportions reporting limited search skills as a perceived barrier changed from 61% pre-workshop, to 53% post-workshop, and 24% at follow-up. The proportions reporting limited appraisal skills as a perceived barrier changed from 60% pre-workshop, to 65% post-workshop, and 41% at follow-up.
After eight months, there was also a non-significant decrease in the proportion of participants that reported difficulty accessing journals (see Table Table77).
Table Table88 summarises the frequency of searching, reading, and critical appraisal activities, and time spent teaching others about evidence-based practice. Only a small proportion of participants used the skills and knowledge acquired at the 2-day workshop although 23 participants (20.2%) completed their assignment.
The only statistically significant difference was a decrease – not an increase as hypothesised – in the proportion of therapists engaged in searching between Time 1 and Time 5. Nonetheless, searching was the most popular activity followed by reading without appraisal. Between 23% and 41% of participants searched electronic databases twice or more over eight weeks, and between 10% and 30% searched at least once over eight weeks. However, for critical appraisal, only 3% to 11% of participants engaged in this activity at least once over eight weeks. The majority, between 83% and 89%, did not participate in any critical appraisal in the eight-week time periods. Research utilisation behaviours were low initially and remained low.
The aim of this study was to evaluate the effect of a 2-day interactive workshop plus follow-up support – a 'multi-faceted' intervention – on the knowledge and skills, attitudes and behaviour of occupational therapists.
There were several key findings. First, improvements in knowledge were statistically and educationally significant, and these changes were maintained at follow-up. Second, attitudes to, and confidence with searching and appraisal improved over time, with more participants feeling confident searching for, than appraising evidence at all stages of the study. Third, the frequency of evidence-seeking behaviour and appraisal changed little over eight months.
Based on the Adapted Fresno Test, there were significant changes in knowledge over time. Furthermore, changes in the Adapted Fresno Test total score were consistent with other measures of knowledge: more therapists correctly answered the knowledge test questions and reported better knowledge and skills over time.
Nonetheless, it is possible that a learning effect occurred with each of these measures and contributed to the positive findings. Therefore, the findings should be interpreted with caution. In addition, some of the change reported for the Adapted Fresno Test may reflect measurement error, because of less-than-perfect interrater reliability for four of the seven questions. Scoring guidelines for Questions 1, 3, 4 and 5 require further refinement before the test can be published and used by researchers.
The results indicate that participants found appraisal more challenging to learn than searching. A large proportion (65%) still felt their appraisal skills were a barrier to evidence-based practice even after the 2-day workshop. However, participants learned about different research methods, both qualitative and quantitative, as well as how to interprete p-values and confidence intervals in a relatively short time. Most recognised that they would have to practice these skills back at work, if they wished to confidently appraise research on their own. They also realised how much time this additional learning would take out of their already busy schedule.
Unfortunately, activity diaries show that few participants found time to read or appraise research, in the eight months that followed. Over 60% did not read, and over 80% did not appraise any research literature. Additional studies, both quantitative and qualitative, are needed to determine the most effective way of establishing reading and appraisal habits. None of the current study participants managed to establish a regular journal club in their workplace over eight months. A recent study of physiotherapy departments  found that journal clubs existed in 42% of responding departments in England, but only 19% of those surveyed in Australia. However, relatively few departments used an evidence-based format involving appraisal. Currently, there are descriptive examples of multidisciplinary  and cross-regional  journal clubs in the allied health and nursing literature, but no evaluations of outcome. Without regular events like a journal club, allied health professionals will potentially lose these skills, and individuals will be less likely to move on and implement evidence in practice.
Disappointingly, our findings about activity levels concur with those of Forsetlund and colleagues , who targetted public health physicians in Norway. In their randomised controlled trial, education was followed by extensive support including newsletters, an electronic discussion list, assignments and free access to several databases. Despite the lengthy follow-up period (1.5 years), there was little difference in the frequency of searching or the application of evidence over time, or between intervention and control groups. Knowledge improved, but not the use of evidence.
Forsetlund and colleagues proposed that despite their negative findings, changes in knowledge and attitude may still be important pre-requisites for evidence-seeking behaviour, like a developmental stage. Further, they proposed that 1.5 years after learning essential knowledge and skills may still be too early to observe evidence being used in practice.
The findings of our study support the views of Forsetlund and colleagues. Developing skills for evidence-based practice involves a major change in values and priorities, habits and routines at an individual and organisational level. In the current study, it was not anticipated that occupational therapists would apply evidence in daily practice with patients after eight months, only that they would look for, appraise and summarise best evidence. Using Prochaska and DiClemente's staged model of change , most participants in the current study moved from the stage of contemplation to the stage of action. In real terms, they moved from thinking about, and taking an interest in evidence-based practice, to attending a workshop, planning and then (for some at least), working on a CAAT. However, study participants had to overcome barriers such as lack of time and large workloads, and make time to apply their skills and knowledge back at work. While more than 50% of the sample engaged in searching post-workshop, less than 20% reached the stage of critical appraisal, and only 20% completed their CAT within the eight months.
As with all research, this study had limitations. First, no control group was used for comparison. Therefore, we do not know if changes in knowledge would have occurred anyway, without the intervention. However, that seems unlikely based on other studies that used a control group [6,16] and reported similar knowledge outcomes.
Second, all participants were self-selected occupational therapists from one state in Australia. Randomly selected participants may have been less motivated to learn, and the results less positive. However, care was taken during recruitment to minimise recruitment bias and obtain a representative sample. Table Table22 indicates that the sample was demographically representative of occupational therapists across NSW with over two- thirds working full-time in the Sydney metropolitan area . Equal proportions of participants worked in the public and private sector, also consistent with OT Australia NSW membership statistics at the time of data collection. Nonetheless, occupational therapists in the study chose to participate and the self-selected sample from one state in Australia needs to be considered when interpreting results.
A third limitation relates to the primary outcome measure. While the interrater reliability of the Adapted Fresno Test total score is acceptable (0.87 to 0.88), further work is required to improve the scoring system and reliability of four sub-test questions. For this reason, only mean differences and confidence intervals for the total score have been reported.
A fourth limitation is the possibility of a learning effect from repeated administration of the outcome measures. Repeated administration of measures, particularly those focussing on knowledge, may have over-estimated the treatment effect.
Finally, the follow-up rate in the current study was low. Only 51 participants (44.7%) returned their survey and Adapted Fresno Test for analysis at follow-up, introducing another methodological bias. Those who did not return their documentation are likely to have been less engaged, and possibly less knowledgeable and confident than those who responded. This loss to follow-up is also likely to over-estimate the effect of the intervention and needs consideration when interpreting results.
The focus of this study was on change – in knowledge, skills, attitudes and behaviour. The intervention helped occupational therapists to improve their knowledge, skills and confidence in relation to the first three steps of evidence-based practice. Furthermore, knowledge and skills were retained over eight months.
The study provides practical strategies for measuring change in skills and knowledge, and the frequency of evidence-seeking behaviour. Some of these measures can be used to evaluate the effectiveness of undergraduate and graduate education programs.
Disappointingly, the intervention had little impact on behaviour over eight months. Most participants were unable to establish a regular pattern of searching, reading and appraisal even when their behaviour was monitored.
The ongoing challenge for educators, researchers and managers is how to help clinicians establish new routines and priorities around evidence-based practice. The reality is that behaviour change of this nature probably takes years, not months. Allied health educators and learners may find it helpful to examine outcomes from the current study, discuss the process and stages of behaviour change, and plan realistic longer term goals.
The author(s) declare that they have no competing interests.
Annie McCluskey designed, conducted, analysed and wrote up the results of this study. Meryl Lovarini developed teaching materials for the workshop, provided support to participants during the study, assisted with data collection, analysis and writing up.
The pre-publication history for this paper can be accessed here:
Valerie Cupitt and Bianca Bishop contributed to analysis of the activity diaries and Adapted Fresno Test, respectively, as part of their undergraduate honours research in 2002/2003.
Funding was received from the Motor Accidents Authority of New South Wales (No. 01/061); this agency did not influence, or participate in decisions about, study design or manuscript submission.