Presenting Evidence to Patients Online: What Do Web Users Think of Consumer Summaries of Cochrane Musculoskeletal Reviews?
Background: The Internet has the potential to be an effective medium for delivering health care knowledge to consumers. While computer usability research makes recommendations about how to present Web-based information generally, there remains no clear guidance on how to present specific forms of health care research evidence online in a way that facilitates understanding and good health care decision making.
Objective: The two goals of this study were to describe the Cochrane Musculoskeletal Group’s (CMSG’s) process for developing online patient-focused summaries of systematic reviews and to evaluate the impressions of these summaries formed by users.
Methods: A process for summarizing the results of systematic reviews via consumer summaries has evolved over 15 years. An evaluation of this approach took the form of Internet surveys on the Arthritis Society of Canada website and surveys of members of the Canadian Arthritis Patient Alliance (CAPA). Respondents provided information on background, relationship to the decision, their satisfaction with and preparation for decision making, and suggestions for improvements to the summaries. Survey data were collected between August 1, 2005, and February 28, 2006.
Results: A total of 261 respondents completed the survey. The majority (226/261 or 87%) of respondents reported having an arthritis-related condition. The consumer summary approach was generally reviewed favorably by respondents, with most agreeing that the summary provided appropriate information (177/261 or 68%), would be useful to others (160/261 or 61%), was well laid out (159/261 or 61%), was easy to learn from (157/261 or 60%), and was useful to the reader (153/261 or 59%). Areas of potential improvement were indicated by relatively fewer respondents agreeing that they could easily find all the information they wanted (118/261 or 45%), by a substantial proportion being unable to judge whether the providers of the information are reliable (80/261 or 31%), and by a similar proportion being unable to determine whether the information presented was the best available (68/261 or 26%).
Conclusions: The CMSG has developed an approach to summarizing the results of often-technical systematic reviews into public-friendly consumer summaries. Our online survey showed that this approach was generally well liked but identified specific areas for improvement. Feedback from this survey will help to reshape and improve the current template for consumer summaries used by the CMSG.
J Med Internet Res 2011;13(1):e5)
- Decision support techniques;
- health care surveys;
- patient education;
- patient satisfaction;
- patient-centered care;
- musculoskeletal diseases
A key aspect of the rapidly changing face of health care is the explosion of knowledge targeted at health care consumers. In part, because of advancing information technology allowing access to them, knowledge producers are increasingly seeing consumers as an important target group. In a recent review of 56 Canadian organizations producing practice guidelines between 2000 and 2005, it was found that 630 unique guidelines had been developed; of these, 42.7% included consumer versions or were intended for consumers . Furthermore, many major governmental and nonprofit organizations such as the UK National Institute for Health and Clinical Excellence (NICE) [ ], the US Agency for Healthcare Research and Quality (AHRQ) [ ], the Journal of the American Medical Association [ ], the Cochrane Collaboration [ ], along with many patient condition-specific organizations (eg, Arthritis Society of Canada [ ]) now disseminate research evidence directly to health care consumers, primarily via the Internet. Indeed, targeting patients can be an effective strategy to reduce the gap between research knowledge and clinical practice [ , ].
Development of consumer-targeted health knowledge is not only on the rise because of “push” from knowledge producers but also because of “pull” by consumers . Consumers are often highly motivated to maximize the quality of their own care and are demanding greater involvement in decisions surrounding their own health care [ ]. This is manifested as a demand for health care knowledge, with the Internet serving as an important mode of delivery. For example, about 80% of Canadians over the age of 16 now report using the Internet, with health information being the most commonly reported search topic [ , ].
The Internet has many advantages as a medium for delivering health care knowledge to consumers. For knowledge producers, the Web can provide wide distribution at relatively little cost. Its potential for interactivity can allow both an improved learning environment and data gathering alongside dissemination activities. Furthermore, knowledge can be updated with relative ease and low cost when compared to print or other media. For consumers, the Web is accessible, free, convenient, and allows for learning at a pace the individual finds most comfortable.
Despite these advantages, effective dissemination of online health care knowledge can be limited by variable quality and indeterminate reliability . Relevant literatures exist, but have not been exploited. Computer usability research makes general recommendations about presenting Web-based information [ , ], but these lessons have often not been followed in presenting health information online. Similarly, considerable research from the patient decision-support and human decision-making literatures [ , ] has not been translated into specific recommendations for facilitating good decision making based on online health information. We propose that specific approaches to presenting health care research knowledge online to consumers need to be developed and evaluated.
The Cochrane Musculoskeletal Group (CMSG) has produced consumer summaries of over 100 of its systematic reviews of health care interventions for arthritis and other related conditions. These consumer summaries are targeted at patients but have only recently become widely available on the Internet and are now accessible by over half of the world’s population through country-level subscriptions to the Cochrane Library and consumer-targeted websites (eg, Arthritis Society of Canada  and Arthritis Victoria in Australia [ ]).
While these summaries were developed with extensive consumer input, they were primarily developed in a paper format, which may not have translated well into effective online information tools. Guidance for producing effective online tools comes from at least two sources. First, the human-computer interaction literature identifies components of effective online information, as well as means to measure them . For example, the extent to which a computer website is favorably rated by users has been categorized into 5 domains of satisfaction including aesthetics, likeability, usability, emotion, and expectation [ , ]. Second, a substantial literature has grown up around the Ottawa Decision Support Framework [ , , ] focused on measuring constructs related to good-quality decision making. Based on this literature, we sought to examine the extent to which the CMSG summaries were seen to be relevant to such constructs.
This paper had two primary objectives: (1) to provide a narrative describing the work of the CMSG in creating online consumer summaries of the evidence from systematic reviews and (2) to evaluate the impressions of these summaries formed by users of the Arthritis Society of Canada website via an Internet survey. This knowledge will contribute toward establishing guidelines about how to summarize and present research evidence to consumers on the Internet.
The Cochrane Musculoskeletal Group Consumer Summaries
The CMSG consumer summaries have evolved over the last 15 years. Initially developed on an ad hoc basis, they have been revised and standardized based on the recommendations of a variety of guidelines for creating patient information [- ], evidence from research [ ], and user feedback. Summaries are regularly distributed and feedback sought from consumer members of the CMSG at annual CMSG meetings and from consumers and research professionals at workshops at Cochrane Colloquium meetings [ , ].
What has resulted is the standard 1-page summary now used by the CMSG (see example,). Each summary usually consists of no more than 400 words and typically takes consumers approximately 5 minutes to read. Each summary is divided into short sections with illustrative, standardized questions as headers. An introductory section (section 1) provides background on what and who was studied and mentions the Cochrane Collaboration as the source of the information. Section 2 answers questions about the intervention and the condition, for example, “What are osteoarthritis and glucosamine?” Section 3 answers questions about the effectiveness of the intervention, for example, “How well does glucosamine work?” Section 4 answers questions about safety, explicitly addressing both benefits and harms, for example, “How safe is glucosamine?” The final section provides a single-statement summary of the overall meaning of the results and provides a Web link to the description of the level of evidence underlying these statements (ie, platinum, gold, silver, and bronze levels of evidence). This method for grading scientific evidence was derived by Tugwell et al and incorporates the types of studies and quality of evidence into the ranking [ ]. This section answers the question, “What is the bottom line?”
Sections are kept short, and information is provided in a logical flow to facilitate interpretation by a wide range of audiences. While formal readability evaluations have been carried out at various points during the development process of these summaries, questions about the validity of formal readability evaluations have been noted . We found that for these summaries the technical terminology describing the diseases and treatments in CMSG reviews often inflated readability scores, while eliminating this terminology to reduce readability scores often caused more problems than it solved for users. Rather than relying on formal readability scores, we opted to ensure that all technical terms were clearly defined, and that all language was clear and readable. We saw readability analysis, therefore, as a means to end (a readable document dealing with technical issues) rather than an end itself (a document with a particular readability score).
We also strove to maintain the active voice throughout, based on the recommendations of a number of guidelines [- ]. When available, outcome probabilities are presented in natural frequencies with consistent denominators (eg, “72 out of 100”), explicitly describe the time frame to which the results apply (eg, “after 6 months of treatment, 10 patients out of 100 will improve will improve”), and present the numbers in multiple ways (eg, “this means that 12 more people will improve”) to facilitate understanding [ ]. describes what we see as key components of successful summaries as informed by the experience of the CMSG development process. While these individual recommendations have not been empirically tested in the current context, their combination gave rise to the summaries evaluated here.
In total, over 100 consumer summaries of systematic reviews for treatments of arthritis-related conditions have been produced and made available by the CMSG between 1993 and 2005. All are available online through the Arthritis Society of Canada website  and were the focus of our online survey. This site contains a wide variety of information about the many conditions related to arthritis, tips for living well, drug information, discussion forums, self-management programs, and research information. The consumer summaries are located in the latter section of the website and can be reached in 2 clicks from the home page. The summaries are organized by type of arthritis, and links are provided to related information on the website. To evaluate and improve the summaries, we conducted a Web-based survey of a subset of visitors to the Arthritis Society Website who read 1 or more CMSG summaries.
Survey: Overall Approach
Over the years, CMSG summaries have been subjected to many rounds of focus groups, interviews, and other forms of qualitative testing and evaluation. We chose to conduct an evaluation using a Web-based survey for 2 main reasons. First, the majority of this work was carried out on paper-based summaries, and we wanted to know how well the summaries translated to an online environment. Second, most of this qualitative testing was carried out on a select sample of people who were closely aligned with the CMSG, and we wanted to elicit the impressions of a wider range of types of respondents. In addition, while our survey should not be considered theory-derived, the choice of constructs was primarily informed by principles of computer usability [, ] and the Ottawa Decision Support Framework [ , ], around which constructs relevant to good quality decision making have been developed.
This study was approved by the Ottawa Hospital Research Ethics Board. We designed an exploratory Web-based survey to evaluate user impressions of the CMSG consumer summaries. After the author group identified all the key constructs to be included in the survey, reviews of the literature were carried out for validated measures of the relevant constructs, and, where appropriate, such measures were included in the original survey or versions modified for Web-based administration. An iterative process of design, evaluation by the author team, and redesign resulted in a draft version of the survey that was programmed for the Web. The survey was then pilot tested at the Carleton University Human Oriented Technology Laboratory by 5 senior students trained in issues of computer usability. These students completed the survey and provided feedback on how to improve the aesthetic qualities, layout, content, and navigational ease of the survey.
The final version of the survey consisted of a cover letter and 5 sections and included both open-ended questions and closed-ended questions with pull-down menu or check box-type response options.
The cover letter of the survey included the names of principal investigators on the project, the purpose of the study, and the length of time the survey was expected to take. It also stated that completion and submission of the survey would serve as tacit consent that the subject’s responses could be used in the study. Details of duration and location of data storage were not provided unless specifically requested by the participant.
The survey consisted of 5 sections. The first addressed summary-specific issues such as how long participants had spent reading the summary. Subsequent sections included items on user experience, satisfaction with the way the information prepared them for decision making, experience with computers, background/demographic information, and suggestions for improvement. The complete survey consisted of 1 screen displaying a total of 53 items.
Summary-specific items measured the time spent reading the summary as an indirect measure of the care with which reviewers read the items (categorized in 5-minute increments). Also measured was confidence in participants’ understanding of the key points of the summary (on a 5-point scale from “not at all” to “very” confident) as a key prerequisite of good decision making as described by the Ottawa Decision Support Framework [, ]. Finally, items targeting the extent to which the respondent was closely related to the issue included how participants had discovered the treatment summaries (ie, by exploring arthritis.ca, through a search engine, a recommendation of a physician, friends/family, or a support group, or other); participants’ main reasons for reading the summary (ie, for personal relevance, general interest, or other), and interest in arthritis (ie, “I have arthritis,” “I know someone with arthritis,” or other). The summary the respondent had read was logged automatically.
Computer user experience was measured using 14 items () related to 5 domains of user satisfaction [ ] including (1) aesthetics, (2) likeability, (3) usability, (4) emotion, and (5) expectation. Each item was rated on a 6-point scale from “strongly disagree” to “strongly agree” with an option of “no opinion/don’t know.” Also, 4 items (the summary is boring, the summary is frustrating, learning from the summary seems hard, I had to read too much) were phrased in the opposite direction from the rest of the items, and reported item means were reversed for ease of comparison.
Satisfaction in how the information materials prepared them for decision making was measured using 11 items adapted from Graham and O’Connor . Items elicited respondents’ feelings of support and preparation to make decisions ( ). Items were rated on a 5-point scale from “not at all” to “a great deal.” Finally, an open-ended question asked for advice on how the consumer summaries could be improved. Note that items from both the user experience and satisfaction with preparation for decision-making instruments were analyzed separately, as neither scale had been validated in the form administered in this survey. Our goal for these items was to describe people’s impressions of the consumer summaries rather than target potential underlying constructs. Analysis and validation of underlying constructs for these items will be the subject of separate investigation.
We included 6 computer expertise items selected from a scale reported by Liaw et al , which asked respondents to self-report on their experience with (1) computers in general, (2) the Internet/World Wide Web, (3) Internet search engines, (4) word processing software, (5) database software, and (6) computer programming languages. Participants were asked to rate each of these 6 items on a 5-point scale ranging from 0, “none at all” to 4, “a great deal.” The 6 items were summed to produce a scale score ranging from 0 to 24. We then categorized this scale into thirds to indicate respondents with low, middle, and high levels of self-reported computer experience.
Background questions included demographic questions about sex, age, education, location of residence, and employment status.
Sampling and Recruitment
Sample size for this descriptive survey was based on estimates of the amount of traffic to the website and the response rate to the online questionnaire. We chose to aim for 300 respondents to yield a wide range of opinions on each of the 10 summaries attached to the survey. In prior years, approximately 17,000 users visited the relevant section of the Arthritis Society of Canada website. From this pool, previous (1-question) surveys administered on this website had typically yielded 200 to 500 responses per month. Based on an estimated completion rate of 30% of that number, we planned to collect 60 to 150 responses per month and take between 2 and 5 months to collect the data.
Recruitment to the survey proved a bigger challenge than expected. As a result, we engaged in 3 relatively separate recruitment strategies. First, we used a convenience sampling strategy in which the administrator of the Arthritis Society of Canada website posted a link on the home page of the site encouraging any visitors to the site to consider completing a voluntary, Web-based, open version of the survey. Respondents were assured of the confidentiality of their data, but this version did not specifically state that ethics board approval had been granted. Second, the society distributed a letter on behalf of the study authors to the provincial Arthritis Society educational team leaders asking them to encourage their local members and contacts to complete the Web-based survey. Finally, an electronic invitation to participate in a closed version of the survey was sent to all 463 members of the Canadian Arthritis Patient Alliance (CAPA), an advocacy group closely related to the Arthritis Society of Canada. Members of CAPA typically have arthritis or have a strong interest in arthritis advocacy.
The Web survey was linked to 10 of the most popular CMSG summaries on the Arthritis Society of Canada website. All 10 summaries were standardized with regard to format and content as of July 2005 (see).
Respondents chose the summary that was most relevant to them either by clicking on it during their Internet session or by being directed by an email recruitment letter to a list of the 10 target summaries. A link to the survey itself was appended to the bottom of each selected summary, and instructions on the survey indicated that participants should complete the survey about the summary they had just read. Summary topics addressed issues around various treatments for osteoarthritis, rheumatoid arthritis, and shoulder and elbow pain. Each summary generally described the effects of a single intervention, such as drug interventions (eg, methotrexate or glucosamine) or nonpharmacologic treatments (eg, exercise or ultrasound).
To the extent possible, administration of the closed version of the survey to CAPA members was governed by Dillman’s Tailored Design Method . A prenotification to complete the survey, an invitation to visit the summaries and complete the survey, and 2 reminder emails were sent via email. This version of the survey consisted of 1 screen displaying a total of 71 items including most questions from the earlier version (with the exception of 2 arthritis.ca website feedback items) plus 1 additional scale, the Medical Data Interpretation Test [ ], which was excluded from analysis due to technical problems. An appendix with the CAPA prenotification, CAPA cover letter, CAPA survey, Web link on the Arthritis Society of Canada website, Web survey, and the 10 consumer summaries can be obtained from the authors. Correspondence assured respondents of the confidentiality of their data and stated that the survey had been approved by the Ottawa Hospital Research Ethics Board.
All survey items were provided in a consistent order for all participants; no randomization of questions was carried out. All participants saw the same questions; no adaptive questioning or question branching was required. Completion of items was not enforced and the options “not applicable,” “no opinion,” and “rather not say” were included for all items comprising preexisting scales. Respondents were not asked to review responses before submitting the survey. No incentives were provided for respondents to complete the survey.
Data Collection and Statistical Analysis
Survey data were collected between August 1, 2005, and February 28, 2006. Closed-ended items were analyzed using frequencies and descriptive statistics using SPSS, version 16 (SPSS Inc, Chicago, IL, USA) and SAS, version 9.1 (SAS Institute Inc, Cary, NC, USA). Missing data were assessed on a question-by-question basis. Open-ended comments about how to improve the summaries were assembled into Microsoft Excel and examined for themes. In the next step, 3 coders (authors JB, AL, NS) reviewed all comments provided by respondents and coded them into 1 or more of 12 themes determined by an initial scan of the responses. Disagreements over how comments should be categorized into themes were resolved by consensus. We employed the Checklist for Reporting Results of Internet E-Surveys (CHERRIES) reporting guideline to inform our report of this study .
Over the 7-month period from August 2005 through February 2006 that the survey was available on the Arthritis Society website, 162 site visitors responded to the survey; the number of hits to the survey-relevant summaries during this time period was not collected. Of the invitations we sent to the CAPA members, we obtained 99 responses out of 395 successfully delivered emails (25%). After comparisons of demographic characteristics showed no important differences between the website and CAPA respondents, responses from the 2 groups were combined for a total of 261 responses.
describes the background and summary-specific characteristics of the respondents. The majority (176/261 or 67%) of respondents were female, and 56% (147/261) were between the ages of 45 and 64 years of age. Most were well educated, with 77% (202/261) of respondents reporting at least some postsecondary education, and 35% (92/261) were employed full-time. Respondents were distributed across western (British Columbia, Alberta, Saskatchewan, Manitoba; 49/261 or 19%), central (Ontario, Quebec; 61/261 or 23%), and eastern Canada (New Brunswick, Nova Scotia, Prince Edward Island, Newfoundland; 115/261 or 44%) and included respondents from all 10 provinces, but none from the 3 northern territories. Approximately 7% (17/261) of respondents were not from Canada.
Respondents showed variable computer experience, but 48% (125/261) reported a moderate level of experience. A substantial majority (226/261 or 87%) reported that they themselves had some form of arthritis. Of the 10 different target summaries, 48% (124/261) of respondents chose a summary focused on rheumatoid arthritis, while another 41% (107/261) chose osteoarthritis. Nearly half (111/261 or 48%) reported spending less than 5 minutes reading the summaries. The most common ways in which the summaries were discovered were by simple exploration of the Arthritis Society of Canada website (60/261 or 23%), upon physician recommendation (55/261 or 21%), or through support groups like CAPA (44/261 or 17%), while the most common reasons for reading the summaries were for personal relevance (169/261 or 65%) or general interest (51/261 or 20%).
describes the number and percentage of the 261 respondents reviewing each consumer summary and self-reported time spent reviewing each summary. Of the 10 consumer summaries, 3 accounted for 54% (141/261) of respondents. A significant proportion of respondents spent less than 5 minutes reviewing each summary (summary-specific range of the percent of respondents who spent less than 5 minutes reviewing the summary, 25% to 70%).
describes responses to the 14 user experience items. A number of items indicated favorable impressions by respondents, with a majority either strongly or moderately agreeing with statements that the summary used language that was appropriate (196/261 or 75%), provided expected information (178/261 or 68%), (didn’t) involve too much reading (172/261 or 66%), would be of use to many others (160/261 or 61%), was well laid out (159/261 or 61%), (wasn’t) hard to learn from (157/261 or 60%), (wasn’t) frustrating (154/261 or 59%), was useful to me (153/261 or 59%) and (wasn’t) boring (149/261 or 57%). Somewhat fewer respondents agreed that they could find all the information they wanted (118/261 or 45% strongly or moderately agreed). Finally, several items suggested that many respondents felt unable to judge the credibility or reliability of the summaries. Relatively few agreed that the summary clearly provided information on the providers of the information (112/261 or 43%). Furthermore, large proportions of respondents selected “no opinion/don’t know” to whether the information is up-to-date (54/261 or 21%), the providers of the information are reliable (80/261 or 31%), and the information presented was the best available (68/261 or 26%).
describes responses to the items related to preparation for decision making. In general, the responses indicated that the summary would improve preparation for decision making, particularly to help to identify questions to ask the physician (157/261 or 60%). Relatively few agreed with the statement that the summary helps you know that your values affect your decisions (103/261 or 40%).
To probe in more detail issues around people’s impressions of the summaries, we asked respondents to provide us with feedback about how we might improve the summaries. Of the 261 respondents, 131 provided 1 or more comments.presents the themes identified, the relative frequencies with which those themes were mentioned, and shows examples of comments relevant to each theme. The most commonly cited themes were to provide more detail about the treatment and options, risks and the research; to provide additional interactivity or functionality to the summaries; to make the language clearer or simpler; and to increase the use of pictures, graphs or colors. Several people specifically requested the summaries use point form and provide more bottom lines about the treatments, and several suggested that more detailed information about the credibility of the information should be provided.
The 15-year evolution of the CMSG consumer summaries has resulted in a successful standardized presentation format that enables brief but clear presentation of research evidence for a wide range of treatments and interventions for musculoskeletal disorders. Because these consumer summaries were carefully tested as paper-based tools but hadn’t been tested as Web-based tools, we decided to evaluate how they performed on the Web and what specific areas needed work in order to improve them as Web-based tools. Our survey showed that the tools were generally rated favorably and identified specific areas for improvement, which we discuss below.
Amount of Detail
One key finding was that many respondents reported wanting additional information to be available from the summaries. Less than half of respondents (118/261 or 45%) agreed that they could easily find all the information they wanted. Open-ended comments also revealed that many respondents wanted more details about specific risks, about the types of studies comprising the research, and about the type of participants in the studies (presumably so that they could compare themselves to the study participants). Many respondents also wanted more information about the condition and its various treatment options.
We note that our sample of respondents was likely quite sophisticated in terms of its existing knowledge on arthritis-related issues. Over 77% (202/261) of respondents had some postsecondary education, and many came through recruitment from the Canadian Arthritis Patient Alliance group, members of which are likely to be actively engaged in issues of musculoskeletal disorders. This sample may, therefore, have been quite well informed already about issues around their illness (226/261 or 87% had personally experienced some arthritis condition) and, therefore, preferred relatively detailed information.
A key challenge when presenting any health information online is dealing with the variability in user needs. Nearly half (111/261 or 48%) spent less than 5 minutes reading the consumer summaries, and the clear majority (200/261 or 77%) reported spending 10 minutes or less reading them. The standardized format used here was designed primarily to yield clear, concise summaries of systematic reviews. Yet many users will wish to use these summaries as springboards for more detailed information searches. This need for more flexible, interactive information presentation was evident in our survey findings: 2 commonly cited themes were to increase use of interactivity and to include more in the way of graphs, charts, and other information-rich display formats. To facilitate such uses, it seems likely that rather than trying to create a one-size-fits-all solution that would be too detailed for some and not detailed enough for others, the challenge for developers is to provide a flexible, interactive approach that can allow users to tailor for themselves the amount and type of information they review . This approach is becoming increasingly feasible on the Web, and suggests a clear avenue for future research.
We have begun this work in at least 3 ways. First, we are exploring the utility of providing links within these summaries to other, more detailed descriptions of the systematic reviews. Second, the consumer summaries are housed within the larger Arthritis Society website that includes a wealth of information on all aspects of musculoskeletal disorders; efforts to link from the summaries to this additional information may well improve areas identified in this survey. Third, we are currently conducting research on how to adapt more detailed decision support tools such as patient decision aids for use online, discussed in more detail below.
Layout and Language
While the majority of respondents felt our consumer summaries incorporated clear language and were well laid out, some respondents did feel that the language or formatting could be improved. Within the CMSG, we have avoided evaluating the summaries using readability algorithms due to the number of technical, often complex terms at the heart of the reviews. Instead, our goal has been to use relatively few content words per sentence (ie, use lower lexical density) and a clear, logical progression from background information to the effects of treatment . While this approach appears to be a qualified success, there may be opportunity for linking terms; such an approach can allow optional, more detailed information and definitions to be provided without adding length or clutter. We are currently exploring the use of such techniques in another study.
Credibility of the Source
Many people were unable to judge the credibility or reliability of the summaries or did not know who the providers of the information were. This is a concern since assessment of the credibility of online information is a key component of evaluating health information . Our summaries included a statement that the research is based on a Cochrane review, a reference to the review at the bottom of the summary, and a link to a website, About Cochrane, that describes the processes involved in writing a review. Clearly some respondents did not see or make use of this information; more investigation of how to make this information more salient to users will be important.
Preparation for Decision Making
While user experience was generally positive, fewer than half of the respondents felt that the summary helped them recognize that a decision needs to be made, think about pros and cons, know how their values affect their decision, organize their own thoughts, or prepare to make a better decision. It is unsurprising that these summaries should not have all information necessary to prepare people for a decision since they are limited chiefly to providing information about the treatment options and the pros and cons of the treatments . We are currently assessing whether patient decision aids, that is, decision support tools designed to help people make specific and deliberative choices among options, may be useful when presented online. While their effectiveness has been demonstrated in a variety of other presentation formats [ ], it remains an open question how such tools can most effectively be employed via the medium of the Web and for what situations such tools may be most useful. These tools not only present information on the options and outcomes relevant to the person’s health status but often also include exercises to help patients explicate factors such as how they value the different options, preference for role in decision making, or choice predisposition. Patient decision aids can be much more detailed and, therefore, longer than consumer summaries. Use of these more detailed tools may be warranted in situations where a decision cannot be made on the basis of a consumer summary.
The response rate and makeup of our sample of respondents is one clear limitation of the present study. The Web subsample was collected over a period of 7 months from summaries that see hundreds of visitors per month. While we have no information on nonresponders, we have to assume that our relatively slow accrual rate suggests that we were only obtaining data from a small, select group of visitors to the site. The CAPA survey subgroup likely exhibits similar biases. Despite our best efforts, logistical limitations prevented us from obtaining a high response rate from the CAPA survey sample. The response rate (99/395 or 25%) suggests that we may have a very select sample of the CAPA group, which itself is likely quite different from the target population of all arthritis patients using online information. In total, we must assume that our sample is biased with respect to our overall target population. We have chosen to interpret our results not as a sample representative of all arthritis sufferers, but as one of a relatively sophisticated sample of well-educated patients. Future work will be needed to assess whether these findings generalize to the wider population of arthritis patients who use online information.
A second clear limitation of the current work is the lack of a control group against which to compare the survey findings. We chose the single-group design in order to evaluate the CMSG model for presenting consumer information online as it is currently being implemented on the Canadian Arthritis Society website, and we chose to use the information derived from it to inform future controlled studies. Our use of previously validated measures (eg, computer experience and satisfaction with preparation for decision making) gives us confidence in the constructs we have measured, but ongoing work using controlled designs will assess the extent to which the levels of these constructs can be improved upon using other approaches.
The relationship between the Cochrane Musculoskeletal Group and the Arthritis Society has created an excellent opportunity for research producers to target those people who would benefit most from this research information. The current work focuses specifically on engaging in this knowledge translation process in an online environment and makes clear that while we are on the right track, there is more work to do in order to understand how best to communicate systematic review information online. We have begun this work and provided some initial recommendations about how consumer summaries should look. Feedback from this survey will help to reshape and improve on the current presentation format for consumer summaries used by the CMSG. Our results should also provide initial guidelines to other developers of patient information who wish to reach consumers via the Internet.
We would like to thank the Arthritis Society of Canada and the Canadian Arthritis Patient Alliance for their support of this survey and, in particular, Rob Watts, manager of Internet resources, Arthritis Society of Canada; Anne Dooley, cochair of research, Canadian Arthritis Patient Alliance; and Colleen Maloney, vice president, Canadian Arthritis Patient Alliance. Also thanks to Anton Saarimaki who programmed the survey for the Web and to Dr Ann Cranney for clinical input. This work was funded by an operating grant from the Canadian Institutes of Health Research (CIHR; MOP-68973). Dr Brehaut is a CIHR New Investigator in Knowledge Translation.
Conflicts of Interest
- Kryworuchko J, Stacey D, Bai N, Graham ID. Twelve years of clinical practice guideline development, dissemination and evaluation in Canada (1994 to 2005). Implement Sci 2009;4:49 [FREE Full text] [CrossRef] [Medline]
- National Institute for Health and Clinical Excellence (NICE). URL: http://www.nice.org.uk/ [accessed 2010-02-02] [WebCite Cache]
- AHRQ Agency for Healthcare Research and Quality. URL: http://www.ahrq.gov/ [accessed 2010-02-02] [WebCite Cache]
- JAMA. JAMA Patient Page URL: http://jama.ama-assn.org/cgi/collection/patient_page [accessed 2010-02-02] [WebCite Cache]
- The Cochrane Collaboration. URL: http://www.cochrane.org/ [accessed 2010-02-02] [WebCite Cache]
- The Arthritis Society. URL: http://www.arthritis.ca/splash/default.asp?s=1&returnurl=/arthritis%20home/default.asp?s=1 [accessed 2011-01-17] [WebCite Cache]
- Schuster MA, McGlynn EA, Brook RH. How good is the quality of health care in the United States? Milbank Q 1998;76(4):517-63, 509. [Medline]
- Grol R. Successes and failures in the implementation of evidence-based guidelines for clinical practice. Med Care 2001 Aug;39(8 Suppl 2):II46-II54. [Medline]
- Lavis JN, Lomas J, Hamid M, Sewankambo NK. Assessing country-level efforts to link research to action. Bull World Health Organ 2006 Aug;84(8):620-628 [FREE Full text] [Medline]
- Kassirer JP. Incorporating patients' preferences into medical decisions. N Engl J Med 1994 Jun 30;330(26):1895-1896. [CrossRef] [Medline]
- Statistics Canada. 2010 May 10. The Daily Canadian Internet Use Survey URL: http://www.statcan.gc.ca/daily-quotidien/100510/dq100510a-eng.htm [accessed 2010-05-05] [WebCite Cache]
- Statistics Canada. Internet users at home URL: http://www40.statcan.gc.ca/l01/cst01/comm29a-eng.htm?sdi=internet [accessed 2010-05-05] [WebCite Cache]
- Eysenbach G, Powell J, Kuss O, Sa ER. Empirical studies assessing the quality of health information for consumers on the world wide web: a systematic review. JAMA 2002;287(20):2691-2700 [FREE Full text] [Medline]
- Hammerich I, Harrison C. Developing online content: the principles of writing and editing for the Web. New York: John Wiley & Sons; 2002.
- Lindgaard G. Usability testing and system evaluation: a guide for designing useful computer systems. London: Chapman & Hall; 1994.
- Elwyn G, O'Connor AM, Bennett C, Newcombe RG, Politi M, Durand MA, et al. Assessing the quality of decision support technologies using the International Patient Decision Aid Standards instrument (IPDASi). PLoS One 2009;4(3):e4705 [FREE Full text] [CrossRef] [Medline]
- Hibbard JH, Peters E. Supporting informed consumer health care decisions: data presentation approaches that facilitate the use of information in choice. Annu Rev Public Health 2003;24:413-433. [CrossRef] [Medline]
- Arthritis Victoria. URL: http://www.arthritisvic.org.au/ [accessed 2010-02-02] [WebCite Cache]
- Lindgaard G, Dudek C. What is this evasive beast we call user satisfaction? Interacting with Computers 2003;15:429-454. [CrossRef]
- O'Connor AM, Bennett CL, Stacey D, Barry M, Col NF, Eden KB, et al. Decision aids for people facing health treatment or screening decisions. Cochrane Database Syst Rev 2009(3):CD001431. [CrossRef] [Medline]
- O'Connor AM, Tugwell P, Wells GA, Elmslie T, Jolly E, Hollingworth G, et al. A decision aid for women considering hormone therapy after menopause: decision support framework and evaluation. Patient Educ Couns 1998 Mar;33(3):267-279. [Medline]
- Duman M. In: Duman M, editor. Producing Patient Information: How to Research, Develop and Produce Effective Information Resources. London: King's Fund; 2003.
- Australian Government National Health and Medical Research Council. 2000. How to present the evidence for consumers: preparation of consumer publications URL: http://www.nhmrc.gov.au/publications/synopses/cp66syn.htm [accessed 2010-10-10] [WebCite Cache]
- DISCERN Online: Quality criteria for consumer health information. URL: http://www.discern.org.uk/ [accessed 2010-02-02] [WebCite Cache]
- Moynihan R, Bero L, Ross-Degnan D, Henry D, Lee K, Watkins J, et al. Coverage by the news media of the benefits and risks of medications. N Engl J Med 2000 Jun 1;342(22):1645-1650 [FREE Full text] [CrossRef] [Medline]
- Santesso N, Maxwell L, Tugwell PS, Wells GA, O'connor AM, Judd M, et al. Knowledge transfer to clinicians and consumers by the Cochrane Musculoskeletal Group. J Rheumatol 2006 Nov;33(11):2312-2318. [Medline]
- Tugwell PS, Santesso NA, O'Connor AM, Wilson AJ, Effective Consumer Investigative Group. Knowledge translation for effective consumers. Phys Ther 2007 Dec;87(12):1728-1738 [FREE Full text] [CrossRef] [Medline]
- Tugwell P, Shea B, Boers M, Brooks P, Simon L, Strand V, et al. In: Wells G, editor. Evidence-based rheumatology. London: BMJ; 2004.
- Clerehan R, Buchbinder R, Moodie J. A linguistic framework for assessing the quality of written patient information: its use in assessing methotrexate information for rheumatoid arthritis. Health Educ Res 2005 Jun;20(3):334-344 [FREE Full text] [CrossRef] [Medline]
- Fisk D, Rogers WA, Charness N, Czaja SJ, Sharit J. In: Sharit J, editor. Designing for older adults: principles and creative human factors approaches. Boca Raton, FL: CRC Press; 2004.
- Nielsen J. In: Nielsen J, editor. Designing Web Usability: The Practice of Simplicity. Indianapolis: New Riders Publishing; 2000.
- World Wide Web Consortium. Web Accessibility Initiative (WAI) URL: http://www.w3.org/WAI/ [accessed 2010-02-02] [WebCite Cache]
- Graham ID, O'Connor AM. User Manual - Preparation for Decision Making URL: http://decisionaid.ohri.ca/docs/develop/User_Manuals/UM_PrepDM.pdf [accessed 2010-02-10] [WebCite Cache]
- Liaw SS. An Internet survey for perceptions of computers and the World Wide Web: relationship, prediction, and difference. Computers in Human Behavior 2002;18:17-35. [CrossRef]
- Dillman D. In: Dillman D, editor. Mail and Internet surveys: the tailored design method. New York: John Wiley & Sons, Inc; 2000.
- Schwartz LM, Woloshin S, Welch HG. Can patients interpret health information? An assessment of the medical data interpretation test. Med Decis Making 2005;25(3):290-300. [CrossRef] [Medline]
- Eysenbach G. Improving the quality of Web surveys: the Checklist for Reporting Results of Internet E-Surveys (CHERRIES). J Med Internet Res 2004 Sep 29;6(3):e34 [FREE Full text] [CrossRef] [Medline]
- Hibbard JH, Greene J, Tusler M. Improving the outcomes of disease management by tailoring care to the patient's level of activation. Am J Manag Care 2009 Jun;15(6):353-360 [FREE Full text] [Medline]
- O'Connor AM, Llewellyn-Thomas HA, Flood AB. Modifying unwarranted variations in health care: shared decision making using patient decision aids. Health Aff (Millwood) 2004;Suppl Web Exclusives:VAR63-VAR72. [CrossRef] [Medline]
|AHRQ: (US) Agency for Healthcare Research and Quality|
|CAPA: Canadian Arthritis Patient Alliance|
|CHERRIES: Checklist for Reporting Results of Internet E-Surveys|
|CSMG: Cochrane Musculoskeletal Group|
|NICE: (UK) National Institute for Health and Clinical Excellence|
Edited by G Eysenbach; submitted 17.03.10; peer-reviewed by A Gagliardi, M Sampson, R Bannuru; comments to author 21.04.10; revised version received 14.06.10; accepted 10.07.10; published 18.01.11
©Jamie C Brehaut, Nancy Santesso, Annette M O'Connor, Alison Lott, Gitte Lindgaard, Ania Syrowatka, Ian D Graham, Peter S Tugwell. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 18.01.2011.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.