Advertisement: Preregister now for the Medicine 2.0 Congress



Article Thumbnail
Members can download this full article for Adobe PDF Format.
Membership provides unlimited access to all PDF files.
Or, ask your department head to become an institutional member.

For tax purposes please select your country and if applicable state/province of residence:
Buy Now (Pay-per-download for non-members):
Download Price (USD): $22.00

Original Paper

Online Consumer Surveys as a Methodology for Assessing the Quality of the United States Health Care System

Christina Bethell1, PhD; John Fiorillo2, MA; David Lansky3, PhD; Michael Hendryx4, PhD; James Knickman5, PhD

1Kaiser Permanente Center for Health Research, Portland OR, USA
2(formerly) The Robert Wood Johnson Foundation, Princeton, NJ, USA; (currently) The Health Strategy Group Inc, Chatham NY, USA
3FACCT—Foundation for Accountability, Portland OR, USA
4Washington State University, Spokane WA, USA
5The Robert Wood Johnson Foundation, Princeton NJ, USA

Corresponding Author:
Christina Bethell, PhD

Kaiser Permanente Center for Health Research
3800 N. Interstate Avenue
Portland OR 97227
USA
Phone: +1 503 528 3912
Fax: +1 503 335 2424
Email:



ABSTRACT

Background: Interest in monitoring the quality of health care in the United States has increased in recent years. However, the policy objectives associated with collecting this information are constrained by the limited availability of timely and relevant data at a reasonable cost. Online data-collection technologies hold the promise of gathering data directly and inexpensively from large, representative samples of patients and consumers. These new information technologies also permit efficient, real-time assessment in such areas as health status, access to care, and other aspects of the care experience that impact health outcomes.
Objective: This study investigates the feasibility, validity, and generalizability of consumer online surveys to measure key aspects of health care quality in the United States.
Methods: Surveys about the health and health care experiences of a general adult population and of adults with diabetes were administered online and by telephone. The online survey drew from a sample frame of nearly 1 million consumers and used a single e-mail notification. The random-digit-dial methodology included 6 follow-up calls. Results from the online sample were compared to the telephone sample and to national benchmark data.
Results: Survey responses about quality of care collected using online and telephone methods were commensurate once they were weighted to represent the demographic distribution of the 2000 United States Census. Expected variations in health and health care quality across demographic and socioeconomic groups were largely observed, as were hypothesized associations among quality indicators and other variables. Fewer individuals were required to be contacted to achieve target sample sizes using online versus telephone methods. Neither method yielded representative cohorts of nonwhite individuals.
Conclusions: Conclusions about the level and variations in health care quality in the United States are similar using data collected in this study compared to data collected using other telephone-based survey methods. As is typical for national telephone surveys conducted by the National Center for Health Statistics, stratified sampling and weighting of survey responses is necessary for results to be generalizable. Online methods are more appropriate for understanding health care quality than for conducting epidemiologic assessments of health in the United States.

(J Med Internet Res 2004;6(1):e2)
doi:10.2196/jmir.6.1.e2

KEYWORDS

Quality of health care; data collection; Internet; health care surveys; consumers



Introduction

Recent years have seen a marked increase of interest in monitoring the quality of health care in the United States. Congress has mandated the annual release of a National Healthcare Quality Report, which will include results from consumer-reported surveys on health care quality [1]. Congress, a presidential commission, and the National Quality Forum have all called for publication of consumer-centered quality performance information, and the administrator of the Medicare program has indicated the government's intention of releasing performance data for nursing homes, hospitals, and perhaps even physicians [2,3]. State Medicaid and State Children's Health Insurance Programs (SCHIP) are required to assess and report on quality of care provided to consumers enrolled in these programs [4,5].

Need for Timely and Efficient Collection of Quality Information

These policy objectives are constrained by the limited availability of timely and relevant data at a reasonable cost. Often, information strategies for health care quality must rely on datasets defined and populated for other reasons, such as documentation of financial transactions, public health surveillance, or contractual oversight and audit. Seldom do such assessment systems address the health care quality concerns of patients and consumers, and rarely do they capture their experiences or attitudes.

Consequently, a tension exists between the policy objective of evaluating our success at creating a more-responsive health system that achieves priority health goals and our dependence on an information infrastructure unable to capture the necessary data to determine whether these goals are being achieved. Two trends offer some hope of resolving this tension. First, scientists have developed and validated an extensive library of patient survey instruments over the past 20 years. Tools now permit us to measure the performance of the health system along the dimensions of health care outcomes and the provision of clinically-appropriate, patient-centered care [6- 12]. Second, new information technologies hold the promise of gathering data directly and inexpensively from large, representative samples of patients and consumers. Online data-collection technologies also permit efficient, real-time assessment in such areas as health status, access to care, and other aspects of the care experience that affect health outcomes [13- 14]. Given the potential efficiencies and expediency of collecting data online, as well as growing limitations in telephone-based and/or mail-based surveys, it is clearly worthwhile—perhaps vital—that we develop and test online methods for capturing consumer-reported information on quality of health care.

The Challenges of Web-Based Patient Surveys

All modes of consumer-survey administration entail challenges of measurement error, nonresponse error, and, particularly, coverage error. Online methods may be helpful in reducing some of these sources of error, but may also encounter new challenges in other sources of error.

Measurement Error

Web-based surveys introduce a new mode of interaction with respondents. The online experience involves both technical and contextual changes that may cause variations from how the same individuals would answer questions if presented in person, on the telephone, or by mail. Among technical differences are the presentation of questions and responses on computer screens, and variations in browser layouts, colors, text, and communication speeds. Contextual factors include users' ability to review and change prior answers, look ahead to other content, "multi-task," or start and stop during a session. Studies evaluating Web-based survey-mode effects have generally shown them to more closely resemble self-administered mail surveys than interviewer-administered telephone surveys, though with lower item nonresponse [13,15] and the potential for immediate data analysis and feedback to sponsors and respondents [16,17]. To ensure consistent user experiences and reduce measurement error, a consensus set of procedural recommendations analogous to those for mailed and telephone administered surveys is emerging for conducting Web-based surveys [18- 20].

Non-Response Error

Continuing changes in consumer telephone behavior have increased and redefined nonresponse error in phone surveys [21,22]. The common use of answering machines and of technologies for caller identification and unknown-caller blocking all contribute to nonresponse bias for telephone surveys. Consumer resistance to receiving telephone calls by telemarketers is reflected in the Do-Not-Call registry recently required by Congress and implemented by the Federal Trade Commission [23]. While survey researchers conducting surveys for not-for-profit or public-interest purposes are not prevented from calling individuals in this registry, the overarching resistance and resentment expressed by consumers regarding calls made to their home during evenings and weekends could generalize to a resistance to respond to calls to conduct these types of surveys.

Although researchers have begun to study the extent and causes of nonresponse error to e-mail and Web-based surveys, it is not well-documented and remains an especially-serious concern of methodologists when considering the use of the Web to conduct population-based surveys intended for use in policy contexts [21,23,24]. Documented reasons for nonresponse range from traditional questions of content interest to respondents' use of multiple e-mail accounts and defunct or infrequently-accessed e-mail accounts [21,25,26]. The emerging consensus procedures focus on the importance of repeat contacts, tracking nondelivery to e-mail accounts, and incentives to maximize response rates.

Incomplete Coverage

Errors are introduced in national telephone surveys because of households without telephones. Similarly, incomplete Internet access, or coverage, introduces error in national estimates based on online survey data. While the majority of Americans have access to the Web, there remain significant economic, cultural, and educational disparities. As of mid-2000, the US (United States) government estimated that only13% of persons with annual incomes under $15000 had Internet access in their homes, compared with 78% of those with incomes over $75000 [27]. Overall, whites enjoy greater Web access than do African-Americans and Hispanics [27]. These disparities may be rapidly changing, however. According to a 2001 UCLA report, 72% of all Americans have Internet access, including 65% with less than a high school education [28]. Certain demographics can be specifically targeted through Web technology—for example, customers of America Online, users of eHealth services, or populations sharing an e-mail domain who can be sent a request to complete a survey. Examples could be university-affiliated populations or employees of large companies. Also, larger proportions of low-income and elderly Americans increasingly use the Internet and can be sampled in online surveys [27].

Researchers advocate various ways of responding to the noncoverage and nonresponse challenges inherent in all survey administration modes, including Web-based surveys:

  1. adjusting nonrepresentative completed samples according to characteristics known of both the starting sample and underlying populations [29- 30]
  2. coupling participant recruitment using random-digit dialing with the efficiencies and interactivity of the Web [31]
  3. restricting use of online data collection to studies of fully-covered populations (eg, university surveys and subscribers to specific Web sites)
  4. delaying use of Web surveys until coverage improves and methodological developments take place [32].

The Robert Wood Johnson Foundation (RWJF) and the Foundation for Accountability (FACCT) recognized the potential value of conducting Internet-based surveys to calculate national estimates of health care quality. In 2000 and 2001, as part of the Robert Wood Johnson Foundation National Strategic Indicators Project (NSIP), online surveys were fielded to assess health care quality for children, teens, and adults with and without chronic conditions. Survey topics were identified as relevant for each of the 5 Foundation for Accountability Consumer Information Framework (CIF) domains (the basics, staying healthy, getting better, living with illness, and changing needs) and types of measures (health outcomes, appropriate clinical care, and patient-centered experience of care) [33]. The Consumer Information Framework has been adapted for use in the Congressionally-mandated National Health Care Quality Report [1] to structure the identification and communication of quality information to the public.

In this paper, we report the extent to which data derived from a national online sample of the general adult population and from a sample of adults with diabetes meet initial criteria for use in characterizing the performance of health care systems.

Four research questions are preliminary to the overall feasibility and validity of using Web-based surveys to estimate health care quality:

  • Are online survey response rates (derived from a sampling frame recruited using opt-in Internet methods) of sufficient size and representation to estimate health care indicators for the US population?
  • Do estimates of key health status and health care system quality variables demonstrate face validity, compared to other national studies?
  • Do these estimates demonstrate concurrent validity, such that demographic and other correlates of health status and health care system performance match those observed in national telephone-based surveys?
  • Are survey scale data collected online psychometrically reliable?

Methods

This study reports both online and telephone-administered health care survey results for adults age 18 and over, as well as for adults with diabetes. Also used are data from adult respondents to the 1999 Behavioral Risk Factor Surveillance Survey (BRFSS) and the 1998 National Health Interview Survey (NHIS) administered by the US Centers for Disease Control and Prevention (CDC).

Data Collection

Online Surveys

A market research firm, Common Knowledge, Inc, recruited a panel of approximately 1 million individuals, using Internet advertisements intended to attract a group with diverse demographic and psychographic characteristics. Approximately 70% of the panel was recruited online, the remaining 30% through traditional direct-mail and telephone contact. Panelists were invited to participate in only one study per month to prevent "professional" survey takers from responding and to minimize respondent fatigue.

Two waves of sampling and data collection took place for the general-adult and adult-diabetes online surveys. In the first wave, separate stratified random samples were drawn, each representing the US population along the dimensions of age, sex, and education using 4 age groups (18-24, 25-44, 45-64, over 65) and 4 educational groups (less than high school, high school/GED [General Equivalency Diploma], some college, college or more). A standard self-reported screening tool was used to identify individuals age 18 and over and those with diabetes. For the general adult survey, 13400 invitations were sent in the first wave of data collection. Diabetes-qualified respondents were screened as part of a larger effort to identify several chronic illnesses. Once a person qualified for one condition they were routed to complete the survey for persons with that condition until target sample sizes for each condition were achieved. As such, no sample-wide qualification rate for adult diabetes is available. A second wave of 1400 invitations oversampled individuals with Spanish surnames or who lived in zip-code areas with disproportionate numbers of African-Americans and/or Hispanics. An online survey research firm, E-valuations, Inc, sent invitations and collected data for both waves, using the sampling design and surveys developed by the Foundation for Accountability and the Robert Wood Johnson Foundation. Each respondent was given a unique 5-digit access code to ensure that the survey was taken only once. Those who completed it were entered into a drawing for a $250 cash prize. No reminder e-mails were sent, nor were nonworking or dormant e-mail addresses tracked.

Telephone Surveys

Adults age 18 and over constituted the sampling frame for the 2 telephone surveys. Wirthlin Associates, Inc identified individuals by means of traditional random telephone-survey methods, and used the sampling design and surveys developed by the Foundation for Accountability and the Robert Wood Johnson Foundation to conduct the surveys. Candidate telephone numbers were randomly selected and call attempts made until the target completed sample sizes of 400 for each survey was reached.

Measures

This study evaluates the Internet methodology for both the general-adult and adult-diabetes samples, using demographic variables and the following topics. Sources of survey items for each topic are provided in the reference associated with each of these topics:

  1. days lost because of poor physical health problems [34]
  2. self-assessed overall health status [35]
  3. health insurance status/affordability of care (Robert Wood Johnson Foundation, oral communication, in person and by telephone, 2000)
  4. presence of a regular personal doctor [34]
  5. utilization of health care services [36]
  6. smoking behavior [34]
  7. doctors advising smokers to quit [36]
  8. drinking behavior [34]
  9. routine retinal eye exams (diabetes sample only) [34].

We selected these variables based on the availability of external benchmarks and representation of a range of health and health care quality topics.

The psychometric reliability of the following survey scales constructed using several survey items was also assessed (these are the multi-item survey scales referred to below in the "Data Analysis" part of "Methods"). A reference for each multi-item survey scale is provided in the reference associated with each of these scales:

  1. getting medical care quickly [37]
  2. getting dental care quickly [37]
  3. shared decision making (diabetes only) [38]
  4. self-care education and support (diabetes only) [39].

Data Analysis

We calculated response rates for the online general adult survey as the ratio of the completed sample size to the number of e-mail invitations needed to achieve this sample. The response rate for the online adult-diabetes survey was the proportion of the people completing the survey who were positively identified as having diabetes. Neither rate accounts for nonworking or dormant e-mail addresses. Telephone response rates were the ratio of completed sample size to the number of randomly-selected, working residential phone numbers that had to be called to achieve this sample size.

Survey responses for adults with diabetes were weighted using diabetes-specific age and sex distributions from the 1999 Behavioral Risk Factor Surveillance Survey [34]. General adult survey responses were weighted for age, sex, educational level, and presence of a chronic condition using distributions from the 2000 National Health Interview Survey (Robert Wood Johnson Foundation, oral communication, in person and by telephone, 2000). These distributions were used in lieu of those available from the US Bureau of the Census through the Current Population Survey (CPS) because chronic-condition status was not available from the Current Population Survey [40].

We compared weighted results from the online and telephone surveys for variables listed above to available benchmarks using either the 1999 Behavioral Risk Factor Surveillance Survey or the 1998 National Health Interview Survey. For online, telephone, and benchmarking dataset samples, we used regression analysis methods to evaluate patterns of variation across population subgroups for selected health and health care quality variables. Dependent variables for the general adult sample included health insurance status, having a regular doctor or nurse, physician counseling to quit smoking (for smokers), and poor health days in the last month. Dependent variables for the adult-diabetes sample included receipt of a routine retinal exam, use of health care services, smoking behavior, and poor health days in the last month. Independent variables were age, sex, race, education, and income, plus health insurance status and having a regular doctor or nurse, except where health insurance or regular doctor or nurse was used as a dependent variable. We compared results across samples in terms of the overall explanatory value of independent variables using the Cox and Snell generalized coefficient of determination [41]. The direction, general magnitude, and significance of the effect of each explanatory variable were also compared across samples for each dependent variable.

Each of the 4 multi-item survey scales (see Measures, in Methods, above) were evaluated for psychometric reliability using standardized estimates of Cronbach alpha [41]. SPSS version 9.0 was used to conduct data analysis [42].


Results

Response Rates, Response Bias, and Representativeness

Of the approximately 13400 e-mail invitations sent for the online general adult population survey, 2324 individuals responded and completed at least 80% of the survey, resulting in a 17.3% raw response rate. Based on industry norms, we estimate that at least 10% to15% of e-mail addresses are nonworking or dormant. Assuming this, the true response rate for the online general adult survey is 19% to 20%. For the general adult population telephone survey, approximately 4300 working, residential phone numbers had to be dialed to achieve the target sample size of 400. This resulted in an estimated 9.3% response rate after adjusting for nonworking and nonresidential phone numbers. Completed survey samples for the online and telephone adult-diabetes surveys were 1048 and 397 respectively.


[view this table]
Table 1. Demographic indicators for Robert Wood Johnson Foundation indicator survey: general adult population

Table 1 and Table 2 summarize the demographic characteristics of the unweighted and weighted online and telephone survey samples. Overall, respondents to the online general adult survey match the distribution of the sampled population, with some underrepresentation of individuals age 18 to 24 and overrepresentation of individuals age 45 to 64 and individuals reporting more than a high school education. Both the unweighted online and telephone general-adult completed survey samples underrepresent nonwhite individuals, those with less than a high school education, and those with incomes over $75000. Compared to the Current Population Survey, both general adult samples overrepresent individuals with a college education (or more) and incomes of $15000 to $35000. The telephone general-adult sample was more likely to underrepresent those with less than a high school education and overrepresent those with a college education. Similar results were found in both adult-diabetes samples (Table 2). However, while the telephone diabetes survey sample dramatically underrepresented individuals under age 44, and overrepresented those over age 65 and with incomes over $75000, this was not the case for the online adult-diabetes survey sample. Neither the online nor telephone methods resulted in samples properly representing racial groups with diabetes.


[view this table]
Table 2. Demographic indicators for Robert Wood Johnson Foundation indicator survey: adult-diabetes population

Comparison to Other National Studies

Table 3 and Table 4 compare results from both the general-adult and adult-diabetes online surveys to those obtained from the telephone surveys and benchmark data reported in other national studies. For the general adult population, the weighted online-survey results are not significantly different from those derived from the Behavioral Risk Factor Surveillance Survey and the National Health Interview Survey on 7 of the 12 health statuses, access to care, utilization of care, and clinically-appropriate health and health care quality indicators, including: (1) presence of health insurance, (2) having a regular doctor or nurse, and (3) receipt of advice to quit smoking for smokers. For the sample of persons with diabetes, results from the online survey were not significantly different from the BRFSS or NHIS benchmarks on 7 of the 13 indictors used, including (1) self assessed health status, (2) presence of health insurance, (3) having a routine checkup, (4) getting a retinal eye exam at least once in the least year, (5) receipt of advice to quit smoking for smokers, and (6) routine retinal exams for diabetics.

In the general-adult population survey, we observed higher proportions of individuals reporting 7 or more poor health days, fair or poor health status, and smoking.

In addition to comparing point estimates produced by this online survey to those produced by national benchmark datasets, we also evaluated how these datasets compare in terms of identifying variations and disparities in the health and health care quality across demographic subgroups as well as according to characteristics such as health insurance status and presence of a regular doctor. Table 5 and Table 6 present results from logistic regression analyses conducted to evaluate patterns of variation observed using data collected online versus data collected by telephone and versus telephone-based national benchmark datasets (BRFSS and NHIS).

The independent variables included in this analysis had similar explanatory power for dependent variables from the general adult survey whether data were collected using online or telephone methods. Specifically, at the low end, the demographic and health care related independent variables explained 5% to13% of the variation observed in reports of days lost because of poor health for the national dataset sample (5%), the telephone sample (13%), and the general adult online sample (9%), respectively. At the high end, these variables explained 25% to 34% of variation observed in the presence of health insurance across all datasets. For the adult-diabetes samples, on the low end, the independent variables used here accounted for less than 5% of the variation observed in rates of high utilization of health care. On the high end, these variables accounted for 11% to17% of variation observed in rates of smoking for all 3 adult-diabetes samples compared.

Along with the overall explanatory value of independent variables, we observed consistency across the general adult population datasets in terms of the approximate magnitude and significance of effect of specific independent variables. Having a regular doctor and income showed the most consistent and statistically significant effects (P < .05). Age and educational level, meanwhile, were the most consistently significant for the dependent variables evaluated for the adult-diabetes samples. No instances were found in which a variable was significant in one sample and also significant in the opposite direction in another. We did find cases of a variable being significant in one sample, but not in another. In most cases, this is attributed to chance or smaller sample size.


[view this table]
Table 3. Comparing online scores to external benchmarks and telephone data on selected health care indicators: general adult population (adjusted to the benchmark by gender and age)


[view this table]
Table 4. Comparing online scores to external benchmarks and telephone data on selected health care indicators: adult-diabetes population (adjusted to the benchmark by gender and age)


[view this table]
Table 5. Logistic regression results for the general adult population—comparison of odds ratios estimated using online data versus 1999 BRFSS (Behavioral Risk Factor Surveillance Survey) data and telephone supplemental data (cell values are odds ratios calculated using logistic regression analysis methods [43])


[view this table]
Table 6. Logistic regression results for the adult diabetes population—comparison of odds ratios estimated using online data versus 2 external benchmarks, the 1999 BRFSS (Behavioral Risk Factor Surveillance Survey) or 1998 NHIS (National Health Interview Survey) data and telephone supplemental data (cell values are odds ratios calculated using logistic regression analysis methods [43])

Scale Reliability

Cronbach alpha internal consistency scores were .72 or above for each of the 4 multi-item scales observed here (.72-.95), demonstrating their psychometric reliability when online administration is used (Table 7).


[view this table]
Table 7. Cronbach alpha reliability scores for multi-item scales using online data


Discussion

This study found evidence that online health care surveys originally designed for mail or telephone administration maintained both psychometric reliability and concurrent validity in results across demographic and other subgroups. More specifically, estimates of access to care, utilization of care, application of clinically appropriate care, and consumer experiences of care were similar to those derived from more traditional methods of obtaining representative samples of the US population.

We were able to achieve a sample representative of the US population in terms of age, sex, and education using a readily-available, opt-in sampling frame that employs relatively low-cost recruitment methods. Basic statistical weighting methods further aligned the responding population sample on these variables. Prior information on the affiliation of individuals included in the Web panel prevented stratified sampling based on race or income. Consequently, we cannot determine whether differences between our completed survey sample and the US population in the proportion of persons representing each racial and income group are due to response biases or inadequate representation in the original sampling frame for this study. Since a great deal of concern focuses on health care for lower-income individuals, it is important keep in mind that this group was, in fact, overrepresented when compared to the US population.

Given the importance of equitably representing the range of racial and economic groups, Web-based panels used for public information about health care quality should strive to include these variables so that stratified sampling may occur and/or assessments of response bias can take place. Here, oversampling methods often used in other national studies were successful in attenuating potential biases in results caused by lower rates of representation among nonwhite racial groups.

Response rates for Internet-based, telephone, or mailed surveys must be calculated in comparable ways and take into account differences in follow-up steps with nonrespondents. In this study, while analogous administration steps were used for both the online and telephone surveys, more-robust follow-up strategy was used for the telephone survey (6 follow-up calls for telephone and no follow-up steps for online survey). In spite of this, the online response rate was higher than for telephone when comparable calculations were used. This finding is true even when nonworking and nonresidential numbers are removed from the telephone sample and similarly nonworking or dormant e-mail addresses are not removed from the online sample. Given the unique sampling and administration processes employed for both surveys, these findings may not be observed in cases where relatively-simple online methods are compared to more-complex and more-costly sampling and administration methods typical of national studies such as the National Health Interview Survey and the National Medical Expenditures Panel Survey. An important question to examine further is whether such extensive follow-up methods are required to generate public information about health care quality and whether Internet-based methods outlined here may be suitable, especially as Web access continues to expand for all population groups.

Overall, findings from this study demonstrate that many of the sampling and survey administration challenges inherent in telephone and mail modes of data collection are also present for Internet-based methods. In turn, the survey administration, statistical sampling, and weighting approaches used to ensure that data collected via telephone or through mailed surveys yield adequate and representative samples, are also required for data collected via the Internet.

Internet-based data collection is appealing in its potential for allowing information to be collected in a timely and efficient manner. These efficiencies are eroded, however, if costly strategies are required to recruit panels from which sampling may occur and/or when the survey administration process includes extensive nonresponder follow-up and tracking steps. The methods used in this study were selected to be low burden in terms of the sampling frame and administration. This was done in order to begin to explore whether the benefit of obtaining data in a timely and potentially-interactive manner using the Internet can be achieved without incurring costs that diminish the value of doing so when compared to traditional telephone methods used by most nationally-recognized studies.

As these and other issues regarding the use of the Internet to conduct health and health care quality surveys are evaluated, it is worth recalling that our comfort with telephone surveys dates only from the late 1970s, when relatively-sophisticated methodologies were established involving random-digit dialing and multiple contact strategies [24]. In fact, the rise of the telephone survey in the late 1960s and early 1970s was attended by similar methodological concerns as those now associated with Web surveys - and took a decade of research and refinement to resolve. In recent years, the growing use of unlisted numbers, cell phones, call waiting, caller identification, and answering machines have induced a steady decline in response rates and growing disparities in the populations willing to be contacted by telephone. For example, Gallagher et al found that only an elaborate and expensive combination of mail, phone, and door-to-door solicitations produced a respondent pool fully representative of the low-income community [44]. As a result, Dillman has argued that only self-administered surveys - whether made available by mail, interactive voice response, or the Internet - are likely to be successful in the coming years [19].

Results of these analyses suggest that weighted online sampling offers an imperfect but promising avenue for collecting large-scale representative survey data. Overall, conclusions about the level and variations in health care quality in the United States are similar whether based on data collected online or data collected using more elaborate and costly survey methods.

All forms of survey-based data collection involve certain sampling and mode effect biases. Tradeoffs in the biases entailed in online versus telephone based surveys need to be carefully considered by policymakers. As Internet access increases along with the propensity for individuals to resist telephone solicitations, online survey methods may increasingly represent an efficient, real-time alternative for assessing health and health care quality in the United States.


Acknowledgments

This work has been funded by The Robert Wood Johnson Foundation.


Conflicts of Interest

None declared.


References

  1. . AHRQ leading effort to develop national report on health care quality [press release]. 2000 Nov 3.   URL: http://www.ahrq.gov/news/press/pr2000/natreppr.htm [accessed 2002 Jul 26] [WebCite Cache]
  2. . Centers for Medicare and Medicaid Services. 2002.   URL: http://www.cms.gov/providers/nursinghomes/nhi/ [accessed 2002 Jul 26] [WebCite Cache]
  3. Hurtado MP, Swift EL, Corrigan JM. Envisioning the National Healthcare Quality Report. Washington, DC: National Academy Press; 2001.   URL: http://www.nap.edu/catalog/10073.html
  4. . Medicaid program, Medicaid managed care; final rules. Federal Register 2002 Jun 14;67(115):40988-41116.
  5. . State Children's Health Insurance Program (SCHIP).   URL: http://www.cms.hhs.gov/schip [accessed 2002 Jul 26] [WebCite Cache]
  6. Safran DG, Kosinski M, Tarlov AR, Rogers WH, Taira DH, Lieberman N, et al. The Primary Care Assessment Survey: tests of data quality and measurement performance. Med Care 1998 May;36(5):728-739. [Medline] [CrossRef]
  7. Cleary PD, Edgman-levitan S. Health care quality. Incorporating consumer perspectives. JAMA 1997 Nov 19;278(19):1608-1612. [Medline] [CrossRef]
  8. Fowler FJ, Cleary PD, Magaziner J, Patrick DL, Benjamin KL. Methodological issues in measuring patient-reported outcomes: the agenda of the Work Group on Outcomes Assessment. Med Care 1994 Jul;32(7 Suppl):JS65-JS76. [Medline]
  9. Wasson J, Keller A, Rubenstein L, Hays R, Nelson E, Johnson D. Benefits and obstacles of health status assessment in ambulatory settings. The clinician's point of view. The Dartmouth Primary Care COOP Project. Med Care 1992 May;30(5 Suppl):MS42-MS49. [Medline]
  10. Ware JE. Measuring patients' views: the optimum outcome measure. BMJ 1993 May 29;306(6890):1429-1430. [Medline]
  11. Epstein AM. Rolling down the runway: the challenges ahead for quality report cards. JAMA 1998 Jun 3;279(21):1691-1696. [Medline] [CrossRef]
  12. Bethell C, Read D, Hochheimer J. CAHPS 2.0H Child survey with children with special health care needs module inserted: HEDIS work-up. Presented to the National Committee on Quality Assurance, Washington, DC, by the Child and Adolescent Health Measurement Initiative, CAHMI 2000.
  13. Kwak N, Radler BT. Using the Web for public opinion research: a comparative analysis between data collected via mail and the Web. Presented at: Annual conference of American Association of Public Opinion Research; 2000; Portland, OR.
  14. Yun GW, Trumbo CW. Comparative response to a survey executed by post, e-mail, and Web form. J Computer-Mediated Comm 2000;6 [FREE Full text] [WebCite Cache]
  15. . Internet vs. telephone data collection: does method matter?. 2000.   URL: http://www.burke.com/whitepapers/PDF/B.WhitePaperVol2-2000-Iss5.pdf [accessed 2002 Jul 26] [WebCite Cache]
  16. . Compare your care: a Web-based strategy for consumer activation. Quality counts; final report to Robert Wood Johnson Foundation. Portland, OR: Foundation for Accountability; 2001.
  17. . Home page.   URL: http://www.nichq.org/ [accessed 2002 Jul 26] [WebCite Cache]
  18. Fischbacher C, Chappel D, Edwards R, Summerton N. Health surveys via the Internet: quick and dirty or rapid and robust? J R Soc Med 2000 Jul;93(7):356-359. [Medline]
  19. Dillman DA. Mail and other self-administered surveys in the 21st century: the beginning of a new era. 1998 Mar.   URL: http://survey.sesrc.wsu.edu/dillman/papers/svys21st.pdf [accessed 2003 Oct 15] [WebCite Cache]
  20. Dillman DA. Mail and Internet Surveys: The Tailored Design Method. New York: Wiley; Nov 19, 1999.
  21. Atrostic BK, Bates N, Burt G, Silberstein A, Franklin , Winters F. Nonresponse in federal household surveys: new measures and new insights. Presented at: International Conference on Survey Nonresponse; 1999 Oct; Portland, OR   URL: http://www.fcsm.gov/committees/ihsng/portland_3__120299.pdf [WebCite Cache]
  22. Dillman DA, Carley-Baxter LR. Structural determinants of mail survey response rates over a 12-year period, 1988-1999. In: Proceedings of the Survey Research Methods Section. Alexandria, VA: American Statistical Association; 2000.   URL: http://www.amstat.org/sections/srms/Proceedings/papers/2000_063.pdf
  23. . National Do-Not-Call Registry.   URL: https://www.donotcall.gov/ [accessed 2003 Oct 29] [WebCite Cache]
  24. Dillman DA, Phelps G, Tortora R, Swift K, Kohrell J, Berck J. Response rate and measurement differences in mixed mode surveys using mail, telephone, interactive voice response and the Internet [draft].   URL: http://survey.sesrc.wsu.edu/dillman/papers/Mixed%20Mode%20ppr%20_with%20Gallup_%20POQ.pdf%20 [accessed 2003 Oct 29] [WebCite Cache]
  25. Wiebe EF, Eyerman J, Loft J. Evaluating nonresponse in a Web-enabled survey on health and aging. Presented at: Annual conference of American Association for Public Opinion Research; 2001; Montreal, Canada.
  26. Bauman S, Jobity N, Airey J, Atak H. Invites, intros, and incentives: lessons from a Web survey. Presented at: Annual conference of American Association for Public Opinion Research; 2000; Portland, OR.
  27. . Falling through the net: toward digital inclusion. 2000 Oct.   URL: http://search.ntia.doc.gov/pdf/fttn00.pdf [WebCite Cache]
  28. Lebo H. Surveying the digital future. The UCLA Internet report 2001: surveying the digital future. Los Angeles, CA: UCLA Center for Communication Policy; Nov 2001.   URL: http://www.ccp.ucla.edu/pdf/UCLA-Internet-Report-2001.pdf
  29. Terhanian G, Bremer J, Smith R, Thomas R. Correcting data from online surveys for the effects of nonrandom selection and nonrandom assignment. In: Harris Interactive White Paper 2000.
  30. CyBulski KA, Sinclair M, Potter F, et al. Adjusting for nonresponse among Medicaid households that could not be located or were located but did not participate in the Minnesota Managed Care Survey. Presented at: American Statistical Association 161st Annual Meeting; 2001; Atlanta, GA.
  31. McCready WC. A national random sample comparison between CATI and Web TV. Presented at: Annual Conference of American Association of Public Opinion Research; 2000; Portland, OR.
  32. Crabtree S. Charting new ground: current approaches to the problems of Web-based polling. Gallup Management Journal. 2000 Jul 31.   URL: http://web.archive.org/web/20000818091337/http:/www.gallup.com/poll/managing/mr000731.asp [accessed 2004 Jan 15] [WebCite Cache]
  33. Bethell CD, Lansky D, Read D. Reporting quality information to consumers, FACCT — The Foundation for Accountability. In: Report to the Health Care Financing Administration 1997.
  34. . Behavioral risk factor surveillance system survey data. Atlanta, GA: National Center for Chronic Disease Prevention and Health Promotion, Behavioral Surveillance Branch; 1999.
  35. Ware JE, Kosinski M, Keller SD. SF-36 Physical & Mental Health Summary Scales: A User's Manual. Boston, MA: The Health Institute, New England Medical Center; 1994.
  36. . Questionnaires from the National Health Interview Survey. Hyattsville, MD: National Center for Health Statistics; 1998.
  37. . Fact Sheet. In: CAHPS® and the National CAHPS® Benchmarking Database (AHRQ Publication No. 00-PO47). Rockville, MD: Agency for Healthcare Research and Quality; Apr 2000.   URL: http://www.ahrq.gov/qual/cahpfact.pdf
  38. Tarlov AR, Ware JE, Greenfield S, Nelson EC, Perrin E, Zubkoff M. The Medical Outcomes Study. An application of methods for monitoring the results of medical care. JAMA 1989 Aug 18;262(7):925-930. [Medline] [CrossRef]
  39. . FACCT Diabetes Survey. Portland, OR: The Foundation for Accountability; 1999.
  40. . Current Population Survey.   URL: http://www.bls.gov/cps/home.htm [accessed 2002 May 22] [WebCite Cache]
  41. Cronbach LJ. Coefficient alpha and the internal structure of tests. Psychometrica 1951;16:297-334.
  42. . SPSS. Version 9.   URL: http://www.spss.com/ [accessed 2003 Dec 26] [WebCite Cache]
  43. Dawson-saunders B, Trapp RG. Basic and Clinical Biostatistics (Lange Medical Books). Norwalk, CT: Appleton & Lange; Jan 15, 1994.
  44. Gallagher PM, Fowler FJ, Stringfellow VL. The nature of nonresponse in a Medicaid survey: causes and consequences. Nonresponse in federal household surveys: new measures and new insights. Presented at: International Conference on Survey Nonresponse; 1999; Portland, OR   URL: http://www.jpsm.umd.edu/icsn/papers/Gallagher.htm [WebCite Cache]


Abbreviations

BRFSS: Behavioral Risk Factor Surveillance Survey
NHIS: National Health Interview Survey
US: United States



Edited by G. Eysenbach; submitted 12.05.03; peer-reviewed by M Murray; comments to author 24.06.03; revised version received 06.09.03; accepted 12.09.03; published 20.01.04

Please cite as:
Bethell C, Fiorillo J, Lansky D, Hendryx M, Knickman J
Online Consumer Surveys as a Methodology for Assessing the Quality of the United States Health Care System
J Med Internet Res 2004;6(1):e2
URL: http://www.jmir.org/2004/1/e2/
doi: 10.2196/jmir.6.1.e2
PMID: 15111268

Export Metadata:
END, compatible with Endnote
BibTeX, compatible with BibDesk, LaTeX
RIS, compatible with RefMan, Procite, Endnote, RefWorks

Add this article to your Mendeley library
Add this article to your CiteULike library



Copyright

© Christina Bethell, John Fiorillo, David Lansky, Michael Hendryx, James Knickman. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 20.1.2004. Except where otherwise noted, articles published in the Journal of Medical Internet Research are distributed under the terms of the Creative Commons Attribution License (http://www.creativecommons.org/licenses/by/2.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited, including full bibliographic details and the URL (see "please cite as" above), and this statement is included.