Published on 10.05.10 in Vol 12, No 2 (2010): Apr-Jun
Effects of a Financial Incentive on Health Researchers’ Response to an Online Survey: a Randomized Controlled Trial
Background: Nonresponse to questionnaires can affect the validity of surveys and introduce bias. Offering financial incentives can increase response rates to postal questionnaires, but the effect of financial incentives on response rates to online surveys is less clear.
Objective: As part of a survey, we aimed to test whether knowledge of a financial incentive would increase the response rate to an online questionnaire.
Methods: A randomized controlled trial of 485 UK-based principal investigators of publicly funded health services and population health research. Participants were contacted by email and invited to complete an online questionnaire via an embedded URL. Participants were randomly allocated to groups with either “knowledge of” or “no knowledge of” a financial incentive (£10 Amazon gift voucher) to be provided on completion of the survey. At the end of the study, gift vouchers were given to all participants who completed the questionnaire regardless of initial randomization status. Four reminder emails (sent from the same email address as the initial invitation) were sent out to nonrespondents at one, two, three, and four weeks; a fifth postal reminder was also undertaken. The primary outcome measure for the trial was the response rate one week after the second reminder. Response rate was also measured at the end of weeks one, two, three, four, and five, and after a postal reminder was sent.
Results: In total, 243 (50%) questionnaires were returned (232 completed, 11 in which participation was declined). One week after the second reminder, the response rate in the “knowledge” group was 27% (66/244) versus 20% (49/241) in the “no knowledge” group (χ21 = 3.0, P = .08). The odds ratio for responding among those with knowledge of an incentive was 1.45 (95% confidence interval [CI] 0.95 - 2.21). At the third reminder, participants in the “no knowledge” group were informed about the incentive, ending the randomized element of the study. However we continued to follow up all participants, and from reminder three onwards, no significant differences were observed in the response rates of the two groups.
Conclusions: Knowledge of a financial incentive did not significantly increase the response rate to an online questionnaire. Future surveys should consider including a randomized element to further test the utility of offering incentives of other types and amounts to participate in online questionnaires.
Trial Registration: ISRCTN59912797; http://www.controlled-trials.com/ISRCTN59912797 (Archived by WebCite at http://www.webcitation.org/5iPPLbT7s)
J Med Internet Res 2010;12(2):e13
Nonresponse to questionnaires can affect the validity of surveys and introduce bias.
The offer of financial incentives has been a widely used method to increase response rates to postal questionnaires. A Cochrane systematic review of 481 randomized controlled trials (RCTs) evaluating 110 different ways of increasing response rates to postal questionnaires in a wide range of populations found that odds of response can be doubled through the use of monetary incentives . Other factors shown to increase the odds of response included a topic of interest, pre notification, follow-up contact, unconditional incentives, shorter questionnaires, provision of a second copy of the questionnaire at follow-up, mention of an obligation to respond, and university sponsorship [ ].
However, this evidence base relates to postal questionnaires, and although a number of systematic reviews [, ] and meta-analyses [ ] have been conducted, the available evidence base relating to use of incentives in electronic questionnaires is less substantive. The Cochrane review included 32 RCTs that evaluated 27 different ways of increasing response rates to electronic questionnaires in a wide range of populations [ ]. Although the one included RCT that evaluated monetary incentives found no significant effect, a further six RCTs found that use of other financial incentives (such as Amazon gift vouchers) doubled the odds of response. Limited evidence from social and market research also suggests that the offer of some form of monetary or financial incentive can increase the odds of a person responding and completing a web survey [ ].
Theoretical frameworks have been used to explain the potential influence of incentives on response rates. Social exchange theory  proposes that the actions of individuals are influenced by the balance between the rewards they expect to obtain and the costs they perceive may occur as a consequence; this exchange paradigm has become a key concept in marketing [ ]. A systematic review of the design and conduct of questionnaire surveys suggests that making exchange theory operational (in order to maximize response)involves minimizing the physical, mental, emotional, and economic costs of response, maximizing the tangible and intangible rewards for response, and establishing trust that those rewards will be delivered [ ]. In contrast, Leverage-saliency theory [ ] proposes that a potential participant’s decision to respond to a survey is influenced by the importance placed on key factors such as interest in the topic, [ ] available time; the credibility of the research source, and the benefits (tangible or otherwise) the individual perceives will result from participation. The theory postulates that potential participants with a strong interest in the topic are more likely to respond; incentives can act as leverage for those potential participants for whom influencing factors (such as topic of interest) are deemed less important.
Our study was undertaken as part of a survey to assess what steps researchers in the fields of health service and population health within the United Kingdom are taking to disseminate the findings of their research. Addressing deficiencies in the dissemination and transfer of research-based knowledge into routine clinical practice is high on the policy agenda both in the United Kingdom [- ] and internationally [ ]. Research dissemination and knowledge transfer is also highly relevant to the United Kingdom applied health research community. The main funder, the National Institute for Health Research (NIHR), is seeking to maximize the impact of its £800 million investment in applied health research [ ]. The NIHR has expectations that researchers will work to ensure that research is made available, can be used to support decision making, and will ultimately improve the quality and delivery of health care.
The population of interest for this survey is university-based and has high levels of Internet and email access. In addition, the major public funders of public health and health services research in the United Kingdom operate electronic online submission processes and use email as the principal mode of communication with grant holders and applicants. Given this, we decided to adopt a Web-based survey approach as it represented the most efficient and low cost mode of delivery.
However, there is some evidence that Web-based surveys can result in lower response rates (around 10%) compared with other survey modes [, , ]. Because of this, we decided to offer an incentive (gift vouchers from the online retailer Amazon) to participants to respond. Although a variety of incentives to increase response rates have been tested in a wide range of professional populations (including nine previous studies involving faculty members at universities [ ]), to our knowledge there is no evidence based on a randomized trial relating to our specific population of interest. In addition, the Cochrane review included three randomized evaluations of Amazon gift vouchers that showed mixed effects [ ]. Given this, we decided to test—using a randomized controlled trial nested within a survey—whether knowledge of a financial incentive would increase the response rate to the online questionnaire.
In July 2008, after obtaining ethical approval for the study from the University of York IRISS Ethics Committee, we contacted 10 UK programs and agencies that fund health services and public health research. The agencies were invited to provide (secure and encrypted) email contact details for UK-based principal investigators of health services and public health research completed in the last five years (2003-2008). Five agencies (the Scottish Chief Scientist Office, Economic and Social Research Council, Medical Research Council, NIHR Health Technology Assessment Programme and Wellcome Trust) responded and provided details. Principal investigator details for one non responding agency (NIHR Service Delivery and Organisation Programme) were publicly available and were obtained from their website. Two agencies (British Heart Foundation and Joseph Rowntree Foundation) indicated that they fund very little public health and health services research and so were excluded from the survey. The Department of Health Policy Research Programme and Cancer Research UK responded stating that they were unable to provide details of principal investigators.
We identified 743 principal investigators from the six funding agencies. Duplicates were removed from the list resulting in a total survey sample of 536 potential participants. Email addresses for identified principal investigators were then checked and compiled.
Study design and randomisation
Potential participants were randomized to receive either “knowledge of” or “no knowledge of” a financial incentive—in this instance gift vouchers (from the online retailer Amazon) to the value of £10. Amazon gift vouchers (distributed via the Amazon email gift certificate facility) were sent to all participants who completed the questionnaire regardless of the study group to which they were randomized.
Random allocation of participants using computer-generated numbers was undertaken independently by a statistician at the Medical Research Council (MRC) General Practice Research Framework.
On October 13, 2008, both groups were contacted by email (). Participants were told the purpose of the study and invited to complete an online questionnaire via an embedded URL. The online questionnaire was hosted on the SurveyMonkey website [ ] and was based on an instrument previously used to assess the practices of intramural MRC Research Units in an earlier phase of the project. The questionnaire comprised a combination of 36 open and closed questions that could be completed in 20 to 30 minutes. The questionnaire was piloted prior to use.
The email sent to the participants in the “knowledge” group stated that those who completed the online questionnaire would receive a £10 Amazon gift voucher. The study design specified that four reminder emails (sent from the same email address as the initial invitation) would be sent out to nonrespondents at one, two, three, and four weeks following the initial invitation; a fifth postal reminder would be sent to nonrespondents if the response rate was considered to be low. Participants who completed the online questionnaire were deemed to have given their consent. Questionnaires not returned by December 31, 2008, were deemed to be nonresponses.
As this RCT was nested within a larger survey, the primary concern was to maximize response rates. Given this, it was determined thatif the difference in the response rate between the two groups was such that it was likely to adversely affect the main aims of the survey, then knowledge of the incentive would be provided to the “no knowledge” group, but not before the third reminder. At the third reminder, we provided “knowledge of” the incentive to the “no knowledge” group to limit any adverse effects on total response to the survey.
A combination of IP address and questionnaire responses were used to identify multiple responses from a single participant . Where multiple responses from a single participant occurred, the most recently completed questionnaire was retained for analysis. Noninvited responses from individuals not part of the study sample were excluded from the analysis.
The primary outcome measure for the trial was rate of response one week after the second reminder. Rate of response was also measured at the end of weeks one, two, three, four, and five, and after the postal reminder. Data were entered and analysed in SPSS version 15.0 (SPSS Inc, Chicago, IL, USA). We compared the response rates in each group using the chi-square statistic.
Of the 536 identified email addresses, 51 were undeliverable resulting in a sample of 485. A total 243 (50%) questionnaires were returned (232 completed; 11 in which participation was declined).illustrates the flow of responses to the study.
As a measure of completion , 100% of the 232 participants who completed questionnaires answered the questions on the first page, and 95% (220/232) answered the final question. Excluded from the analyses were 4 questionnaires completed by noninvited individuals. Multiple responses were submitted by 2 participants; the most recently submitted questionnaire was included in the analyses in each case.
shows the cumulative response rate over time by group. shows the cumulative percentage response over time, again by group. The primary outcome measure for the trial was rate of response one week after the second reminder. The cumulative response rate in the ”knowledge” group was 27% (66/244) versus 20% (49/241) in the “no knowledge” group. This difference was not statistically significant (χ21=3.0, P=.08). The odds ratio for those with knowledge of an incentive that responded was 1.45 (95% confidence interval [CI] 0.95 - 2.21).
|Knowledge Group (n = 244)||No Knowledge Groupa (n = 241)||χ2 Significance|
|First response||31 (13%)||15 (6%)||P=.01|
|Reminder 1||49 (20%)||36 (15%)||P=.13|
|Reminder 2||66 (27%)||49 (20%)||P=.08|
|Reminder 3||81 (33%)||61 (25%)||P=.06|
|Reminder 4||94 (38%)||77 (32%)||P=.13|
|Postal Reminder||122 (50%)||110 (46%)||P=.33|
aNo knowledge group informed about incentive from reminder 3 onwards
At the third reminder, participants in the “no knowledge” group were informed about the incentive, ending the randomized trial nested within the survey. As this was a survey, we continued to follow up all respondents, and for transparency purposes,presents further data on the cumulative response rates. No significant differences were observed in response rates between the two groups from reminder 3 onwards.
Statement of principal findings
Knowledge of a financial incentive did not significantly improve the response rate to this online questionnaire. However, one week after the second reminder—the point before the “no knowledge” group were informed about the incentive—a difference of 7% was apparent.
Comparison with other studies
In terms of overall response, our rate of 50% compares favourably with those reported for other Web surveys. For example, in a review of comparisons of Web survey versus other survey modes, only 6 of 45 Web surveys managed to obtain a response rate higher than 50% . In a second meta-analysis, which reported an 11% difference in response rates in favor of postal over Web modes, only 10 of 39 comparisons obtained a Web survey response rate higher than 50% [ ]. Previous randomized evaluations of our choice of incentive (an Amazon gift voucher) [ ] have shown mixed effects in different populations and settings. However, of three previous studies similar to ours included in the Cochrane review, researchers compared the effects of: a $5 cash incentive versus a $5 gift voucher [ ]; no incentive versus entry into a lottery for $50, $100, $150, or $200 gift vouchers [ ]; and unconditional $15 or $25 gift vouchers versus $15 or $25 gift vouchers conditional on completion of the survey [ ].
Strengths and limitations of study
In developing our survey, we adhered to recommendations for the design of email questionnaires . These included deriving an appropriate sample, using an embedded URL, using incentives, and sending the request for information from a recognized academic source. One recommendation beyond our control was that the research be perceived to be relevant to the population surveyed. As stated above, there is renewed emphasis on increasing the uptake and transfer of publicly funded research into policy and practice, and those responding indicated that dissemination of the results of research was highly relevant to their work. However, we had no way of knowing beforehand whether the topic or goal would be deemed relevant or of interest by those we surveyed.
In our study, we utilized a 36-item questionnaire and stated that it would take participants up to 30 minutes to complete. Shorter postal questionnaires are associated with increased response rates . It may be that the perceived return (£10) for time invested in completing the 36 items was deemed inadequate compensation by some participants, especially if considered in relation to their incomes as professional researchers. We do not know whether an increase in the financial incentive relative to participant income would have made any difference in this instance. Another consideration relates to the nature of the incentive offered. Receipt of the gift voucher was dependent on the participant completing the questionnaire. There is evidence that response rates can be higher when an incentive is given up front unconditionally rather than given conditional on completion [ ]. The use of unconditional versus conditional incentives merits further investigation.
In this study, members of the population of interest have high levels of Internet and email access. Yet, around a fifth of all returned completed questionnaires were paper copies that had been mailed out as part of the postal reminder. This decision to adopt a mixed mode approach in the event of a low response rate appears sensible in light of feedback from two of the respondents. They indicated that they found it hard to find the time to respond to Web surveys, and as they were often out of the office, it was easier to complete a survey that used a paper-and-pen format. Although we recognize that our experience relates to a very specific population and suggest some caution in generalizing these findings to other populations, designers of future Web surveys may wish to consider using this mixed mode approach.
This randomized study was undertaken as part of a wider survey to assess what steps public health and health services researchers working across the United Kingdom are taking to disseminate the findings of their research. This nesting approach offered a cheap and efficient method of adding to our knowledge of the utility of different survey modes. However, undertaking such an approach was not without potential challenges. Normally in randomized studies, one would compare an intervention against standard practice when the outcome is unknown. But in this instance our primary concern was to maximize response rates to the wider survey. In doing so it was possible we limited the duration of the intervention making it difficult to determine what the true effect of the incentive would have been over a longer time period. Future web surveys should consider nesting a randomized element to further test the utility of incentives but should also consider whether the time frame for response is adequate to determine the true effect.
Our trial can help researchers planning future Web-based surveys. It would appear that immediate responses within two weeks of initial contact to a Web-based survey might be improved by the offer of a small financial incentive. Hence, we would recommend small financial incentives to those researchers requiring quick responses to Web-based questionnaires. However our findings suggest that this effect may dissipate over time. Researchers should consider that even in specific populations with high levels of access to the Internet, there might be advantages in using mixed methods (ie, use of both web and paper questionnaires) in terms of participant preferences and in increasing response rates.
This research was funded by MRC Population Health Sciences Research Network.
Conflicts of Interest
All authors contributed to the conception, design and analysis of the study. All authors were involved in the writing of the first and subsequent versions of the paper. The views expressed in this paper are those of the authors alone.
- Edwards PJ, Roberts I, Clarke MJ, DiGuiseppi C, Wentz R, Kwan I, et al. Methods to increase response to postal and electronic questionnaires. Cochrane Database Syst Rev 2009;3. [CrossRef] [Medline]
- VanGeest JB, Johnson TP, Welch VL. Methodologies for improving response rates in surveys of physicians: a systematic review. Eval Health Prof 2007 Dec;30(4):303-321. [Medline] [CrossRef]
- Göritz AS. Incentives in Web studies: Methodological issues and a review. International Journal of Internet Science 2008;1:58-70.
- Homans GC. Social Behavior as Exchange. Amer J Sociol 1958;63(6):597-606. [CrossRef]
- Bagozzi RP. Marketing as exchange. J Mark 1975;39:32-39.
- McColl E, Jacoby A, Thomas L, Soutter J, Bamford C, Steen N, et al. Design and use of questionnaires: a review of best practice applicable to surveys of health service staff and patients. Health Technol Assess 2001;5(31):1-256 [FREE Full text] [Medline]
- Groves RM, Singer E, Corning A. Leverage-saliency theory of survey participation: description and an illustration. Public Opin Q 2000;64(3):299-308. [Medline] [CrossRef]
- Groves RM, Presser S, Dipko S. The role of topic interest in survey participation decisions. Public Opin Q 2004;68(1):2-31. [CrossRef]
- Cooksey D. A Review of UK Health Research Funding. London, England: HM Treasury; 2006.
- ; Department of Health. Best Research for Best Health: A new national health research strategy. London, England: Department of Health; 2006.
- Tooke J. Report of the High Level Group on Clinical Effectiveness. A report to Sir Liam Donaldson Chief Medical Officer. London, England: Department of Health; 2007.
- ; World Health Organization. World Report on Knowledge for Better Health: Strengthening Health Systems. Geneva, Switzerland: World Health Organization; 2004.
- ; Department of Health, Research and Development Directorate. Delivering Health Research. National Institute for Health Research Progress Report 2008/09. London, England: Department of Health; 2009.
- Manfreda KL, Bosnjak M, Berzelak J, Haas I, Vehovar V, Berzelak N. Web surveys versus other survey modes. A Meta-Analysis Comparing Response Rates. International Journal of Market Research 2008;50:79-104.
- Shih TH, Fan X. Comparing response rates from web and mail surveys: A meta-analysis. Field methods 2008;20(3):249-271. [CrossRef]
- SurveyMonkey.com. URL: http://www.surveymonkey.com [WebCite Cache]
- Eysenbach G. Improving the quality of Web surveys: the Checklist for Reporting Results of Internet E-Surveys (CHERRIES). J Med Internet Res 2004 Sep 29;6(3):e34 [FREE Full text] [Medline] [CrossRef]
- Michaelidou N, Dibb S. Using e-mail questionnaires for consumer research: best practice in tackling non-response. Journal of Targeting, Measurement and Analysis for Marketing 2006;14(4):289-296. [CrossRef]
- Birnholtz JP, Horn DB, Finholt TA, Bae SJ. The effect of cash, electronic,and paper gift certificates asrespondent incentives for a web based survey of technologically sophisticated respondents. Soc Sci Comput Rev 2004;22:355-362. [CrossRef]
- Porter SR, Whitcomb ME. The impact of lottery incentives on student survey response rates. Research in Higher Education 2003;44:389-407.
- Downes-Le Guin T, Janowitz P, Stone R, Khorram S. Use of pre-incentives in an Internet survey. Journal of Online Research 2002. [CrossRef]
Edited by H Potts, G Eysenbach; submitted 21.04.09; peer-reviewed by M Dobrow, M Callegaro; comments to author 09.07.09; revised version received 02.11.09; accepted 01.12.09; published 10.05.10
©Paul M Wilson, Mark Petticrew, Mike Calnan, Irwin Nazareth. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 10.05.2010
This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.