Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Advertisement

Citing this Article

Right click to copy or hit: ctrl+c (cmd+c on mac)

Published on 24.08.18 in Vol 20, No 8 (2018): August

Preprints (earlier versions) of this paper are available at http://preprints.jmir.org/preprint/10351, first published Mar 08, 2018.

This paper is in the following e-collection/theme issue:

    Original Paper

    Participant Recruitment and Retention in Remote eHealth Intervention Trials: Methods and Lessons Learned From a Large Randomized Controlled Trial of Two Web-Based Smoking Interventions

    1Fred Hutchinson Cancer Research Center, Seattle, WA, United States

    2Kaiser Permanente Washington Health Research Institute, Seattle, WA, United States

    3University of Washington, Seattle, WA, United States

    Corresponding Author:

    Noreen L Watson, PhD

    Fred Hutchinson Cancer Research Center

    1100 Fairview Avenue North, M3-B232

    Seattle, WA, 98109

    United States

    Phone: 1 2066672942

    Email: nlwatson@fredhutch.org


    ABSTRACT

    Background: Despite having many advantages, online eHealth trials are not without challenges—notably, participant recruitment, and outcome data retention. Moreover, publications from these trials rarely provide detailed information on the methods used for recruitment and retention or discuss implications of the methods for future studies.

    Objective: To address this need for empirical guidance regarding recruitment and outcome data retention planning, we aim to describe the methods and lessons learned from the recruitment and retention procedures used in a large randomized trial of 2 Web-based smoking cessation interventions.

    Methods: To ensure a demographically and geographically diverse participant sample, we used the recruitment strategies (1) traditional, (2) Web-based, and (3) online survey panel methods and adaptively modified each in response to recruitment success. At baseline, participants indicated how they heard about the study and answered demographic questions. To maximize trial retention at each of the 3-, 6-, and 12-month assessment points, 4 survey modalities (first Web, followed by phone, mail, and postcard) were sequentially timed over a 30-day period. Participants received US $25 for submitting their responses, regardless of modality, and received an additional US $10 bonus for completing the Web survey within 24h of electronic notification.

    Results: We randomized 2637 smokers in 16 months and achieved 88% retention at 12-months. Participants (79.26% female, 72.60% Caucasian) were recruited from all 50 states. The majority of participants were recruited through Facebook (49.43%), followed by the survey panel (20.85%), free internet sources (14.54%), traditional media (11.34%), and Google ads (3.84%). Descriptively, participant demographics varied by recruitment source. Of the completed follow-up surveys, most were completed by Web (92%). Retention rates did not vary by recruitment source.

    Conclusions: Continuous monitoring and refinement of multiple recruitment methods, particularly of online advertising campaigns, allowed us to maximize the effectiveness of recruitment strategies in recruiting a large, diverse sample of smokers. Likewise, offering multiple follow-up survey modalities in sequential order along with time-dependent bonus incentives enabled us to obtain outcome data from a very high level of enrolled participants for the duration of the trial protocol. These strategies may be similarly useful in other trials.

    Trial Registration: ClinicalTrials.gov NCT01812278; https://clinicaltrials.gov/ct2/show/NCT01812278 (Archived by WebCite at http://www.webcitation.org/71gy5GLvO)

    J Med Internet Res 2018;20(8):e10351

    doi:10.2196/10351

    KEYWORDS



    Introduction

    In recent years, access to the internet has climbed exponentially worldwide [1]. In parallel with this global trend, more eHealth interventions are being developed and tested in online trials. In fact, as of June 2018, searching “eHealth” in PubMed yields over 29,000 results. Considering the potential for high reach and opportunity to provide low-cost interventions, one of the prominent allures of eHealth interventions is the possibility for them to make population-level health impacts. Moreover, conducting eHealth intervention research online offers a multitude of advantages including automated data collection, high control over intervention content and format, low cost, maximizing external validity, and potential for rapid recruitment of large numbers of participants [2,3]. However, online eHealth trials are not without challenges—most notably, recruitment of research participants and outcome data retention.

    Although not unique to online trials [4,5], many online eHealth studies have difficulty obtaining adequate sample size [2,6-9] and recruiting a representative sample of their target population [10-12]. Such difficulties with participant recruitment can result in extended recruitment time, increased costs for recruitment, as well as inadequate statistical power if accrual targets are not met [13]. In fact, failure to meet accrual targets is the primary reason for premature trial termination [4,14,15].

    Another major methodological challenge faced by online eHealth trials is keeping participants engaged in the study protocol and achieving adequate rates of outcome data retention [2,9,16-19]. Low retention rates increase the risk of selection bias (particularly if there are differential retention rates between arms), threaten validity, and lead to loss of statistical power [5,17,20]. Moreover, using tobacco treatment studies as an example, methods commonly used to address issues of low retention (eg, imputing all missing outcome data as smoking) are also problematic in that they can lead to inflated type-I and type-II errors [21]. For these reasons, drawing conclusions from studies with low retention rates, even when using conventional methods of imputation [21], can be misleading.

    At present, best-practice standards do not exist for how to recruit or retain participants in remotely conducted eHealth intervention trials and most trials do not detail their recruitment and retention methods sufficiently so that others can learn from their successes or mistakes. Among the remotely conducted trials of eHealth interventions that report a more significant amount of detail, there is considerable variability in the amount and type of information provided on the trial recruitment and retention strategies [3,22-26]. As a result, much of the information is not generalizable, it is difficult to determine which methods are most effective under what circumstances, and little is known about how recruitment sources might affect participant characteristics or data retention. Greater transparency of these issues has, therefore, been strongly encouraged in the existing literature [6,11,14,27,28] as detailed reports could inform plans for recruitment and retention in future trials of eHealth interventions.

    Thus, the primary aim of this paper is to provide a detailed description of the recruitment and outcome data retention methods used in a successful online eHealth trial (WebQuit), including lessons learned that might be useful for future eHealth trials. The WebQuit trial compared the effectiveness of 2 online smoking cessation programs in a diverse sample of smokers (N=2637) recruited across the US; 88% of participants completed the one-year follow-up survey [29]. To inform future strategies, we also examine the effects of recruitment source on participant characteristics and outcome data retention as well as the effects of participant characteristics on data retention.


    Methods

    Overview of WebQuit Trial Design

    The WebQuit trial was conducted to compare the effectiveness of 2 Web-based interventions for smoking cessation among adult smokers [29]. The websites evaluated in the trial were (1) a Web intervention based on Acceptance and Commitment Therapy (WebQuit.org), and (2) the National Cancer Institute’s Smokefree.gov website, which is the most accessed cessation website worldwide. The interventions were designed to be stand-alone interventions. Thus, the trial involved minimal contact with study personnel and did not provide pharmacotherapy to study participants. In addition, to access to their randomly assigned program, participants in both arms could receive up to four messages per day (via text or email) designed to encourage engagement with their assigned website, unless they opted out. These messages were sent for the first 28 days after randomization. Participant follow-up data were collected at 3, 6, and 12 months after randomization.

    Target Population.

    Participants (N=2637) were adult smokers living in the US. The eligibility requirements for the study included: (1) ≥ 18 years old, (2) smoke ≥5 cigarettes per day for the last year, (3) desire to quit smoking within 30 days, (4) have access to high-speed internet and email, (5) not participating in other cessation interventions or treatment, (6) never having used Smokefree.gov, (7) never having participated in one of our previous studies, (8) have no other household member participating in the study, (9) willingness to be randomized to treatment, complete 3 outcome surveys, and to provide contact information for themselves and 2 relatives, (10) live in the US, and (11) the ability to read in English. To recruit a diverse sample with adequate representation of smokers identifying as racial and ethnic minorities, we aimed to recruit a sample comprised of at least 25% smokers identifying as a racial/ethnic minority (ie, smokers who do not identify as non-Hispanic Caucasian). The target sample size for the study, which was based on having 80% power to detect a two-tailed significant difference between quit rates estimated for the 2 arms from our pilot study and relapse rates [29], was met.

    Recruitment Strategy

    We recruited participants for 18 months, from March 2014 to August 2015. We implemented a multi-pronged recruitment strategy that encompassed traditional, Web-based, and online survey panel methods (detailed below). Recruitment methods were monitored on an ongoing basis and modified as needed based on recruitment success. Our recruitment strategy was based on methods used in the pilot trial of WebQuit [23,30]. Also, we consulted with the Dana-Farber/Harvard Cancer Center Health Communication Core (DF/HCC) to develop advertisements, a recruitment website, and logo that would be relevant, sensitive, and appealing to our target population and distinguishable from other tobacco cessation websites. The DF/HCC also provided ongoing consultation regarding online advertising strategies for Facebook and Google. To ensure that our data retention operations team would have the capacity to handle the procedures for collecting follow-up data (discussed below), we limited the number of participants able to enroll in the study per month. We were able to closely monitor the study flow and ensure data quality by putting restraints on how quickly we recruited participants. All sources of recruitment directed interested individuals to a study website. The purpose of the website was to establish the credibility of the study and communicate the purpose of the study, how to enroll, and what would be asked of participants. Individuals who consented to be screened for the study were directed to a short screening survey to determine basic eligibility criteria. Individuals deemed eligible by the screening survey were immediately sent an email directing them to an online informed consent.

    Traditional Recruitment Methods

    Traditional methods for recruitment included press releases as well as newspaper, radio, and television interviews with the principal investigator (JBB). Participants recruited through targeted mailed invitations to known smokers within a large regional health plan (approximately 10,000 letters mailed in batches from December 2014 to August 2015), hospitals, or word of mouth by friends, family, or health care providers were also considered as being recruited through traditional methods.

    Free Internet Methods

    Participants were recruited through several free Web-based sources including Craigslist, Reddit, visibility of our recruitment website on search engines, and information about the study seen or shared on local and national websites (eg, news websites, fredhutch.org, the Penny Hoarder, Twitter). For recruitment on Craigslist, 296 advertisements were made throughout the recruitment period, with most posts made during the first 10 months. Two advertisements were made per day until all predetermined areas across 18 states were posted in. Predetermined areas were selected based on smoking prevalence [31], economic status, and rural areas (Alaska, Arizona, Arkansas, Indiana, Kentucky, Louisiana, Michigan, Mississippi, Missouri, Nevada, North Dakota, Ohio, Oklahoma, South Carolina, South Dakota, Tennessee, Virginia, and West Virginia). Additional postings were made in 18 cities with high concentrations of smokers [32] (eg, Wilkes Barre-Scranton, PA; Birmingham, AL; Grand Rapids, MI) and in 24 cities with high percentages of Black and African American populations (eg, Miami Gardens, FL; Birmingham, AL; Baltimore, MD). Some areas received up to three posts over the recruitment period.

    In addition to a general study advertisement, a season-specific advertisement was placed in several cities around January 2015. Ads were placed in the “Community” tab under both the “General Community” and “Volunteers” subsections of all cities. An example of the wording used in the Craigslist ads can be found in Figure 1. For recruitment on Reddit, 9 posts with similar wording were made throughout the recruitment period in different Reddit subgroups (eg, SampleSize, stopsmoking, addiction).

    Facebook Advertisements

    In consultation with the DF/HCC, 17 Facebook ads were created that varied in the images and wording used (see Figure 2 for an example ad) to be relevant to our target population. Most images were of cigarettes, the Fred Hutchinson Cancer Research Center logo, or people. Images of people varied by gender, race/ethnicity, the age of the person/people in the photo, and in how many people were pictured (1 to 3). Wording varied around eight content categories: (1) health (“You heard your doctor. And you’re ready to quit smoking.”), (2) readiness to quit (“You’re ready to quit smoking and we’re ready to help!”), (3) relationships (“You cherish your time with your children. So you’re going to quit smoking.”), (4) freedom from cigarettes (“You’re ready to be free, You’re ready to quit smoking.”), (5) research (“Earn up to $105 to quit with us—free, online quit smoking study from the Fred Hutchinson Cancer Research Center!”), (6) financial (“You’re ready to spend money on what YOU want, not on cigarettes!”), (7) appearance (“You’re ready to look better, smell better—BE better! Are you ready to quit smoking?”), and (8) help/altruism (“You’re ready to quit smoking. We’re ready to share free skills and support to help you quit!”).

    Creating multiple ads allowed us to determine which were most effective in real time. For the first 16 days of the ad campaign, advertisements were run in two-day intervals to determine which permutations of images and wording were most successful.

    Afterward, the top 3 ads were run one at a time so that Facebook’s embedded algorithm could optimize ad performance. One of the most successful ads read, “You heard your doctor. And you’re ready to quit smoking. We have a great opportunity—earn up to $105!” Ads were turned on and off in response to which were yielding the highest rates of randomization into the trial at the lowest cost per randomization. To increase the likelihood of reaching our desired population, we set several targeting parameters including ages 18-65 or older, English speaking, US, and people who identified relevant interests (eg, cigarette, quit smoking, electronic cigarette). Facebook ads ran from March 2014 until April 2015. Per recommendations from the DF/HCC and Facebook, we continued to monitor the performance of the ad campaign as response drop-offs are common over time. Minor adjustments were made as needed to boost ad performance.

    Figure 1. Content used for general Craigslist advertisements.
    View this figure
    Figure 2. Example Facebook ad with image.
    View this figure
    Google Advertisements

    Google ads were created with the same wording as the Facebook ads. Twenty keywords determined by the study team and DF/HCC that were associated with the study theme (eg, “quit smoking,” “help to quit smoking,” “stop smoking,” “how do I stop smoking,” “tips to quit smoking”) were rotated through to optimize response. Similar to the strategy used for Craigslist postings, we set parameters to target areas based on high prevalence of smoking, economic status, and rural areas. However, in monitoring conversions from Google ads, it quickly became apparent that the ads were not performing as well as Facebook ads (ie, lower rates of randomization). As a result, and to maximize recruitment efforts and resources, we discontinued the Google ads after approximately two months (March-May 2014).

    Online Survey Panel

    To help boost recruitment of minority smokers, we utilized the online survey panel company Survey Sampling International (SSI) [33] beginning in April 2015. Such companies can send study announcements to members of their online panels who meet specific criteria. For this study, we requested that they recruit racial/ethnic minority smokers living in the US. The panels are comprised of verified individuals who have voluntarily signed up to participate in online surveys and research in exchange for incentives. SSI was able to identify members of their panel who were likely to be eligible for the study and targeted the audience based on behavioral and demographic characteristics. Respondents recruited from the survey panel answered a brief screener to ensure they were the target audience before being directed to the study website.

    In addition to utilizing SSI to boost enrollment of minority smokers, we also limited enrollment on non-minority smokers beginning in February 2015 because we would otherwise not have reached our goal of enrolling at least 25% minority smokers. Specifically, the screening survey was programmed with an algorithm that would randomly reject a set proportion of non-Hispanic Caucasian smokers that were otherwise eligible. For example, if the rejection rate was 30%, non-minority smokers were randomly assigned a number between 0 and 1. If that number were less than 0.30, they would be ineligible. To ensure we met our recruitment goal, the rate of rejection varied over time.

    All activities for participant recruitment were coordinated by our project manager in consultation with the study team. On average, the project manager spent an estimated hour per day on recruitment activities throughout the recruitment period, with more time spent at the beginning of the recruitment period. Primary recruitment-related activities included communicating with outside vendors (DF/HCC, SSI, media) and setting up contracts as necessary, reviewing enrollment reports and communicating the enrollment status to key stakeholders, making daily posts to Craigslist (as discussed above), monitoring and tweaking Google and Facebook ads, and communications with our data retention operations team.

    Enrollment

    All enrollment procedures occurred online by way of the study website noted above. Interested individuals completed a screening survey to assess eligibility criteria. Antifraud measures were also implemented to decrease the likelihood of fraudulent participation (eg, enrolling more than once, changing survey responses to become eligible). Specifically, these measures included reviewing internet protocol addresses for duplicates or non-US origin, CAPTCHA authentication, and review of survey logs for suspicious response times (ie, completing the screening assessment in <90 seconds or completing the baseline survey in <10 minutes). Study staff contacted individuals with suspicious responses to confirm their information. If their information could not be confirmed, they were not enrolled in the study [29]. Eligible individuals were sent an e-mail from the study inviting them to return to the study website to provide informed consent, complete a baseline survey, and provide contact information. Eligible individuals had 14 days to complete the online enrollment process. Automated reminder emails were sent to eligible individuals on days 5 and 11 if they had not returned to the website. On day 7, a personalized email from study staff was sent in case the emails from the study email address were sent to participants’ spam folder. After completing the baseline assessment, participants had 30 days to return to the study website for randomization. Up to three weekly email reminders were sent from the study email address to participants who had not returned to the website. Ultimately, 2637 participants were randomized into the trial.

    Measures

    In the initial screening survey, participants reported how they learned about the study by selecting from one of 13 response options, including an “other” category in which they were able to write out an answer. For this manuscript, response options were grouped into the 5 recruitment methods described above: (1) traditional, (2) free-internet, (3) Facebook ads, (4) Google ads, and (5) online survey panel. We were able to classify all but 9 participants into 1 of the recruitment categories. While we were able to confirm participants recruited through the survey panel, all other responses are self-reported, which is typical for research on recruitment methods [22,34,35]. The baseline assessment also included questions about demographic characteristics and validated self-report screening measures of the following mental health conditions: depression (Center for Epidemiologic Studies Depression scale) [36], generalized anxiety (Generalized Anxiety Disorder 7-item scale) [37], panic disorder (Autonomic Nervous System Questionnaire) [38], posttraumatic stress disorder (PTSD; PTSD Checklist) [39], and social anxiety disorder (Mini-Social Phobia Inventory) [40]. A geographic classification of participants was determined by linking the participants’ zip codes to Rural-Urban Commuting Area (RUCA) codes [41]. There are 10 primary classifications based on population density, urbanization, and daily commuting. Definitions of the 10 RUCA codes can be found on the US Department of Agriculture website [41]. Zip codes associated with RUCA codes 1-6 were classified as metropolitan/micropolitan while those associated with RUCA codes 7-10 were classified as small town/rural areas.

    Retention Strategy

    Follow-up data were collected at 3, 6, and 12 months after randomization. To maximize data retention at each follow-up assessment, participants had up to 30 days to complete the assessments. Four survey modalities were sequentially timed until the survey was completed. Sequentially timing survey modalities has been shown to improve response rates compared to offering multiple survey modalities in parallel [42]. The modalities were (1) Web, (2) telephone, (3) mailed survey, and (4) a postcard with selected outcomes.

    For each follow-up assessment, we utilized the following strategy until participants completed the survey. Two weeks before the Web-based follow-up survey was available participants were mailed a survey invitation with a US $2 preincentive. Participants were then sent up to three automated emails with a link to the Web version of the survey on days 0 (exactly 3, 6, or 12 months after randomization), 5, and 9. Afterward, participants had the opportunity to complete the survey via phone. Study staff called participants up to eight times, once per day on days 10-17. On day 18, if participants had not completed the survey, study staff mailed a paper version of the survey with a prestamped and addressed return envelope. If participants did not respond to any of the previous modalities by day 30, they were mailed a postcard that only inquired about primary outcomes and a few selected secondary outcomes.

    We incentivized participants with US $25 for completing a survey, regardless of modality. Additionally, to encourage timely responses, participants who completed the Web-based survey within 24 hours of any email received a US $10 bonus. Thus, participants received up to US $105 in incentives for completing follow-up surveys.

    Analyses

    Descriptives regarding participant demographics across recruitment sources are reported. To examine the association between participant characteristics and recruitment source and 12-month data retention, we used logistic regression models with a covariate for treatment arm and accounted for multiple comparisons by adjusting P values to control the false discovery rate [43]. To assess differences in data retention across recruitment sources, we conducted chi-square tests for total response rates at the 3-, 6-, and 12-month follow-up assessments.


    Results

    Participants

    We recruited and randomized 2637 smokers. Of these, 2628 could be classified into 1 of the 5 recruitment categories and were thus retained for these analyses. The mean age of the sample at baseline was 46.15 (SD 13.36) years, and most of the sample was female (2083/2628, 79.26%). A large proportion of participants identified as Caucasian (1908, 72.60%), 278 (10.58%) identified as Black or African American, and 442 (16.82%) identified as another race (Asian, Native American, Native Hawaiian, or more than one race). A total of 222 (8.44%) identified their ethnicity as Hispanic. The remaining demographic characteristics of the overall sample can be found in Table 1.

    Recruitment and Demographic Variation by Recruitment Source

    Most of the sample was recruited from Facebook (1299/2628, 49.43%), followed by the survey panel (548, 20.85%), free internet sources (382, 14.54%), traditional methods (298, 11.35%), and Google ads (101, 3.84%). Using these recruitment channels, we recruited participants from all 50 states (Figure 3).

    Most demographic characteristics of the participants varied across recruitment sources (Table 1). Facebook advertisements recruited the oldest smokers with a mean of 52.87 (SD 10.54) years of age, while the survey panel recruited the youngest with a mean of 35.67 (SD 9.81) years of age. Although most smokers from all recruitment sources were women, traditional recruitment sources yielded the highest percentage of males (85/298, 28.52%), while Facebook ads resulted in the lowest (198/1299, 15.24%). The highest proportion of Black and African American smokers were recruited by both Google ads (18/101, 17.82%) and the online survey panel (98/548, 17.88%). As in the targeted recruitment plan, the online survey panel recruited the highest proportion of smokers identifying as Hispanic ethnicity (111/548, 20.26%) and as a race other than Caucasian or Black/African American (196, 35.77%). Google ads recruited the highest proportion of smokers with a high school education or less (37/101, 36.63%) while traditional recruitment sources recruited the highest proportion of smokers with a bachelor’s degree or more (77/298, 25.84%). Free internet sources recruited the highest percentage of participants who identified as lesbian, gay, or bisexual (56/382, 14.66%); with the lowest percentage recruited from Facebook (84/1229, 6.83%). Regarding income, participants recruited from Google ads were most likely to report an income that is greater than US $20,000 (36/101, 35.64%) while participants recruited from the survey panel were most likely to report their income as higher than US $20,000 (432/548, 78.83%). Although most participants in the study lived in metropolitan or micropolitan areas, Facebook recruited the highest proportion of participants from small towns or rural areas (155/1299, 11.93%), whereas traditional methods recruited the lowest proportion from these areas (15/298, 5.03%). All recruitment sources yielded similar proportions of smokers who screened positive for one or more mental health conditions.

    Table 1. Demographic characteristics by recruitment source.
    View this table
    Figure 3. Geographic location of participants from all 50 states; each participant is represented by a single dot.
    View this figure

    In assessing baseline participant characteristics associated with data retention (see Table 2), only gender emerged as a significant predictor, with males being less likely to complete the 12-month follow-up assessment than females (81% versus 89%, respectively; OR=0.52, 95% CI 0.40-0.67, P<.001). Data retention rates did not differ by recruitment sources at any follow-up assessment (P>.05, data not shown).

    Costs for Participant Recruitment Advertisements

    Excluding costs for personnel, the total cost (US $) of recruitment was approximately $84,083.59, or $31.89 per randomized participant. These costs included $1,995.00 for press releases, $4,054.00 for costs associated with mailed letters (ie, postage, printing, mailing supplies, graphic design), $49,791.49 for Facebook ads, $3,506.00 for Google ads, and $7,644.00 for services provided by SSI. The cost per randomized participant for each recruitment source from highest to lowest was $40.51 for Facebook, $34.71 for Google, $20.30 for traditional sources, and $13.95 for the survey panel.

    Rates of Participant Recruitment

    Collapsing across all recruitment sources, we enrolled an average of 146 participants per month for the 18-month recruitment period. Recruitment sources varied in rates of participant recruitment (see Table 3).

    Outcome Data Retention

    Outcome data retention rates for all assessment points and modalities can be found in Table 4. Overall, data retention rates were 88.85%, 89.16%, and 88.17% for the 3-, 6-, and 12-month follow-ups, respectively.

    Collapsing across follow-up assessments and assessment modality, a total of 6995 follow-up assessments were completed. The majority of surveys (6386/6995, 91.29%) were completed online. Of the surveys completed online, most (4261/6386, 66.91%) were completed within 24 hours of an email, earning the US $10 bonus incentive. An additional 894 (14.04%) of online surveys were completed prior to any phone calls; the remaining 1231 (19.33%) were completed after phone calls that began on day 10 of the follow-up period. Of the surveys not completed online, (160/609, 26.27%) were completed by phone, 289 (47.45%) were completed via mailed paper versions, and 160 (26.27%) by postcard. In other words, paper surveys accounted for 289/6995 (4.13%) of all survey responses and the phone and postcard surveys each accounted for 160 (2.28%).

    Table 2. Baseline predictors of 12-month data retention.
    View this table
    Table 3. Rates of recruitment by source.
    View this table
    Table 4. Outcome retention rates at 3-, 6-, and 12-month follow-ups by survey modality.
    View this table

    Discussion

    Study Objectives

    Despite the growing popularity of remotely conducted eHealth trials and difficulties regarding participant recruitment and data retention, most studies do not provide detailed accounts or implications regarding such methodologies. As a result, researchers are left with little guidance when planning these methods for eHealth trials. We sought to add to the literature by explicating the recruitment and retention methods used in the WebQuit trial and discussing implications for future online eHealth intervention trials.

    Recruitment and Implications

    By implementing a flexible, multi-modal strategy, we recruited and randomized 2637 geographically and demographically diverse adult smokers across the US into a Web-based smoking cessation trial in 18 months. This strategy enabled us to: (1) reallocate resources to methods that were most effective (eg, discontinuing Google ads when Facebook ads were performing better), (2) choose which advertisements to use by monitoring advertisement performance (eg, comparing response rates to Facebook ads), and (3) implement alternative strategies as needed (eg, using an online survey panel to boost recruitment of racial/ethnic minorities).

    Unlike findings from our pilot trial [23], participant characteristics varied across recruitment sources. While some differences were intended (eg, a higher proportion of racial and ethnic minority smokers from the online survey panel), other differences were not expected (eg, a more significant proportion of sexual minority smokers from free internet sources; a greater proportion of Black and African American smokers from Google ads). Other studies (eHealth and otherwise) have also found sources of recruitment to be differentially associated with demographic characteristics [22,34,35,44-46]. However, not all studies found the same differences, which may be attributable to differences in recruitment sources used, target population, type of research, and more granular details of recruitment methods (eg, images and words used in advertisements). The variability in participants’ demographic characteristics by recruitment source has implications worth considering for future trials and suggests that restricting recruitment to a single recruitment source may limit the sample diversity and, therefore, the generalizability of trial findings.

    A further examination of the recruitment methods lends useful insights for future eHealth trials. For example, although previous studies, including the WebQuit pilot [23], have successfully recruited participants with Google ads [30,47,48], relative to Facebook ads with the same text, Google ads significantly underperformed in terms of the number of participants recruited into the present study, leading to our decision to stop implementing these ads early on. Others have also reported poorer performance of Google relative to Facebook ads [49]. Interestingly, however, Google ads recruited the highest proportions of smokers with lower education, income, and higher proportions of smokers who identified as Black or African American, even though only minimal targeting parameters were used to display some ads in areas based on smoking prevalence, economic status, and rural areas. This suggests that while non-targeted Facebook ads may recruit more substantial numbers of people in shorter amounts of time, Google ads may be more effective at recruiting specific subgroups of individuals with only minimal targeting. Future studies should compare these methods systematically to determine if unique targeting strategies can be used to make the platforms equally effective regarding the rate of recruitment and participant demographics.

    Nearly half of the participants in this study were recruited from Facebook. Participants recruited from Facebook were predominantly Caucasian (1116/1299, 85.91%), and female (1101, 84.76%)—characteristics also found in other studies of Facebook-recruited participants [10,12,50]. However, relative to our other sources, Facebook recruited the lowest proportion of young adults aged 18-24 years (24, 1.85%), the highest proportion of older participants aged 45-64 years (902, 69.43%) those aged greater than 65 years (144, 11.09%), and the largest proportion (155, 11.93%) of participants living in small towns or rural areas. Contrary to previous reports that Facebook (and other social media platforms) tends to recruit younger samples [11,50,51], these findings add further support [10,52] that Facebook can recruit older participants, even without targeting for specific age groups. In the context of smoking cessation research, this suggests that, without targeting younger age groups, Facebook ads may be more effective for recruiting an older demographic of smokers wanting to quit. It may also be that Facebook’s algorithm for optimizing ad performance detected that ads worked best among people aged 45-64, then displayed ads to this group more frequently. These findings also suggest that, even without unique targeting parameters, Facebook may effectively reach some hard-to-reach populations, such as the 19% of the US population living in rural areas [53]. Although not done for this study, research suggests that further targeting and adjusting Facebook ad campaigns may increase the likelihood of reaching the desired population [49,54]. For example, different targeting parameters, advertisement images, and wording may be used strategically to recruit highly-specific subsets of participants [27].

    Our original sources of recruitment were not recruiting enough racial/ethnic minority smokers, prompting decisions to use an additional source to recruit minority participants (ie, online survey panel) and program our recruitment website to limit enrollment of non-minority smokers. By implementing this combination of strategies, we met our goal of recruiting at least 25% racial/ethnic minority smokers. A closer look at the other characteristics of participants recruited by the survey panel indicates that they tended to be younger and more likely to have an income higher than US $20,000 relative to participants from other sources. This may be due to the nature of individuals who participate in online survey panels. However, since the only specifications we provided SSI were smokers who identified as a racial/ethnic minority, it is possible that providing additional specifications regarding whom we were seeking to recruit would have produced a different sample.

    Most participants in this trial were women. Although some studies have successfully recruited samples comprised entirely of males for eHealth research through targeted advertisement campaigns [27,49,54], it is quite common among eHealth intervention trials to have a greater proportion of women compared to men [3,9,24,25,30,55-58]. This is not surprising as women are more likely to utilize eHealth programs [59]. Future studies, particularly those in which specific demographic variables (eg, gender) are deemed essential, should carefully monitor enrollment of participants with selected demographics to help ensure the desired sample is recruited. Much as we limited enrollment of Caucasian participants in this study, future studies might consider limiting enrollment of participants with certain demographics and creating targeted advertising campaigns to recruit participants meeting specific demographic criteria (eg, males). In making decisions regarding which demographic variables to monitor closely and put enrollment limitations on, we encourage researchers to determine who their target population is because a sample’s representativeness is dependent on the characteristics of the target population. For example, characteristics of a representative sample would be different for each of the following target populations: the US adult population, the population of adult smokers, the population of adult smokers who are interested in quitting, and the population of adult smokers interested in quitting with an eHealth intervention.

    Overall, when selecting methods for recruitment, researchers should consider many factors including target population, cost, level of reach, targeting abilities, level of ongoing effort required, possible rate of recruitment, and demographic characteristics likely to be recruited from a particular source. For example, in the current study, although the cost per participant enrolled from traditional and free internet sources was relatively low, these methods had limited reach and targeting ability and relatively slow rates of recruitment. In contrast, advertisements through Facebook and Google had a much broader reach, greater ability to target certain populations, faster rates of recruitment, and required low levels of ongoing effort after an initial learning curve. In summary, we highly recommend not only implementing a multi-modal recruitment plan to increase sample diversity, but also monitoring enrollment of participants with characteristics deemed essential for the research question(s). Such practices will not only help ensure that the desired sample is recruited, but can also help researchers determine when alternative recruitment strategies should be implemented to obtain the desired sample.

    Retention Strategy and Implications

    Yielding an 88.17% 12-month outcome data retention rate, our sequential, multi-modal participant retention strategy was highly effective at obtaining participant follow-up data. We believe 2 key factors contributed to the high rates of data retention. The first is offering bonus incentives for participants who complete their survey early on. Offering a US $10 bonus to the base incentive resulted in 60.91% (4,261/6,995) of all surveys being completed within 24 hours of receiving 1 of the 3 emails for the Web-based survey. The second is sequentially offering different survey modalities [42], thereby offering multiple opportunities and alternative ways of completing the follow-up surveys. Although very few participants completed the surveys via phone, phone calls appear to have prompted many participants to complete their survey online. Interestingly, of the surveys not completed online, more were completed by mail (the third modality offered) than by phone (the second modality). Future research should empirically examine the possibility of reducing participant- and personnel-related costs associated with these methods (eg, reducing the number of phone calls; sending paper versions of the surveys earlier), and still achieve the high rates of data retention found in this study.

    Data retention was not related to recruitment source or participant characteristics measured, with the exception that men were less likely to provide outcome data than women. This gender difference in outcome data retention is consistent with some previous eHealth studies [26,60], though results are mixed [3,57,61]. Reasons for this discrepancy are unclear and are beyond the scope of the present study. Despite this difference, 81.10% of men were retained in the present study at 12-months, which is still quite high for remotely conducted eHealth intervention trials. Future research is needed to understand under what conditions men might be less likely to complete follow-up assessments in eHealth intervention trials and to empirically test strategies to improve their retention rates under those conditions.

    Limitations

    The key limitation of these findings is that the study was not designed to compare the effectiveness or cost-effectiveness of the recruitment or retention strategies. Thus, the findings here are meant to describe the methods we used to recruit and retain adult smokers in the WebQuit trial [29], as well as discuss possible methodological implications for future studies. The field would greatly benefit from empirical research designed to test the efficacy and necessity of different recruitment and retention strategies for remotely conducted eHealth trials as well as from more detailed reports regarding recruitment and retention methods to provide generalizable knowledge. Second, online recruitment methods are rapidly evolving. Thus, it is essential to keep in mind that recruitment for this study occurred from 2014-2015. Higher demand for and advancements in technology-based advertising campaigns may limit the generalizability of these findings in the market today. Differences in recruitment data from the WebQuit pilot study conducted in 2010 [23,30] as compared to the present study help emphasize this point. For example, in contrast to the present study, Google ads outperformed Facebook ads in our pilot study. This change may, in part, be due to updates in Facebook advertising options and their proprietary targeting algorithm between the two studies, which have been updated in many ways since this study was completed and can change without notice. Relatedly, our findings regarding Facebook recruitment may not generalize to future trials as user demographics and use trends evolve. Third, because we recruited adult smokers ready to quit smoking, the findings presented here may not generalize to other eHealth trials seeking to recruit other populations. Fourth, although we were able to validate participants recruited from the online panel, all other reports of recruitment source were self-reported, which is subject to recall bias. Finally, as discussed above, despite being diverse in many other ways, our sample was predominately female, which is consistent with a pattern in eHealth trials overall. While our sample may be representative of the population of smokers interested in quitting with an eHealth intervention, it may not be representative of the entire population of smokers.

    Conclusions

    Continuous monitoring and refinement of multiple recruitment methods, particularly of online advertising campaigns, was key to our success in recruiting a large, diverse sample of smokers from across the US. Relatedly, offering multiple follow-up survey modalities in sequential order along with time-dependent bonus incentives enabled us to retain most enrolled participants for the duration of the 12-month protocol. Our findings suggest that recruitment sources are associated with demographic differences among participants, but not with differential rates of outcome data retention. Based on the overall success of our participant recruitment and data retention efforts, our experience may serve as an example to others interested in conducting randomized, online clinical intervention trials.

    Acknowledgments

    We would like to thank the study participants, the invaluable involvement of our study staff at the Fred Hutchinson Cancer Research Center, and the staff at the Kaiser Permanente Washington Health Research Institute for their assistance in recruiting participants. We also thank the National Cancer Institute (R01CA166646 to JBB; R01CA192849 to JBB) and the National Institute on Drug Abuse (R01DA038411 to JBB). NLW’s work on the project was supported by a grant from NIDA (F32 DA041100-02). We would also like to thank the Dana-Farber/Harvard Cancer Center for the use of the Health Communication Core, which provided design services for the trial recruitment website. Dana-Farber/Harvard Cancer Center is supported in part by an NCI Cancer Center Support Grant (# NIH 5 P30 CA06516). Funders had no role in the trial conduct or interpretation of results.

    Conflicts of Interest

    In July 2016, JBB was a consultant to Glaxo Smith Kline, the makers of a nicotine replacement therapy. He now serves on the Scientific Advisory Board of Chrono Therapeutics, the makers of a nicotine replacement therapy device. JLH has received research support from Pfizer, the makers of a smoking cessation medication. Other authors have no declarations.

    References

    1. Poushter J. Pew Research Center.: Pew Research Center; 2016 Feb 22. Smartphone Ownership and Internet Usage Continues to Climb in Emerging Economies   URL: http:/​/www.​pewglobal.org/​2016/​02/​22/​smartphone-ownership-and-internet-usage-continues-to-climb-in-emerging-economies/​ [accessed 2018-08-16] [WebCite Cache]
    2. Murray E, Khadjesari Z, White IR, Kalaitzaki E, Godfrey C, McCambridge J, et al. Methodological challenges in online trials. J Med Internet Res 2009;11(2):e9 [FREE Full text] [CrossRef] [Medline]
    3. Anguera JA, Jordan JT, Castaneda D, Gazzaley A, Areán PA. Conducting a fully mobile and randomised clinical trial for depression: access, engagement and expense. BMJ Innov 2016 Jan;2(1):14-21 [FREE Full text] [CrossRef] [Medline]
    4. Kasenda B, von EE, You J, Blümle A, Tomonaga Y, Saccilotto R, et al. Prevalence, characteristics, and publication of discontinued randomized trials. JAMA 2014 Mar 12;311(10):1045-1051. [CrossRef] [Medline]
    5. Gul RB, Ali PA. Clinical trials: the challenge of recruitment and retention of participants. J Clin Nurs 2010 Jan;19(1-2):227-233. [CrossRef] [Medline]
    6. Danaher BG, Seeley JR. Methodological issues in research on web-based behavioral interventions. Ann Behav Med 2009 Aug;38(1):28-39 [FREE Full text] [CrossRef] [Medline]
    7. Haines-Saah RJ, Kelly MT, Oliffe JL, Bottorff JL. Picture Me Smokefree: a qualitative study using social media and digital photography to engage young adults in tobacco reduction and cessation. J Med Internet Res 2015;17(1):e27 [FREE Full text] [CrossRef] [Medline]
    8. Powell J, Newhouse N, Martin A, Jawad S, Yu L, Davoudianfar M, et al. A novel experience-based internet intervention for smoking cessation: feasibility randomised controlled trial. BMC Public Health 2016 Dec 11;16(1):1156 [FREE Full text] [CrossRef] [Medline]
    9. van Spijker BA, Werner-Seidler A, Batterham PJ, Mackinnon A, Calear AL, Gosling JA, et al. Effectiveness of a Web-Based Self-Help Program for Suicidal Thinking in an Australian Community Sample: Randomized Controlled Trial. J Med Internet Res 2018 Feb 14;20(2):e15 [FREE Full text] [CrossRef] [Medline]
    10. Carter-Harris L, Bartlett ER, Warrick A, Rawl S. Beyond Traditional Newspaper Advertisement: Leveraging Facebook-Targeted Advertisement to Recruit Long-Term Smokers for Research. J Med Internet Res 2016;18(6):e117 [FREE Full text] [CrossRef] [Medline]
    11. Topolovec-Vranic J, Natarajan K. The Use of Social Media in Recruitment for Medical Research Studies: A Scoping Review. J Med Internet Res 2016 Nov 7;18(11):e286 [FREE Full text] [CrossRef] [Medline]
    12. Thornton L, Batterham PJ, Fassnacht DB, Kay-Lambkin F, Calear AL, Hunt S. Recruiting for health, medical or psychosocial research using Facebook: Systematic review. Internet Interventions 2016 May;4:72-81. [CrossRef]
    13. Sully BGO, Julious SA, Nicholl J. A reinvestigation of recruitment to randomised, controlled, multicenter trials: a review of trials funded by two UK funding agencies. Trials 2013;14:166 [FREE Full text] [CrossRef] [Medline]
    14. Carlisle B, Kimmelman J, Ramsay T, MacKinnon N. Unsuccessful trial accrual and human subjects protections: an empirical analysis of recently closed trials. Clin Trials 2015 Feb;12(1):77-83 [FREE Full text] [CrossRef] [Medline]
    15. Pica N, Bourgeois F. Discontinuation and Nonpublication of Randomized Clinical Trials Conducted in Children. Pediatrics 2016 Sep;138(3):1-11 [FREE Full text] [CrossRef] [Medline]
    16. Arean PA, Hallgren KA, Jordan JT, Gazzaley A, Atkins DC, Heagerty PJ, et al. The Use and Effectiveness of Mobile Apps for Depression: Results From a Fully Remote Clinical Trial. J Med Internet Res 2016 Dec 20;18(12):e330 [FREE Full text] [CrossRef] [Medline]
    17. Eysenbach G. The law of attrition. J Med Internet Res 2005;7(1):e11 [FREE Full text] [CrossRef] [Medline]
    18. Melville KM, Casey LM, Kavanagh DJ. Dropout from Internet-based treatment for psychological disorders. Br J Clin Psychol 2010 Nov;49(Pt 4):455-471. [CrossRef] [Medline]
    19. Riper H, Kramer J, Smit F, Conijn B, Schippers G, Cuijpers P. Web-based self-help for problem drinkers: a pragmatic randomized trial. Addiction 2008 Feb;103(2):218-227. [CrossRef] [Medline]
    20. Fewtrell MS, Kennedy K, Singhal A, Martin RM, Ness A, Hadders-Algra M, et al. How much loss to follow-up is acceptable in long-term randomised trials and prospective studies? Arch Dis Child 2008 Jun;93(6):458-461. [CrossRef] [Medline]
    21. Hall SM, Delucchi KL, Velicer WF, Kahler CW, Ranger-Moore J, Hedeker D, et al. Statistical analysis of randomized trials in tobacco treatment: longitudinal designs with dichotomous outcome. Nicotine Tob Res 2001 Aug;3(3):193-202. [CrossRef] [Medline]
    22. Gordon JS, Akers L, Severson HH, Danaher BG, Boles SM. Successful participant recruitment strategies for an online smokeless tobacco cessation program. Nicotine Tob Res 2006 Dec;8 Suppl 1:S35-S41. [Medline]
    23. Sandy M, Butler A. Microbial iron acquisition: marine and terrestrial siderophores. Chem Rev 2009 Oct;109(10):4580-4595 [FREE Full text] [CrossRef] [Medline]
    24. Ritterband LM, Thorndike FP, Ingersoll KS, Lord HR, Gonder-Frederick L, Frederick C, et al. Effect of a Web-Based Cognitive Behavior Therapy for Insomnia Intervention With 1-Year Follow-up: A Randomized Clinical Trial. JAMA Psychiatry 2017 Jan 01;74(1):68-75. [CrossRef] [Medline]
    25. Squiers L, Brown D, Parvanta S, Dolina S, Kelly B, Dever J, et al. The SmokefreeTXT (SFTXT) Study: Web and Mobile Data Collection to Evaluate Smoking Cessation for Young Adults. JMIR Res Protoc 2016 Jun 27;5(2):e134 [FREE Full text] [CrossRef] [Medline]
    26. Geraghty AWA, Torres LD, Leykin Y, Pérez-Stable EJ, Muñoz RF. Understanding attrition from international Internet health interventions: a step towards global eHealth. Health Promot Int 2013 Sep;28(3):442-452 [FREE Full text] [CrossRef] [Medline]
    27. Choi SH, Mitchell J, Lipkus I. Lessons Learned From an Online Study with Dual-smoker Couples. Am J Health Behav 2017 Jan;41(1):61-66. [Medline]
    28. Lane TS, Armin J, Gordon JS. Online Recruitment Methods for Web-Based and Mobile Health Studies: A Review of the Literature. J Med Internet Res 2015;17(7):e183 [FREE Full text] [CrossRef] [Medline]
    29. Bricker JB, Mull KE, McClure JB, Watson NL, Heffner JL. Improving quit rates of web-delivered interventions for smoking cessation: full-scale randomized trial of WebQuit.org versus Smokefree.gov. Addiction 2017 Dec 13:914-923. [CrossRef] [Medline]
    30. Bricker J, Wyszynski C, Comstock B, Heffner JL. Pilot randomized controlled trial of web-based acceptance and commitment therapy for smoking cessation. Nicotine Tob Res 2013 Oct;15(10):1756-1764 [FREE Full text] [CrossRef] [Medline]
    31. Nguyen K, Marshall L, Hu S, Neff L. State-specific prevalence of current cigarette smoking and smokeless tobacco use among adults aged ≥ 18 years--United States, 2011-2013. Morbidity and Mortality Weekly Report 2015;64(19):532-536.
    32. The Daily Beast. 2011 Jan 03. 30 cities with worst smoking problems, from Memphis to Honolulu   URL: https:/​/www.​thedailybeast.com/​30-cities-with-worst-smoking-problems-from-memphis-to-honolulu?ref=scroll [accessed 2018-08-16] [WebCite Cache]
    33. Survey Sampling International.   URL: https://www.surveysampling.com/ [accessed 2018-06-22] [WebCite Cache]
    34. Brodar KE, Hall MG, Butler EN, Parada H, Stein-Seroussi A, Hanley S, et al. Recruiting Diverse Smokers: Enrollment Yields and Cost. Int J Environ Res Public Health 2016 Dec 16;13(12):1-9 [FREE Full text] [CrossRef] [Medline]
    35. Frandsen M, Thow M, Ferguson SG. The Effectiveness Of Social Media (Facebook) Compared With More Traditional Advertising Methods for Recruiting Eligible Participants To Health Research Studies: A Randomized, Controlled Clinical Trial. JMIR Res Protoc 2016;5(3):e161 [FREE Full text] [CrossRef] [Medline]
    36. Radloff LS. The CES-D Scale: A Self-Report Depression Scale for Research in the General Population. Applied Psychological Measurement 1977 Jun 01;1(3):385-401. [CrossRef]
    37. Spitzer RL, Kroenke K, Williams JBW, Löwe B. A brief measure for assessing generalized anxiety disorder: the GAD-7. Arch Intern Med 2006 May 22;166(10):1092-1097. [CrossRef] [Medline]
    38. Stein MB, Roy-Byrne PP, McQuaid JR, Laffaye C, Russo J, McCahill ME, et al. Development of a brief diagnostic screen for panic disorder in primary care. Psychosom Med 1999;61(3):359-364. [Medline]
    39. Lang AJ, Wilkins K, Roy-Byrne PP, Golinelli D, Chavira D, Sherbourne C, et al. Abbreviated PTSD Checklist (PCL) as a guide to clinical response. Gen Hosp Psychiatry 2012;34(4):332-338 [FREE Full text] [CrossRef] [Medline]
    40. Connor KM, Kobak KA, Churchill LE, Katzelnick D, Davidson JR. Mini-SPIN: A brief screening assessment for generalized social anxiety disorder. Depress Anxiety 2001;14(2):137-140. [Medline]
    41. United States Department of Agriculture. Rural-Urban Commuting Area Codes   URL: https://www.ers.usda.gov/data-products/rural-urban-commuting-area-codes.aspx [accessed 2018-06-22] [WebCite Cache]
    42. Dillman DA, Phelps G, Tortora R, Swift K, Kohrell J, Berck J, et al. Response rate and measurement differences in mixed-mode surveys using mail, telephone, interactive voice response (IVR) and the Internet. Social Science Research 2009 Mar;38(1):1-18. [CrossRef]
    43. Benjamini Y, Hochberg Y. Controlling the false discovery rate: A practicalpowerful approach to multiple testing. Journal of the Royal Statistics Society B 1995;57(1):289-300.
    44. Graham AL, Milner P, Saul JE, Pfaff L. Online advertising as a public health and recruitment tool: comparison of different media campaigns to increase demand for smoking cessation interventions. J Med Internet Res 2008;10(5):e50 [FREE Full text] [CrossRef] [Medline]
    45. Laws RA, Litterbach EV, Denney-Wilson EA, Russell CG, Taki S, Ong K, et al. A Comparison of Recruitment Methods for an mHealth Intervention Targeting Mothers: Lessons from the Growing Healthy Program. J Med Internet Res 2016 Sep 15;18(9):e248 [FREE Full text] [CrossRef] [Medline]
    46. Lindner P, Nyström MB, Hassmén P, Andersson G, Carlbring P. Who seeks ICBT for depression and how do they get there? Effects of recruitment source on patient demographics and clinical characteristics. Internet Interventions 2015 May;2(2):221-225. [CrossRef]
    47. Morgan AJ, Jorm AF, Mackinnon AJ. Internet-based recruitment to a depression prevention intervention: lessons from the Mood Memos study. J Med Internet Res 2013;15(2):e31 [FREE Full text] [CrossRef] [Medline]
    48. Gross MS, Liu NH, Contreras O, Muñoz RF, Leykin Y. Using Google AdWords for international multilingual recruitment to health research websites. J Med Internet Res 2014 Jan 20;16(1):e18 [FREE Full text] [CrossRef] [Medline]
    49. Prescott TL, Phillips IG, DuBois LZ, Bull SS, Mustanski B, Ybarra ML. Reaching Adolescent Gay, Bisexual, and Queer Men Online: Development and Refinement of a National Recruitment Strategy. J Med Internet Res 2016;18(8):e200 [FREE Full text] [CrossRef] [Medline]
    50. Whitaker C, Stevelink S, Fear N. The Use of Facebook in Recruiting Participants for Health Research Purposes: A Systematic Review. J Med Internet Res 2017 Aug 28;19(8):e290 [FREE Full text] [CrossRef] [Medline]
    51. Guillory J, Wiant KF, Farrelly M, Fiacco L, Alam I, Hoffman L, et al. Recruiting Hard-to-Reach Populations for Survey Research: Using Facebook and Instagram Advertisements and In-Person Intercept in LGBT Bars and Nightclubs to Recruit LGBT Young Adults. J Med Internet Res 2018 Jun 18;20(6):e197 [FREE Full text] [CrossRef] [Medline]
    52. Cowie JM, Gurney ME. The Use of Facebook Advertising to Recruit Healthy Elderly People for a Clinical Trial: Baseline Metrics. JMIR Res Protoc 2018 Jan 24;7(1):e20 [FREE Full text] [CrossRef] [Medline]
    53. United States Census Bureau. 2016 Dec 08. New Census Data Show Differences Between Urban and Rural Populations   URL: https://www.census.gov/newsroom/press-releases/2016/cb16-210.html [accessed 2018-08-16] [WebCite Cache]
    54. Reiter PL, Katz ML, Bauermeister JA, Shoben AB, Paskett ED, McRee A. Recruiting Young Gay and Bisexual Men for a Human Papillomavirus Vaccination Intervention Through Social Media: The Effects of Advertisement Content. JMIR Public Health Surveill 2017 Jun 02;3(2):e33 [FREE Full text] [CrossRef] [Medline]
    55. Abroms LC, Boal AL, Simmens SJ, Mendel JA, Windsor RA. A randomized trial of Text2Quit: a text messaging program for smoking cessation. Am J Prev Med 2014 Sep;47(3):242-250. [CrossRef] [Medline]
    56. Bricker JB, Copeland W, Mull KE, Zeng EY, Watson NL, Akioka KJ, et al. Single-arm trial of the second version of an acceptance & commitment therapy smartphone application for smoking cessation. Drug Alcohol Depend 2016 Nov 04;170:37-42. [CrossRef] [Medline]
    57. Christoforou M, Sáez FJA, Tsakanikos E. Two Novel Cognitive Behavioral Therapy-Based Mobile Apps for Agoraphobia: Randomized Controlled Trial. J Med Internet Res 2017 Nov 24;19(11):e398 [FREE Full text] [CrossRef] [Medline]
    58. Cha S, Erar B, Niaura RS, Graham AL. Baseline Characteristics and Generalizability of Participants in an Internet Smoking Cessation Randomized Trial. Ann Behav Med 2016 Oct;50(5):751-761. [CrossRef] [Medline]
    59. Kontos E, Blake KD, Chou WS, Prestin A. Predictors of eHealth usage: insights on the digital divide from the Health Information National Trends Survey 2012. J Med Internet Res 2014;16(7):e172 [FREE Full text] [CrossRef] [Medline]
    60. Murray E, White IR, Varagunam M, Godfrey C, Khadjesari Z, McCambridge J. Attrition revisited: adherence and retention in a web-based alcohol trial. J Med Internet Res 2013;15(8):e162 [FREE Full text] [CrossRef] [Medline]
    61. Jahangiry L, Shojaeizadeh D, Montazeri A, Najafi M, Mohammad K, Yaseri M. Adherence and Attrition in a Web-Based Lifestyle Intervention for People with Metabolic Syndrome. Iran J Public Health 2014;43(9):1248-1258 [FREE Full text] [Medline]


    Abbreviations

    DF/HCC: Dana-Farber/Harvard Cancer Center Health Communication Core
    LGB: lesbian, gay, or bisexual
    MHC: mental health condition
    RUCA: Rural-Urban Commuting Area
    SSI: Survey Sampling International


    Edited by G Eysenbach; submitted 08.03.18; peer-reviewed by C Short, N Chavannes; comments to author 29.03.18; revised version received 25.07.18; accepted 01.08.18; published 24.08.18

    ©Noreen L Watson, Kristin E Mull, Jaimee L Heffner, Jennifer B McClure, Jonathan B Bricker. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 24.08.2018.

    This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.