Background: In recent years, there has been a proliferation of third-party Web-based services available to consumers to interpret raw DNA from direct-to-consumer genetic testing companies. Little is known about who uses these services and the downstream health implications. Identifying this hard-to-reach population of consumers for research raised questions about the most effective recruitment methods to undertake. Past studies have found that Web-based social media survey distribution can be cost-effective for targeting hard-to-reach populations, yet comparative efficacy information across platforms is limited.
Objective: The aim of this study was to identify the most effective Web-based strategies to identify and recruit the target population of direct-to-consumer genetic testing users who also made use of third-party interpretation services to analyze their raw genetic data. Web-based survey recruitment methods varying by social media platform and advertising method were compared in terms of cost-effectiveness and demographics of survey respondents.
Methods: A total of 5 Web-based survey distribution conditions were examined: 4 paid advertising services and 1 unpaid service. For the paid services, a 2x2 quasi-experimental design compared social media platforms (Facebook vs Twitter) and advertising tracking metrics (by click vs by conversion). The fifth unpaid comparison method consisted of study postings on the social media platform, Reddit, without any paid advertising. Links to identical Web-based versions of the study questionnaire were posted for 10 to 14 days for each of the distribution conditions, which allowed tracking the number of respondents that entered and completed the questionnaire by distribution condition.
Results: In total, 438 individuals were recruited to the study through all conditions. A nearly equivalent number of participants were recruited from paid campaigns on Facebook (n=159) and Twitter (n=167), with a smaller sample recruited on Reddit (n=112). Significantly more participants were recruited through conversion-tracking (n=222) than through click-tracking campaigns (n=104; Z=6.5, P<.001). Response rates were found to be partially driven by organic sharing of recruitment materials among social media users. Conversion tracking was more cost-effective than click tracking across paid social media platforms. Significant differences in terms of gender and age distributions were noted between the platforms and between the tracking metrics.
Conclusions: Web-based recruitment methods were effective at recruiting participants from a hard-to-reach population in a short time frame. There were significant differences in the effectiveness of various paid advertising techniques. Recruitment through Web-based communities also appeared to perform adequately, yet it may be limited by the number of users accessible in open community groups. Future research should evaluate the impact of organic sharing of recruitment materials because this appeared to play a substantial role in the observed effectiveness of different methods.
To date, there have been a number of inquiries into using social media for research recruitment and there has been little consensus in terms of results. A systematic review of 30 existing studies on social media recruitment found mixed evidence with regard to the efficacy of survey recruitment on social media but did find that such methods were consistently found to be effective when specifically targeting hard-to-reach populations—those that are difficult to find or involve in research and public health programs because of their geographical location or socioeconomic situation [, ]. However, the review also suggested that this methodology has not been studied often enough to generate firm conclusions as to its efficacy, arguing that further research, particularly research examining the cost-efficacy of different recruitment techniques and demographic differences in the resulting samples, is necessary [ ]. This study partially fills this gap in the research by directly comparing multiple analogous advertisement recruitment methods on Twitter and Facebook along with the unpromoted posts on Reddit to recruit survey respondents from the same hard-to-reach population.
Social media has been defined, in a public health context, as websites that allow users to create profiles and use those profiles to connect and interact with other users . Although there are dozens to hundreds of different forms of social media, at present, most of the documented social media recruitment efforts for population health research have used Facebook [ - ]. Facebook is a large social media platform with approximately 1.56 billion daily active users [ ]. Studies have found success in reaching target audiences by sharing posts within Facebook communities, enlisting respondents in snowball sampling campaigns, and purchasing paid advertising on the platform targeting specific demographics. Although there have been fewer studies focusing on the platform, Twitter—a somewhat smaller platform with 126 million daily active users [ ]—has also been used for recruitment purposes both through researchers tweeting and retweeting recruitment tweets [ , ] and advertisements [ ].
There has been little research to compare the effectiveness of recruitment from across different social media platforms, although some studies have sought to use multiple platforms for recruitment without making direct statistical comparisons [, ]. One study did compare 2 social media platforms (Twitter and Facebook) as well as another method (distributing quick response codes through mail) but did not use directly analogous recruitment methods across conditions or identify platform-level differences [ ]. As a result, direct comparisons of relative effectiveness between the platforms themselves remain a challenge.
Survey research employing social media for participant recruitment has also yet to consider the multiple recruitment strategies available on a given platform. Popular platforms, such as Facebook, enable both cost per click advertisement sales that charge advertisers each time a user clicks on an advertisement and cost per conversion sales: Advertisers are billed on the basis of specific, predefined actions that follow from a user clicking through the advertisement, such as purchasing a product or completing a questionnaire. These tracking metrics may yield different results when it comes to reaching target audiences as well as achieving a cost-effective survey sample.
This study sought to better understand the differences in survey participant recruitment between social media platforms, as well as within-platform differences resulting from different tracking methods. Targeted survey participants were a hard-to-reach population of users of direct-to-consumer genetic testing (DTC-GT) services (eg, AncestryDNA and 23andMe) who had subsequently used third-party interpretation tools to analyze their raw genetic data. The goal of the study was to compare the cost-effectiveness as well as the demographic characteristics of the sample across different platforms and between different advertising tracking metrics.
To enable a more rigorous comparison between different social media platforms, this study conducted advertising campaigns on both Facebook and Twitter—a platform deemed to possess sufficient similarities to Facebook in terms of advertising affordances and presentation of content so that comparisons can be made. In addition, click-based and conversion-based tracking metrics were used on each platform. To allow for further comparisons across social media platforms, an additional condition contrasted unpaid posts to community message boards on Reddit with the advertising campaigns on Facebook and Twitter.
This study addressed the following questions regarding platform differences, cost-efficacy, and paid versus unpaid uses of social media in survey recruitment:
Q1: Among paid social media campaigns, which social media platform is most effective at generating survey responses from the hard-to-reach population of DTC-GT users who had also used third-party interpretation tools?
Q2: Among paid social media campaigns, which advertising tracking method is most effective at generating survey responses from the hard-to-reach population of DTC-GT users who had also used third-party interpretation tools?
Q3: Do surveys conducted via paid social media campaigns on Facebook and Twitter generate more survey responses from the hard-to-reach population of DTC-GT users who had also used third-party interpretation tools compared with surveys posted on open (unpaid) Web-based communities?
Q4: What demographic differences exist between survey respondents who are recruited using (1) different platforms and (2) different advertising tracking methods?
This study compared the effectiveness and cost of different social media recruitment methodologies that comprised both paid and unpaid advertising structures across different platforms (Facebook, Twitter, and Reddit). Despite its large user base (1 billion active users), Instagram was not included because of the lack of a well-defined community of interest, which was the basis for targeting advertisements toward the relevant population. The target population for the survey was defined as US residents who had undergone genetic testing via direct-to-consumer (DTC) companies and who subsequently used third-party Web-based DNA interpretation services.
Paid Recruitment Methods
A 2x2 factorial design was used to test the comparative effectiveness and cost of different platforms and advertising tracking metrics for paid recruitment. Facebook and Twitter were selected as the platforms to be compared based on their large US resident user bases. Both platforms have proprietary content distribution networks that distribute paid advertising content to their users. Advertised content appeared as promoted status cards or tweets in the news feed of the targeted users (and ), intermingled with user-generated content.
There are slight differences in the way each content distribution network allows for targeting of specific user demographics. An effort was made to mirror the approach taken to targeting users across both sites. On Facebook, the potential audience of the campaign was defined as users living in the United States with an interest in 23andMe, a major DTC-GT company. Facebook targets paid advertising campaigns by identifying interests from information users have added to their Timeline, keywords associated with the Pages they like or apps they use, ads they have clicked on, and other similar sources . All users aged 18 years and older were included, resulting in a potential audience for the campaign of 740,000 Facebook users at the time of launch.
On Twitter, the potential audience of the campaign was defined as US-based Twitter users who were followers of @23andMe’s Twitter account, as well as users with interests similar to followers of @23andMe. According to Twitter, “[Follower targeting] works by displaying your Twitter Ads campaigns to people who follow specific usernames or are similar to the followers of those usernames” . This resulted in a potential audience size for the campaign of between 178,000 and 267,000 Twitter users at the time of launch.
A total of US $1000 was budgeted for the paid campaigns, divided evenly between Facebook and Twitter. Automatic bidding was the default on both platforms and was used in all 4 conditions. This feature dynamically adjusts the cost of advertising based on availability and demand, as well as the bidding parameters set by other advertisers. There are minor differences in the way each advertising platform handles bidding for advertisements: Twitter requires advertisers to specify a daily budget and provides an optional total budget setting for automatic bidding, after which the campaign will end. The daily budget for each condition was set at US $25 per day with a total budget of US $250 per campaign. Facebook does not require a daily budget setting; however, the total budget for this campaign was also set at US $250. Each campaign was allowed to run until the total budget was exhausted: Twitter advertisements ran for 10 days each, whereas Facebook advertisements were displayed for 14 days each. Although both Facebook and Twitter provide advertisers some control over the time of the day when advertisements are displayed, it was not specified on either platform in this study.
On both social media platforms, 2 advertising campaigns were conducted using different payment structures corresponding to different tracking methods. Both platforms allow the advertiser to either pay for each click through to the advertiser’s landing page (cost per click) or to pay for each iteration of a defined conversion action after the user has clicked through to the landing page (cost per conversion). For the purposes of the study, a conversion was defined as the user reaching the end of the questionnaire.
Both advertising platforms claim to iteratively optimize the targeting of a given advertising campaign based on the tracking metric used. Thus, a campaign for which the advertiser is billed per click is purportedly targeted in such a way as to maximize the likelihood that a given user who is shown the content will click on it. Conversely, a campaign for which the advertiser is billed per conversion is purportedly targeted in such a way as to maximize the likelihood that a given user will complete the conversion action after having clicked through.
Unpaid Recruitment Method
A parallel recruitment campaign was conducted on Reddit, a social news and community discussion site, to assess the viability of recruitment through unpaid posts to relevant Web-based communities. Reddit was selected because of the presence of several relevant community groups (see the table provided in) as well as the open structure of the site, which allows any user to post to any public group or subreddit, subject to community moderation. In total, 13 relevant subreddits were identified, although r/Health was not used because of community guidelines that prohibited the posting of content other than news. Identical posts were made on each of the remaining 12 subreddits seeking respondents for the survey (see the textbox provided in ).
The dataset was screened for duplicate responses using the internet protocol (IP) address and demographic profile of respondents, where responses from the same IP address within a 24-hour period or responses from the same IP address with a matching demographic profile were flagged as duplicates. This resulted in 17 responses being removed from the subsequent analysis. No responses were found to have been duplicated more than once, suggesting that these were likely the result of user error rather than a systematic effort.
A chi-square test was conducted to determine the extent to which the proportion of observed frequencies among the 4 paid campaigns conformed to a discrete uniform distribution, which would suggest the absence of a measurable difference in recruitment effectiveness between conditions. Posthoc pairwise Z tests were performed between all campaigns with a Bonferroni correction for multiple comparisons.
The recruitment budget for each condition was fixed at US $250, such that more cost-effective methods would yield a greater total number of responses over the study period. On the basis of this fixed budget, the cost-efficacy of each paid campaign was calculated in terms of the cost per survey response and cost per 1000 impressions. Each impression marks a time when the recruitment materials were displayed to a user, regardless of whether that user had seen the materials before or interacted with them in any way.
A total of 4 demographic variables were collected in the survey: age, gender, education, and race and ethnicity. Chi-square tests of homogeneity were performed to determine the statistical significance of differences in the distributions of gender and ethnicity. Participants who reported their gender as neither male nor female were excluded from the analysis of gender distributions because of the absence of reliable information on the expected proportion of nonbinary gender identifying individuals in the population. Age distributions were compared using a one-way analysis of variance. Posthoc pairwise comparisons were conducted with a Bonferroni correction, as appropriate. Kruskal-Wallis H tests were used to compare the education level of respondents. To maximize the response rate, demographic questions were not required to complete the survey. For demographic analyses only, participants who did not report the demographic characteristic of interest were excluded.
Participant demographics are presented in. Notably, because demographic questions were optional in the questionnaire, a substantial portion of respondents who completed the rest of the questionnaire elected not to answer them.
The mean age of those who reported this (n=266) was 46 years at the time of the survey. Among respondents who reported their gender (n=298), the majority (204/298, 68.5%) were female. The median level among those who reported their level of education (n=296) was a 4-year college degree across all conditions. Among respondents who reported their race or ethnicity (n=294), the majority (238/294, 81.0%) were white.
|Demographic variables||Participants (N=438), n (%)||Participants (excluding missing), n (%)|
|18-24||24 (5.5)||24 (9.0)|
|25-44||101 (23.1)||101 (38.0)|
|45-64||109 (24.9)||109 (41.0)|
|65 and older||32 (7.3)||32 (12.0)|
|Did not report||172 (39.3)||—a|
|Female||204 (46.6)||204 (68.5)|
|Male||93 (21.2)||93 (31.2)|
|Other||1 (0.2)||1 (0.3)|
|Did not report||140 (32.0)||—|
|Less than high school||3 (0.7)||3 (1.0)|
|High school/GEDb||11 (2.5)||11 (3.7)|
|Some college||63 (14.4)||63 (21.3)|
|2-year college degree||39 (8.9)||39 (13.2)|
|4-year college degree||91 (20.8)||91 (30.7)|
|Advanced degree (postgraduate)||89 (20.3)||89 (30.1)|
|Did not report||142 (32.4)||—|
|Race and Ethnicity||N=294|
|White/Caucasian||238 (54.3)||238 (81.0)|
|African American||7 (1.6)||7 (2.4)|
|Hispanic/Latino||10 (2.3)||10 (3.4)|
|Asian||8 (1.8)||8 (2.7)|
|Multiethnic||23 (5.3)||25 (7.8)|
|Other||8 (1.8)||8 (2.7)|
|Did not report||144 (32.9)||—|
aValid percentage excludes respondents who did not report for a given demographic variable.
bGED refers to those respondents who reported completing the General Education Development tests as their highest level of educational attainment.
A total of 540 responses were received in the survey; however, 17 duplicate responses were identified during data cleaning, and an additional 88 respondents did not report having used a DTC-GT or did not report being aware of any third-party genetic interpretation companies and were subsequently excluded from the final sample (N=438). There were significant differences in the frequency of survey responses between the different experimental conditions (χ23=84.2; P<.001). Seefor frequencies and for pairwise comparisons. Nearly equal samples were collected from paid campaigns on Facebook (159/438, 36.3%) and Twitter (167/438, 38.1%). A significant but somewhat smaller sample of participants was recruited through the parallel unpaid campaign on Reddit (112/438, 25.6%). Of the participants recruited through paid campaigns (n=326), significantly more were recruited through the conversion-tracking campaigns (222/326, 68.1%) than through the click-tracking campaigns (104/326, 31.9%; Z=6.5; P<.001). The difference between conversion-based and click-based tracking metrics was much more pronounced on Twitter than on Facebook (Z=6.7, P<.001); correspondingly, of the 5 recruitment methodologies used, the Twitter-Conversion campaign recruited the greatest number of participants (142/438, 32.4%) and the Twitter-Click campaign recruited the fewest (25/438, 5.7%).
|Condition 1 (n)||Condition 2 (n)||Z score||P value|
|Facebook-Click (79)||Twitter-Click (25)||5.3||<.001|
|Facebook-Conversion (80)||Twitter-Conversion (142)||3.3||<.001|
|Facebook-Click (79)||Twitter-Conversion (142)||4.3||<.001|
|Facebook-Conversion (80)||Twitter-Click (25)||5.4||<.001|
|Twitter-Click (25)||Twitter-Conversion (142)||9.1||<.001|
|Facebook-Click (79)||Facebook-Conversion (80)||0.1||.94|
Conversion-tracking campaigns on both Facebook and Twitter were more cost-effective at garnering survey respondents, averaging US $3.13 and US $1.76 per response, respectively. Click-based campaigns cost an average of US $3.16 and US $10.00 per response on the same platforms. There was a substantial difference in the cost of impressions between platforms in the click-based conditions, with Facebook charging US $1.91 per 1,000 impressions compared with US $9.90 on Twitter. There was, however, only a nominal difference in the cost of impressions among conversion-based conditions.
Of the 4 paid advertising conditions, the Twitter-Conversion campaign was the most cost-effective in terms of generating survey responses, followed by the Facebook-Conversion, Facebook-Click, and Twitter-Click campaigns. The Facebook-Click campaign was the most cost-effective in generating broad audience exposure, followed by the Twitter-Conversion, Facebook-Conversion, and Twitter-Click campaigns ().
Demographic Comparisons Across Platforms and Tracking Methods
Differences Between Platforms
There were notable differences in the demographic characteristics of survey respondents recruited on each of the 3 platforms (Facebook, Twitter, and Reddit), particularly by age, gender, and race and ethnicity. There was no significant difference in the level of education reported by respondents recruited across different platforms (H2= 4.61, P=.10).
Among those who reported their age (n=266), there were significant differences in the average age of respondents recruited on different platforms (F2,263=58.18; P<.001). The age distributions for respondents recruited on each platform are presented in. There was not a significant difference between the age of respondents recruited on Facebook (mean 49.13) and Twitter (mean 53.11); however, respondents recruited on Reddit were, on average, significantly younger than either of the other 2 groups (mean 34.23).
There were significant differences in the ratio of female to male respondents between those who were recruited on different platforms (χ22=53.0; P<.001). The gender distributions for each recruitment platform are presented in. Female respondents made up the majority on both Facebook (65/82, 79%) and Twitter (102/122, 83.6%), but were in the minority among those recruited on Reddit (37/94, 39%).
contains a complete reporting of respondent race and ethnicity by recruitment platform. The difference in the proportion of white to nonwhite respondents across platforms approached significance (χ22=5.7; P=.06). The proportion of white respondents was higher on Twitter (105/121, 86.8%) than on Facebook (58/79, 73%) or Reddit (75/94, 80%).
|Category||Age (years), mean (SD)||F test (df1,df2)||P value|
|Recruitment platform (n)|
|Facebook (69)||49.13 (12.64)||58.18 (2,263)||<.001|
|Twitter (110)||53.11 (12.95)||58.18 (2,263)||<.001|
|Reddit (87)||34.23 (11.89)||58.18 (2,263)||<.001|
|Overall (266)||45.90 (15.00)||—a||—|
|Tracking metric (n)|
|Click (50)||51.93 (12.67)||0.35 (1,177)||.56|
|Conversion (129)||51.81 (13.08)||0.35 (1,177)||.56|
|Overall (179)||51.58 (12.94)||—||—|
|Category||Race and ethnicity|
|Facebook (N=79)||58 (73)||2 (3)||5 (6)||1 (1)||8 (10)||5 (6)|
|Twitter (N=121)||105 (86.8)||3 (2.5)||4 (3.3)||2 (1.7)||5 (4.1)||2 (1.7)|
|Reddit (N=94)||75 (80)||2 (2)||1 (1)||5 (5)||10 (11)||1 (1)|
|Overall (N=294)||238 (81.0)||7 (2.4)||10 (3.4)||8 (2.7)||23 (7.8)||8 (2.7)|
|Click (N=57)||44 (77)||2 (4)||2 (4)||0 (0)||6 (11)||3 (5)|
|Conversion (N=143)||119 (83.2)||3 (2.1)||7 (4.9)||3 (2.1)||7 (4.9)||4 (2.8)|
|Overall (N=200)||163 (81.5)||5 (2.5)||9 (4.5)||3 (1.5)||13 (6.5)||7 (3.5)|
aRespondents who selected more than 1 option for race and ethnicity.
Differences Between Tracking Methods
The demographic characteristics of survey respondents recruited on paid advertising platforms (ie, Facebook and Twitter) using click-tracking were compared with those recruited using conversion-tracking. Respondents recruited from Reddit were excluded from this analysis, as no tracking method was used on this platform. The demographic differences between tracking methods were generally less substantial than those observed between platforms. The only significant difference noted was in the ratio of female to male respondents between those recruited using different tracking methods (χ21=4.5; P=.03). Gender distributions for each tracking method are summarized in. Female respondents made up the majority in both cases, but were more prevalent among those recruited using conversion-tracking (124/145, 85.5%), compared with those recruited using click-tracking (43/59, 73%).
There was no significant difference in the age distribution of respondents recruited using conversion-tracking compared with click-tracking (F1,177=0.35, P=.56) or in their level of education (H1=0.02, P=.88). Similarly, no significant differences were found between tracking methods in the proportion of white to nonwhite respondents recruited (χ21=1.0, P=.32).
This study set out to test and evaluate the use of social media platforms as a recruitment tool for research on a hard-to-reach population. To do so, it directly compared paid and unpaid recruitment campaigns implemented on multiple social media platforms (Facebook, Twitter, and Reddit) and employed different advertising tracking metrics (click-based and conversion-based). Only a handful of studies have directly compared multiple methods of survey recruitment on social media; thus, this study represents a novel contribution to the development of Web-based survey methodology in general and recruitment approaches for hard-to-reach populations in particular.
Nearly identical sample sizes were obtained via paid Facebook (n=159) and Twitter (n=167) advertising, as well as the sample obtained via unpaid posting on Reddit (n=112). Although survey recruitment on social media for population health research has predominantly taken place on Facebook, this finding suggests that targeted advertising via other social media platforms may also be viable. Although the overall user base of Twitter is often noted to be substantially smaller than that of Facebook, and this was reflected in the reach of recruitment campaigns on that platform, this did not appear to constrain the effectiveness of Twitter as a platform for recruiting participants, with both Facebook and Twitter yielding comparable numbers of participants over both tracking metrics. Similarly, unpaid posting to community groups may also prove productive in achieving a broader sample.
The difference in the effectiveness of survey recruitment on Facebook and Twitter when considering both tracking metrics was found to be negligible; however, significant differences in the effectiveness of different tracking metrics across platforms were observed. Conversion-tracking campaigns recruited more than twice the number of respondents recruited by click-tracking campaigns, given the same budget. These results suggest that the use of different tracking metrics has important implications in determining the success of survey recruitment campaigns and warrants further investigation.
Overall, the use of conversion-tracking on Twitter was found to be the most cost-effective combination of tracking metric and platform conditions. The effectiveness of this approach may have been partially driven by organic growth (ie, individual users reposting recruitment materials from their own accounts). The number of survey responses garnered by this campaign exceeded the number of clicks detected on the advertisement, suggesting that this version of the recruitment materials was shared beyond the initial target audience for the advertising campaign. In an open-response question attached to the Reddit version of the questionnaire that asked respondents to identify the subreddit through which they had been recruited, 4 respondents indicated that recruitment materials had been forwarded to them by a friend or family member. Although these are the clearest indications for organic growth among the study conditions, it is possible that similar redistribution may have occurred in other cases as well.
This observation should serve as a reminder that all Web-based survey recruitment materials have the potential to be redistributed beyond the initial target audience, or otherwise go viral, unless steps are taken to prevent this. This potential may be useful in recruiting a larger sample or if recruiting entirely on social media platforms that do not allow targeted advertising. Here, researchers may wish to adopt from the existing literature on predictors of advertising message virality [, ]. There is also significant cause for concern in contexts where nontarget audiences may be undesirable because of the scope or subject matter of the survey. Web-based surveys dealing with contentious topics have, in recent years, been redistributed in partisan discussion groups with the goal of sending a political message through the community’s collective response [ , ]. Surveys regarding health issues where significant public controversy exists might likewise be subject to purposive redistribution with the intent of affecting the results, even if initially targeted at a more limited audience. Future research should seek to identify the extent of organic sharing of survey recruitment materials and distinguish between the data collected from targeted and nontargeted respondents.
Demographic differences among study participants were observed between social media platforms and tracking methods. It should be noted that observed differences in the demographic makeup of samples apply only to the subset of participants in each who answered optional demographic questions. Reddit attracted a sample that was approximately 15 years younger on average than that recruited from either Facebook or Twitter. In addition, the sample recruited from Reddit included far more male respondents than that of either Facebook or Twitter. No significant difference between platforms was observed for either education levels or race and ethnicity.
In terms of tracking methods, conversion-tracking resulted in a sample that included more women than that recruited through click-tracking. No other demographic differences were observed. It is possible, given the higher number of female respondents observed among those recruited on Facebook and Twitter, that conversion-based targeting may have skewed the sample even further toward female respondents by iteratively targeting female users at a higher rate than male users.
The demographic breakdown of survey participants closely matches that of other surveys conducted with DTC testing consumers [, ], reflecting early adopters of this technology who are primarily white and highly educated. Although this study appeared to represent more females than males, past surveys have shown gender variation across DTC companies themselves, which may reflect demographic differences in user base [ ]. Similarly, although recruitment on Reddit, compared with Facebook and Twitter, resulted in a very different set of respondents, it is unlikely that any of the samples is more intrinsically representative of anything beyond the respective platform’s user base. As such, conducting recruitment on multiple platforms likely facilitated access to a more demographically diverse set of respondents, which yielded a final sample population that was more consistent with the past studies.
A major limitation for any study of modern Web-based advertising is the issue of algorithm dynamics or the changes made by engineers to improve the commercial service and by consumers in using that service . Research findings on the effectiveness of specific software tools are intrinsically limited by the potential of such tools to evolve and change in unpredictable ways. The platform features used in this study to select an audience may not be viable in the near future, and Facebook or Twitter may update their advertisement targeting algorithm to select interested groups more effectively or more narrowly than researchers intend.
Limitations also arise from differences in the affordances of advertising platforms for defining a target audience as well as parameters for the display and timing of advertisements. Although the practical implications of these distinctions may be negligible, they do undermine the ability of researchers to directly compare the performance of advertising materials on different platforms or otherwise require researchers to isolate the most salient points of comparison: for example, in this study, differences in the way Facebook and Twitter handled bidding for advertising space meant that campaigns could either be restricted to equal budgets or to equivalent timeframes, with the former ultimately being deemed more relevant to the research questions at hand.
There were also substantial differences in the affordances of each platform for displaying advertising materials: the amount and size of displayed text, as well as the availability and size of graphics, are constrained both by the technological limitations of the platform (eg, Twitter’s character limit) as well as community norms and expectations. This study adjusted the advertising materials displayed in each condition to best take advantage of the affordances of that platform: for example, more extensive copy was displayed to Reddit users before clicking on the recruitment link than users on Facebook or Twitter. Although this allowed the experiment to conform more closely to the norms of each platform, and thus supported its ecological validity, it does introduce further limitations on the direct comparability of results across platforms.
When conducting survey research, numerous considerations can influence the desired sample size. Although unpaid recruitment in Web-based community groups may perform comparably to recruitment via paid advertisements on social media, it should be noted that the number of potential respondents reached through paid advertising is more readily scalable, given a sufficient budget. By contrast, the potential audience in Web-based communities is limited by the number of active users, which may be quite small for hard-to-reach populations of interest. Such communities typically frown on repeated posting, further limiting the audience that may be reached to those who are active in the period immediately following the initial post. The inherent limits on the scope of populations that can be reached through Web-based communities may, therefore, render unpaid Web-based recruitment less effective than paid advertising for achieving larger sample sizes.
In addition, paid advertising platforms allowed for audiences to be targeted based on location. In cases of organic sharing of those advertisements, as well as recruitment in community groups, no similar controls are available. Future research should compare survey data from targeted audiences against those reached through organic growth. Likewise, an assessment of how particular tracking metrics may lead to targeting of particular demographic or interest groups is necessary to fully understand the implications of using Web-based advertising for survey recruitment. As audience targeting and tracking algorithms continue to develop, longitudinal sampling of the same small population may be useful to evaluate whether algorithm dynamics have significant effects on Web-based survey recruitment.
These limitations only stress the need for additional comparative studies of survey recruitment through different Web-based advertising platforms and tracking metrics. By understanding and evaluating the results of Web-based distribution, researchers can be aware of the effectiveness and limitations of various targeting and tracking approaches. Likewise, by comparing the characteristics of respondents from multiple recruitment campaigns, it is possible to test the effectiveness of the different methods in reaching target populations. The results of this study suggest that conversion-tracking metrics support more cost-effective survey recruitment than conventional designation of audience parameters accompanied by click-based tracking. However, algorithmic targeting of advertisements also poses problems for the reliability and reproducibility of survey research as sampling mechanisms may change in unpredictable ways.
The results of this study indicate that there are meaningful differences between different approaches to Web-based survey recruitment. Advertisements on social media are a pragmatic method for survey recruitment, particularly within hard-to-reach populations and are most effective when combined with conversion-based tracking metrics. Recruitment through Web-based community groups is an effective complementary approach for reaching such populations and may give access to a more diverse sample than advertising alone. These tools must be used with due intentionality and an awareness of limitations so as to avoid potential pitfalls. Future research is needed to fully understand the effect of organic sharing and algorithm dynamics on the constitution of Web-based samples.
Conflicts of Interest
Multimedia Appendix 1
Reddit Campaign Details.DOCX File, 14KB
- Topolovec-Vranic J, Natarajan K. The use of social media in recruitment for medical research studies: a scoping review. J Med Internet Res 2016 Dec 7;18(11):e286 [FREE Full text] [CrossRef] [Medline]
- Atkinson R, Flint J. Social Research Update - University of Surrey. 2001. Accessing Hidden and Hard-To-Reach Populations: Snowball Research Strategies URL: http://sru.soc.surrey.ac.uk/SRU33.PDF [accessed 2019-06-25]
- Thornton LK, Harris K, Baker AL, Johnson M, Kay-Lambkin FJ. Recruiting for addiction research via Facebook. Drug Alcohol Rev 2016 Dec;35(4):494-502. [CrossRef] [Medline]
- King DB, O'Rourke N, DeLongis A. Social media recruitment and online data collection: a beginner’s guide and best practices for accessing low-prevalence and hard-to-reach populations. Can Psychol 2014;55(4):240-249. [CrossRef]
- Lohse B, Wamboldt P. Purposive facebook recruitment endows cost-effective nutrition education program evaluation. JMIR Res Protoc 2013 Aug 15;2(2):e27 [FREE Full text] [CrossRef] [Medline]
- Baltar F, Brunet I. Social research 2.0: virtual snowball sampling method using Facebook. Internet Res 2012 Jan 27;22(1):57-74. [CrossRef]
- Baltar F, Gorjup MT. Online mixted sampling: an application in hidden populations. Intang Cap 2012 Apr 28;8(1):123-149. [CrossRef]
- Bhutta CB. Not by the book: Facebook as a sampling frame. Sociol Methods Res 2012 Mar 21;41(1):57-88. [CrossRef]
- Ramo DE, Prochaska JJ. Broad reach and targeted recruitment using Facebook for an online survey of young adult substance use. J Med Internet Res 2012 Feb 23;14(1):e28 [FREE Full text] [CrossRef] [Medline]
- Investor Relations. 2019. Facebook Reports First Quarter 2019 Results URL: https://investor.fb.com/investor-news/press-release-details/2019/Facebook-Reports-First-Quarter-2019-Results/default.aspx [accessed 2019-05-23]
- Shareholder Letter: Twitter. 2018. Q4 and Fiscal Year 2018 Letter to Shareholders URL: https://s22.q4cdn.com/826641620/files/doc_financials/2018/q4/Q4-2018-Shareholder-Letter.pdf [accessed 2019-05-23]
- O'Connor A, Jackson L, Goldsmith L, Skirton H. Can I get a retweet please? Health research recruitment and the Twittersphere. J Adv Nurs 2014 Mar;70(3):599-609. [CrossRef] [Medline]
- Horrell LN, Kneipp SM, Gonzales CM, Williams KS. Using Social Media to Enhance Study Recruitment: How Facebook and Twitter May Engage Vulnerable Populations. In: Proceedings of the 27th International Nursing Research Congress. 2016 Presented at: NRC'16; July 21-25, 2016; Cape Town, South Africa URL: https://sigma.nursingrepository.org/handle/10755/620661
- Guillory J, Kim A, Murphy J, Bradfield B, Nonnemaker J, Hsieh Y. Comparing Twitter and online panels for survey recruitment of e-cigarette users and smokers. J Med Internet Res 2016 Dec 15;18(11):e288 [FREE Full text] [CrossRef] [Medline]
- Yuan P, Bare MG, Johnson MO, Saberi P. Using online social media for recruitment of human immunodeficiency virus-positive participants: a cross-sectional survey. J Med Internet Res 2014 May 1;16(5):e117 [FREE Full text] [CrossRef] [Medline]
- Quach S, Pereira JA, Russell ML, Wormsbecker AE, Ramsay H, Crowe L, et al. The good, bad, and ugly of online recruitment of parents for health-related focus groups: lessons learned. J Med Internet Res 2013 Nov 14;15(11):e250 [FREE Full text] [CrossRef] [Medline]
- Gu LL, Skierkowski D, Florin P, Friend K, Ye Y. Facebook, Twitter, & QR codes: an exploratory trial examining the feasibility of social media mechanisms for sample recruitment. Comput Human Behav 2016 Jul;60:86-96. [CrossRef]
- Facebook Business. URL: https://www.facebook.com/business/products/ads/ad-targeting [accessed 2017-02-23]
- Twitter Business. Follower Targeting on Twitter URL: https://business.twitter.com/en/targeting/follower.html [accessed 2017-02-23]
- Alhabash S, McAlister AR. Redefining virality in less broad strokes: predicting viral behavioral intentions from motivations and uses of Facebook and Twitter. New Media Soc 2014 Feb 24;17(8):1317-1339. [CrossRef]
- Kosinski M, Matz SC, Gosling SD, Popov V, Stillwell D. Facebook as a research tool for the social sciences: opportunities, challenges, ethical considerations, and practical guidelines. Am Psychol 2015 Sep;70(6):543-556. [CrossRef] [Medline]
- Reddit. 2017. The Trump Administration is Sending Out a Survey (Primarily to His Supporters) About Accountability of the Mainstream Media. Fill It Out Here! URL: https://www.reddit.com/r/BlueMidterm2018/comments/5uj54n/the_trump_administration_is_sending_out_a_survey/ [accessed 2017-07-10]
- Reddit. 2017. Mainstream Media Accountability Survey Put Out by Trump. r/all is Trying to Swarm the Survey URL: https://www.reddit.com/r/The_Donald/comments/5ul4m1/mainstream_media_accountability_survey_put_out_by/ [accessed 2017-07-10]
- Kaufman DJ, Bollinger JM, Dvoskin RL, Scott JA. Risky business: risk perception and the use of medical services among customers of DTC personal genetic testing. J Genet Couns 2012 Jun;21(3):413-422. [CrossRef] [Medline]
- Roberts JS, Gornick MC, Carere DA, Uhlmann WR, Ruffin MT, Green RC. Direct-to-consumer genetic testing: user motivations, decision making, and perceived utility of results. Public Health Genomics 2017;20(1):36-45 [FREE Full text] [CrossRef] [Medline]
- Lazer D, Kennedy R, King G, Vespignani A. Big data. The parable of Google Flu: traps in big data analysis. Science 2014 Mar 14;343(6176):1203-1205. [CrossRef] [Medline]
|DTC-GT: direct-to-consumer genetic testing|
|IP: internet protocol|
Edited by G Eysenbach; submitted 28.11.18; peer-reviewed by A Kim, K Himes, B Bie, F Timimi, J Cassaniti; comments to author 31.03.19; revised version received 24.05.19; accepted 04.06.19; published 30.07.19Copyright
©Tiernan J Cahill, Blake Wertz, Qiankun Zhong, Andrew Parlato, John Donegan, Rebecca Forman, Supriya Manot, Tianyi Wu, Yazhu Xu, James J Cummings, Tricia Norkunas Cunningham, Catharine Wang. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 30.07.2019.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.