<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id><journal-id journal-id-type="publisher-id">jmir</journal-id><journal-id journal-id-type="index">1</journal-id><journal-title>Journal of Medical Internet Research</journal-title><abbrev-journal-title>J Med Internet Res</abbrev-journal-title><issn pub-type="epub">1438-8871</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v27i1e68092</article-id><article-id pub-id-type="doi">10.2196/68092</article-id><article-categories><subj-group subj-group-type="heading"><subject>Viewpoint</subject></subj-group></article-categories><title-group><article-title>Increasing Rigor in Online Health Surveys Through the Reduction of Fraudulent Data</article-title></title-group><contrib-group><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Ng</surname><given-names>Wen Zhi</given-names></name><degrees>BSocSci</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Erdembileg</surname><given-names>Sundarimaa</given-names></name><degrees>BA</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Liu</surname><given-names>Jean C J</given-names></name><degrees>BSc, PhD</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Tucker</surname><given-names>Joseph D</given-names></name><degrees>MD, PhD, AM</degrees><xref ref-type="aff" rid="aff4">4</xref><xref ref-type="aff" rid="aff5">5</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Tan</surname><given-names>Rayner Kay Jin</given-names></name><degrees>BSocSci, PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib></contrib-group><aff id="aff1"><institution>Saw Swee Hock School of Public Health, National University of Singapore, National University Health System</institution><addr-line>12 Science Drive 2, #10-01</addr-line><addr-line>Singapore</addr-line><country>Singapore</country></aff><aff id="aff2"><institution>Yale-NUS College, National University of Singapore</institution><addr-line>Singapore</addr-line><country>Singapore</country></aff><aff id="aff3"><institution>Health and Social Sciences Cluster, Singapore Institute of Technology</institution><addr-line>Singapore</addr-line><country>Singapore</country></aff><aff id="aff4"><institution>London School of Hygiene and Tropical Medicine</institution><addr-line>London</addr-line><country>United Kingdom</country></aff><aff id="aff5"><institution>UNC School of Medicine, University of North Carolina at Chapel Hill</institution><addr-line>Chapel Hill</addr-line><addr-line>NC</addr-line><country>United States</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Cahill</surname><given-names>Naomi</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Blumberg</surname><given-names>Stephen</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Yogeshappa</surname><given-names>Vedamurthy Gejjegondanahalli</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Rayner Kay Jin Tan, BSocSci, PhD, Saw Swee Hock School of Public Health, National University of Singapore, National University Health System, 12 Science Drive 2, #10-01, Singapore, 117549, Singapore, 65 91878576; <email>rayner.tan@nus.edu.sg</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>21</day><month>8</month><year>2025</year></pub-date><volume>27</volume><elocation-id>e68092</elocation-id><history><date date-type="received"><day>28</day><month>10</month><year>2024</year></date><date date-type="rev-recd"><day>26</day><month>06</month><year>2025</year></date><date date-type="accepted"><day>26</day><month>06</month><year>2025</year></date></history><copyright-statement>&#x00A9;Wen Zhi Ng, Sundarimaa Erdembileg, Jean C J Liu, Joseph D Tucker, Rayner Kay Jin Tan. Originally published in the Journal of Medical Internet Research (<ext-link ext-link-type="uri" xlink:href="https://www.jmir.org">https://www.jmir.org</ext-link>), 21.8.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://www.jmir.org/">https://www.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://www.jmir.org/2025/1/e68092"/><abstract><p>Online surveys have become a key tool of modern health research, offering a fast, cost-effective, and convenient means of data collection. It enables researchers to access diverse populations, such as those underrepresented in traditional studies, and facilitates the collection of stigmatized or sensitive behaviors through greater anonymity. However, the ease of participation also introduces significant challenges, particularly around data integrity and rigor. As fraudulent responses&#x2014;whether from bots, repeat responders, or individuals misrepresenting themselves&#x2014;become more sophisticated and pervasive, ensuring the rigor of online surveys has never been more crucial. This article provides a comprehensive synthesis of practical strategies that help to increase the rigor of online surveys through the detection and removal of fraudulent data. Drawing on recent literature and case studies, we outline several options that address the full research cycle from predata collection strategies to validation post data collection. We emphasize the integration of automated screening techniques (eg, CAPTCHAs and honeypot questions) and attention checks (eg, trap questions) for purposeful survey design. Robust recruitment procedures (eg, concealed eligibility criteria and 2-stage screening) and a proper incentive or compensation structure can also help to deter fraudulent participation. We examine the merits and limitations of different sampling methodologies, including river sampling, online panels, and crowdsourcing platforms, offering guidance on how to select samples based on specific research objectives. Post data collection, we discuss metadata-based techniques to detect fraudulent data (eg, duplicate email or IP addresses, response time analysis), alongside methods to better screen for low-quality responses (eg, inconsistent response patterns and improbable qualitative responses). The escalating sophistication of fraud tactics, particularly with the growth of artificial intelligence (AI), demands that researchers continuously adapt and stay vigilant. We propose the use of dynamic protocols, combining multiple strategies into a multipronged approach that can better filter for fraudulent data and evolve depending on the type of responses received across the data collection process. However, there is still significant room for strategies to develop, and it should be a key focus for upcoming research. As online surveys become increasingly integral to health research, investing in robust strategies to screen for fraudulent data and increasing the rigor of studies is key to upholding scientific integrity.</p></abstract><kwd-group><kwd>online surveys</kwd><kwd>web-based surveys</kwd><kwd>web-based research</kwd><kwd>data quality</kwd><kwd>data integrity</kwd><kwd>data validation</kwd><kwd>fraudulent responses</kwd><kwd>fraud</kwd><kwd>survey fraud</kwd><kwd>methodological rigor</kwd><kwd>recruitment strategies</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Surveys are a key tool in research, offering insight into participants&#x2019; minds and allowing researchers to better understand their motivations and preferences [<xref ref-type="bibr" rid="ref1">1</xref>]. Traditionally, survey data was collected primarily through paper questionnaires, potentially distributed through the mail or face-to-face, and eventually grew to include telephone surveys [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>]. However, as use of the internet has grown and response rates for traditional survey modes dropped, the primary choice for surveys has shifted to online surveys, alternatively known as web-based surveys, which focus on distributing a survey form or instrument hosted on the internet, even though actual distribution methods can range from more traditional ways like mail to newer methods like social media advertising [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref4">4</xref>].</p><p>The use of online surveys in internet-based health research has become more common, largely due to its ability to efficiently reach a large audience in a cost-effective manner and its low barriers to participation [<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref6">6</xref>]. For example, online surveys eliminate the need for face-to-face interaction and remove transportation and logistical barriers, offering participants the chance to complete surveys at their own convenience and comfort, reducing the burden on them [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref8">8</xref>]. Furthermore, researchers can easily reach underrepresented populations, as well as create samples that ignore geographic boundaries [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref10">10</xref>]. Depending on the design, it also gives participants the ability to remain anonymous and the freedom to respond honestly without stigma or judgment [<xref ref-type="bibr" rid="ref11">11</xref>]. Online surveys can elicit significantly more reports on socially undesirable behaviors in comparison to surveys administered in person, highlighting the benefit of online surveys when it comes to sensitive topics [<xref ref-type="bibr" rid="ref12">12</xref>]. COVID-19 has also accelerated the use of digital methods for research purposes, prompting the transition of surveys from in-person to online formats [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref14">14</xref>].</p><p>However, the rise of online surveys comes with an increased risk of fraudulent behavior [<xref ref-type="bibr" rid="ref8">8</xref>], which can be defined as &#x201C;individuals, groups or computer processes (ie, bots) participating in online, internet or web-based data collection methods at a statistically significant level, such that data are or would be measurably distorted&#x201D; [<xref ref-type="bibr" rid="ref15">15</xref>]. Fraudulent behavior comes in many forms, which leads to further complications when screening for it. For example, incentives are a popular method to encourage participation, but they increase both authentic response rates and repeated responders [<xref ref-type="bibr" rid="ref16">16</xref>]. Individuals may choose to misrepresent themselves to match eligibility criteria (alias scammers) or submit multiple surveys in order to receive more incentives (repeat responders) [<xref ref-type="bibr" rid="ref17">17</xref>]. However, incentives are not the only motivational factor, as response distorters could spam responses in an effort to misinform results due to their own agenda or political motivation [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>]. There has also been an increase in easily obtainable sophisticated software applications, otherwise known as bots, which are designed to automatically fill out surveys, allowing individuals to quickly flood surveys with responses [<xref ref-type="bibr" rid="ref20">20</xref>]. Furthermore, there are also careless responders, where individuals do not give sufficient attention to questions or fail to read them comprehensively, leading to answers that may not accurately reflect reality [<xref ref-type="bibr" rid="ref21">21</xref>].</p><p>Another consideration when crafting an online survey is the sampling method, which has the potential to significantly introduce errors and bias into data collection, reducing the representativeness and generalizability of results [<xref ref-type="bibr" rid="ref22">22</xref>]. When carrying out any research, it is important to clearly define the population of interest and tailor the sampling method to effectively engage that target population [<xref ref-type="bibr" rid="ref23">23</xref>]. Sampling methods can largely be split into two categories: (1) probability samples, in which elements from a population are randomly selected and elements have a nonzero known probability of being selected, and (2) nonprobability samples, which are nonrandom and the probability of each element being selected may be unknown [<xref ref-type="bibr" rid="ref1">1</xref>]. Probability sampling is generally a more accurate method to establish the characteristics of an entire population. In comparison, nonprobability sampling is typically used when access to the full population is limited or not needed, such as when researchers only want to focus on a specific subgroup of characteristics. If an inappropriate sampling method is chosen, researchers run the risk of collecting data that is not an accurate representation of the target population [<xref ref-type="bibr" rid="ref24">24</xref>].</p><p>Despite the difficulty of doing so, ensuring that data collected is accurate is critical, as the inclusion of even a small amount of fraudulent data can create or mask statistically significant differences, decreasing the rigor of the study [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref25">25</xref>]. The validity and integrity of results then come into question and may even artificially create relationships between uncorrelated factors [<xref ref-type="bibr" rid="ref26">26</xref>]. For example, a study found that of the 1281 respondents, only 197 were real respondents, and the inclusion of fake respondents into the results significantly altered the findings&#x2014;both creating and masking relationships between factors [<xref ref-type="bibr" rid="ref27">27</xref>]. Without a sufficiently rigorous data screening process, bias and noise could be introduced to the dataset through the inclusion of fraudulent data, leading to incorrect conclusions of study results and potentially biasing future research or policy recommendations [<xref ref-type="bibr" rid="ref8">8</xref>]. Furthermore, fraudulent behavior not only compromises the quality of the research done but also increases research costs through the disbursement of incentives to individuals who had already completed the study or were ineligible and the need for increased time spent on identifying fraudulent responses [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref11">11</xref>].</p><p>Data quality comprises many different attributes, such as reliability, accuracy, and integrity; yet, different disciplines and platforms will give rise to different levels of importance placed on each attribute when conducting research [<xref ref-type="bibr" rid="ref28">28</xref>]. One of the key threats to online surveys is the rise in fraudulent responses; yet, there is a lack of articles exploring the impact of fraudulent data on online surveys and how to best guard against them [<xref ref-type="bibr" rid="ref27">27</xref>]. The purpose of this article is to examine the methods in which researchers can increase the rigor in online surveys by taking an integrated approach to recruitment, survey instrument design, and metadata checks in order to reduce the inclusion of fraudulent data. The measures that are commonly used by researchers now can be largely split into 2 different phases of research: before data collection and post data collection. Subsequently, each phase gave 2 main areas of focus, (1) survey design and (2) recruitment strategies for predata collection, and (3) metadata checks and (4) data quality for postdata collection. A summary of the methods mentioned in this article can be found in <xref ref-type="fig" rid="figure1">Figure 1</xref>.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Measures to increase rigor of online health surveys [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref6">6</xref>-<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref22">22</xref>-<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref28">28</xref>-<xref ref-type="bibr" rid="ref49">49</xref>].</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e68092_fig01.png"/></fig></sec><sec id="s2"><title>Predata Collection: Guarding Against Fraudulent Data</title><p>Presurvey methods that help to guard against fraudulent data can either be implemented through survey design or recruitment strategies. Survey design strategies are primarily meant to make screening for fraudulent data easier. Contrarily, recruitment strategies aim to prevent fraudulent responders&#x2019; access to the survey.</p><sec id="s2-1"><title>Survey Design: Automatic Checks</title><p>One basic automatic check to protect against bots is CAPTCHAs, which is based on user actions such as checkbox clicks and image challenges [<xref ref-type="bibr" rid="ref8">8</xref>]. Newer versions like Google&#x2019;s reCAPTCHA V3 use sophisticated algorithms to evaluate user interactions based on various criteria, such as typing speed and IP address, which are then scored based on the study&#x2019;s scoring system and expectations in order to identify possible bots [<xref ref-type="bibr" rid="ref29">29</xref>]. Honeypot questions are an additional automated defense against bots that can be easily included within surveys alongside CAPTCHAs; these are questions that are designed specifically for bot detection and hidden from actual participants using custom code [<xref ref-type="bibr" rid="ref8">8</xref>]. Thus, receiving answers to a honeypot question would indicate bot activity.</p></sec><sec id="s2-2"><title>Survey Design: Survey Questions</title><p>Another widely used strategy to identify and exclude inattentive respondents is attention checks. The most commonly suggested method is to ask for a specific response, otherwise known as a trap question or instructional manipulation check [<xref ref-type="bibr" rid="ref30">30</xref>]. For example, participants could be asked to pick the italicized option [<xref ref-type="bibr" rid="ref31">31</xref>], or a specific choice from a list, eg, please select the option &#x201C;strongly agree&#x201D; [<xref ref-type="bibr" rid="ref32">32</xref>]. Current literature advises attention checks to be used throughout the survey to better identify responders who are giving low-quality responses [<xref ref-type="bibr" rid="ref30">30</xref>]. The wording and quantity of trap questions should strike a balance between filtering out low-quality responses and ensuring that genuine respondents are not accidentally filtered out [<xref ref-type="bibr" rid="ref32">32</xref>]. An alternative method is to make use of speed bump questions, which require participants to properly read through the question and use reason to arrive at an answer, filtering out bots or respondents who answer hastily or automatically [<xref ref-type="bibr" rid="ref33">33</xref>]. Speed bump questions can look like this, &#x201C;The man couldn&#x2019;t lift his son because he was so weak. Who was weak, the man or his son?&#x201D; However, researchers should consider the type of research they are doing before making use of attention checks, as they have been found to increase deliberation, potentially affecting cognitive task performance [<xref ref-type="bibr" rid="ref34">34</xref>]. On the other hand, other research has found that they do not affect participants&#x2019; responses on scales or scale validity [<xref ref-type="bibr" rid="ref35">35</xref>]. Researchers should thus monitor responses or trial survey questions to see if responses are significantly affected before rolling out their surveys officially.</p><p>Another strategy researchers can consider using is to leverage domain-specific knowledge to detect respondents who are not from the correct target audience. This is especially effective in fields where many industry-specific terms are used, and researchers can make use of this by including qualitative response formats [<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref31">31</xref>]. For instance, one survey targeting experienced programmers included simple programming tasks and questions as assessments [<xref ref-type="bibr" rid="ref30">30</xref>], while another survey used specific terminology and jargon as soft-checks to verify expertise [<xref ref-type="bibr" rid="ref31">31</xref>]. Subsequently, identifying inconsistencies or suspicious responses compared to established norms, such as unlikely health symptoms for the given demographic, can identify unfit respondents [<xref ref-type="bibr" rid="ref36">36</xref>].</p><p>Researchers have also found success in using duplicate questions presented in diverse formats. Researchers have advocated for requesting the same data point in multiple formats [<xref ref-type="bibr" rid="ref29">29</xref>] or incorporating specific items into questionnaires that essentially ask the same question more than once [<xref ref-type="bibr" rid="ref30">30</xref>]. For example, one study detected bot-based responses by including duplicate gender identity questions, which would trigger subsequent logic checks and reveal bots that followed code rather than survey logic [<xref ref-type="bibr" rid="ref31">31</xref>]. Inconsistent responses to duplicate questions serve as indicators of low-quality responses, so researchers should take note of mismatches in answers. Researchers should also disguise these checks by distributing questions across different sections of the questionnaire and phrasing them differently to prevent easy identification.</p></sec><sec id="s2-3"><title>Recruitment Strategies: Eligibility Screening</title><p>To ensure the integrity of survey responses, researchers can implement rigorous eligibility screening processes and streamline recruitment procedures. For example, researchers can consider using 2-stage recruitment, in which participants first have a public eligibility screener to assess whether they meet eligibility criteria before a second personalized link is sent to them for the main survey [<xref ref-type="bibr" rid="ref33">33</xref>]. The eligibility screener can further implement features like CAPTCHAs or browser cookies to prevent duplicate responses [<xref ref-type="bibr" rid="ref33">33</xref>]. Eligibility screeners could also request contact information so that researchers can contact participants personally to confirm their identity or ask for documents that could confirm information critical to eligibility [<xref ref-type="bibr" rid="ref37">37</xref>]. In order to prevent individuals from tailoring their responses to meet eligibility criteria, researchers can consider recruiting participants without explicitly disclosing eligibility criteria upfront [<xref ref-type="bibr" rid="ref29">29</xref>]. Furthermore, having duplicated questions in the eligibility screener and main survey will allow researchers to look for inconsistencies in responses, which is another way to expose suspicious respondents [<xref ref-type="bibr" rid="ref8">8</xref>]. Researchers can consider requiring participants to go through steps before accessing the survey, such as using specialized links or passwords, but must also ensure that this process is as straightforward and user-friendly as possible; otherwise, it risks increasing dropouts [<xref ref-type="bibr" rid="ref38">38</xref>].</p></sec><sec id="s2-4"><title>Recruitment Strategies: Compensations</title><p>While incentives can promote higher numbers of survey responses, they also run an increased risk of attracting fraudulent responses [<xref ref-type="bibr" rid="ref33">33</xref>]. To mitigate such risks, researchers can consider structuring participant reimbursements in a way that reduces the financial rewards of fraudulent responses without undermining overall study participation. For example, framing the incentive as a raffle instead of a guaranteed reimbursement reduced the number of bot responses from 633 to 23 in one particular study [<xref ref-type="bibr" rid="ref29">29</xref>]. This may be explained by the way in which bots are programmed to seek out guaranteed financial incentive surveys, and thus avoid surveys that include random chance incentives [<xref ref-type="bibr" rid="ref29">29</xref>]. Furthermore, researchers should appropriately match incentives to study requirements and payment norms in the area, as incentives that were seen as higher than others attracted more fraudulent responses [<xref ref-type="bibr" rid="ref39">39</xref>]. Lastly, separating the reimbursement form and letting participants be directed to it only after finishing the main survey can help ensure only genuine participants receive the incentive [<xref ref-type="bibr" rid="ref29">29</xref>].</p></sec><sec id="s2-5"><title>Recruitment Strategies: Sampling Methods</title><p>The way in which researchers derive their sample is key, as it is the basis for the entire study. Probability sampling is not typically used with online surveys, simply because it is difficult to establish a probability sample of the whole population on the internet; the population of interest may not necessarily be online, and even those online may not use the internet frequently enough to be captured [<xref ref-type="bibr" rid="ref40">40</xref>,<xref ref-type="bibr" rid="ref41">41</xref>]. However, there are workarounds if a probability sample is required, such as randomly sampling and contacting participation via another method (eg, telephone or mail) and doing the actual survey online, or defining a target population (eg, health care workers in a specific geographical area) and inviting all of them to complete the online survey [<xref ref-type="bibr" rid="ref40">40</xref>]. While probability sampling often leads to a more representative sample of the whole population, it may not necessarily be the most ideal sampling method, as not all research requires the entire population, and it is difficult to obtain representative numbers of smaller subpopulations [<xref ref-type="bibr" rid="ref23">23</xref>]. Instead, researchers can consider nonprobability sampling, which is ideal for sampling hard-to-reach or hidden populations, where the members are unknown and make up a small proportion of the whole population, meaning that it is difficult to get a sufficiently large sample through probability sampling [<xref ref-type="bibr" rid="ref22">22</xref>]. Regardless of whether probability or nonprobability sampling is used, the majority of those selected to be sampled should complete the survey; if not, there is a high potential for the sample to be skewed in an unpredictable direction due to potential selection bias and the introduction of large amounts of sampling error [<xref ref-type="bibr" rid="ref23">23</xref>].</p><p>One of the most common sampling methods used in online surveys is river sampling, which is a nonprobability method based upon convenience sampling. In river sampling, researchers place a survey link somewhere it is likely to be seen by members of the target population, such as a web page or through email [<xref ref-type="bibr" rid="ref24">24</xref>]. However, while cost-effective and convenient, it suffers from coverage bias, where not every subgroup is represented equally on digital platforms, leading to results being most representative of active users of the chosen platform, and not necessarily the target population. It is a great method to quickly recruit large numbers of respondents, but is particularly vulnerable to attacks by fraudulent respondents due to the easy access to the survey [<xref ref-type="bibr" rid="ref24">24</xref>]. Despite its limitations, river sampling can be very beneficial if the study is focused on a narrower population, such as undergraduate students at a university [<xref ref-type="bibr" rid="ref1">1</xref>].</p><p>Another common sampling method used in online surveys is commercial panels, which can be either probability-based (eg, KnowledgePanel or AmeriSpeak) or nonprobability-based (eg, Qualtrics Panels and Dynata) [<xref ref-type="bibr" rid="ref28">28</xref>]. Commercial online panel providers include participants who are prerecruited and have agreed to take part in research and differ based on how their participants are recruited. Probability-based panels tend to rely on &#x201C;offline&#x201D; recruitment, such as random digit dialing or address-based sampling to invite participants to join, while non-probability-based panels run on a volunteer or opt-in basis [<xref ref-type="bibr" rid="ref42">42</xref>]. In general, commercial online panels are an alternative option that offer respondent pools that better mirror the demographic compositions of the general population compared to river sampling, but at a greater monetary cost [<xref ref-type="bibr" rid="ref24">24</xref>]. Commercial panel providers ensure quality by verifying their respondents beforehand, placing the onus of respondent verification on them rather than the researcher [<xref ref-type="bibr" rid="ref1">1</xref>]. Furthermore, researchers can impose quotas on specific demographic criteria, forcing participants to better approximate the target population. Researchers must keep in mind that even with quotas, online panel providers may still be unable to fully reach hidden populations, as subpopulations within the participant pool may differ in the frequency of how often they actually access the panel and complete surveys, resulting in results that reflect those who are most online rather than being fully representative of the target population [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref43">43</xref>]. The quality of online commercial panels also depends on how it is managed, and researchers should take care in checking how panel participants are recruited or managed before committing to a specific panel in order to maintain a high level of accuracy in the data collected [<xref ref-type="bibr" rid="ref30">30</xref>].</p><p>Lastly, crowdsourcing platforms such as Amazon&#x2019;s Mechanical Turk and Prolific serve as non&#x2013;probability-based online marketplaces where potential participants are &#x201C;hired&#x201D; to participate in research, but differ from panel providers in that their participant pools are not necessarily curated [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref44">44</xref>]. This means that researchers need to expend extra effort in screening participants for specific target populations, and larger sample sizes are required in order to sufficiently power analysis [<xref ref-type="bibr" rid="ref45">45</xref>]. Another benefit is that large sample sizes can be generated quickly compared to other methods. While not always ideal for research that requires a sample that is representative of the general population, researchers should leverage the unique strengths of crowdsourcing platforms when studying nondemographic subpopulations or hard-to-reach subpopulations [<xref ref-type="bibr" rid="ref46">46</xref>]. Such platforms are suitable for experimental work, allowing researchers to establish relationships and effect sizes between different conceptual variables, even if samples are not representative [<xref ref-type="bibr" rid="ref47">47</xref>]. Although the conclusions drawn are limited, the information is still useful to begin testing hypotheses in diverse samples, making inferences into subpopulation characteristics, and showing that specific phenomena exist [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref48">48</xref>].</p><p>A summary of the sampling platforms and methods for online surveys with their respective potential advantages and disadvantages can be found in <xref ref-type="table" rid="table1">Table 1</xref>.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Sampling platforms and methods for online surveys with respective potential advantages and disadvantages.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom" colspan="2">Sampling method and survey platforms</td><td align="left" valign="bottom">Potential advantages</td><td align="left" valign="bottom">Potential disadvantages</td><td align="left" valign="bottom">Implications for research</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="5">Probability sampling</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Random sampling: recruitment through an online sampling frame and randomization of participants</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Generalizability of findings to an entire population</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>An online sampling frame may not be available or obtainable for a given population</p></list-item><list-item><p>Not all surveys require populations representation</p></list-item><list-item><p>Large probability sampling approaches may lead to underrepresentation of underserved populations</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Suitable approach for research among online populations with a defined sampling frame, with a goal of characterizing the true spread or prevalence of phenomena to be investigated</p></list-item></list></td></tr><tr><td align="left" valign="top" colspan="5">Nonprobability sampling</td></tr><tr><td align="left" valign="top" rowspan="2"/><td align="left" valign="top">River sampling: recruitment through sites where target populations visit</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Cost-effective and convenient approach</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Coverage bias: Not every subgroup or target population of interest may be equally represented on digital platforms</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Suitable approach for research among harder-to-reach or underserved populations</p></list-item><list-item><p>May require additional weighting or offline sampling approaches to better characterize target populations</p></list-item></list></td></tr><tr><td align="left" valign="top">Crowdsourcing platforms: recruitment through online marketplaces where potential participants are &#x2018;hired&#x2019; to participate in research</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Larger sample sizes can be generated quickly compared to other methods</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Participant characteristics are not typically curated to be representative of a target population</p></list-item><list-item><p>More screening procedures and larger sample sizes are required to sufficiently power analyses</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Suitable approach for research that uses inferential statistics, controlling for relevant confounders that are present in a wide range of participants</p></list-item></list></td></tr><tr><td align="left" valign="top" colspan="5">Either probability or nonprobability sampling depending on recruitment</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Online panels: recruitment through prerecruited individuals who have agreed to take part in research</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Allow for more representative samples through quota sampling</p></list-item><list-item><p>Allow for methods that verify the identity of participants</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Greater monetary cost</p></list-item><list-item><p>In spite of quotas, certain populations remain underserved through online panels</p></list-item><list-item><p>Participant pool may reflect only the most active platform users and not all members of online panels</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Suitable approach for research among the general population, especially with the potential for government-backed verification and quota sampling</p></list-item><list-item><p>Limitations that reflect differences between participants who participate in online panels versus those who don&#x2019;t need to be accounted for</p></list-item></list></td></tr></tbody></table></table-wrap></sec></sec><sec id="s3"><title>Postdata Collection: Strategies to Identify Fraudulent Data</title><p>Measures to reduce the inclusion of fraudulent data in online surveys postdata collection involve analyzing metadata to flag suspicious responses and removing low-quality data to ensure the final dataset is more reliable and accurate.</p><sec id="s3-1"><title>Metadata Checks: Duplicate Emails</title><p>One of the primary methods of metadata checks is to flag responses that share identical email addresses with previously enrolled participants, or similar email addresses with slight variations in the order of the letters or numbers [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref8">8</xref>]. One example removed any email address that had numbers exceeding 4 digits, as it is a sign of a bot-generated email [<xref ref-type="bibr" rid="ref33">33</xref>]. To further ensure the quality of such checks, researchers can let the email addresses be checked for any discrepancies by a third researcher after conducting the previous 2 steps of removal based on protocol or duplicates [<xref ref-type="bibr" rid="ref29">29</xref>]. Reimbursements should be done after all the checks have been conducted. In the case that a genuine participant was excluded erroneously, the research team&#x2019;s contacts should be available for the participant to contact.</p></sec><sec id="s3-2"><title>Metadata Checks: IP Addresses and VPN</title><p>While IP addresses can be shared among legitimate respondents in communal spaces or households, patterns of identical IP addresses across multiple submissions can indicate fraud. Researchers can consider implementing a feature to flag multiple submissions from the same IP address [<xref ref-type="bibr" rid="ref33">33</xref>]. However, this alone should not be used as an automatic rejection measure, but considered as an aid in the review of responses for potential data quality issues. Although IP addresses can act as a proxy for the legitimacy of a response, such as the geographic location, it is not completely accurate, as participants can fake their IP address or use the same computer [<xref ref-type="bibr" rid="ref11">11</xref>].</p><p>Furthermore, participants may use a virtual private network (VPN) or a virtual private server to change their IP address [<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref31">31</xref>]. The use of VPNs and proxies to mask respondents&#x2019; true locations poses a significant challenge in ensuring data validity. Such challenges can be addressed by designing tests to detect discrepancies between the time zone reported by the participant&#x2019;s browser and the IP address&#x2019;s inferred time zone [<xref ref-type="bibr" rid="ref31">31</xref>]. If the time zones differed significantly, the response can then be flagged as potentially suspicious. However, researchers must consider that the average internet user is also more likely to make use of VPNs now, meaning that IP addresses may not necessarily be a signal of a low-quality response [<xref ref-type="bibr" rid="ref15">15</xref>]. Despite its complications, IP address duplication checks are one of the most common data quality checks [<xref ref-type="bibr" rid="ref39">39</xref>].</p></sec><sec id="s3-3"><title>Metadata Checks: Survey Timing and Duration</title><p>Another aspect of metadata checks is the analysis of survey timing and duration. Rapid survey submissions, where multiple surveys are completed within an unrealistically short timeframe, can often indicate bot activity [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref11">11</xref>]. Suspicious submissions can be identified by comparing the start and stop times of surveys, flagging those completed within a minute of each other [<xref ref-type="bibr" rid="ref8">8</xref>]. Survey duration can also be used as a key indicator of fraudulent responses, as overly rapid completion times are unrealistic for genuine participants [<xref ref-type="bibr" rid="ref39">39</xref>]. Studies have shown that valid respondents typically spend more time on surveys, with response time distributions skewed to the right on a histogram visualization [<xref ref-type="bibr" rid="ref31">31</xref>]. Pretests or soft launches can help establish realistic time boundaries for survey completion, further enhancing the detection of improbable durations [<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref49">49</xref>].</p></sec><sec id="s3-4"><title>Data Quality: Duplicate and Improbable Free-Text Responses</title><p>By incorporating multiple qualitative questions and making at least one a requirement for survey submission, researchers can more easily identify automatic responses [<xref ref-type="bibr" rid="ref33">33</xref>]. One possible protocol is that a response is seen as identical if it is repeated 100 times or more for single words, 10 times or more for 2-word entries, and 3 times or more for entries of 3 or more words [<xref ref-type="bibr" rid="ref8">8</xref>]. In one study, researchers found that qualitative survey questions, despite being optional, were instrumental in identifying 88 (13.3%) bot responses through exact duplicate answers unlikely to occur by chance [<xref ref-type="bibr" rid="ref29">29</xref>]. The effectiveness of qualitative questions in detecting bots is supported by research, which found that bots struggled with open-ended questions requiring a minimum response length, suggesting that incorporating multiple such questions can enhance bot detection [<xref ref-type="bibr" rid="ref29">29</xref>].</p></sec><sec id="s3-5"><title>Data Quality: Suspicious Response Patterns and Data Inconsistencies</title><p>Detecting straight-lining, consistencies, or patterns in the response of a participant across a series of questions is crucial for identifying low-effort or automated responses. Thus, it is important to check for uniform answers across grid questions or obvious patterns, such as sequential responses [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref11">11</xref>]. These patterns can indicate a lack of genuine engagement with the survey content. Similarly, verifying selection patterns in multiple response questions and detecting low differentiation in answers also identifies low-quality responses [<xref ref-type="bibr" rid="ref38">38</xref>]. Furthermore, detecting inconsistencies in responses helps identify respondents who may not be providing truthful or accurate information [<xref ref-type="bibr" rid="ref29">29</xref>]. It is important to exclude participants with inconsistent responses, such as mismatched ages and birth dates, as this ensures that survey data accurately represents the target population [<xref ref-type="bibr" rid="ref11">11</xref>]. For example, one study excluded participants whose answers to questions about sex, gender, and sexuality were contradictory, eg, &#x201C;I have had insertive vaginal sex with multiple female partners,&#x201D; yet &#x201C;none of my partners have vaginas&#x201D; [<xref ref-type="bibr" rid="ref11">11</xref>]. Implementing comprehensive checks for straight-lining, response patterns, and internal consistency helps filter out low-quality data, ensuring that the final dataset is reliable and valid.</p><p>Fraudulent data can occur through a variety of ways, and thus poses distinct challenges for the prevention and detection of such responses [<xref ref-type="bibr" rid="ref39">39</xref>]. It is important to consider which methods would be best suited for each study, and using multiple methods in a multipronged approach can help to account for the shortcomings of any single method [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref11">11</xref>]. Researchers can consider taking a dynamic approach with evolving screening protocols based on the type of responses received across the study in order to keep up with ever-changing fraud attempts [<xref ref-type="bibr" rid="ref38">38</xref>]. In addition, when deciding the threshold required to identify a response as fraudulent or low quality, researchers could assign points to indicators of suspicious activity, which reduces the likelihood of incorrectly excluding valid data [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref38">38</xref>]. Having a point system allows specific responses to be flagged for further investigation and establishes a cutoff score for invalid responses [<xref ref-type="bibr" rid="ref30">30</xref>]. This would also follow best practices in which responses would be removed based on multiple issues rather than a single failure.</p></sec></sec><sec id="s4"><title>Case Study</title><p>An example of a study that was able to maintain a high level of rigor throughout its run was the Restore study by Dewitt et al [<xref ref-type="bibr" rid="ref38">38</xref>]. The study wanted to recruit gay, bisexual, and other men who have sex with men who were treated for prostate cancer. The challenges they faced with recruitment were because it was a minority within a minority group, and it largely impacted an older demographic, which meant that the process of recruitment had to be as streamlined as possible due to the unclear level of internet familiarity the target population would have. Ultimately, they focused recruitment on a community partner&#x2019;s email list, who focused on providing support for survivors of cancer, as well as Facebook groups of prostate cancer community organizations, which would be a form of targeted river sampling. They hosted their survey on Qualtrics, a web-based survey platform, and used built-in protection features such as &#x201C;Prevent Ballot Box Stuffing,&#x201D; which prevents multiple submissions based on browser cookies, and &#x201C;Prevent Indexing,&#x201D; which blocks the survey from being indexed and found on search engines. In addition, they made use of multiple data quality control measures, such as eligibility screeners and manual data validation using both automated and hybrid protocols that flagged suspicious survey entries to the researchers. For example, survey metadata was checked (eg, short response timing, IP addresses that did not match geographic locations), and open-ended responses were evaluated to see if they were reasonable or possible. They ran into issues with a spam attack through the Facebook link and noticed quickly due to a sudden influx of low-quality survey attempts. A further problem was that the answer patterns began to evolve, where initial spam attempts were clearly fraudulent, but later entries began to provide more probable responses. Later entries were only identifiable by manually checking the process analytics and interpreting each survey response&#x2019;s data. After some review, the researchers created a new copy of the survey and imposed a stricter recruitment protocol with more validation steps and focused their recruitment through community partners&#x2019; email lists. They did not provide reimbursement to surveys identified as invalid but provided a method for participants to contact them to verify their answers and receive their compensation if they were unwittingly identified as invalid, but not a single invalid respondent followed up on this. The restore study is a great case study of a dynamic protocol that was able to respond to shifting patterns of fraudulent responses and successfully maximize the rigor of their study.</p></sec><sec id="s5"><title>Challenges and Opportunities of Artificial Intelligence</title><p>One issue of concern is that bots are becoming increasingly sophisticated and are able to better replicate human behavior, allowing them to potentially overwhelm measures put into place and more easily gain access to online surveys [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref50">50</xref>]. CAPTCHAs and open-ended questions can now be solved by high-end bots, and the rise of artificial intelligence (AI) software like ChatGPT means that open-ended questions can also be answered automatically [<xref ref-type="bibr" rid="ref50">50</xref>,<xref ref-type="bibr" rid="ref51">51</xref>]. In a study to see if AI-generated texts could be differentiated from actual people, AI-generated texts were only accurately recognized 40.45% of the time, meaning that even manual evaluation of answers may not be sufficient to weed out invalid responses [<xref ref-type="bibr" rid="ref52">52</xref>]. The level of sophistication that fraudulent responders can achieve using AI and bots only continues to grow, and they pose an incredible threat to data quality when conducting online surveys as they can easily submit large numbers of responses while masquerading as multiple valid respondents [<xref ref-type="bibr" rid="ref27">27</xref>]. As such, ensuring that surveys do not fall into the wrong hands is even more critical. Researchers should embrace the fact that different sampling methods reach different demographics and even consider multimodal and blended samples where multiple mediums of recruitment and communication complement each other to obtain higher quality samples [<xref ref-type="bibr" rid="ref24">24</xref>]. Researchers themselves can leverage technological growth, such as by using automated fraud detection systems on web-based survey platforms. For example, Qualtrics includes RelevantID, which uses machine learning to evaluate behavioral data, browser interactions, and other metadata to identify the likelihood of fraud [<xref ref-type="bibr" rid="ref8">8</xref>]. However, these automated systems should be integrated with manual review or other fraud detection methods to maximize effectiveness.</p></sec><sec id="s6"><title>Ethical Considerations</title><p>In addition, researchers need to consider ethical issues when deciding what data should be collected for screening measures. For example, one of the most common methods used in assessing data quality is to collect IP addresses, or personal details such as email addresses and phone numbers; yet, these undermine one of the key advantages of online surveys, which is anonymity [<xref ref-type="bibr" rid="ref27">27</xref>]. If a study is particularly sensitive or collects details of criminalized behavior, collecting IP addresses and personal identifiers may introduce additional complications and concerns about data breaches or loss of confidentiality and anonymity. One possible way to overcome this is to have personal details collected in a separate survey, such that key identifiers are not able to be tied to a specific response. Second, the methods reported online are intrinsically limited because studies are unable to definitively categorize which responses are valid or invalid, so we cannot truly evaluate the precision of techniques or their effectiveness. Furthermore, methods are often not verified as per randomized controlled trials and are mostly developed through empirical research when researchers experience fraudulent data in their studies. It would be of value for research to be done comparing different measures against some unfalsifiable assessment of real data in order to accurately determine the effectiveness of such measures. There is still significant room for strategies to develop, and it should be a key focus for upcoming research due to the rise in online surveys being a premier choice for survey-based research.</p></sec><sec id="s7" sec-type="conclusions"><title>Conclusion</title><p>Methodological rigor is key for any study regardless of its modality, but online surveys require researchers to consider broader methodological considerations as compared to offline research [<xref ref-type="bibr" rid="ref53">53</xref>]. Although a myriad of ways in which researchers can increase the rigor of their online surveys are presented in this article, there is no singular &#x201C;best&#x201D; method as every research method has its own limitations [<xref ref-type="bibr" rid="ref54">54</xref>]. Online research appears superficially similar to offline research, but its characteristics affect the entire research process from initial design to reporting the findings in ways that influence their implementation [<xref ref-type="bibr" rid="ref55">55</xref>,<xref ref-type="bibr" rid="ref56">56</xref>]. Ultimately, researchers must consider the wide range of factors unique to the requirements of their own studies and make use of a combination of various resources and methods to increase the quality of research published.</p></sec></body><back><fn-group><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">VPN</term><def><p>virtual private network</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Stantcheva</surname><given-names>S</given-names> </name></person-group><article-title>How to run surveys: A guide to creating your own identifying variation and revealing the invisible</article-title><source>Annu Rev Econom</source><year>2023</year><month>09</month><day>13</day><volume>15</volume><issue>1</issue><fpage>205</fpage><lpage>234</lpage><pub-id pub-id-type="doi">10.1146/annurev-economics-091622-010157</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Alam</surname><given-names>I</given-names> </name><name name-style="western"><surname>Khusro</surname><given-names>S</given-names> </name><name name-style="western"><surname>Rauf</surname><given-names>A</given-names> </name><name name-style="western"><surname>Zaman</surname><given-names>Q</given-names> </name></person-group><article-title>Conducting surveys and data collection: From traditional to mobile and SMS-based surveys</article-title><source>Pakjstatoperres</source><year>2014</year><volume>10</volume><issue>2</issue><fpage>169</fpage><pub-id pub-id-type="doi">10.18187/pjsor.v10i2.758</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ebert</surname><given-names>JF</given-names> </name><name name-style="western"><surname>Huibers</surname><given-names>L</given-names> </name><name name-style="western"><surname>Christensen</surname><given-names>B</given-names> </name><name name-style="western"><surname>Christensen</surname><given-names>MB</given-names> </name></person-group><article-title>Paper- or web-based questionnaire invitations as a method for data collection: Cross-sectional comparative study of differences in response rate, completeness of data, and financial cost</article-title><source>J Med Internet Res</source><year>2018</year><month>01</month><day>23</day><volume>20</volume><issue>1</issue><fpage>e24</fpage><pub-id pub-id-type="doi">10.2196/jmir.8353</pub-id><pub-id pub-id-type="medline">29362206</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Andrade</surname><given-names>C</given-names> </name></person-group><article-title>The limitations of online surveys</article-title><source>Indian J Psychol Med</source><year>2020</year><month>11</month><volume>42</volume><issue>6</issue><fpage>575</fpage><lpage>576</lpage><pub-id pub-id-type="doi">10.1177/0253717620957496</pub-id><pub-id pub-id-type="medline">33354086</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Evans</surname><given-names>JR</given-names> </name><name name-style="western"><surname>Mathur</surname><given-names>A</given-names> </name></person-group><article-title>The value of online surveys: a look back and a look ahead</article-title><source>INTR</source><year>2018</year><month>08</month><day>6</day><volume>28</volume><issue>4</issue><fpage>854</fpage><lpage>887</lpage><pub-id pub-id-type="doi">10.1108/IntR-03-2018-0089</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bybee</surname><given-names>S</given-names> </name><name name-style="western"><surname>Cloyes</surname><given-names>K</given-names> </name><name name-style="western"><surname>Ellington</surname><given-names>L</given-names> </name><name name-style="western"><surname>Baucom</surname><given-names>B</given-names> </name><name name-style="western"><surname>Supiano</surname><given-names>K</given-names> </name><name name-style="western"><surname>Mooney</surname><given-names>K</given-names> </name></person-group><article-title>Bots and nots: Safeguarding online survey research with underrepresented and diverse populations</article-title><source>Psychol Sex</source><year>2022</year><volume>13</volume><issue>4</issue><fpage>901</fpage><lpage>911</lpage><pub-id pub-id-type="doi">10.1080/19419899.2021.1936617</pub-id><pub-id pub-id-type="medline">36439051</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bauermeister</surname><given-names>J</given-names> </name><name name-style="western"><surname>Pingel</surname><given-names>E</given-names> </name><name name-style="western"><surname>Zimmerman</surname><given-names>M</given-names> </name><name name-style="western"><surname>Couper</surname><given-names>M</given-names> </name><name name-style="western"><surname>Carballo-Di&#x00E9;guez</surname><given-names>A</given-names> </name><name name-style="western"><surname>Strecher</surname><given-names>VJ</given-names> </name></person-group><article-title>Data quality in web-based HIV/AIDS research: Handling invalid and suspicious data</article-title><source>Field methods</source><year>2012</year><month>08</month><day>1</day><volume>24</volume><issue>3</issue><fpage>272</fpage><lpage>291</lpage><pub-id pub-id-type="doi">10.1177/1525822X12443097</pub-id><pub-id pub-id-type="medline">23180978</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bonett</surname><given-names>S</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>W</given-names> </name><name name-style="western"><surname>Sexton Topper</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Assessing and improving data integrity in web-based surveys: Comparison of fraud detection systems in a COVID-19 study</article-title><source>JMIR Form Res</source><year>2024</year><month>01</month><day>12</day><volume>8</volume><issue>1</issue><fpage>e47091</fpage><pub-id pub-id-type="doi">10.2196/47091</pub-id><pub-id pub-id-type="medline">38214962</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Watson</surname><given-names>NL</given-names> </name><name name-style="western"><surname>Mull</surname><given-names>KE</given-names> </name><name name-style="western"><surname>Heffner</surname><given-names>JL</given-names> </name><name name-style="western"><surname>McClure</surname><given-names>JB</given-names> </name><name name-style="western"><surname>Bricker</surname><given-names>JB</given-names> </name></person-group><article-title>Participant recruitment and retention in remote eHealth Intervention trials: Methods and lessons learned from a large randomized controlled trial of two web-based smoking interventions</article-title><source>J Med Internet Res</source><year>2018</year><month>08</month><day>24</day><volume>20</volume><issue>8</issue><fpage>e10351</fpage><pub-id pub-id-type="doi">10.2196/10351</pub-id><pub-id pub-id-type="medline">30143479</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McInroy</surname><given-names>LB</given-names> </name><name name-style="western"><surname>Beer</surname><given-names>OWJ</given-names> </name></person-group><article-title>Adapting vignettes for internet-based research: eliciting realistic responses to the digital milieu</article-title><source>Int J Soc Res Methodol</source><year>2022</year><month>05</month><day>4</day><volume>25</volume><issue>3</issue><fpage>335</fpage><lpage>347</lpage><pub-id pub-id-type="doi">10.1080/13645579.2021.1901440</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Teitcher</surname><given-names>JEF</given-names> </name><name name-style="western"><surname>Bockting</surname><given-names>WO</given-names> </name><name name-style="western"><surname>Bauermeister</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Hoefer</surname><given-names>CJ</given-names> </name><name name-style="western"><surname>Miner</surname><given-names>MH</given-names> </name><name name-style="western"><surname>Klitzman</surname><given-names>RL</given-names> </name></person-group><article-title>Detecting, preventing, and responding to &#x201C;fraudsters&#x201D; in internet research: ethics and tradeoffs</article-title><source>J Law Med Ethics</source><year>2015</year><volume>43</volume><issue>1</issue><fpage>116</fpage><lpage>133</lpage><pub-id pub-id-type="doi">10.1111/jlme.12200</pub-id><pub-id pub-id-type="medline">25846043</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gnambs</surname><given-names>T</given-names> </name><name name-style="western"><surname>Kaspar</surname><given-names>K</given-names> </name></person-group><article-title>Disclosure of sensitive behaviors across self-administered survey modes: a meta-analysis</article-title><source>Behav Res Methods</source><year>2015</year><month>12</month><volume>47</volume><issue>4</issue><fpage>1237</fpage><lpage>1259</lpage><pub-id pub-id-type="doi">10.3758/s13428-014-0533-4</pub-id><pub-id pub-id-type="medline">25410404</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bratan</surname><given-names>T</given-names> </name><name name-style="western"><surname>Aichinger</surname><given-names>H</given-names> </name><name name-style="western"><surname>Brkic</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Impact of the COVID-19 pandemic on ongoing health research: an ad hoc survey among investigators in Germany</article-title><source>BMJ Open</source><year>2021</year><month>12</month><day>6</day><volume>11</volume><issue>12</issue><fpage>e049086</fpage><pub-id pub-id-type="doi">10.1136/bmjopen-2021-049086</pub-id><pub-id pub-id-type="medline">34872995</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Adi Syani</surname><given-names>P</given-names> </name><name name-style="western"><surname>Rahiem</surname><given-names>MDH</given-names> </name><name name-style="western"><surname>Subchi</surname><given-names>I</given-names> </name><name name-style="western"><surname>Suryani</surname><given-names>R</given-names> </name><name name-style="western"><surname>Kurniawan</surname><given-names>F</given-names> </name></person-group><article-title>COVID-19: accelerating digital transformation for university&#x2019;s research administration</article-title><conf-name>2020 8th International Conference on Cyber and IT Service Management (CITSM)</conf-name><fpage>1</fpage><lpage>6</lpage><pub-id pub-id-type="doi">10.1109/CITSM50537.2020.9268913</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Johnson</surname><given-names>MS</given-names> </name><name name-style="western"><surname>Adams</surname><given-names>VM</given-names> </name><name name-style="western"><surname>Byrne</surname><given-names>J</given-names> </name></person-group><article-title>Addressing fraudulent responses in online surveys: Insights from a web&#x2010;based participatory mapping study</article-title><source>People and Nature</source><year>2024</year><month>02</month><volume>6</volume><issue>1</issue><fpage>147</fpage><lpage>164</lpage><pub-id pub-id-type="doi">10.1002/pan3.10557</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bowen</surname><given-names>AM</given-names> </name><name name-style="western"><surname>Daniel</surname><given-names>CM</given-names> </name><name name-style="western"><surname>Williams</surname><given-names>ML</given-names> </name><name name-style="western"><surname>Baird</surname><given-names>GL</given-names> </name></person-group><article-title>Identifying multiple submissions in Internet research: preserving data integrity</article-title><source>AIDS Behav</source><year>2008</year><month>11</month><volume>12</volume><issue>6</issue><fpage>964</fpage><lpage>973</lpage><pub-id pub-id-type="doi">10.1007/s10461-007-9352-2</pub-id><pub-id pub-id-type="medline">18240015</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chandler</surname><given-names>JJ</given-names> </name><name name-style="western"><surname>Paolacci</surname><given-names>G</given-names> </name></person-group><article-title>Lie for a dime: when most prescreening responses are honest but most study participants are impostors</article-title><source>Soc Psychol Personal Sci</source><year>2017</year><volume>8</volume><issue>5</issue><fpage>500</fpage><lpage>508</lpage><pub-id pub-id-type="doi">10.1177/1948550617698203</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cimpian</surname><given-names>JR</given-names> </name><name name-style="western"><surname>Timmer</surname><given-names>JD</given-names> </name><name name-style="western"><surname>Birkett</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Marro</surname><given-names>RL</given-names> </name><name name-style="western"><surname>Turner</surname><given-names>BC</given-names> </name><name name-style="western"><surname>Phillips</surname><given-names>GL</given-names>  <suffix>II</suffix></name></person-group><article-title>Bias from potentially mischievous responders on large-scale estimates of lesbian, gay, bisexual, or questioning (LGBQ)&#x2013;heterosexual youth health disparities</article-title><source>Am J Public Health</source><year>2018</year><month>11</month><volume>108</volume><issue>S4</issue><fpage>S258</fpage><lpage>S265</lpage><pub-id pub-id-type="doi">10.2105/AJPH.2018.304407</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Grey</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Konstan</surname><given-names>J</given-names> </name><name name-style="western"><surname>Iantaffi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Wilkerson</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Galos</surname><given-names>D</given-names> </name><name name-style="western"><surname>Rosser</surname><given-names>BRS</given-names> </name></person-group><article-title>An updated protocol to detect invalid entries in an online survey of men who have sex with men (MSM): how do valid and invalid submissions compare?</article-title><source>AIDS Behav</source><year>2015</year><month>10</month><volume>19</volume><issue>10</issue><fpage>1928</fpage><lpage>1937</lpage><pub-id pub-id-type="doi">10.1007/s10461-015-1033-y</pub-id><pub-id pub-id-type="medline">25805443</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Storozuk</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ashley</surname><given-names>M</given-names> </name><name name-style="western"><surname>Delage</surname><given-names>V</given-names> </name><name name-style="western"><surname>Maloney</surname><given-names>EA</given-names> </name></person-group><article-title>Got bots? Practical recommendations to protect online survey data from bot attacks</article-title><source>TQMP</source><year>2020</year><volume>16</volume><issue>5</issue><fpage>472</fpage><lpage>481</lpage><pub-id pub-id-type="doi">10.20982/tqmp.16.5.p472</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ward</surname><given-names>MK</given-names> </name><name name-style="western"><surname>Meade</surname><given-names>AW</given-names> </name></person-group><article-title>Dealing with careless responding in survey data: Prevention, identification, and recommended best practices</article-title><source>Annu Rev Psychol</source><year>2023</year><month>01</month><day>18</day><volume>74</volume><fpage>577</fpage><lpage>596</lpage><pub-id pub-id-type="doi">10.1146/annurev-psych-040422-045007</pub-id><pub-id pub-id-type="medline">35973734</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Berndt</surname><given-names>AE</given-names> </name></person-group><article-title>Sampling methods</article-title><source>J Hum Lact</source><year>2020</year><month>05</month><volume>36</volume><issue>2</issue><fpage>224</fpage><lpage>226</lpage><pub-id pub-id-type="doi">10.1177/0890334420906850</pub-id><pub-id pub-id-type="medline">32155099</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tyrer</surname><given-names>S</given-names> </name><name name-style="western"><surname>Heyman</surname><given-names>B</given-names> </name></person-group><article-title>Sampling in epidemiological research: issues, hazards and pitfalls</article-title><source>BJPsych Bull</source><year>2016</year><month>04</month><volume>40</volume><issue>2</issue><fpage>57</fpage><lpage>60</lpage><pub-id pub-id-type="doi">10.1192/pb.bp.114.050203</pub-id><pub-id pub-id-type="medline">27087985</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lehdonvirta</surname><given-names>V</given-names> </name><name name-style="western"><surname>Oksanen</surname><given-names>A</given-names> </name><name name-style="western"><surname>R&#x00E4;s&#x00E4;nen</surname><given-names>P</given-names> </name><name name-style="western"><surname>Blank</surname><given-names>G</given-names> </name></person-group><article-title>Social media, web, and panel surveys: Using non&#x2010;probability samples in social and policy research</article-title><source>Policy &#x0026; Internet</source><year>2021</year><month>03</month><volume>13</volume><issue>1</issue><fpage>134</fpage><lpage>155</lpage><pub-id pub-id-type="doi">10.1002/poi3.238</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Arias</surname><given-names>VB</given-names> </name><name name-style="western"><surname>Garrido</surname><given-names>LE</given-names> </name><name name-style="western"><surname>Jenaro</surname><given-names>C</given-names> </name><name name-style="western"><surname>Mart&#x00ED;nez-Molina</surname><given-names>A</given-names> </name><name name-style="western"><surname>Arias</surname><given-names>B</given-names> </name></person-group><article-title>A little garbage in, lots of garbage out: Assessing the impact of careless responding in personality survey data</article-title><source>Behav Res Methods</source><year>2020</year><month>12</month><volume>52</volume><issue>6</issue><fpage>2489</fpage><lpage>2505</lpage><pub-id pub-id-type="doi">10.3758/s13428-020-01401-8</pub-id><pub-id pub-id-type="medline">32462604</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huang</surname><given-names>JL</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>M</given-names> </name><name name-style="western"><surname>Bowling</surname><given-names>NA</given-names> </name></person-group><article-title>Insufficient effort responding: examining an insidious confound in survey data</article-title><source>J Appl Psychol</source><year>2015</year><month>05</month><volume>100</volume><issue>3</issue><fpage>828</fpage><lpage>845</lpage><pub-id pub-id-type="doi">10.1037/a0038510</pub-id><pub-id pub-id-type="medline">25495093</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Agans</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Schade</surname><given-names>SA</given-names> </name><name name-style="western"><surname>Hanna</surname><given-names>SR</given-names> </name><name name-style="western"><surname>Chiang</surname><given-names>SC</given-names> </name><name name-style="western"><surname>Shirzad</surname><given-names>K</given-names> </name><name name-style="western"><surname>Bai</surname><given-names>S</given-names> </name></person-group><article-title>The inaccuracy of data from online surveys: A cautionary analysis</article-title><source>Qual Quant</source><year>2024</year><month>06</month><volume>58</volume><issue>3</issue><fpage>2065</fpage><lpage>2086</lpage><pub-id pub-id-type="doi">10.1007/s11135-023-01733-5</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Peer</surname><given-names>E</given-names> </name><name name-style="western"><surname>Rothschild</surname><given-names>D</given-names> </name><name name-style="western"><surname>Gordon</surname><given-names>A</given-names> </name><name name-style="western"><surname>Evernden</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Damer</surname><given-names>E</given-names> </name></person-group><article-title>Data quality of platforms and panels for online behavioral research</article-title><source>Behav Res Methods</source><year>2022</year><month>08</month><volume>54</volume><issue>4</issue><fpage>1643</fpage><lpage>1662</lpage><pub-id pub-id-type="doi">10.3758/s13428-021-01694-3</pub-id><pub-id pub-id-type="medline">34590289</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Griffin</surname><given-names>M</given-names> </name><name name-style="western"><surname>Martino</surname><given-names>RJ</given-names> </name><name name-style="western"><surname>LoSchiavo</surname><given-names>C</given-names> </name><etal/></person-group><article-title>Ensuring survey research data integrity in the era of internet bots</article-title><source>Qual Quant</source><year>2022</year><volume>56</volume><issue>4</issue><fpage>2841</fpage><lpage>2852</lpage><pub-id pub-id-type="doi">10.1007/s11135-021-01252-1</pub-id><pub-id pub-id-type="medline">34629553</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Brazhkin</surname><given-names>V</given-names> </name></person-group><article-title>&#x201C;I have just returned from the moon:&#x201D; online survey fraud</article-title><source>SCM</source><year>2020</year><month>03</month><day>27</day><volume>25</volume><issue>4</issue><fpage>489</fpage><lpage>503</lpage><pub-id pub-id-type="doi">10.1108/SCM-12-2019-0466</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Mink</surname><given-names>J</given-names> </name><name name-style="western"><surname>Xiong</surname><given-names>A</given-names> </name><name name-style="western"><surname>Song</surname><given-names>L</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>G</given-names> </name></person-group><article-title>Beyond bot detection: combating fraudulent online survey takers</article-title><year>2022</year><month>04</month><day>25</day><access-date>2025-08-08</access-date><conf-name>WWW &#x2019;22</conf-name><conf-loc>Virtual Event, Lyon France</conf-loc><publisher-name>Association for Computing Machinery</publisher-name><fpage>699</fpage><lpage>709</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://dl.acm.org/doi/proceedings/10.1145/3485447">https://dl.acm.org/doi/proceedings/10.1145/3485447</ext-link></comment><pub-id pub-id-type="doi">10.1145/3485447.3512230</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Meade</surname><given-names>AW</given-names> </name><name name-style="western"><surname>Craig</surname><given-names>SB</given-names> </name></person-group><article-title>Identifying careless responses in survey data</article-title><source>Psychol Methods</source><year>2012</year><month>09</month><volume>17</volume><issue>3</issue><fpage>437</fpage><lpage>455</lpage><pub-id pub-id-type="doi">10.1037/a0028085</pub-id><pub-id pub-id-type="medline">22506584</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Calderon</surname><given-names>G</given-names> </name><name name-style="western"><surname>Hager</surname><given-names>ER</given-names> </name><etal/></person-group><article-title>Identifying and preventing fraudulent responses in online public health surveys: Lessons learned during the COVID-19 pandemic</article-title><source>PLOS Glob Public Health</source><year>2023</year><volume>3</volume><issue>8</issue><fpage>e0001452</fpage><pub-id pub-id-type="doi">10.1371/journal.pgph.0001452</pub-id><pub-id pub-id-type="medline">37610999</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hauser</surname><given-names>DJ</given-names> </name><name name-style="western"><surname>Schwarz</surname><given-names>N</given-names> </name></person-group><article-title>It&#x2019;s a trap! Instructional manipulation checks prompt systematic thinking on &#x201C;tricky&#x201D; tasks</article-title><source>Sage Open</source><year>2015</year><month>04</month><day>1</day><volume>5</volume><issue>2</issue><fpage>2158244015584617</fpage><pub-id pub-id-type="doi">10.1177/2158244015584617</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kung</surname><given-names>FYH</given-names> </name><name name-style="western"><surname>Kwok</surname><given-names>N</given-names> </name><name name-style="western"><surname>Brown</surname><given-names>DJ</given-names> </name></person-group><article-title>Are attention check questions a threat to scale validity?</article-title><source>Applied Psychology</source><year>2018</year><month>04</month><volume>67</volume><issue>2</issue><fpage>264</fpage><lpage>283</lpage><pub-id pub-id-type="doi">10.1111/apps.12108</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pratt-Chapman</surname><given-names>M</given-names> </name><name name-style="western"><surname>Moses</surname><given-names>J</given-names> </name><name name-style="western"><surname>Arem</surname><given-names>H</given-names> </name></person-group><article-title>Strategies for the identification and prevention of survey fraud: Data analysis of a web-based survey</article-title><source>JMIR Cancer</source><year>2021</year><month>07</month><day>16</day><volume>7</volume><issue>3</issue><fpage>e30730</fpage><pub-id pub-id-type="doi">10.2196/30730</pub-id><pub-id pub-id-type="medline">34269685</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Roehl</surname><given-names>J</given-names> </name><name name-style="western"><surname>Harland</surname><given-names>D</given-names> </name></person-group><article-title>Imposter participants: Overcoming methodological challenges related to balancing participant privacy with data quality when using online recruitment and data collection</article-title><source>TQR</source><year>2022</year><volume>10</volume><pub-id pub-id-type="doi">10.46743/2160-3715/2022.5475</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dewitt</surname><given-names>J</given-names> </name><name name-style="western"><surname>Capistrant</surname><given-names>B</given-names> </name><name name-style="western"><surname>Kohli</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Addressing participant validity in a small internet health survey (the restore study): Protocol and recommendations for survey response validation</article-title><source>JMIR Res Protoc</source><year>2018</year><month>04</month><day>24</day><volume>7</volume><issue>4</issue><fpage>e96</fpage><pub-id pub-id-type="doi">10.2196/resprot.7655</pub-id><pub-id pub-id-type="medline">29691203</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nur</surname><given-names>AA</given-names> </name><name name-style="western"><surname>Leibbrand</surname><given-names>C</given-names> </name><name name-style="western"><surname>Curran</surname><given-names>SR</given-names> </name><name name-style="western"><surname>Votruba-Drzal</surname><given-names>E</given-names> </name><name name-style="western"><surname>Gibson-Davis</surname><given-names>C</given-names> </name></person-group><article-title>Managing and minimizing online survey questionnaire fraud: Lessons from the Triple C Project</article-title><source>Int J Soc Res Methodol</source><year>2024</year><volume>27</volume><issue>5</issue><fpage>613</fpage><lpage>619</lpage><pub-id pub-id-type="doi">10.1080/13645579.2023.2229651</pub-id><pub-id pub-id-type="medline">39494158</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ball</surname><given-names>HL</given-names> </name></person-group><article-title>Conducting online surveys</article-title><source>J Hum Lact</source><year>2019</year><month>08</month><volume>35</volume><issue>3</issue><fpage>413</fpage><lpage>417</lpage><pub-id pub-id-type="doi">10.1177/0890334419848734</pub-id><pub-id pub-id-type="medline">31084575</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Van Selm</surname><given-names>M</given-names> </name><name name-style="western"><surname>Jankowski</surname><given-names>NW</given-names> </name></person-group><article-title>Conducting online surveys</article-title><source>Qual Quant</source><year>2006</year><month>06</month><volume>40</volume><issue>3</issue><fpage>435</fpage><lpage>456</lpage><pub-id pub-id-type="doi">10.1007/s11135-005-8081-8</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kocar</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kaczmirek</surname><given-names>L</given-names> </name></person-group><article-title>A meta-analysis of worldwide recruitment rates in 23 probability-based online panels, between 2007 and 2019</article-title><source>Int J Soc Res Methodol</source><year>2024</year><month>09</month><day>2</day><volume>27</volume><issue>5</issue><fpage>589</fpage><lpage>604</lpage><pub-id pub-id-type="doi">10.1080/13645579.2023.2242202</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Im</surname><given-names>EO</given-names> </name><name name-style="western"><surname>Chee</surname><given-names>W</given-names> </name></person-group><article-title>Quota sampling in internet research: practical issues</article-title><source>Comput Inform Nurs</source><year>2011</year><month>07</month><volume>29</volume><issue>7</issue><fpage>381</fpage><lpage>385</lpage><pub-id pub-id-type="doi">10.1097/NCN.0b013e3181f9dc45</pub-id><pub-id pub-id-type="medline">20975541</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Behrend</surname><given-names>TS</given-names> </name><name name-style="western"><surname>Sharek</surname><given-names>DJ</given-names> </name><name name-style="western"><surname>Meade</surname><given-names>AW</given-names> </name><name name-style="western"><surname>Wiebe</surname><given-names>EN</given-names> </name></person-group><article-title>The viability of crowdsourcing for survey research</article-title><source>Behav Res Methods</source><year>2011</year><month>09</month><volume>43</volume><issue>3</issue><fpage>800</fpage><lpage>813</lpage><pub-id pub-id-type="doi">10.3758/s13428-011-0081-0</pub-id><pub-id pub-id-type="medline">21437749</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bentley</surname><given-names>JW</given-names> </name></person-group><article-title>Improving the statistical power and reliability of research using Amazon Mechanical Turk</article-title><source>Accounting Horizons</source><year>2021</year><month>12</month><day>1</day><volume>35</volume><issue>4</issue><fpage>45</fpage><lpage>62</lpage><pub-id pub-id-type="doi">10.2308/HORIZONS-18-052</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Smith</surname><given-names>NA</given-names> </name><name name-style="western"><surname>Sabat</surname><given-names>IE</given-names> </name><name name-style="western"><surname>Martinez</surname><given-names>LR</given-names> </name><name name-style="western"><surname>Weaver</surname><given-names>K</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>S</given-names> </name></person-group><article-title>A convenient solution: Using MTurk To sample from hard-to-reach populations</article-title><source>Ind Organ Psychol</source><year>2015</year><month>06</month><volume>8</volume><issue>2</issue><fpage>220</fpage><lpage>228</lpage><pub-id pub-id-type="doi">10.1017/iop.2015.29</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jerit</surname><given-names>J</given-names> </name><name name-style="western"><surname>Barabas</surname><given-names>J</given-names> </name></person-group><article-title>Are nonprobability surveys fit for purpose?</article-title><source>Public Opin Q</source><year>2023</year><month>11</month><day>21</day><volume>87</volume><issue>3</issue><fpage>816</fpage><lpage>840</lpage><pub-id pub-id-type="doi">10.1093/poq/nfad037</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Aguinis</surname><given-names>H</given-names> </name><name name-style="western"><surname>Villamor</surname><given-names>I</given-names> </name><name name-style="western"><surname>Ramani</surname><given-names>RS</given-names> </name></person-group><article-title>MTurk research: Review and recommendations</article-title><source>J Manage</source><year>2021</year><month>04</month><volume>47</volume><issue>4</issue><fpage>823</fpage><lpage>837</lpage><pub-id pub-id-type="doi">10.1177/0149206320969787</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>DePalma</surname><given-names>MT</given-names> </name><name name-style="western"><surname>Rizzotti</surname><given-names>MC</given-names> </name><name name-style="western"><surname>Branneman</surname><given-names>M</given-names> </name></person-group><article-title>Assessing diabetes-relevant data provided by undergraduate and crowdsourced web-based survey participants for honesty and accuracy</article-title><source>JMIR Diabetes</source><year>2017</year><month>07</month><day>12</day><volume>2</volume><issue>2</issue><fpage>e11</fpage><pub-id pub-id-type="doi">10.2196/diabetes.7473</pub-id><pub-id pub-id-type="medline">30291072</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Pace</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Threats to online surveys: Recognizing, detecting, and preventing survey bots</article-title><source>Soc Work Res</source><year>2022</year><month>11</month><day>14</day><volume>46</volume><issue>4</issue><fpage>343</fpage><lpage>350</lpage><pub-id pub-id-type="doi">10.1093/swr/svac023</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lebrun</surname><given-names>B</given-names> </name><name name-style="western"><surname>Temtsin</surname><given-names>S</given-names> </name><name name-style="western"><surname>Vonasch</surname><given-names>A</given-names> </name><name name-style="western"><surname>Bartneck</surname><given-names>C</given-names> </name></person-group><article-title>Detecting the corruption of online questionnaires by artificial intelligence</article-title><source>Front Robot AI</source><year>2023</year><volume>10</volume><fpage>1277635</fpage><pub-id pub-id-type="doi">10.3389/frobt.2023.1277635</pub-id><pub-id pub-id-type="medline">38371744</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>H&#x00E4;m&#x00E4;l&#x00E4;inen</surname><given-names>P</given-names> </name><name name-style="western"><surname>Tavast</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kunnari</surname><given-names>A</given-names> </name></person-group><article-title>Evaluating large language models in generating synthetic HCI research data: a case study</article-title><year>2023</year><month>04</month><day>19</day><conf-name>Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems</conf-name><conf-loc>Hamburg, Germany</conf-loc><publisher-name>Association for Computing Machinery</publisher-name><fpage>1</fpage><lpage>19</lpage><pub-id pub-id-type="doi">10.1145/3544548.3580688</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lawlor</surname><given-names>J</given-names> </name><name name-style="western"><surname>Thomas</surname><given-names>C</given-names> </name><name name-style="western"><surname>Guhin</surname><given-names>AT</given-names> </name><etal/></person-group><article-title>Suspicious and fraudulent online survey participation: Introducing the REAL framework</article-title><source>Methodological Innovations</source><year>2021</year><month>09</month><volume>14</volume><issue>3</issue><pub-id pub-id-type="doi">10.1177/20597991211050467</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Eckerd</surname><given-names>S</given-names> </name><name name-style="western"><surname>DuHadway</surname><given-names>S</given-names> </name><name name-style="western"><surname>Bendoly</surname><given-names>E</given-names> </name><name name-style="western"><surname>Carter</surname><given-names>CR</given-names> </name><name name-style="western"><surname>Kaufmann</surname><given-names>L</given-names> </name></person-group><article-title>On making experimental design choices: Discussions on the use and challenges of demand effects, incentives, deception, samples, and vignettes</article-title><source>J of Ops Management</source><year>2021</year><month>03</month><volume>67</volume><issue>2</issue><fpage>261</fpage><lpage>275</lpage><pub-id pub-id-type="doi">10.1002/joom.1128</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Roberts</surname><given-names>JK</given-names> </name><name name-style="western"><surname>Pavlakis</surname><given-names>AE</given-names> </name><name name-style="western"><surname>Richards</surname><given-names>MP</given-names> </name></person-group><article-title>It&#x2019;s more complicated than it seems: Virtual qualitative research in the COVID-19 era</article-title><source>Int J Qual Methods</source><year>2021</year><month>01</month><volume>20</volume><fpage>16094069211002959</fpage><pub-id pub-id-type="doi">10.1177/16094069211002959</pub-id></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Regmi</surname><given-names>PR</given-names> </name><name name-style="western"><surname>Waithaka</surname><given-names>E</given-names> </name><name name-style="western"><surname>Paudyal</surname><given-names>A</given-names> </name><name name-style="western"><surname>Simkhada</surname><given-names>P</given-names> </name><name name-style="western"><surname>van Teijlingen</surname><given-names>E</given-names> </name></person-group><article-title>Guide to the design and application of online questionnaire surveys</article-title><source>Nepal J Epidemiol</source><year>2016</year><month>12</month><volume>6</volume><issue>4</issue><fpage>640</fpage><lpage>644</lpage><pub-id pub-id-type="doi">10.3126/nje.v6i4.17258</pub-id><pub-id pub-id-type="medline">28804676</pub-id></nlm-citation></ref></ref-list></back></article>