Published on in Vol 27 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/67303, first published .
Women’s Preferences and Willingness to Pay for AI Chatbots in Women’s Health: Discrete Choice Experiment Study

Women’s Preferences and Willingness to Pay for AI Chatbots in Women’s Health: Discrete Choice Experiment Study

Women’s Preferences and Willingness to Pay for AI Chatbots in Women’s Health: Discrete Choice Experiment Study

1School of Public Health, Peking University, 38 Xueyuan Road, Haidian District, Beijing, China

2Department of Gynaecology and Obstetrics, The Fourth Central Hospital of Baoding City, Hebei, China

3School of Basic Medical Sciences, Hebei University, Hebei, China

4Jitang College, North China University of Science and Technology, Hebei, China

5Center for Evidence-Based Chinese Medicine, Beijing University of Chinese Medicine, Beijing, China

*these authors contributed equally

Corresponding Author:

Xinying Sun, PhD


Background: Over 96% of adult women face health issues, with 70% experiencing conditions like infections. Mobile health education is increasingly popular but faces challenges in personalization and readability. Artificial intelligence (AI) chatbots provide tailored support, and a discrete choice experiment can help in understanding user preferences to improve chatbot design.

Objective: This study aims at exploring the preferences of women toward AI chatbots to improve health education communication and user experience.

Methods: A discrete choice experiment was conducted, identifying 6 main attributes of AI chatbots: response accuracy, legibility, service cost, background information collection, information utility, and content provision. A total of 957 female participants from a hospital in Hebei Province participated, choosing between 2 hypothetical chatbots or opting for neither (a no-choice option). The conditional logit model was used to estimate user preferences.

Results: A total of 957 participants were included in the analysis. The results showed that participants preferred a chatbot with 100% response accuracy (β=0.940, P<.001; 95% CI 0.624 to 1.255), very easy to understand information (β=0.907, P<.001; 95% CI 0.634 to 1.180), a service fee of CN ¥0/month (β=−0.095, P<.001; 95% CI −0.108 to −0.082; a currency exchange rate of US $1=CN ¥7.09 was applicable), practical information utility (β=1.085, P<.001; 95% CI 0.832 to 1.338), and provision of disease-related knowledge (β=0.752, P<.001; 95% CI 0.485 to 1.018). Whether or not to allow the collection of background information (only question and answer information) has no significant impact on women’s choice preferences. Additionally, participants were willing to pay an additional CN ¥9.916 (95% CI 6.843 to 12.292) for 100% response accuracy, CN ¥9.567 (95% CI 6.843 to 12.292) for “very easy to understand” information, and CN ¥11.451 (95% CI 8.704 to 14.198) for the “very practical” information utility. Additionally, they were willing to pay CN ¥7.931 (95% CI 4.975 to 10.886) for “knowledge of diseases” compared to “gender knowledge” (CN ¥2.602, 95% CI −0.551 to 5.756). The relative importance of the chatbot attributes indicated that information utility (1.085/3.858, 28.12%) and response accuracy (0.940/3.858, 24.37%) were the most influential factors in participants’ preferences.

Conclusions: AI chatbots designed for female users should focus on high response accuracy, clear content, free access, privacy protection, practical information, and disease knowledge to attract users and enhance health education.

J Med Internet Res 2025;27:e67303

doi:10.2196/67303

Keywords



According to the United Nations [1], women make up 49.75% of the global population, yet their health needs are often overlooked. In response, the World Health Organization [2] identified 6 key priorities for women’s health in 2021, aiming to address the numerous challenges faced by women and girls in accessing adequate health care. These priorities included ensuring universal access to sexual and reproductive health services and tackling noncommunicable diseases, which remain a leading cause of preventable death among women worldwide (World Health Organization) [2]. The agenda also focused on eliminating violence against women and preventing noncommunicable diseases, including obesity. Furthermore, neglecting women’s health could jeopardize the well-being of future generations.

Compared to other health conditions, issues related to women’s health are generally more preventable and sensitive. Effective health education programs play a key role in preventing these conditions, while accurate health knowledge acquisition helps mitigate disease progression [3]. Combining systematic education with evidence-based knowledge dissemination not only benefits the women themselves but also has significant implications for society as a whole.

There are various methods to implement health education for women. These include regular physical examinations with structured educational components, distributing scientifically validated informational materials, and offering medical consultations integrated with knowledge-sharing sessions. Knowledge dissemination is also achieved through special television programs and radio segments [4] that translate complex medical information into accessible content. With the advancement of technology, the paradigm of health knowledge acquisition has been transforming. Mobile health education, based on apps and social media platforms, facilitates knowledge transfer while developing self-care skills among women [5-10]. In contrast to traditional methods, mobile health education offers greater accessibility, personalization, and cost-effectiveness, overcoming time and geographical barriers [11-13].

Mobile health education still faces challenges in pedagogical design and knowledge delivery. Educational effectiveness depends on content diversity, while knowledge retention requires information accuracy. For educational interventions targeting vulnerable groups, designing pedagogically sound content with actionable knowledge points is crucial. In this context, artificial intelligence (AI) chatbots emerge as a promising solution that may combine digital education with personalized knowledge delivery [14]. AI, which comprises programs and algorithms that perform tasks traditionally requiring human intelligence, is increasingly applied in health care, with potential in both specialized and general settings [15,16]. AI chatbots, as one of the key AI apps, offer digital platforms for delivering health information and services [17]. Using natural language processing [18], these systems convert user inquiries into a machine-readable format, leveraging a broad spectrum of medical data to establish a knowledge base that supports health-related interactions [19]. AI chatbots could offer appropriate health advice and facilitate health discussions, thereby supporting communication and patient empowerment [20]. With the expansion of mobile internet access, AI chatbots are being explored as a practical and accessible tool for women with limited health care access, enhancing patient-centered care and self-management [21]. Their use notably increased during the COVID-19 pandemic, demonstrating some positive support in various health domains [17].

Specifically, AI chatbots have the potential to restructure health education frameworks by providing tailored guidance, while establishing reliable knowledge repositories on public platforms [22,23]. For instance, Rosie, a health education chatbot, is a great example. Rosie aims to provide reliable health information to new mothers, helping them address various issues during pregnancy and postpartum. Through interactions with the community, Rosie has not only enhanced users’ access to health information but also helped to narrow the health disparities between different racial groups to some extent [24]. Maeda et al [25] conducted a randomized controlled trial with 927 women to assess the effects of a fertility education chatbot, finding it significantly improved fertility knowledge and preconception behaviors. The chatbot also reduced anxiety, demonstrating its potential to educate and empower women without increasing stress. Additionally, a review by Kim [26] examined the broader effects of AI chatbots on women’s health, including mental health, cancer self-care, and preconception intentions. The meta-analysis revealed that chatbot interventions significantly reduced anxiety, showing the positive impact of such technologies on physical, physiological, and cognitive health outcomes. This suggests that AI chatbots can serve as an effective tool in addressing various health concerns and providing digital therapeutic support to women.

However, despite these advancements, there has been insufficient exploration of how these AI tools can be customized for specific gynecological needs. AI chatbots that fully consider the needs and preferences of users may better enhance the effectiveness of interventions [27]. Previous studies have shown that people tend to care about the accuracy of medical chatbot responses and are concerned about whether their privacy is being violated [28-30]. A review of mental health chatbots indicated that patients generally find the usefulness and ease of use of these chatbots to be positive, but were less satisfied with the language expression and response quality of the chatbots. Some chatbots were considered slow in response, with answers that were often superficial, brief, and sometimes even confusing [31]. Moreover, another review revealed that the personalization of chatbots, such as in response results, content, user interfaces, delivery channels, and features, was an important factor in improving user satisfaction [32]. In summary, research on AI needs and preferences has mostly focused on the general population or adolescents, with some involvement in the development and research of mental health chatbots [27,31,33,34]. However, most studies have concentrated on English-language chatbots and lack in-depth investigation into preference differences among various demographic groups, which could potentially limit the benefits that different populations can derive from these technologies.

This study aims to explore Chinese women’s preferences regarding the attributes of AI chatbots. To improve user experience and the accuracy of AI chatbot information, it is crucial to understand user preferences. A suitable method for measuring these preferences is the discrete choice experiment (DCE), a robust survey technique that presents respondents with multiple choices among hypothetical treatments [35]. The choices made by respondents are influenced by the alternatives presented. By offering options that consist of various attributes at differing levels and repeatedly asking participants to select their preferred option or the one that maximizes utility [36], we seek to understand how users may value different aspects of AI chatbot use in women’s health. Quantifying these preferences enables the identification of preferred attributes and provides insight into how women perceive and value AI chatbots for their health needs [37].


Study Design

In this study, we investigated women’s preferences for attributes of the AI chatbots using the DCE method. DCE is a questionnaire-based stated preference approach to study people’s preferences for certain characteristics or conditions by asking subjects to make choices about hypothetical scenarios or alternatives they prefer [38,39], DCE can be used to obtain quantitative data about people’s preferences, gain in-depth insights and predictive power, and is often applied in market research, social sciences, medicine, and other fields. Selected attributes in the DCE scheme can contain price attributes for different price levels to estimate willingness to pay (WTP).

Development of the DCE

In this study, to ensure that the selected attributes comprehensively reflect user preferences for AI chatbots and have broad representativeness, we followed a systematic process involving a preliminary literature review, expert consultations, and a presurvey.

First, we conducted a preliminary literature review by searching multiple academic databases (such as Google Scholar, Web of Science, PubMed, etc) using relevant keywords like “AI chatbots,” “user preferences,” and “digital health,” and focusing on studies published between 2015 and 2022. We selected studies that focused on female participants, including both quantitative and qualitative research that explored the characteristics of AI chatbots and users’ responses to them. This ensured that the literature reviewed was relevant to the female population, providing a solid theoretical foundation for identifying potential attributes that could influence user choices [28,30,40-44]. Based on this review, we moved to the expert consultation phase.

We conducted 3 rounds of expert consultations, with each round having specific objectives and outcomes. In the first round, we brought together a multidisciplinary team of 10 experts from fields such as law, ethics, AI, and health data management. During this round, we presented the initial set of attributes identified from the literature review and asked the experts to review and comment on their relevance, clarity, and comprehensiveness. Their feedback led to the removal of certain attributes that were deemed too general or not actionable. In the second round, the attributes were refined based on the feedback from the first round, and the experts were asked to further discuss the attributes’ levels, ensuring that they were clear and easy to understand for the target audience. This round also involved developing concrete examples for some attributes to help experts understand their real-world applications. In the third and final round, the experts were asked to validate the revised attributes and provide final input on their importance and whether they sufficiently reflected user preferences. The outcome of this round was the identification of 6 core attributes, with final adjustments made to the descriptions and levels to make them as actionable and understandable as possible.

After the expert consultations, we proceeded with a presurvey to test the applicability of the selected attributes within the target group. The survey was conducted with 100 women, who were recruited from a hospital in Hebei Province. The purpose of this presurvey was to gauge how well the attributes were understood, ranked, and valued by potential users. Through this survey, we collected feedback on the clarity of the attribute descriptions and the relative importance of each attribute. The results indicated that some attribute descriptions were unclear or too complex, requiring further revision. Based on the presurvey feedback, we modified the wording of the attributes to ensure that they were easily understandable and accurately captured the participants’ preferences.

Through these three comprehensive validation steps—literature review, expert consultation, and presurvey—we ultimately identified six key attributes that would guide the design of the AI chatbot: (1) response accuracy, (2) legibility, (3) service cost, (4) whether to allow background collection of information (question and answer information only), (5) information utility, and (6) provision of information content.

Each attribute and its level are shown in Table 1.

Table 1. Attributes and levels of the chatbot.
AttributeAttribute levela
Response accuracy60%, 70%, 80%, 90%, or 100%
LegibilityVery easy to understand, easier to understand, harder to understand, or difficult to understand
Service cost (CN ¥)b0/month, 5/month, 10/month, 15/month, or 20/month
Whether to allow the collection of background information (only question and answer information)Yes or no
Information utilityVery practical, more practical, less practical, or very impractical
Provision of information contentKnowledge of diseases, daily health case, knowledge of maternity, or gender knowledge

aRefers to the different levels or variations of each attribute used in the discrete choice experiment.

bA currency exchange rate of US $1=CN ¥7.09 was applicable.

Based on the identified attribute levels, we created choice sets containing various attribute levels for respondents to select from through orthogonal tests. Given that there are 2 to 5 attribute levels for each of the 6 attributes in our study, using a full factorial design would result in 3200 combinations (5×4×5×2×4×4=3200), which is clearly impractical for generating such a large number of choice sets. Consequently, we used a fractional factorial design to determine the optimal number of choice sets. Based on the principles of orthogonality, balance, and minimal overlap, 25 proposals were ultimately generated. Selecting one of them as a fixed reference, along with the remaining 24 options, forms a total of 24 selection sets. Each selection set contains 3 conceptual options (“choose option A,” “choose option B,” or “choose neither”). Randomly dividing all selection sets into 4 groups, each group consists of 7 selection sets (6 random selection sets and 1 repeated selection set). Respondents selected their preferred combination of attributes for the AI chatbots in each scenario or opted for none, thereby minimizing bias resulting from forced choices. According to the sample size formula for DCEs: n=(500c)/(ta), where “c” represents the largest number of levels for an attribute, “t” denotes the number of choice sets in a block, and “a” indicates the number of alternatives. For this study, the values were as follows: “c” was set to 5, “t” was set to 6, and “a” was set to 2. Therefore, the minimum required sample size for this study was determined to be 209 participants. Figure 1 shows an example of a DCE program.

Figure 1. An example scenario of the choice-based conjoint in the questionnaire. A currency exchange rate of US $1=CN ¥7.09 was applicable. Q&A: question and answer.

Study Population and Data Collection

This study began in December 2022 and involved a questionnaire survey of women at a hospital in Hebei Province, lasting for 2.5 months. The recruitment period started on December 20, 2022, and ended on February 15, 2023. Convenience sampling was used, with gynecologists recruiting eligible female patients from both the gynecology outpatient department and the gynecology inpatient ward. Research staff explained the purpose of this study to eligible women and invited them to participate. Research information was collected through electronic questionnaires, which were divided into 2 sections: the first section gathered basic demographic information, including age, education level, and usual residence, while the second section assessed preferences regarding AI chatbots. The electronic questionnaire was distributed via the WJX platform[45]. After patients agreed to participate and completed the informed consent form, they filled out the questionnaire online by scanning a QR code. The inclusion and exclusion criteria are as follows:

The inclusion criteria are as follows: (1) age ≥18 years; (2) sex: female; (3) possesses the nationality of the People’s Republic of China; (4) own a smartphone and have the ability to use WeChat, including following WeChat public accounts; (5) ability to complete web-based questionnaires independently or with assistance from another; (6) possess basic literacy skills to enable normal communication and interaction; and (7) voluntary participation in the study, with willingness to complete an informed consent form.

The exclusion criteria are as follows: (1) individuals with severe cognitive impairments or mental health conditions, (2) individuals who are participating in other similar research projects, and (3) individuals who are unwilling to cooperate.

A total of 1281 women were invited to participate, and 1216 women agreed to participate, yielding a participation rate of 94.9%. The final sample size of 957 participants was determined after excluding unqualified questionnaires based on the exclusion criteria. This process is depicted in Figure 2, which illustrates the flowchart of this study’s recruitment and analysis.

Figure 2. Flowchart of study process. AI: artificial intelligence; DCE: discrete choice experiment; CLOGIT: conditional logit model.

Ethical Considerations

This study has been approved by the Ethics Committee of the Baoding No. 4 Central Hospital (2022013). All study participants have voluntarily participated and signed informed consent forms, ensuring that they are fully aware of this study’s objectives, methods, and any potential risks. This consent was obtained before participation, per the ethical standards for human subject research. As this study involves human participants, appropriate ethical oversight was conducted, and all research activities were approved by the Ethics Committee. To ensure privacy and confidentiality, all study data are anonymized and deidentified to protect the identity of the participants. The data will be stored securely and will only be accessible to authorized personnel involved in this research. No compensation is provided to the participants for their involvement in this study. To maintain the confidentiality and anonymity of the participants, no identifiable images or data of individual participants will be included in this paper or supplementary materials.

Statistical Analysis

Statistical analyses were performed using IBM SPSS Statistics (version 26.0; IBM Corporation) and Stata (version 15.0; StataCorp LLC). Descriptive statistics of demographic variables were performed by frequency counts (composition ratios). Conditional logit models (CLOGITs) were used to quantify the relative levels of attribute preferences used by the AI chatbots via Stata (version 15.0). Different levels of each attribute were dummy-coded, and one of the levels was set as the reference level. In a DCE study, the calculated results provide important statistical information. These results include coefficients, P values, SEs, and 95% CI. We also calculated this study’s participants’ WTP for different attribute levels of the AI chatbots to more intuitively reflect the strength of the respondents’ attribute preferences for the AI chatbots.


Characteristics of Respondents

The sociodemographic characteristics of this study’s sample are summarized in Table 2.

The age distribution showed that the largest group was aged between 26 and 35 years (n=431, 45.04%), followed by those aged 36 to 45 years (n=280, 29.26%). For further details on participant demographics, please refer to Table 2.

Table 2. General characteristics of the subjects (N=957).
ItemsCases, n (%)
Occupation
 Employed726 (75.86)
 Student27 (2.82)
 Retired11 (1.15)
 Unstable occupation, freelancer, unemployed, and not in employment193 (20.17)
Location
 Urban633 (66.14)
 Rural324 (33.86)
Age (years)
 18‐25159 (16.61)
 26‐35431 (45.04)
 36‐45280 (29.26)
 >4587 (9.09)
Education level
 Junior high school or lower266 (27.8)
 Special school or senior high school211 (22.05)
 Junior college or higher480 (50.16)
Marital status
 Single115 (12.02)
 Married (including first marriage, remarried, or remarried after divorce)823 (86)
 Divorced16 (1.67)
 Widowed3 (0.31)
Incomeab
 ≤100066 (6.9)
 1001‐3000338 (35.32)
 3001‐5000386 (40.33)
 >5000167 (17.45)

aPer capita monthly household income.

bA currency exchange rate of US $1=CN ¥7.09 was applicable.

Percent Importance of AI Chatbot Attributes

The percent importance reflects the extent to which each attribute influences the differences in preferences. It is calculated based on the range of preference weights across each attribute’s levels, which provides a measure of its impact on decision-making. The sum of all percent importance is 100%, and a higher value corresponds to a greater influence on preferences. The information utility of the chatbot is considered the most important attribute (1.805/3.858, 28.123%), while cost and background information collection are the lowest, at 2.462% (0.095/3.858) and 2.074% (0.080/3.858), respectively.

DCE Results

The results of the CLOGIT model analysis indicate that the participants exhibited a preference for attributes such as 100% response accuracy, very easy to understand readability, CN ¥0/month (a currency exchange rate of US $1=CN ¥7.09 was applicable) service cost, very practical information utility, and provision of information content as knowledge of the disease in the AI chatbots. Table 3 presents detailed information on the preference attributes and their coefficients.

Table 3. CLOGITa results of participants’ preferences for attributes of AIb chatbots.
Attributes and levelsCoefficientP valueSE95% CI
Response accuracy (%)
 60c
 70 0.561 .001 0.1680.232 to 0.891
 80 0.511 .002 0.1670.184 to 0.838
 90 0.637 <.001 0.1610.321 to 0.953
 100 0.94 <.001 0.1610.624 to 1.255
Legibility
 Harder to understandc
 Difficult to understand. 0.207 .21 0.163–0.113 to 0.528
 Easier to understand 0.725 <.001 0.1370.456 to 0.993
 Very easy to understand. 0.907 <.001 0.1390.634 to 1.180
 Service cost –0.095 <.001 0.007–0.108 to –0.082
Whether or not to allow the collection of background information (only question and answer information)
 Yesc
 No 0.080 .37 0.090–0.095 to 0.254
Information utility
 Very impracticalc
 Less practical 0.180 .24 0.153–0.121 to 0.481
 More practical 0.584 <.001 0.1640.262 to 0.907
 Very practical 1.085 <.001 0.1290.832 to 1.338
Provision of information content
 Gender knowledgec
 Knowledge of maternity 0.247 .11 0.153–0.141 to 0.447
 Daily health care 0.664 <.001 0.1370.396 to 0.932
 Knowledge of diseases 0.751 <.001 0.1360.485 to 1.018

aCLOGIT: conditional logit models.

bAI: artificial intelligence.

cReference level.

WTP Results

WTP is a valid indicator of how much money a person is willing to sacrifice to choose 1 diagnostic attribute level over another. Table 4 presents the participants’ WTP for different attributes of the AI chatbot. The findings indicate that participants are willing to pay more for a women’s health AI chatbot that provides accurate answers, clear content, does not collect background information, has high information availability, and focuses on disease-related knowledge.

Table 4. Participants’ WTPa for AIe chatbots.
Attributes and levelsWTP (CN ¥)c95% CI
Response accuracy (%)
 60d
 705.9242.628 to 9.22
 805.3892.034 to 8.744
 906.7213.637 to 9.806
 1009.9166.843 to 12.292
Legibility
 Harder to understandd
 Difficult to understand.2.188–1.154 to 5.529
 Easier to understand7.6464.883 to 10.409
 Very easy to understand.9.5676.843 to 12.292
Whether or not to allow the collection of background information (only question and answer information)
 Yesd
 No0.839–1.017 to 2.695
Information utility
 Very impracticald
 Less practical1.9–1.278 to 5.078
 More practical6.1672.701 to 9.633
 Very practical11.4518.704 to 14.198
Provision of information content
 Gender knowledged
 Knowledge of maternity2.602–0.551 to 5.756
 Daily health care7.0064.049 to 9.963
 Knowledge of diseases7.9314.975 to 10.886

aWTP: willingness to pay.

bAI: artificial intelligence.

cA currency exchange rate of US $1=CN ¥7.09 was applicable.

dReference level.


Principal Findings

This study used a DCE to explore women’s preferences for AI chatbots. The results revealed significant variations in user preferences across different chatbot attributes, particularly for information usefulness, response accuracy, readability, and content provision, whereas cost and data collection were of relatively lower importance.

The scalability, accessibility, ease of use, and rapid information dissemination of AI chatbots offer supplementary benefits to public health efforts [46], addressing issues such as capacity constraints, social distancing requirements, and misinformation [47]. Research suggests that people sometimes prefer interacting with AI chatbots over doctors in certain contexts, as these chatbots can respond more quickly, provide high-quality feedback [48], and offer empathetic interactions. The application of AI in medicine is expanding across various domains, including medical image analysis [49], drug interaction detection [50], high-risk patient identification [51], and medical record coding [52]. In health care education, tools like ChatGPT have been used to facilitate personalized learning, encourage critical thinking, and support problem-based learning [53]. The widespread use of AI has made chatbots popular for accessing health information.

This study found that information utility is one of the most valued attributes of AI chatbots among participants. Participants appeared to prefer chatbots that provide practical and relevant information, aligning with the findings of Kim [26] in their systematic review and meta-analysis on women’s health, which suggested that AI chatbots may positively impact women’s health by reducing anxiety and depression, promoting healthy behaviors, and offering health education. Although the review highlights that the provision of practical and relevant information appears to be a key factor in improving health outcomes for women, our current study found that the P value of the “less practical” level did not reach significance, indicating that the practicality of information only has a significant impact on user choices when it is very high or very low. This suggests that users expect the information provided by chatbots to have practical application value. Therefore, when designing chatbots, developers should consider focusing on providing practical information that can meet users’ actual needs and be immediately applied to their daily lives.

The results also indicated that response accuracy and readability significantly influence participants’ preferences. Participants seemed to favor AI chatbots with high accuracy and comprehensible content, which aligns with the systematic review by Aggarwal et al [54]. Aggarwal et al [54] included 15 empirical studies on AI chatbots that facilitate health behavior changes, including interventions like healthy lifestyles, smoking cessation, and medication adherence. While some studies demonstrated the efficiency of AI chatbots, mixed results were observed regarding their feasibility, acceptability, and usability [55,56]. These findings suggest that improving chatbot accuracy and readability is crucial for effectively promoting behavior change. Accurate health information is likely key to building user trust [57], and this study further confirms that participants valued accurate responses. Readability also plays a significant role in users’ choices, particularly among those with low health literacy or unfamiliarity with complex medical terms, as easily understandable information enhances comprehension and adherence [58-60]. In this study, the P values for the “difficult to understand” attribute were not statistically significant, suggesting that the readability of AI chatbot information might not substantially influence user preferences within certain thresholds. This may indicate that users prefer levels of readability that are “easier to understand” or “very easy to understand.” When information reaches a level deemed difficult to comprehend, users might be disinclined to engage further with the AI chatbot. These findings suggest that AI chatbots should aim to optimize language and interface design based on accurate content to ensure comprehensibility and cater to diverse women’s needs.

Regarding content preferences, participants showed a greater interest in disease knowledge, consistent with the systematic review by Younis et al [61], which categorized AI apps across health care domains such as patient support and education. AI chatbots appear to provide personalized health education and disease prevention information, offering content relevant to users’ needs. The preference for pregnancy and childbirth knowledge did not seem to be significant, indicating that these types of information do not have a clear advantage in user selection preferences. This may be because women are more concerned with the prevention, diagnosis, and treatment of diseases themselves, and may not necessarily have a direct need for information related to pregnancy and childbirth as much as disease knowledge. However, it is important to note that the high preference for disease knowledge could also be influenced by the fact that participants were recruited from a gynecology unit of a hospital, where individuals are likely to have one or more health conditions and a stronger interest in learning about diseases. We cannot conclusively assume that disease knowledge would rate as highly in a nonhospital sample, and this should be taken into consideration when interpreting the findings. While emphasizing the importance of providing information on disease knowledge, it is crucial not to overlook the demand for content in other areas. Women’s health is a multifaceted field that includes knowledge related to gender, pregnancy, childbirth, daily health care, and various diseases. Therefore, when developing chatbots, it is essential to consider all these aspects comprehensively to address the diverse needs of users.

On the topic of service cost, participants appeared to prefer free AI chatbots. This preference is consistent with other systematic reviews [62], which noted that free or low-cost services are more easily accepted by users, particularly when economic pressures are high. In a fast-paced era, the accessibility of free AI chatbots allows users to seek answers anytime and anywhere, increasing their reliance and usage. The preference for free services spans all age groups and socioeconomic backgrounds, suggesting that financial factors may strongly influence the choice for cost-free AI tools.

Finally, data collection did not significantly influence participants’ preferences. However, Kim [26] emphasized the importance of privacy protection in AI apps for women’s health. While data collection can help improve service quality, ensuring data security remains crucial in health care settings. Users generally value the protection of personal health data, and limiting data collection to nonidentifiable question and answer content can help alleviate concerns about data misuse or breaches.

This study provides valuable insights but has several limitations. First, it used convenience sampling and only included gynecological outpatients from a hospital in China. Therefore, the representativeness of the sample is limited. Since these female patients were surveyed in a medical setting, they are likely to be more interested in learning about diseases. As a result, the findings may not fully reflect the preferences and needs of a broader female population regarding women’s health chatbots. Therefore, future research should test these chatbot attributes in different cultural and social contexts to assess whether these preferences are universal or context-dependent. Second, the assumptions in the questionnaire might not fully reflect real-world preferences. Demographic factors such as age, education, and health literacy can influence how users engage with AI chatbots. Therefore, future studies should consider including a more diverse sample to capture a wider range of preferences. Finally, other variables, such as trust in technology and health care system perceptions, were not explored in this study but may impact user preferences. Expanding research to include these factors and diverse populations will help refine chatbot design to better meet the needs of different groups. Future studies should explore these aspects to provide a more comprehensive understanding of AI chatbot preferences across various contexts.

Conclusions

This DCE highlights key preferences for AI chatbots in health care, indicating that users prioritize high response accuracy (100%), legibility, free services, and the practical utility of information, particularly about disease knowledge. These findings suggest that the development of AI chatbots should emphasize clarity, accessibility at no cost, user privacy protection, and the relevance of health content. By incorporating these attributes, user engagement can be enhanced, the dissemination of health education can be improved, and the effectiveness of AI tools in promoting disease management can be maximized.

Acknowledgments

We are grateful to the professionals and students who provided enormous support in the recruitment and data collection of this study, and the participants who took the time to cooperate with the survey. This work was supported by the National Key Research and Development Plan Project (2022YFC3600904).

Data Availability

The datasets used or analyzed during this study are available from the corresponding author on reasonable request.

Authors' Contributions

YW and HM conceived and designed this study. JW wrote the initial draft of the paper. JW, HM, YJ, and JZ reviewed and revised this paper. XS and YW provided theoretical guidance and methodological expertise. HM designed the questionnaire. TL and JL are responsible for this study's implementation. XS is responsible for the study management. All authors provided revisions and approved the final manuscript.

Conflicts of Interest

None declared.

  1. State of world population 2023 - 8 billion lives, infinite possibilities. The United Nations. URL: https:/​/www.​unfpa.org/​publications/​state-world-population-2023-8-billion-lives-infinite-possibilities [Accessed 2025-05-30]
  2. 6 priorities for women and health. The World Health Organization. URL: https://www.who.int/news-room/spotlight/6-priorities-for-women-and-health [Accessed 2025-05-07]
  3. Hibbard JH, Pope CR. Women’s roles, interest in health and health behavior. Women Health. 1987;12(2):67-84. [CrossRef] [Medline]
  4. Yang X, Yang F, Fu W, Guo L, Xu L, Zhang H. The current status of community mental health services in three northern areas of China. Family Med Commun Health. Mar 1, 2014;2(1):1-7. [CrossRef]
  5. Lin F, Sun Q, Peng Z, et al. The innovations in China’s primary health care reform: development and characteristics of the community health services in Hangzhou. Family Med Commun Health. Sep 1, 2015;3(3):52-66. [CrossRef]
  6. Carter MC, Burley VJ, Nykjaer C, Cade JE. Adherence to a smartphone application for weight loss compared to website and paper diary: pilot randomized controlled trial. J Med Internet Res. Apr 15, 2013;15(4):e32. [CrossRef] [Medline]
  7. Cheng Z, Guo S. Automatic text summarization for public health WeChat official accounts platform base on improved TextRank. J Environ Public Health. 2022;2022:1166989. [CrossRef] [Medline]
  8. Chen X, Zhou X, Li H, Li J, Jiang H. The value of WeChat application in chronic diseases management in China. Comput Methods Programs Biomed. Nov 2020;196:105710. [CrossRef] [Medline]
  9. Jiang Y, Liu F, Guo J, et al. Evaluating an intervention program using WeChat for patients with chronic obstructive pulmonary disease: randomized controlled trial. J Med Internet Res. Apr 21, 2020;22(4):e17089. [CrossRef] [Medline]
  10. Sun M, Yang L, Chen W, et al. Current status of official WeChat accounts for public health education. J Public Health (Oxf). Sep 22, 2021;43(3):618-624. [CrossRef]
  11. Naimark JS, Madar Z, Shahar DR. The impact of a web-based app (eBalance) in promoting healthy lifestyles: randomized controlled trial. J Med Internet Res. Mar 2, 2015;17(3):e56. [CrossRef] [Medline]
  12. Lapp B, Zakus D. Factors associated with the utilization of institutional and home birth services among women in Ethiopia: a scoping review. Family Med Commun Health. Dec 1, 2016;4(4):30-43. [CrossRef]
  13. Kao CK, Liebovitz DM. Consumer mobile health apps: current state, barriers, and future directions. PM R. May 2017;9(5S):S106-S115. [CrossRef] [Medline]
  14. Manickam P, Mariappan SA, Murugesan SM, et al. Artificial Intelligence (AI) and Internet of Medical Things (IoMT) assisted biomedical systems for intelligent healthcare. Biosensors (Basel). Jul 25, 2022;12(8):562. [CrossRef] [Medline]
  15. Hendrix N, Hauber B, Lee CI, Bansal A, Veenstra DL. Artificial intelligence in breast cancer screening: primary care provider preferences. J Am Med Inform Assoc. Jun 12, 2021;28(6):1117-1124. [CrossRef] [Medline]
  16. Miller DD, Brown EW. Artificial intelligence in medical practice: the question to the answer? Am J Med. Feb 2018;131(2):129-133. [CrossRef] [Medline]
  17. Wilson L, Marasoiu M. The development and use of chatbots in public health: scoping review. JMIR Hum Factors. Oct 5, 2022;9(4):e35882. [CrossRef] [Medline]
  18. Gilson A, Safranek CW, Huang T, et al. How does ChatGPT perform on the United States Medical Licensing Examination (USMLE)? The implications of large language models for medical education and knowledge assessment. JMIR Med Educ. Feb 8, 2023;9:e45312. [CrossRef] [Medline]
  19. Yeo YH, Samaan JS, Ng WH, et al. Assessing the performance of ChatGPT in answering questions regarding cirrhosis and hepatocellular carcinoma. Clin Mol Hepatol. Jul 2023;29(3):721-732. [CrossRef]
  20. Liu J, Wang C, Liu S. Utility of ChatGPT in clinical practice. J Med Internet Res. Jun 28, 2023;25:e48568. [CrossRef] [Medline]
  21. Milne-Ives M, de Cock C, Lim E, et al. The effectiveness of artificial intelligence conversational agents in health care: systematic review. J Med Internet Res. Oct 22, 2020;22(10):e20346. [CrossRef] [Medline]
  22. Chung K, Cho HY, Park JY. A chatbot for perinatal women’s and partners’ obstetric and mental health care: development and usability evaluation study. JMIR Med Inform. Mar 3, 2021;9(3):e18607. [CrossRef] [Medline]
  23. Tawfik E, Ghallab E, Moustafa A. A nurse versus a chatbot ‒ the effect of an empowerment program on chemotherapy-related side effects and the self-care behaviors of women living with breast cancer: a randomized controlled trial. BMC Nurs. Apr 6, 2023;22(1):102. [CrossRef] [Medline]
  24. Mane HY, Doig AC, Gutierrez FXM, et al. Practical guidance for the development of Rosie, a health education question-and-answer chatbot for new mothers. J Public Health Manag Pract. 2023;29(5):663-670. [CrossRef] [Medline]
  25. Maeda E, Miyata A, Boivin J, et al. Promoting fertility awareness and preconception health using a chatbot: a randomized controlled trial. Reprod Biomed Online. Dec 2020;41(6):1133-1143. [CrossRef] [Medline]
  26. Kim HK. The effects of artificial intelligence chatbots on women’s health: a systematic review and meta-analysis. Healthcare (Basel). Feb 23, 2024;12(5):534. [CrossRef] [Medline]
  27. Maenhout L, Peuters C, Cardon G, Compernolle S, Crombez G, DeSmet A. Participatory development and pilot testing of an adolescent health promotion chatbot. Front Public Health. 2021;9:724779. [CrossRef] [Medline]
  28. Fan X, Chao D, Zhang Z, Wang D, Li X, Tian F. Utilization of self-diagnosis health chatbots in real-world settings: case study. J Med Internet Res. Jan 6, 2021;23(1):e19928. [CrossRef] [Medline]
  29. Chang IC, Shih YS, Kuo KM. Why would you use medical chatbots? interview and survey. Int J Med Inform. Sep 2022;165:104827. [CrossRef] [Medline]
  30. Nadarzynski T, Miles O, Cowie A, Ridge D. Acceptability of artificial intelligence (AI)-led chatbot services in healthcare: a mixed-methods study. Digit Health. 2019;5:2055207619871808. [CrossRef] [Medline]
  31. Abd-Alrazaq AA, Alajlani M, Ali N, Denecke K, Bewick BM, Househ M. Perceptions and opinions of patients about mental health chatbots: scoping review. J Med Internet Res. Jan 13, 2021;23(1):e17828. [CrossRef] [Medline]
  32. Xue J, Zhang B, Zhao Y, et al. Evaluation of the current state of chatbots for digital health: scoping review. J Med Internet Res. Dec 19, 2023;25:e47217. [CrossRef] [Medline]
  33. Mariamo A, Temcheff CE, Léger PM, Senecal S, Lau MA. Emotional reactions and likelihood of response to questions designed for a mental health chatbot among adolescents: experimental study. JMIR Hum Factors. Mar 18, 2021;8(1):e24343. [CrossRef] [Medline]
  34. Kim M, Oh J, Kim D, Shin J, Lee D. Understanding user preferences in developing a mental healthcare AI chatbot: a conjoint analysis approach. International Journal of Human–Computer Interaction. Apr 18, 2025;41(8):4813-4821. [CrossRef]
  35. Laba TL. Using discrete choice experiment to elicit patient preferences for osteoporosis drug treatments: where to from here? Arthritis Res Ther. 2014;16(2):106. [CrossRef] [Medline]
  36. von Wedel P, Hagist C. Physicians’ preferences and willingness to pay for artificial intelligence-based assistance tools: a discrete choice experiment among german radiologists. BMC Health Serv Res. Dec 2022;22(1):398. [CrossRef]
  37. Noben CYG, Stammen LA, Vaassen S, et al. Discrete choice experiment on educating value-based healthcare. Postgrad Med J. Aug 2021;97(1150):515-520. [CrossRef] [Medline]
  38. Ryan M, Bate A, Eastmond CJ, Ludbrook A. Use of discrete choice experiments to elicit preferences. Qual Health Care. Sep 2001;10 Suppl 1(Suppl 1):i55-i60. [CrossRef] [Medline]
  39. Louviere JJ, Lancsar E. Choice experiments in health: the good, the bad, the ugly and toward a brighter future. Health Econ Policy Law. Oct 2009;4(Pt 4):527-546. [CrossRef] [Medline]
  40. Battineni G, Chintalapudi N, Amenta F. AI chatbot design during an epidemic like the novel coronavirus. Healthcare (Basel). Jun 3, 2020;8(2):154. [CrossRef] [Medline]
  41. Chen Q, Gong Y, Lu Y, Tang J. Classifying and measuring the service quality of AI chatbot in frontline service. J Bus Res. Jun 2022;145:552-568. [CrossRef]
  42. Chen JS, Le TTY, Florence D. Usability and responsiveness of artificial intelligence chatbot on online customer experience in e-retailing. IJRDM. Oct 6, 2021;49(11):1512-1531. [CrossRef]
  43. Jiang H, Cheng Y, Yang J, Gao S. AI-powered chatbot communication with customers: dialogic interactions, satisfaction, engagement, and customer behavior. Comput Human Behav. Sep 2022;134:107329. [CrossRef]
  44. Ashfaq M, Yun J, Yu S, Loureiro SMC. I, chatbot: modeling the determinants of users’ satisfaction and continuance intention of AI-powered service agents. Telematics Inf. Nov 2020;54:101473. [CrossRef]
  45. Front page. WJXCN. URL: https://www.wjx.cn/ [Accessed 2025-05-30]
  46. Pearce C. Chronic disease in a digital health environment. Family Med Commun Health. Feb 1, 2018;6(1):20-25. [CrossRef]
  47. Amiri P, Karahanna E. Chatbot use cases in the COVID-19 public health response. J Am Med Inform Assoc. Apr 13, 2022;29(5):1000-1010. [CrossRef] [Medline]
  48. Ayers JW, Poliak A, Dredze M, et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med. Jun 1, 2023;183(6):589-596. [CrossRef] [Medline]
  49. Meskó B, Topol EJ. The imperative for regulatory oversight of large language models (or generative AI) in healthcare. NPJ Digit Med. Jul 6, 2023;6(1):120. [CrossRef] [Medline]
  50. Dunn P, Cianflone D. Artificial intelligence in cardiology: exciting but handle with caution. Int J Cardiol. Oct 1, 2023;388:131117. [CrossRef] [Medline]
  51. Zhong NN, Wang HQ, Huang XY, et al. Enhancing head and neck tumor management with artificial intelligence: integration and perspectives. Semin Cancer Biol. Oct 2023;95:52-74. [CrossRef] [Medline]
  52. Li W, Zhang Y, Chen F. ChatGPT in colorectal surgery: a promising tool or a passing fad? Ann Biomed Eng. Sep 2023;51(9):1892-1897. [CrossRef] [Medline]
  53. Sallam M. ChatGPT utility in healthcare education, research, and practice: systematic review on the promising perspectives and valid concerns. Healthcare (Basel). Mar 19, 2023;11(6):887. [CrossRef] [Medline]
  54. Aggarwal A, Tam CC, Wu D, Li X, Qiao S. Artificial intelligence-based chatbots for promoting health behavioral changes: systematic review. J Med Internet Res. Feb 24, 2023;25:e40789. [CrossRef] [Medline]
  55. Caranfa JT, Bommakanti NK, Young BK, Zhao PY. Accuracy of vitreoretinal disease information from an artificial intelligence chatbot. JAMA Ophthalmol. Sep 1, 2023;141(9):906-907. [CrossRef] [Medline]
  56. Diamandis EP. Chatbot GPT can be grossly inaccurate. Clin Chem Lab Med. Jan 26, 2024;62(2):e25. [CrossRef] [Medline]
  57. Hirosawa T, Harada Y, Yokose M, Sakamoto T, Kawamura R, Shimizu T. Diagnostic accuracy of differential-diagnosis lists generated by generative pretrained transformer 3 chatbot for clinical vignettes with common chief complaints: a pilot study. Int J Environ Res Public Health. Feb 15, 2023;20(4):3378. [CrossRef] [Medline]
  58. Moons P, Van Bulck L. Using ChatGPT and Google Bard to improve the readability of written patient information: a proof of concept. Eur J Cardiovasc Nurs. Mar 12, 2024;23(2):122-126. [CrossRef] [Medline]
  59. Yurdakul OV, Kilicoglu MS, Bagcier F. Evaluating the reliability and readability of online information on osteoporosis. Arch Endocrinol Metab. Nov 1, 2021;65(1):85-92. [CrossRef] [Medline]
  60. Denecke K, Borycki EM, Kushniruk AW. What can we learn from quality requirements in ISO/TS 82304-2 for evaluating conversational agents in healthcare? Stud Health Technol Inform. Nov 3, 2022;299:245-250. [CrossRef] [Medline]
  61. Younis HA, Eisa TAE, Nasser M, et al. A systematic review and meta-analysis of artificial intelligence tools in medicine and healthcare: applications, considerations, limitations, motivation and challenges. Diagnostics (Basel). Jan 4, 2024;14(1):109. [CrossRef] [Medline]
  62. Grassini E, Buzzi M, Leporini B, Vozna A. A systematic review of chatbots in inclusive healthcare: insights from the last 5 years. Univ Access Inf Soc. Mar 2025;24(1):195-203. [CrossRef]


AI: artificial intelligence
DCE: discrete choice experiment
WTP: willingness to pay


Edited by Amaryllis Mavragani; submitted 08.10.24; peer-reviewed by Cynthia Baur, Nathaniel Godwin; final revised version received 21.04.25; accepted 22.04.25; published 10.06.25.

Copyright

© Jing Wang, Hewei Min, Tao Li, Jiaheng Li, Yang Jiang, Jingbo Zhang, Yibo Wu, Xinying Sun. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 10.6.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.