Published on 21.05.20 in Vol 22, No 5 (2020): May
Preprints (earlier versions) of this paper are available at http://preprints.jmir.org/preprint/16795, first published Oct 31, 2019.
An Informatics Framework to Assess Consumer Health Language Complexity Differences: Proof-of-Concept Study
Background: The language gap between health consumers and health professionals has been long recognized as the main hindrance to effective health information comprehension. Although providing health information access in consumer health language (CHL) is widely accepted as the solution to the problem, health consumers are found to have varying health language preferences and proficiencies. To simplify health documents for heterogeneous consumer groups, it is important to quantify how CHLs are different in terms of complexity among various consumer groups.
Objective: This study aimed to propose an informatics framework (consumer health language complexity [CHELC]) to assess the complexity differences of CHL using syntax-level, text-level, term-level, and semantic-level complexity metrics. Specifically, we identified 8 language complexity metrics validated in previous literature and combined them into a 4-faceted framework. Through a rank-based algorithm, we developed unifying scores (CHELC scores [CHELCS]) to quantify syntax-level, text-level, term-level, semantic-level, and overall CHL complexity. We applied CHELCS to compare posts of each individual on online health forums designed for (1) the general public, (2) deaf and hearing-impaired people, and (3) people with autism spectrum disorder (ASD).
Methods: We examined posts with more than 4 sentences of each user from 3 health forums to understand CHL complexity differences among these groups: 12,560 posts from 3756 users in Yahoo! Answers, 25,545 posts from 1623 users in AllDeaf, and 26,484 posts from 2751 users in Wrong Planet. We calculated CHELCS for each user and compared the scores of 3 user groups (ie, deaf and hearing-impaired people, people with ASD, and the public) through 2-sample Kolmogorov-Smirnov tests and analysis of covariance tests.
Results: The results suggest that users in the public forum used more complex CHL, particularly more diverse semantics and more complex health terms compared with users in the ASD and deaf and hearing-impaired user forums. However, between the latter 2 groups, people with ASD used more complex words, and deaf and hearing-impaired users used more complex syntax.
Conclusions: Our results show that the users in 3 online forums had significantly different CHL complexities in different facets. The proposed framework and detailed measurements help to quantify these CHL complexity differences comprehensively. The results emphasize the importance of tailoring health-related content for different consumer groups with varying CHL complexities.
J Med Internet Res 2020;22(5):e16795
The language gap between laypersons (health consumers) and health care professionals has been long recognized as the main hindrance to effective health communication and health information comprehension [- ]. When interpreting health documents written mainly in professional language, consumers often depend on their own language to fill in the comprehension gap (eg, depression vs depressive disorder), which might lead to misinterpretation. Accordingly, it has also been widely agreed that health consumers should be given access to resources in their own languages [ - ]. To improve the readability of health-related content for average health consumers, there has been increasing interest in examining consumer health vocabularies [ , ], health readability measurement [ - ], and automated health text simplification approaches [ - ]. Studies on consumer health vocabularies have largely focused on extracting and building a terminology system of lay health terms used by average health consumers [ , ]. Health readability assessments have focused on developing linguistic metrics to quantify the text complexity of health content generated by health experts and professionals [ , , , ]. On the basis of the findings in both areas, automated health text simplification usually focuses on simplifying difficult texts with respect to 1 or 2 aspects (eg, medical jargon, long sentences) [ , , , , , ].
However, without a comprehensive understanding of the complexity difference between professional health language and consumer health language (CHL), current automated simplification approaches are inadequate to accurately determine what needs to be simplified and to what extent they should be simplified. Also, current simplification approaches assume that consumers share the same CHL preferences and that simplifying text to its lowest complexity can satisfy all users. For example, in synonym replacement tasks, researchers typically identify difficult medical words and then replace them with easier synonyms [, ]. These one-size-fits-all automated simplification approaches ignore the diverse simplification needs of different health customers. Research suggests that consumers with varying health literacy levels have different CHL preferences [ - ]. In addition, contextual and sociocultural factors are found to affect the language preferences of different consumer groups to think, express, and communicate health-related topics [ ]. For example, compared with average health consumers, cancer patients would be more familiar with cancer-related professional health terms (eg, genetic predisposition). Another drawback of this one-size-fits-all approach is that simplifying health content by replacing terms with lay alternatives with the lowest complexity may affect information accuracy and may inadvertently increase the length of the text [ ]. In other words, an adaptive simplification approach that can balance simplicity, accuracy, and sentence length for user groups with various CHL preferences is ideal.
In this paper, CHL has been defined as a system of vocabularies, expressions, and grammar that is commonly used by a group of health consumers in thinking, expressing, and communicating their health-related topics. CHL complexity is defined as a combined measure of varying linguistic metrics, each of which quantifies the complexity of one linguistic feature of a CHL (eg, semantics, syntax, term). The goal of adaptive health text simplification is to simplify the professional health language used in Web-based health content to match the CHL complexities of targeted consumer groups. To quantify the CHL complexity differences for simplification purposes, the linguistic complexities of CHLs used by various health consumer groups should be investigated. The increasing availability of user-generated Web-based health communications (eg, blogs, online communities, social question and answer [Q&A] websites), provides us with ample opportunities to assess CHL complexity through automated text analysis [, , ].
Studies focused on health readability assessment typically quantify the complexity of Web-based health content written by health professionals for health consumers [- ]. Researchers have developed complexity metrics that utilize a combination of various extracted linguistic features to assess the complexity of Web-based health content [ , , ]. The metrics utilized in previous literature can be categorized into 4 groups, namely, text-level complexity (eg, syllables per word) [ , ], syntax-level complexity (eg, distributions of parts of speech [POS]) [ , ], term-level complexity (eg, density of professional medical terms) [ , ], and semantic-level complexity (eg, diversity of semantics) [ ]. Examining how these linguistic features differ among various CHLs can help us gain a more accurate and comprehensive understanding of CHL complexity.
In this proof-of-concept study, we developed an informatics framework (consumer health language complexity [CHELC]) to assess CHL complexity based on existing health text readability metrics and apply this framework to explore complexity differences in CHL in 3 online forums designed for the general public, deaf and hearing-impaired people, and people with autism spectrum disorder (ASD). In previous studies, the latter 2 groups have been found to have relatively low health literacy [- ], different language use behaviors [ , ], and limited access to adaptive health information services [ ]. People with ASD were found to be repetitive and expressive by composing long sentences and words on the Web [ , , ]. Pollard and Barnett [ ] found that even highly educated deaf adults showed significant difficulty in understanding health vocabularies used in the Rapid Estimate of Adult Literacy in Medicine test. In addition, compared with the general population, deaf and hearing-impaired people exhibit significantly lower levels of health literacy and health knowledge [ ]. Accordingly, ASD and deaf and hearing-impaired user groups might use less complex CHL, especially less complex health terms in their expressions. Motivated by these observations, in this study, we explore the use of different measures to assess CHL complexity and provide insights for the development of adaptive health text simplification tools to address the needs of various consumer groups.
We formulated 2 research questions (RQs) in this study:
- RQ1: What is the feasibility of using CHELC, which combines text-level, syntax-level, term-level, and semantic-level measures for examining CHL complexity among users in 3 distinct online forums designed for the general public, people with ASD, and deaf and hearing-impaired people?
- RQ2: How do the CHLs of users in online forums designed for the general public, people with ASD, and deaf and hearing-impaired people differ in complexity on the text level, syntax level, term level, and semantic level?
Consumer Health Language Complexity Measurement Framework
We built CHELC to incorporate a comprehensive array of linguistic complexity metrics developed in previous research. In this framework, we incorporated metrics of text-level, syntax-level, term-level, and semantic-level CHELC scores (CHELCS) to compare various CHLs through a rank-based algorithm. The overall complexity of CHL (CHELCSoverall) was defined as the average value of 4 complexity scores.
We systematically reviewed the metrics that have been utilized in health readability and complexity assessment studies and comprehensively included credible metrics from all facets of linguistic measures. We performed the search on PubMed using the search terms of health readability to retrieve relevant articles and abstracts, which returned 3605 full-text articles to be screened. After excluding duplicates, non-English articles, and articles not about health readability evaluation or assessment, 9 studies with different assessment metrics were identified ().
Considering the overlap between lay and professional health terms, we proposed to use the ratio of core professional term coverage, which is the percentage of health terms that are in the Systematized Nomenclature of Medicine-Clinical Terms (SNOMED-CT) but not in consumer health vocabulary (CHV). In total, we included 8 metrics for text-level, syntax-level, term-level, and semantic-level complexity measurements in the proposed framework CHELC ().
|Health readability measure||Measure specification||Inclusion||Inclusion or exclusion rationale|
|Word length or syllable length [, ]||Average number of characters (eg, syllables) in a given lexical item||No||Already measured in traditional readability metrics|
|Sentence length [, ]||Average number of words in a sentence||No||Already measured in traditional readability metrics|
|Paragraph length [, ]||Average number of sentences in a paragraph||No||Not applicable for CHLa complexity measure|
|Traditional readability metrics [, , , , ]||Flesch-Kincaid grade level, Simple Measure of Gobbledygook, and Gunning fog||Yes||(1) Well-established formulas that are widely utilized in the literature; (2) Combining word, syllable and sentence length; and (3) Flesch-Kincaid grade and Simple Measure of Gobbledygook are the most used readability metrics|
|Ratio of content word [, ]||Ratio of content words (ie, noun, adjective, verb, and adverb) to functional words (ie, pronoun, determiner, preposition, qualifier, conjunction, interjection)||Yes||Indicator for syntax-level complexity measure; validated in previous literature|
|Ratio of nouns [, ]||Ratio of nouns to all types of parts of speech||Yes||Indicator for syntax-level complexity measure; validated in previous literature|
|Average familiarity score of CHVb [, ]||Frequency use of each CHV term to the lay people||Yes||Indicator to tell how lay health terms are used in CHL|
|Coverage in CHV ||Ratio of CHV terms of all terms||No||We used the ratio of professional health terms|
|Coverage in basic medical dictionary ||Health terms that are in basic medical dictionaries||No||Not applicable for CHL complexity measure|
|Coverage in the Unified Medical Language System [, ]||Ratio of Unified Medical Language System terms||Yes||We utilized the Systematized Nomenclature of Medicine-Clinical Terms as the source of professional health terms|
|Term overlap ratio ||A higher overlap indicates a more cohesive and easier to read text; overlapped terms/all terms in the document||No||Not applicable for CHL complexity measure|
|Vocabulary size ||Distinct word counts in the corpus||No||Not applicable for CHL complexity measure|
|Diversity of health topics ||Ratio of semantic types indicated in the Unified Medical Language System||Yes||Indicator for semantic-level complexity measure; validated in previous literature|
aCHL: consumer health language.
bCHV: consumer health vocabulary.
Text-level complexity utilizes the length of lexical units (eg, words, sentences, paragraphs) to indicate the lexical complexity of health texts. The unit may change depending on whether the length is applied to words (average number of syllables/characters per word) , sentences (average words per sentence) [ ], or paragraphs (average sentences per paragraph) [ ]. As a commonly used metric, it assumes that longer lexical units require more cognitive loads, thereby making the text more complex. Most studies have utilized one or more readability formulas (eg, the Flesch-Kincaid grade level [F-K] and Simple Measure of Gobbledygook [SMOG]) to assess text-level complexity, in which word length or sentence length are considered in the grade level ranking or level of difficulty of the health texts [ ].
For text-level complexity, we applied F-K  and SMOG [ ] to quantify the text-level complexity of CHL. The F-K formula assigned a grade level to indicate the minimum schooling (grade) readers should have to understand the text. The formula assumes that the higher the average number of syllables and words per sentence there are, the more complex the text is [ ]. A grade lower than 5.0 indicates that the text is very easy to comprehend. A grade higher than 12.0 indicates greater difficulty and reading level that requires a college degree or above. Similarly, the SMOG formula considers the number of polysyllabic words [ ]. Essentially, the more polysyllabic words, the higher the SMOG score, and the more difficult the texts are.
Syntax-level complexity utilizes POS distribution to evaluate the complexity of health texts . In general, there are 10 commonly used POS types in English, which can be categorized into content words (ie, noun, adjective, verb, adverb) and functional words (ie, pronoun, determiner, preposition, qualifier, conjunction, interjection). Every word in the health text can be assigned a POS tag. A higher proportion of noun words or content words indicates more complex health texts [ ]. Accordingly, we calculated the ratio of (1) noun words to all POS words and (2) content words to functional words used by each user. We assume that the higher the ratio is, the more complex the CHL is.
Term-level complexity focuses on the complexity related to the density of professional or lay terms (eg, myocardial infarction vs heart attack). According to health readability research, the more professional terms and fewer lay terms there are, the more complex are the health texts . By mapping terms to existing controlled vocabularies, previous studies have typically measured the term-level complexity with the prevalence of professional terms or lay terms [ , , ]. Other studies have also utilized the familiarity scores of consumer health terms (provided in CHV) and term cohesiveness (ie, distinct word count or overlapped term ratio) to measure the term-level complexity [ , ,].
To assess the term-level complexity of the health text, we first used the text processing and entity recognition tool MetaMap  to extract health terms that belong to 84 out of 127 semantic types in the Unified Medical Language System (UMLS, a compendium of over 190 medical controlled vocabularies) that are relevant to biomedicine, health, and nutrition [ , ]. Then we evaluated the density of professional terms and lay terms by mapping our extracted health terms to 2 controlled vocabularies in the UMLS: CHV and SNOMED-CT. CHV contains a collection of lay health concepts and expressions commonly used by health consumers in their everyday communications [ ]. We used the 2015AA version, which includes the latest version of CHV with over 116,324 terms [ ]. SNOMED-CT is the world’s largest standardized vocabulary of clinical and medical terms mostly used in health information systems such as electronic health records [ - ]. In this study, CHV was used to evaluate the usage of lay health terms, whereas SNOMED-CT terms were referred to as professional terms. We developed the following 3 measures to evaluate term-level complexity:
- Prevalence of professional terms: we used the ratio of professional terms (number of distinct SNOMED-CT terms) to all health-related terms (number of distinct health terms) to measure the density of professional terms used by each user in a health corpus. We assumed that the higher the ratio is, the more complex is the CHL.
- Prevalence of core professional terms: we first excluded CHV terms from SNOMED-CT terms to obtain the core professional terms (professional health terms that are not commonly used by laypersons), and used the ratio of core professional terms to all health-related terms to measure the density of core professional terms used by each user in a health corpus. We assumed that the higher the ratio is, the more complex is the CHL.
- Familiarity score of CHV terms: it refers to the familiarity of each CHV term to laypersons [ ]. It is also referred to as the combo score in CHV, which combines frequency score (term difficulty based on its frequency in several large text corpora), context score (term difficulty based on its context), and Concept Unique Identifier score (term difficulty derived from how it is close to well-known easy and difficult concepts in the UMLS). We used a modified combo score that ignores easy words from the Dale-Chall list [ , ]. The higher the score is, the easier the term is. We calculated the average familiarity score of terms written by each user. We assumed that users using more complex CHL have a relatively low average familiarity score for the CHV terms.
Semantic-level complexity refers to the complexity of the diversity of the semantics of health texts. Previous studies have found that if the health text includes more diverse health topics, it is more complex . Operationally, the coverage of semantic types in the UMLS was accounted for semantic-level complexity [ ].
We extracted the health terms using MetaMap and counted the average distinct semantic types of the terms used in CHL. We assume that if a user mentioned more distinct semantic types, his or her CHL is more complex.
Consumer Health Language Complexity Scores
We regarded CHL complexity as a 4-faceted variable, which includes metrics related to text-level, syntax-level, term-level, and semantic-level complexity. Each corpus was represented by a vector of 8 metrics for complexity computation. The values of all 8 metrics were generated for every user in the health corpus.
For each metric, the values for users in all health corpora were ranked [, ] using the same mechanism of Wu et al [ ]. In other words, the ranking value for each metric for users was indicated as the complexity differences among users [ , ]. Except for the familiarity score of CHV terms, the higher the metric value is, the more complex the user’s health language is. It should be noted that we ranked the familiarity score of CHV terms in reverse order. All the missing values of metrics were replaced by the mean of the corresponding metric.
In this proof-of-concept study, each metric in a facet was regarded to contribute equally to the complexity score of that facet. As there is no agreed-upon definition of health text complexity, each facet has equal weight when calculating the overall complexity score (CHELCSoverall). The idea of aggregating the metrics is that described by Wu et al . We aggregated the ranks of metrics for each facet using standard aggregate functions with the same weights [ ]. Other researchers can use different weights for each metric or facet based on their definitions of CHL complexity.
Let fij be the jth observed metric value of the ith facet and f’ij be the jth observed metric value of the specific user whose complexity is calculated in the ith facet.
The formula of CHELCSoverall for every user in the health corpora was as follows:
We defined rij, the rank of the jth metric of the ith facet, as the number of users whose fij is not greater (not smaller for metric familiarity score of CHV terms) than f’ij. Note that m represents the number of facets, ni represents the number of metrics in the ith facet, and N is the total number of users.
We calculated the aggregated rank of the metrics for all facets of CHL complexity. We defined rij/N as the normalized rank ranging from 0 to 1. Then the aggregated complexity score of the ith facet is calculated as . The overall complexity score of all facets is calculated as , which is used to represent the overall CHL complexity of every user. All CHELCS range from 0 to 1, and the higher score means the responding user has more complex CHL complexity in all health corpora.
We utilized CHELC, a complexity measure framework that combines text-level (CHELCStext), syntax-level (CHELCSsyntax), term-level (CHELCSterm), semantic-level (CHELCSsemantic), and overall (CHELCSoverall) complexity scores, to compare the CHLs used in online forums targeting 3 user groups: general public, people with ASD, and deaf and hearing-impaired people. We collected data from various online discussion boards and social media to represent the CHL use of our groups of interest. All 3 data sources in this study were chosen because of their popularity in our interest groups and the convenience of data collection.
We chose AllDeaf , a leading online community for deaf and hearing-impaired people who can communicate in English. As of June 2017, AllDeaf had 63,566 members and 114,801 threads. This community has 22 forums in which people can communicate different aspects of everyday life concerns related to deafness, such as sign language, assistive technologies, and health. The majority of the health-related issues are discussed in the forum Lifestyle, Health, Fitness & Food. After manually removing the threads that were unrelated to health (eg, food recipes), we retained 1639 threads and 31,006 posts from that forum, which includes health discussions from 2005 to 2016.
Another data source was Wrong Planet , which is the main English-language online community developed for people with ASD to discuss everyday life topics. It has 37,350 members and 290,067 threads. Similar to AllDeaf, Wrong Planet has 29 forums. Their users mainly discuss health-related topics in the forum Health, Fitness & Sports. After manually removing unrelated threads in that forum, we obtained 2816 threads and 31,194 posts, covering health discussions from 2004 to 2017.
To represent the use of health language by general health consumers, we selected general health discussions in Yahoo! Answers, which is one of the most popular social Q&A sites used by people to discuss health and other life topics. To make the sample size comparable to those collected from AllDeaf and Wrong Planet, we generated a random sample of 8000 questions and their respective answers in the health category, resulting in 34,048 posts from 2009 to 2014.
Data Processing and Analysis
We extracted health-related posts in the 3 forums and calculated CHELCS for each user using text-level (CHELCStext), syntax-level (CHELCSsyntax), term-level (CHELCSterm), semantic-level (CHELCSsemantic), and overall (CHELCSoverall) complexity. As it is not feasible to analyze behavioral patterns for users contributing to few discussions, we only analyzed posts from users who contributed more than 4 sentences per post on average. For the term-level analysis, we only included users who used more than 20 distinct health terms per post. For text- and syntax-level metrics, we generated the scores for each post through a Web-based readability measurement tool  and then calculated the complexity score for each user in the 3 corpora using a rank-based algorithm. For the term coverage and semantic analysis, we analyzed the data in MySQL (Oracle Corporation) and Microsoft Excel. We visualized the distributions using CHELCStext, CHELCSsyntax, CHELCSterm, CHELCSsemantic, and CHELCSoverall for users in each group in Microsoft Excel. Then we employed a 2-sample Kolmogorov-Smirnov test (K-S test) to determine if the CHELCS of the various groups were significantly different. We conducted an analysis of covariance (ANCOVA) to control for possible impacts of sentence number per post on CHELCS when comparing CHL complexity scores of the 3 groups. More detailed comparison results of 3 groups in 8 metrics were presented in , and correlations of CHELCS scores were analyzed in . K-S test and ANCOVA were performed in R software (The R Foundation for Statistical Computing).
Basic Characteristics of the Corpora
As seen in, although we extracted similar numbers of posts from the 3 corpora regardless of the number of sentences, the numbers of posts with more than 4 sentences were different among the 3 groups. Compared with the other online forums, Yahoo! Answers had the fewest number of posts, the most threads, and involved the most users, but had the least number of distinct health terms contributed by the average user. This might be because of the differences between specialized online forums that are closed communities and general social Q&A sites that are open to the public [ ]. However, the 3 corpora did not have major differences in the number of sentences, sentence lengths, and word lengths, implying that platform differences would not significantly impact the overall CHL used in each community. The 3 user groups shared 68 out of 84 health semantic types in the UMLS.
|Basic textual characters||Health corpora|
|AllDeaf (deaf and hearing-impaired people), n||Wrong Planet (people with ASD), n||Yahoo! Answers (general public), n|
|Number of posts||27,545||26,484||12,560|
|Number of threads||1623||2751||3756|
|Number of involved users||788||2978||9544|
|Average number of sentences per post per user||9.21||9.15||9.63|
|Average number of words per sentence per user||12.14||13.99||13.09|
|Average number of syllables per word per user||1.37||1.41||1.35|
|Average number of letters per word per user||4.14||4.23||4.11|
|Distinct health terms per user||199.87||91.63||39.09|
|Mentioned semantics number||71||71||72|
The CHELCStext, which ranges from 0 to 1, indicates the text-level complexity ranking of the individual user among all users in the 3 online forums.shows the distribution of text-level complexity scores of users in 3 corpora.
The 2-sample K-S test results indicate CHELCStext scores of people with ASD, deaf and hearing-impaired people, and the general public were significantly different (Dd-a=0.332, Pd-a<.001; Dd-p=0.108, Pd-p<.001; Da-p=0.228, Pa-p<.001 [d-a refers to score comparison between CHELCStext of deaf and hearing-impaired users and CHELCStext of users with ASD; d-p refers to score comparison between CHELCStext of the deaf and hearing-impaired users and CHELCStext of the general public; a-p refers to score comparison between CHELCStext of userswith ASD and CHELCStext of the general public] ). As seen in, most deaf and hearing-impaired users wrote texts with lower complexity, whereas users with ASD used more complex texts in their posts. General public users did not significantly differ in their use of polysyllabic words.
After controlling for the number of sentences per post, the ANCOVA results (F2=304.5; P<.001) show that users with ASD (mean 0.606) used significantly more complex texts than the other 2 groups (P<.001) and the general public used significantly more complex texts (mean 0.473) than those in the deaf and hearing-impaired group (mean 0.431; P<.001).
The CHELCSsyntax indicates complexity ranking related to the prevalence of content words, especially nouns. As seen in, the peak CHELCSsyntax scores for deaf and hearing-impaired users ranged from 0.6 to 0.7, whereas the peak CHELCSsyntax scores for users with ASD ranged from 0.4 to 0.5. Regarding general public users, they did not show a clear syntax complexity preference. The two-sample K-S tests indicate that CHELCSsyntax scores were significantly different (Dd-a=0.108, Pd-a<.001; Dd-p=0.153, Pd-p<.001; Da-p=0.098, Pa-p<.001).
After controlling for the number of sentences per post, the results (F2=19.206; P<.001) show that deaf and hearing-impaired users used (mean 0.551) significantly more complex syntax than those in the other 2 groups (P<.001), whereas usage of complex syntax was not significantly different between users with ASD (mean 0.506) and the general public (mean 0.494; P=.07).
The CHELCSterm focuses on the complexity of the health terms used in each forum. As seen in, bimodal distributions were observed in all 3 corpora. Most general public users had relatively higher CHELCSterm ranging from 0.2 to 0.9, whereas most users in the other 2 groups had complexity scores lower than 0.7. The two-sample K-S test results indicate that the CHELCSterm scores of users with ASD, deaf and hearing-impaired, and general public users were significantly different in the prevalence of professional terms (Dd-a=0.208, Pd-a=.009; Dd-p=0.523, Pd-p<.001; Da-p=0.590, Pa-p<.001).
After controlling for the number of sentences per post, the ANCOVA results (F2=3822.320; P<.001) show that the general public users (mean 0.568) used significantly more complex health terms than those in the other 2 groups (P<.001), and deaf and hearing-impaired users (mean 0.370) used more complex terms than users with ASD (mean 0.316; P<.001).
The CHELCSsemantic indicates the diversity of semantic types.shows the distribution of the semantic-level complexity scores in the 3 groups. The two-sample K-S test results indicate that the CHELCSsemantic scores for the 3 groups were significantly different (Dd-a=0.141, Pd-a<.001; Dd-p=0.215, Pd-p<.001; Da-p=0.116, Pa-p<.001). As all health corpora were from social media platforms, the semantics that people utilized might be more influenced by the context than personal health literacy.
By controlling the number of sentences per post, results (F2=53.082; P<.001) show that, on average, general public users (mean 0.514) used more semantic types than those in the other 2 groups (P<.001). Users with ASD (mean 0.478) included more semantic types than deaf and hearing-impaired users (mean 0.416; P<.001). In essence, general public users mentioned more diverse health topics than users with ASD and deaf and hearing-impaired users.
shows the CHELCSoverall for users in the 3 forums. The two-sample K-S test results indicate that the overall CHL complexity scores for users in the 3 corpora were significantly different (Dd-a=0.171, Pd-a<.001; Dd-p=0.250, Pd-p<.001; Da-p=0.129, Pa-p<.001).
After controlling the number of sentences for each participant, the ANCOVA result (F2=167.748; P<.001) shows that, on average, general public users (mean 0.512) had more complex CHL than the other 2 groups (P<.001). Users with ASD (mean 0.476) had more complex CHL than deaf and hearing-impaired users (mean 0.442; P<.001).
As health information on the Web often contains medical jargon and complex sentences, general health consumers often find it hard to search for and understand Web-based health information . We argue that health text complexity measurements need to measure the complexity of various CHLs to inform content providers to tailor health information on the Web for health consumers with varying CHL preferences [ , ]. To this end, we developed CHELCS to quantify CHL complexity differences. We applied this measurement to examine CHL complexity differences of health-related posts in 3 online forums targeting the general public, people with ASD, and deaf and hearing-impaired people. In particular, we collected user-generated discussions from 3 online health communities: Yahoo! Answers, Wrong Planet, and AllDeaf. We calculated 8 health readability metrics for each post in the 3 online forums, and calculated text-level (CHELCStext), syntax-level (CHELCSsyntax), term-level (CHELCSterm), semantic-level (CHELCSsemantic), and overall (CHELCSoverall) complexity scores. We then compared the CHL complexity differences for the 3 user groups based on these 5 complexity scores (CHELCS).
The results supported that CHLs of the 3 user groups were significantly different. General public users used more complex health terms and more diverse semantics compared with users with ASD and deaf and hearing-impaired users. Consistent with previous findings, users with ASD used words with more syllables, fewer content or noun words, and less complex health terms [, ]. Deaf and hearing-impaired users used more content words or nouns, fewer complex words, and less diverse semantics [ , ]. CHELCS results indicated that overall, general public users used more complex CHL than those in the other 2 groups. Overall, the findings from CHELCS measurement were consistent with previous findings of CHL differences among people with ASD, deaf and hearing-impaired people, and public groups.
On the basis of our results, when developing algorithms to simplify health content for different user groups, we need to use more lay health terms for deaf and hearing-impaired users and for users with ASD, less complex words for deaf and hearing-impaired users, and more functional words for users with ASD. For example, as the average F-K grade of MedlinePlus articles is around 8 to 10 [, ], deaf and hearing-impaired users may need more textual simplifications than the other 2 groups.
To the best of our knowledge, this is the first framework that harnesses consumer-generated textual data to assess the complexity of language that they are comfortable using in their health communications. An understanding of the various CHL complexities of different user groups can provide better insights for the development of adaptive readability assessment tools and adaptive text simplification services.
Some limitations should be noted. We could not filter out all the users who are not deaf and hearing impaired or users with ASD, which might affect our findings of the 3 user groups to a certain extent. The data were collected from 3 nontopic–specific health forums. The impact of health topics on text complexity was not controlled in this exploratory study. For example, CHLs by patients with chronic conditions may be more complex than the average healthy consumers. As the average user contributed little text content in the forums, the findings might not fully depict the language complexity preference of each user. More datasets, such as patient blogs and social media, need to be explored in future studies.
In this proof-of-concept study, the framework CHELC was developed with 8 metrics validated in previous health readability studies to compare CHL complexity differences. Although these metrics have been validated in previous studies, to the best of our knowledge, they have not been used to compare CHLs of different consumer groups. With a lack of research in this field, there is no agreed-upon definition of CHL complexity with respect to different aspects. Therefore, we cannot find a ground truth dataset or standard to validate CHELCS when estimating CHL complexity differences. In this exploratory study, the evaluation of CHELCS was based on previous research findings of the 3 groups in terms of their language complexity preferences. Although our results were consistent with previous findings, this framework and complexity scores are more informative than conclusive. For example, the scores will be different if more metrics are included in this framework, or if the weights of different metrics are defined differently. Also, to more accurately estimate adaptive simplification efforts, it is critical that future studies further assess the CHELCS difference between Web-based consumer health information sources and various CHLs.
The results of this study demonstrate that differences exist among health consumers with respect to the complexity of their language use when discussing health-related topics. A complexity measurement framework (CHELC) and its accompanying scores (CHELCS) were developed to quantify CHL complexity differences among different user groups. Future studies could further apply CHELCS to other datasets from different user groups. Specifically, there is a clear need for the research on understanding CHL complexity differences that translates to adaptive simplification services for different user groups.
The authors would like to thank Zhiwei Chen for his help with MetaMap. The authors would also like to thank Dr. Sanghee Oh for sharing with them the data collected from Yahoo! Answers. This project was partially supported by the National Institute on Aging of the National Institutes of Health (NIH) under award number R21AG061431 and the University of Florida Clinical and Translational Science Institute, which is supported in part by the NIH National Center for Advancing Translational Sciences under award number UL1TR001427. The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIH.
Conflicts of Interest
Complexity score of seven metrics in consumer health language complexity scores.PDF File (Adobe PDF File), 344 KB
Correlations of consumer health language complexity scores in health corpora.PDF File (Adobe PDF File), 149 KB
- Proulx J, Kandula S, Hill B, Zeng-Treitler Q. Creating Consumer Friendly Health Content: Implementing and Testing a Readability Diagnosis and Enhancement Tool. In: Proceedings of the 2013 46th Hawaii International Conference on System Sciences.: IEEE; 2013 Presented at: HICSS'13; January 7-10, 2013; Wailea, Maui, HI, USA p. 2445-2453. [CrossRef]
- Smith CA, Wicks PJ. PatientsLikeMe: Consumer health vocabulary as a folksonomy. AMIA Annu Symp Proc 2008 Nov 6:682-686 [FREE Full text] [Medline]
- Zeng QT, Tse T. Exploring and developing consumer health vocabularies. J Am Med Inform Assoc 2006;13(1):24-29 [FREE Full text] [CrossRef] [Medline]
- Lewis D, Brennan PF, McCray AT, Bachman J, Tuttle M, Bachman J. If we build it, they will come: Standardized consumer vocabularies. In: Patel VL, Rogers R, Haux R, editors. MEDINFO 2001. Amsterdam, Netherlands: IOS Press; 2001:1530.
- Lee KJ. Literature review in computational linguistics issues in the developing field of consumer informatics: finding the right information for consumer’s health information need. In: Song M, Wu YF, editors. Handbook of Research on Text and Web Mining Technologies. Hershey, Pennsylvania, USA: IGN Global; 2009:758-765.
- Keselman A, Logan R, Smith CA, Leroy G, Zeng-Treitler Q. Developing informatics tools and strategies for consumer-centered health communication. J Am Med Inform Assoc 2008;15(4):473-483 [FREE Full text] [CrossRef] [Medline]
- He Z, Chen Z, Oh S, Hou J, Bian J. Enriching consumer health vocabulary through mining a social Q&A site: A similarity-based approach. J Biomed Inform 2017 May;69:75-85 [FREE Full text] [CrossRef] [Medline]
- Clauson KA, Zeng-Treitler Q, Kandula S. Readability of patient and health care professional targeted dietary supplement leaflets used for diabetes and chronic fatigue syndrome. J Altern Complement Med 2010 Jan;16(1):119-124 [FREE Full text] [CrossRef] [Medline]
- Kandula S, Zeng-Treitler Q. Creating a gold standard for the readability measurement of health texts. AMIA Annu Symp Proc 2008 Nov 6:353-357 [FREE Full text] [Medline]
- Leroy G, Miller T, Rosemblat G, Browne A. A balanced approach to health information evaluation: A vocabulary-based naïve Bayes classifier and readability formulas. J Am Soc Inf Sci 2008;59(9):1409-1419. [CrossRef]
- Kandula S, Curtis D, Zeng-Treitler Q. A semantic and syntactic text simplification tool for health content. AMIA Annu Symp Proc 2010 Nov 13;2010:366-370 [FREE Full text] [Medline]
- Abrahamsson E, Forni T, Skeppstedt M, Kvist M. Medical Text Simplification Using Synonym Replacement: Adapting Assessment of Word Difficulty to a Compounding Language. In: Proceedings of the 3rd Workshop on Predicting and Improving Text Readability for Target Reader Populations. 2014 Presented at: PITR'14; April 26-30, 2014; Gothenburg, Sweden p. 57-65 URL: https://pdfs.semanticscholar.org/ecc0/d5e746cb30d2c4132df760646b18d823631c.pdf [CrossRef]
- Kauchak D, Leroy G. Moving beyond readability metrics for health-related text simplification. IT Prof 2016;18(3):45-51 [FREE Full text] [CrossRef] [Medline]
- Leroy G, Endicott JE, Kauchak D, Mouradi O, Just M. User evaluation of the effects of a text simplification algorithm using term familiarity on perception, understanding, learning, and information retention. J Med Internet Res 2013 Jul 31;15(7):e144 [FREE Full text] [CrossRef] [Medline]
- Leroy G, Helmreich S, Cowie JR, Miller T, Zheng W. Evaluating online health information: beyond readability formulas. AMIA Annu Symp Proc 2008 Nov 6:394-398 [FREE Full text] [Medline]
- Wu DT, Hanauer DA, Mei Q, Clark PM, An LC, Proulx J, et al. Assessing the readability of ClinicalTrials.gov. J Am Med Inform Assoc 2016 Mar;23(2):269-275 [FREE Full text] [CrossRef] [Medline]
- Keselman A, Tse T, Crowell J, Browne A, Ngo L, Zeng Q. Assessing consumer health vocabulary familiarity: an exploratory study. J Med Internet Res 2007 Mar 14;9(1):e5 [FREE Full text] [CrossRef] [Medline]
- Leroy G, Endicott JE. Combining NLP With Evidence-Based Methods to Find Text Metrics Related to Perceived and Actual Text Difficulty. In: Proceedings of the 2nd ACM SIGHIT International Health Informatics Symposium. USA: ACM; 2012 Presented at: IHI'12; January 28 - 30, 2012; Miami, Florida, USA p. 749-754. [CrossRef]
- Moen H, Peltonen L, Koivumäki M, Suhonen H, Salakoski T, Ginter F, et al. Improving layman readability of clinical narratives with unsupervised synonym replacement. Stud Health Technol Inform 2018;247:725-729. [Medline]
- Jacobs RJ, Caballero J, Ownby RL, Kane MN. Development of a culturally appropriate computer-delivered tailored Internet-based health literacy intervention for Spanish-dominant Hispanics living with HIV. BMC Med Inform Decis Mak 2014 Nov 30;14:103 [FREE Full text] [CrossRef] [Medline]
- Claassen AA, van den Ende CH, Meesters JJ, Pellegrom S, Kaarls-Ohms BM, Vooijs J, et al. How to best distribute written patient education materials among patients with rheumatoid arthritis: a randomized comparison of two strategies. BMC Health Serv Res 2018 Mar 27;18(1):211 [FREE Full text] [CrossRef] [Medline]
- Chesser A, Burke A, Reyes J, Rohrberg T. Navigating the digital divide: A systematic review of eHealth literacy in underserved populations in the United States. Inform Health Soc Care 2016;41(1):1-19. [CrossRef] [Medline]
- Shardlow M. A survey of automated text simplification. SpecialIssue 2014;4(1):58-70 [FREE Full text] [CrossRef]
- Yu B, He Z. Exploratory Textual Analysis of Consumer Health Languages for People Who Are D/deaf and Hard of Hearing. In: 2017 IEEE International Conference on Bioinformatics and Biomedicine.: IEEE; 2017 Presented at: BIBM'17; November 13-16, 2017; Kansas City, MO, USA p. 1288-1291. [CrossRef]
- Gemoets D, Rosemblat G, Tse T, Logan R. Assessing readability of consumer health information: an exploratory study. Stud Health Technol Inform 2004;107(Pt 2):869-873. [Medline]
- Walsh TM, Volsko TA. Readability assessment of internet-based consumer health information. Respir Care 2008 Oct;53(10):1310-1315 [FREE Full text] [Medline]
- de Oliveira GS, Jung M, Mccaffery KJ, McCarthy RJ, Wolf MS. Readability evaluation of internet-based patient education materials related to the anesthesiology field. J Clin Anesth 2015 Aug;27(5):401-405. [CrossRef] [Medline]
- Kim H, Goryachev S, Rosemblat G, Browne A, Keselman A, Zeng-Treitler Q. Beyond surface characteristics: a new health text-specific readability measurement. AMIA Annu Symp Proc 2007 Oct 11:418-422 [FREE Full text] [Medline]
- Feng L, Jansche M, Huenerfauth M, Elhadad N. A Comparison of Features for Automatic Readability Assessment. In: Proceedings of the 23rd International Conference on Computational Linguistics. USA: Association for Computational Linguistics; 2010 Presented at: COLING'10; August 23-27, 2010; Beijing, China p. 276-284.
- McKee MM, Paasche-Orlow MK, Winters PC, Fiscella K, Zazove P, Sen A, et al. Assessing health literacy in deaf American sign language users. J Health Commun 2015;20(Suppl 2):92-100 [FREE Full text] [CrossRef] [Medline]
- Zazove P, Meador HE, Reed BD, Gorenflo DW. Deaf persons' English reading levels and associations with epidemiological, educational, and cultural factors. J Health Commun 2013;18(7):760-772. [CrossRef] [Medline]
- Smith SR, Samar VJ. Dimensions of deaf/hard-of-hearing and hearing adolescents' health literacy and health knowledge. J Health Commun 2016;21(sup2):141-154 [FREE Full text] [CrossRef] [Medline]
- Koyama T, Tachimori H, Sawamura K, Koyama A, Naganuma Y, Makino H, et al. Mental health literacy of autism spectrum disorders in the Japanese general population. Soc Psychiatry Psychiatr Epidemiol 2009 Aug;44(8):651-657. [CrossRef] [Medline]
- Chung JW, Min HJ, Kim J, Park JC. Enhancing Readability of Web Documents by Text Augmentation for Deaf People. In: Proceedings of the 3rd International Conference on Web Intelligence, Mining and Semantics. USA: ACM; 2013 Presented at: WIMS'13; June 12-14, 2013; Madrid, Spain p. 1-10. [CrossRef]
- Gabriels RL, Cuccaro ML, Hill DE, Ivers BJ, Goldson E. Repetitive behaviors in autism: relationships with associated clinical features. Res Dev Disabil 2005;26(2):169-181. [CrossRef] [Medline]
- Kushalnagar P, Smith S, Hopper M, Ryan C, Rinkevich M, Kushalnagar R. Making cancer health text on the internet easier to read for deaf people who use American sign language. J Cancer Educ 2018 Feb;33(1):134-140 [FREE Full text] [CrossRef] [Medline]
- Turner M. Annotation: Repetitive behaviour in autism: a review of psychological research. J Child Psychol Psychiatry 1999 Sep;40(6):839-849. [CrossRef] [Medline]
- Benford P, Standen P. The internet: a comfortable communication medium for people with Asperger syndrome (AS) and high functioning autism (HFA)? J Assist Technol 2009;3(2):44-53. [CrossRef]
- Pollard Jr RQ, Barnett S. Health-related vocabulary knowledge among deaf adults. Rehabil Psychol 2009 May;54(2):182-185. [CrossRef] [Medline]
- Ley P, Florio T. The use of readability formulas in health care. Psychol Health Med 1996;1(1):7-28. [CrossRef]
- Zheng J, Yu H. Readability formulas and user perceptions of electronic health records difficulty: A corpus study. J Med Internet Res 2017 Mar 2;19(3):e59 [FREE Full text] [CrossRef] [Medline]
- Leroy G, Helmreich S, Cowie JR. The Effects of Linguistic Features and Evaluation Perspective on Perceived Difficulty of Medical Text. In: Proceedings of the 2010 43rd Hawaii International Conference on System Sciences.: IEEE; 2010 Presented at: HICSS'10; January 5-8, 2010; Honolulu, HI, USA p. 1-10 URL: https://ieeexplore.ieee.org/document/5428381 [CrossRef]
- Flesch RF. How to Write Plain English: A Book for Lawyers and Consumers : With 60 Before-And-After Translations from Legalese. New York, New York, USA: HarperCollins; 1979.
- McLaughlin GH. SMOG grading: A new readability formula. J Read 1969;12(8):639-646 [FREE Full text]
- Aronson AR. Effective mapping of biomedical text to the UMLS Metathesaurus: the MetaMap program. Proc AMIA Symp 2001:17-21 [FREE Full text] [Medline]
- Bodenreider O. The Unified Medical Language System (UMLS): integrating biomedical terminology. Nucleic Acids Res 2004 Jan 1;32(Database issue):D267-D270 [FREE Full text] [CrossRef] [Medline]
- McCray AT, Nelson SJ. The representation of meaning in the UMLS. Methods Inf Med 1995 Mar;34(1-2):193-201. [CrossRef] [Medline]
- Donnelly K. SNOMED-CT: The advanced terminology and coding system for eHealth. Stud Health Technol Inform 2006;121:279-290. [Medline]
- Agrawal A, He Z, Perl Y, Wei D, Halper M, Elhanan G, et al. The readiness of SNOMED problem list concepts for meaningful use of electronic health records. Artif Intell Med 2013 Jun;58(2):73-80. [CrossRef] [Medline]
- SNOMED International. 5-Step Briefing URL: http://www.snomed.org/snomed-ct/what-is-snomed-ct [accessed 2018-06-07]
- National Library of Medicine. UMLS Metathesaurus - CHV (Consumer Health Vocabulary) URL: https://www.nlm.nih.gov/research/umls/sourcereleasedocs/current/CHV/sourcerepresentation.html [accessed 2019-05-03]
- Ke W, Zhang T, Chen J, Wan F, Ye Q, Han Z. Texture Complexity Based Redundant Regions Ranking for Object Proposal. In: Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops.: IEEE; 2016 Presented at: CVPRW'16; June 26 - July 1, 2016; Las Vegas, NV, USA p. 1083-1091. [CrossRef]
- Plumlee MA. The Effect of Information Complexity on Analysts' Use of That Information. Account Rev 2003;78(1):275-296. [CrossRef]
- Kharlamov E, Giacomelli L, Sherkhonov E, Grau B, Kostylev E, Horrocks I. Ranking, Aggregation, and Reachability in Faceted Search with SemFacet. In: Proceedings of the 16th International Semantic Web Conference. 2017 Presented at: ISWC'17; October 21-25, 2017; Vienna, Austria URL: http://ceur-ws.org/Vol-1963/paper650.pdf
- Wagner AJ, Ladwig G, Tran T. Browsing-Oriented Semantic Faceted Search. In: Proceedings of the 22nd International Conference on Database and Expert Systems Applications. 2011 Presented at: DEXA'11; August 29 - September 2, 2011; Toulouse, France URL: https://www.aifb.kit.edu/images/8/89/Dexa2011-facetedSearch.pdf
- Kharlamov E, Giacomelli L, Sherkhonov E, Grau B, Kostylev E, Horrocks I. SemFacet: Making Hard Faceted Search Easier. In: Proceedings of the 2017 ACM on Conference on Information and Knowledge Management. USA: ACM; 2017 Presented at: CIKM'17; November 6 - 10, 2017; Singapore, Singapore p. 2475-2478. [CrossRef]
- Deaf Community. URL: http://www.alldeaf.com/ [accessed 2019-06-01]
- Wrong Planet. URL: https://wrongplanet.net/ [accessed 2019-06-01]
- ReadablePro. A Readability Tool, With Extra Power. URL: https://readable.com/ [accessed 2019-03-13]
- Lerman K, Ghosh R. Arxiv preprints. 2010 Mar 13. Information Contagion: An Empirical Study of the Spread of News on Digg and Twitter Social Networks URL: https://arxiv.org/abs/1003.2664 [accessed 2019-01-05]
|ANCOVA: analysis of covariance|
|ASD: autism spectrum disorder|
|CHELC: consumer health language complexity|
|CHELCS: consumer health language complexity scores|
|CHL: consumer health language|
|CHV: consumer health vocabulary|
|F-K: Flesch-Kincaid grade level|
|NIH: National Institutes of Health|
|POS: parts of speech|
|Q&A: question and answer|
|SNOMED-CT: Systematized Nomenclature of Medicine-Clinical Terms|
|UMLS: Unified Medical Language System|
Edited by G Eysenbach; submitted 31.10.19; peer-reviewed by D He, X Liu, K Chen; comments to author 16.12.19; revised version received 21.01.20; accepted 21.02.20; published 21.05.20
©Biyang Yu, Zhe He, Aiwen Xing, Mia Liza A Lustria. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 21.05.2020.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.