Advertisement: Preregister now for the Medicine 2.0 Congress
Term Identification Methods for Consumer Health Vocabulary Development
Qing T Zeng1, PhD; Tony Tse2, PhD; Guy Divita2,3, MS; Alla Keselman2,4, PhD; Jon Crowell1, MS; Allen C Browne2, MS; Sergey Goryachev1, MS; Long Ngo1, PhD
1Decision Systems Group, Brigham and Women’s Hospital, Harvard Medical School, Boston, MA, USA
2LHNCBC, National Library of Medicine, NIH, DHHS, Bethesda, MD, USA
3Management Systems Designers, Inc, Fairfax, VA, USA
4Aquilent, Inc, Laurel, MD, USA
Harvard Medical School
Decision Systems Group
Brigham and Women's Hospital
Thorn 304, 75 Francis Street
Boston, MA 02115
Phone: +1 617 732 7694
Fax: +1 617 739 3672
Background: The development of consumer health information applications such as health education websites has motivated the research on consumer health vocabulary (CHV). Term identification is a critical task in vocabulary development. Because of the heterogeneity and ambiguity of consumer expressions, term identification for CHV is more challenging than for professional health vocabularies.
Objective: For the development of a CHV, we explored several term identification methods, including collaborative human review and automated term recognition methods.
Methods: A set of criteria was established to ensure consistency in the collaborative review, which analyzed 1893 strings. Using the results from the human review, we tested two automated methods—C-value formula and a logistic regression model.
Results: The study identified 753 consumer terms and found the logistic regression model to be highly effective for CHV term identification (area under the receiver operating characteristic curve = 95.5%).
Conclusions: The collaborative human review and logistic regression methods were effective for identifying terms for CHV development.
(J Med Internet Res 2007;9(1):e4)
Consumer health information; vocabulary; natural language processing
Two important steps in vocabulary development are (1) the identification of candidate strings (ie, words or phrases) in a domain and (2) the determination of which of these should be included in a vocabulary as “valid” terms, also called “termhood determination.” Health vocabulary development, which has a long history, requires significant effort for collecting candidate terms and determining termhood . While vocabularies such as SNOMED (Systematized Nomenclature of Medicine) and ICD-9 (International Classification of Diseases, Ninth Revision) include many health terms, there is no consensus on termhood criteria (ie, what constitutes a “term”) . The decision to include terms in a vocabulary is made for a particular domain for certain tasks (eg, indexing or billing). Thus, the review criteria and procedures used by vocabulary developers, which are often not published, inevitably differ. Terms included in health vocabularies also vary significantly. For instance, in the Unified Medical Language System (UMLS), the same concept is often represented in various source vocabularies by different terms. The terms “head ache” and “cranial pain” are both synonyms of the UMLS concept “headache.” The source vocabulary for “head ache” is DXplain, and the source vocabulary for “cranial pain” is MeSH (medical subject heading).
Research and development of controlled consumer health vocabularies (CHVs) is a relatively new endeavor in the health vocabulary field . In the general biomedical literature, research on consumer understanding of medical words and concepts has focused primarily on relatively short lists of discrete terms in various specialties. In the informatics domain, a few companies (eg, Apelon and WellMed) offer proprietary CHV products, though these products have not been publicly evaluated.
The general goal of our CHV research is to help overcome the vocabulary gap between consumers and health information provided by informatics applications. The specific aim of this paper is to elucidate term identification methods for CHVs. CHV research has largely been driven by the proliferation of health-related materials on the Web, the emergence of electronic personal health records, as well as the growing availability of various consumer health applications (eg, decision support tools). Over the past five years, researchers have found that consumer terms are not well covered by the existing health vocabularies, which mostly represent the language of health professionals [4-9]. Indeed, expressions used by consumers to describe health-related concepts and relationships among such concepts frequently differ on multiple levels (ie, syntactic, conceptual, and explanatory) from those of professionals. Thus, consumer health informatics research and application development will benefit from the development of CHVs.
Developing and validating a comprehensive CHV is challenging because “consumers” constitute a plethora of highly diverse groups. Further, individuals uniquely acquire health-related terms and concepts from formal and informal sources (eg, media exposure) and from personal experiences. Nevertheless, there is strong evidence of the stability of lay health language among particular populations, for specific tasks .
We have been working on an open access and collaborative (OAC) CHV project. The first step in creating the OAC CHV was to identify consumer terms since surface forms, represented as strings in written text, are more tractable than concepts (ie, underlying meanings) or semantic relations, both of which require in-depth understanding of term usage, rhetorical intent, and explanatory models. Because consumer terms are heterogeneous and even less well defined than professional terms , the termhood determination task proved to be particularly challenging. Our term identification effort has been guided by two principles:
1. CHVs consist of actual terms commonly used by consumers (in any particular discourse group).
2. CHV terms must allow for computer processing of consumer language.
Since many professional health vocabulary terms are already used by consumers, though in some cases with different or broader semantics (eg, “diabetes” for diabetes mellitus, types 1 and 2), we focused on consumer terms not yet represented in existing vocabularies (eg, “broken finger” for any type of fracture in the “distal,” “middle,” or “proximal phalanges”).
Because the number of candidate strings is often very large in any domain, researchers have explored the use of corpus-based automated term recognition (ATR) methods for extracting the most promising strings for human review from domain-specific documents [11, 12]. ATRs vary from statistical or information theory–based approaches (eg, t test)  to syntax-based methods (eg, noun phrase extraction and context analysis)  and hybrid mechanisms (eg, C-value formula) [15, 16]. Both the t test and the C-value formula have been used successfully in termhood determination. Such studies reinforce the general notion that strings typically considered as terms share some common characteristics, such as words in a term tend to occur more frequently together, terms are often noun phrases, and terms may be part of several longer strings.
In the biomedical domain, ATR methods have been applied to Medline literature  and clinical reports . While most ATR methods outside the biomedical domain were designed to be general purpose, biomedical ATR methods tend to be more narrowly focused . The type of terms targeted by ATR vary, including gene and protein names in a number of recent studies [18-21].
In this study, we first identified CHV terms through collaborative review of strings derived from query logs of a consumer health site . Because of the considerable variability in lay health expressions, standardized review criteria and procedures to ensure consistency in selecting CHV terms were developed. After obtaining the human-reviewed n-grams (ie, n word strings), we experimented with two ATR methods: logistic regression and the C-value formula. The initial features used in the regression model were informed by existing ATR methods, in particular, the C-value model  and the termhood formula proposed by Wermter and Hahn . We also evaluated the popular C-value method.
Our use of ATRs in this study differs from that in prior studies in the biomedical domain in two aspects: (1) short phrases from query logs were used as the text corpus rather than entire sentences from full-text sources, and (2) “new” CHV terms, not yet part of existing vocabularies, were identified rather than “pre-existing” terms such as UMLS terms.
The term identification study had three components:
- Candidate string extraction from a query log data set of terms that could not be mapped to UMLS
- Collaborative manual review of a subset of the candidate strings and identification of CHV terms
- Application of ATR methods (the C-value formula and logistic regression models) to human-reviewed CHV terms
Candidate String Extraction
We obtained a set of query log files  from the MedlinePlus site covering the period from October 2002 to October 2003, courtesy of the National Library of Medicine (NLM). The log data were preprocessed to filter out all queries that were not in English, appeared to be machine generated (eg, very large numbers of queries from the same IP address within a minute), and that were redundant (ie, from the same host at time intervals of less than 5 minutes).
The preprocessed queries were then mapped to the 2004AA version of the UMLS Metathesaurus using lexical methods (ie, removing non-alphanumeric symbols, stemming, normalization, and truncation). Queries that did not map to the UMLS Metathesaurus were broken into n-grams. N-grams that matched terms in the Metathesaurus were removed, and the remaining n-grams were collected into sets by frequency and number of words.
We used n-gram analysis to find candidate terms from unmapped query strings. The n-gram analysis uses the frequencies of n-grams and text fragments of n words in a text sample to estimate the likelihood that a string is a potential term. In general, the more frequently an n-gram appears in text documents, the increased likelihood that the n-gram is a “useful” term.
Collaborative Manual Review
Six researchers (first six of the authors) reviewed candidate strings (n-grams) collaboratively. First, each reviewer independently reviewed a subset of the n-grams (n = 1 to 4 and frequency > 50) and voted on whether they should be considered CHV terms. Unanimous votes for n-grams that were reviewed by at least three people were entered as “master” votes. Otherwise, termhood was discussed by the entire group until consensus was reached and a master vote was cast. To support reviewers from geographically distributed locations and to calculate votes, a specially designed Web-based application  was utilized (Figure 1).
[view this figure]
|Figure 1. Application to support collaborative manual review of candidate strings|
Through several iterations of votes and discussion, we established the following review criteria:
- CHV terms should be syntactic constituents or phrases such as a noun phrase or adjective phrase (eg, “bypass surgery” is a phrase, but “fever in” is not). Special attention should be given to noun phrases.
- CHV terms should have independent semantics and should not only occur as a part of longer valid terms or as a part of wild card searches (eg, [chicken-, small-] “pox vaccine” is not considered a CHV term).
- CHV terms should be specific to the medical domain (eg, “Google” and “Yahoo” are general words, not CHV terms).
- CHV terms should function as semantic components in addition to functioning as syntactic components (eg, stop words “the” and “a” as well as empty verbs “make” and “take” are not considered CHV terms).
- N-grams representing existing UMLS medical concepts are considered to be CHV terms, but CHV terms may represent non-UMLS concepts.
- Eponymous forms of CHV terms are considered to be CHV terms (eg, “Parkinson’s”).
- CHV terms may include spelling errors, (eg, “Chron's disease”). These misspelled terms are given the label “disparaged.”
- Terms with distinct clinical semantics (eg, “result”) are considered to be CHV terms, regardless of ambiguity and/or vagueness in other domains.
We singled out several types of terms for future investigation and assigned special labels to them:
- meta: A term that is usually used to indicate the category/type of information sought or presented (eg, “picture,” “guideline,” and “tutorial”).
- modifier: A term not typically used by itself, but for limiting or qualifying other terms (eg, “sexually” as in “sexually active”).
- relation: A term not typically used by itself, but used to describe relations among concepts (eg, “caused by” and “results in”). We also include the unary relation “not” in this set.
Currently, we consider terms classified as meta and modifier to be CHV terms, but relations are not considered CHV terms.
Once these review criteria were established, researchers double-checked the previously cast master votes for compliance. A second round of discussion resulted in some adjustments to the votes.
Application of Automated Term Recognition (ATR)
We explored the use of two ATR methods to facilitate candidate selection for human review: (1) the C-value method (C loosely stands for “candidate collection”) and (2) logistic regression.
We applied the C-value method to the strings that had already been reviewed. First, the strings were parsed to filter out single-word strings and strings that were not noun phrases. The C-value was calculated using the formula  given in Textbox 1.
[view this box]
|Textbox 1. The C-value was calculated using this formula|
To create the logistic regression model that predicts the termhood of a candidate string a, we explored syntactic category, frequency of occurrence, string length, word count and number, frequency and termhood status of a’s nesting, and nested strings as variables and used the master vote as outcome. Human-reviewed strings were used as the training and testing data sets. The initial feature variables were as follows:
- part-of-speech (POS) tag (eg, noun or adjective) of the first word
- POS tag of the last word
- noun phrase status (ie, yes/no)
- word count (ie, number of words in a)
- number of distinct a’s nesting string b
- number of repeated b
- percentage of distinct b that are known valid (UMLS) terms
- percentage of repeated b that are known valid (UMLS) terms
- number of distinct a’s nested string c
- number of repeated c
- percentage of distinct c that are known valid (UMLS) terms
- percentage of repeated c that are known valid (UMLS) terms
- frequency of a
- number of distinct host h that a originated from
- average number of distinct queries containing a per host
The frequency distribution of the POS tags (variables 1 and 2) required them to be collapsed into fewer categories for modeling. The original tags came from a Brill-style, rule-based POS tagger developed by Mark Hepple . We first transformed them into a smaller set of tags used by the UMLS SPECIALIST Lexicon of the National Library of Medicine (NLM) . (Details of the transformation rules can be found in .) Several tags appeared with low frequency and were then merged: the tags AUXILARY and MODAL were merged with VERB, and the tags CONJUNCTION, DETERMINER, NUMBER, SYM, UNKNOWN, PRONOUN, and PREP were merged into a new category, OTHER.
The continuous variables (variables 4 to 15) were dichotomized based on the median value. The dichotomized variables were used in the logistic regression to predict or explain the probability of having a term voted “yes” for termhood.
The logistic regression model building was carried out by a stepwise procedure. After calculating the odds ratio estimates, most of the variables were dropped. The remaining variables 1, 2, 3, 6, 10, and 15 were represented in the regression formula as FirstPOS, LastPOS, np_value, repeat_sup_gt_median, repeat_sub_gt_median, and distinct_perhost_gt_median.
For both the C-value formula and the regression model, we calculated the sensitivity and specificity at different thresholds to create the receiver operating characteristic (ROC) curves. To estimate the area under the ROC curve for the logistic regression, we used the c-statistic  (note that this is not the same as C-value). It has the following meaning. From the final multivariable logistic regression model, the predicted probability of the termhood voted “yes” can be computed for each term. For any two terms, one with vote “yes” and one with vote “no,” if the predicted probability for vote “yes” is higher than the predicted probability for vote “no,” then we have a concordant pair. If the predicted probability of vote “no” is higher, then we have a discordant pair. If the pair is neither concordant nor discordant, then it is tied. Let T be the total number of all possible yes-no pairs of all terms. Let C be the number of concordant pairs, and D the number of discordant pairs. The c-statistic is calculated as c = (C + 0.5(T − C − D)) / T.
We identified 18454 candidate n-grams (n = 1 to 5); 7967 were reviewed by at least one reviewer, and 1893 distinct n-grams received master votes (Table 1). Among the n-grams with master votes, 23 were meta, 39 were modifier, and 5 were relation.
[view this table]
|Table 1. Number of n-grams with master votes and number of n-grams voted as CHV terms|
[view this figure]
|Figure 2. The logistic regression model|
The logistic regression model is shown in Figure 2. In this logistic regression model, syntactic information (first 9 variables) and nesting pattern (last 3 variables) determine the termhood. The importance of syntactic information has long been recognized by models like the C-value. Conspicuously, word count and frequency are missing from our model, though longer and more frequent strings are more likely to be considered terms. To a large extent, length and frequency are reflected by the nesting patterns: very short strings are likely to be part of many nesting strings, and less frequent strings are likely to be coincidental combinations of more common words, meaning that it would have more nested strings.
The ROC curves for C-value and the regression model are shown in Figure 3. The area under the ROC curve (AUC) is 70.9% for the C-value method and 95.5% for the regression model. Higher AUC signifies increased distinguishing power: 100% = perfect discriminative ability, 50% = no ability, < 50% = predications were made in the wrong direction. Thus, the AUC results suggest the regression model to be very effective and better than the C-value for identifying CHV terms.
[view this figure]
|Figure 3. Curves for C-value and the regression model|
This paper reports on several term identification methods for the OAC CHV project. We established a set of criteria and procedures to conduct a manual review, resulting in multiple reviewers reaching consensus on 1893 n-grams, including identification of 753 new terms for inclusion in the OAC CHV that were not in the 2004AC version of UMLS.
The OAC termhood criteria were established collaboratively, reflecting the reviewers’ backgrounds in several different fields: controlled vocabulary, health informatics, linguistics, cognitive science, and computer science. While the OAC termhood criteria could be further refined and termhood criteria for health vocabularies are often not published, we believe publishing such criteria could benefit vocabulary research. For instance, many articles evaluate vocabularies and study methods of mapping one vocabulary to another [28-31]. These evaluations and mapping methods could be better guided by the termhood criteria of target vocabularies.
In CHV research, the termhood issue is of particular importance because there has been limited discussion and little consensus on what should be considered a consumer term. Is “sun poisoning” an acceptable term? How about “skin conditions?” As was pointed out in the Introduction, health professional vocabularies do not always agree on the termhood of a phrase. Consumer expressions, however, require more scrutiny because it is harder to determine their semantics and contexts of usage.
We tested two ATR methods (C-value and logistic regression) on the human-reviewed n-grams. The C-value was useful for determining termhood, though it did not have high distinguishing power (AUC = 70.9%). The AUC for the logistic regression model was 95.5%, which is fairly satisfactory.
These results suggest that a specially fitted logistic regression model is better suited than the generic C-value method for the task of identifying CHV terms according to our criteria. The C-value method’s performance problem was partially caused by issues unique to this data set, among them the inclusion of infrequent misspellings and the high frequency of most candidates, which made frequency a less reliable predicator. The imperfection in noun-phrase parsing is not unexpected, though the relatively short query string posed a greater challenge for parsing. Like many vocabularies, OAC includes strings that are single words and are not noun phrases, while C-value is typically calculated for multiword noun phrases.
The logistic regression model demonstrated excellent suitability for OAC termhood determination. It may have to be altered to be used with other corpora or for other types of vocabularies due to the particularities of query-based corpus attributes such as the short length of the documents. Nonetheless, training of predictive models for a particular corpus and vocabulary is a generalizable strategy. Although general principles exist, the determination of which strings are to be considered legitimate vocabulary terms often depends on the domain and the vocabulary developers’ criteria (eg, including verb phrases  or not).
The regression model utilizes syntactic and nesting pattern features; both types of features are well-recognized termhood indicators. A concern often raised about CHV research is that the syntax and semantic of consumer phrases are too unruly to be represented in a computable vocabulary. The fact that many consumer phrases have common term characteristics suggests that they are tractable terms.
Our study has several limitations. Because consumer utterances are not readily available as corpora of medical literature or clinical records, we used query logs that contained relatively few complete sentences. Subsequently, this resulted in many POS and noun phrase analysis errors. As well, we only had researchers and not lay consumers review the candidate terms, due to budget and logistic constraints. However, the analysis was based on utterances from queries submitted by tens of thousands of consumers.
Based on the result of this study, we plan to apply the logistic regression model to the candidate n-grams and select those predicted to be terms for human review. We also plan to add the identified CHV terms to OAC. The authors associated with NLM are interested in investigating similar techniques to aid in identifying candidate terms for inclusion into the SPECIALIST Lexicon of the NLM, and for quality control.
We thank the National Library of Medicine (NLM) for sharing the MedlinePlus query log data. This work is supported by the National Institutes of Health (NIH) grant R01 LM07222 and by the Intramural Research Program of the NIH, NLM.
Conflicts of Interest
- Chute CG. Clinical classification and terminology: some history and current observations. J Am Med Inform Assoc 2000 May;7(3):298-303 [FREE Full text] [Medline]
- Cimino JJ. Desiderata for controlled medical vocabularies in the twenty-first century. Methods Inf Med 1998 Nov;37(4-5):394-403. [Medline]
- Zeng QT, Tse T. Exploring and developing consumer health vocabularies. J Am Med Inform Assoc 2006 Jan;13(1):24-29 [FREE Full text] [Medline] [CrossRef]
- Brennan PF, Aronson AR. Towards linking patients and clinical information: detecting UMLS concepts in e-mail. J Biomed Inform 2003 Aug;36(4-5):334-341. [Medline] [CrossRef]
- Zeng Q, Kogan S, Ash N, Greenes RA, Boxwala AA. Characteristics of consumer terminology for health information retrieval. Methods Inf Med 2002;41(4):289-298. [Medline]
- Mccray AT, Loane RF, Browne AC, Bangalore AK. Terminology issues in user access to Web-based medical information. Proc AMIA Symp 1999:107-111 [FREE Full text] [Medline]
- Tse T. Identifying and Characterizing a “Consumer Medical Vocabulary. ” College Park: University of Maryland; 2003.
- Patrick TB, Monga HK, Sievert ME, Houston Hall J, Longo DR. Evaluation of controlled vocabulary resources for development of a consumer entry vocabulary for diabetes. J Med Internet Res 2001 Aug 28;3(3):E24 [FREE Full text] [Medline] [CrossRef]
- Smith CA, Stavri PZ, Chapman WW. In their own words? A terminological analysis of e-mail to a cancer information service. Proc AMIA Symp 2002:697-701. [Medline]
- Zielstorff RD. Controlled vocabularies for consumer health. J Biomed Inform 2003 Aug;36(4-5):326-333. [Medline] [CrossRef]
- Krauthammer M, Nenadic G. Term identification in the biomedical literature. J Biomed Inform 2004 Dec;37(6):512-526. [Medline] [CrossRef]
- Wermter J, Hahn U. Effective grading of termhood in biomedical literature. AMIA Annu Symp Proc 2005:809-813. [Medline]
- Church K, Gale W, Hanks P, Hindle D. Using statistics in lexical analysis. In: Zernik U, editor. Lexical Acquisition: Exploiting On-Line Resources to Build a Lexicon. Hillsdale, NJ: Lawrence Erlbaum Associates, Inc; 1991.
- Basili R, Pazienza MT, Zanzotto FM. Modelling the syntactic contextual information for term extraction.. In: In: Conference on Recent Advances in Natural Language Processing (RANLP2001). Tzigiv Chark, Bulgaria; Sep 5, 2001.
- Harris MR, Savova GK, Johnson TM, Chute CG. A term extraction tool for expanding content in the domain of functioning, disability, and health: proof of concept. J Biomed Inform 2003 Aug;36(4-5):250-259. [Medline] [CrossRef]
- Frantzi KT, Ananiadou S, Mima H. Automatic recognition of multi-word terms: the C-value/NC-value method. Int J on Digital Libraries 2000;3(2):115-130. [CrossRef]
- Rindflesch TC, Hunter L, Aronson AR. Mining molecular binding terminology from biomedical text. Proc AMIA Symp 1999:127-131 [FREE Full text] [Medline]
- Fundel K, Güttler D, Zimmer R, Apostolakis J. A simple approach for protein name identification: prospects and limits. BMC Bioinformatics 2005;6 Suppl 1(Suppl 1):S15 [FREE Full text] [Medline] [CrossRef]
- Saric J, Jensen LJ, Ouzounova R, Rojas I, Bork P. Extraction of regulatory gene/protein networks from Medline. Bioinformatics 2006 Mar 15;22(6):645-650. [Medline] [CrossRef]
- Saric J, Jensen LJ, Rojas I. Large-scale extraction of gene regulation for model organisms in an ontological context. In Silico Biol 2005;5(1):21-32 [FREE Full text] [Medline]
- Cohen AM, Hersh WR, Dubay C, Spackman K. Using co-occurrence network structure to extract synonymous gene and protein names from MEDLINE abstracts. BMC Bioinformatics 2005 Apr;6(1):103 [FREE Full text] [Medline] [CrossRef]
- Zeng QT, Tse T, Crowell J, Divita G, Roth L, Browne AC. Identifying consumer-friendly display (CFD) names for health concepts. AMIA Annu Symp Proc 2005:859-863.
- Crowell J, Zeng Q, Tse T. A web application to support consumer health vocabulary development. AMIA Annu Symp Proc 2005:932. [Medline]
- Hepple M. Independence and Commitment: Assumptions for Rapid Training and Execution of Rule-based Part-of-Speech Taggers. Proceedings of the 38th Annual Meeting of the Association for Computational Linguistics (ACL-2000). Hong Kong, October 2000. Midtown, NJ: Association for Computational Linguistics; 2000:278-285.
- Browne A, McCray A, Srinivasan S. The Specialist Lexicon. Report No NLM-LHC-93-1. Bethesda, MD: Lister Hill National Center for Biomedical Communications, National Library of Medicine; 2000.
- Notes on Tagger Integration. URL: http://mmtx.nlm.nih.gov/taggerNotes.shtml [accessed 2006 Oct 23] [WebCite Cache]
- Hanley JA, Mcneil BJ. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982 Apr;143(1):29-36. [Medline]
- Aronson AR. Effective mapping of biomedical text to the UMLS Metathesaurus: the MetaMap program. Proc AMIA Symp 2001:17-21. [Medline]
- Cimino JJ. Auditing the Unified Medical Language System with semantic methods. J Am Med Inform Assoc 1998 Jan;5(1):41-51 [FREE Full text] [Medline]
- Bodenreider O, Nelson SJ, Hole WT, Chang HF. Beyond synonymy: exploiting the UMLS semantics in mapping vocabularies. Proc AMIA Symp 1998:815-819. [Medline]
- Humphreys BL, Mccray AT, Cheh ML. Evaluating the coverage of controlled health data terminologies: report on the results of the NLM/AHCPR large scale vocabulary test. J Am Med Inform Assoc 1997 Nov;4(6):484-500 [FREE Full text] [Medline]
|ATR: automated text recognition|
|AUC: area under the curve|
|CHV: consumer health vocabulary|
|NIH: National Institutes of Health|
|NLM: National Library of Medicine|
|OAC: open access and collaborative|
|POS: part of speech|
|ROC: receiver operating characteristic|
|UMLS: Unified Medical Language System|
Edited by G. Eysenbach;
submitted 23.10.06; peer-reviewed by T Patrick; comments to author 21.11.06; revised version received 21.11.06; accepted 22.02.07; published 14.03.07|
Please cite as:
Zeng QT, Tse T, Divita G, Keselman A, Crowell J, Browne AC, Goryachev S, Ngo L
Term Identification Methods for Consumer Health Vocabulary Development
J Med Internet Res 2007;9(1):e4
END, compatible with Endnote
BibTeX, compatible with BibDesk, LaTeX
RIS, compatible with RefMan, Procite, Endnote, RefWorks
Add this article to your Mendeley library
Add this article to your CiteULike library