Research article | Open | Open Peer Review | Published:
Performance evaluation of unified medical language system®'s synonyms expansion to query PubMed
BMC Medical Informatics and Decision Makingvolume 12, Article number: 12 (2012)
PubMed is the main access to medical literature on the Internet. In order to enhance the performance of its information retrieval tools, primarily non-indexed citations, the authors propose a method: expanding users' queries using Unified Medical Language System' (UMLS) synonyms i.e. all the terms gathered under one unique Concept Unique Identifier.
This method was evaluated using queries constructed to emphasize the differences between this new method and the current PubMed automatic term mapping. Four experts assessed citation relevance.
Using UMLS, we were able to retrieve new citations in 45.5% of queries, which implies a small increase in recall. The new strategy led to a heterogeneous 23.7% mean increase in non-indexed citation retrieved. Of these, 82% have been published less than 4 months earlier. The overall mean precision was 48.4% but differed according to the evaluators, ranging from 36.7% to 88.1% (Inter rater agreement was poor: kappa = 0.34).
This study highlights the need for specific search tools for each type of user and use-cases. The proposed strategy may be useful to retrieve recent scientific advancement.
The most important tool to access the medical literature is the PubMed search engine, which allows access to more than 20 millions of biomedical citations. The major part of these citations comes from the MEDLINE bibliographic database, which uses the MeSH thesaurus for indexing . Other citations, i.e. OLDMEDLINE, out of scope or recent citations, are not indexed at the time of user query . The most comprehensive way to find citations in MEDLINE is to use the MeSH thesaurus.
Because one third of Medline queries are performed by members of the general public  and furthermore because most health professionals  are not aware of this thesaurus, they run free-text queries, as they do when using Google™. This allows searching the entire PubMed collection but does not at all exploit the indexing work produced by National Library of Medicine (NLM) building the MeSH thesaurus and indexing millions of citations. Consequently, the US National Center of Biotechnology Information has developed several techniques (Automatic Term Mapping (ATM)) to map end-user queries to the MeSH thesaurus and other search field descriptors (e.g. author's name, publication's name, etc.) . The first ATM aim is to improve information retrieval in structured information: searching indexes (mostly MeSH terms used to index citations in MEDLINE) instead of only the free text. Almost nothing is done to enhance the search for recent citations (not indexed). This is a limiting factor because 1) these citations contain the most recent scientific discoveries and 2) they are the first returned by PubMed, which displays recent articles first, by default.
Although PubMed ATM query is continuously improved, a recent review  has counted 28 different entities that have devoted themselves to develop Web tools for helping users to quickly and effectively search and retrieve relevant publications on MEDLINE. This highlights the need for alternative ways of searching the medical literature. Thirion et al.  have shown that it is possible to improve ATM's performance, mainly in precision (tools available with the Doc'CISMeF search engine http://www.cismef.org) using MeSH synonyms. The aim of this paper was to propose an extension to this previous optimization, using Unified Medical Language System® (UMLS) synonyms, and to assess its performance.
MeSH & UMLS
The MeSH is the terminology, covering the whole area of medicine, used by the NLM for indexing MEDLINE citations. Each MeSH descriptor is named by a preferred term and may have some entry terms or synonyms, e.g. "myocardial infarction" is the preferred term designating the same MeSH descriptor rather than "myocardial infarct", "infarct, myocardial", etc. which are entry terms, or synonyms.
The UMLS contains a metathesaurus gathering many health terminologies/ontologies (T/O), like MeSH. For each T/O, each term is assigned to one or more concepts in UMLS. We defined as UMLS synonyms all the different terms from different T/O gathered under the same UMLS concept (same Concept Unique Identifier), e.g. "myocardial infarction" from the MeSH, "myocardial infarction" from the WHO Adverse Reaction Terminology (WHO-ART) and "heart attack" from the WHO-ART, etc. are UMLS synonyms as they are within the same UMLS concept.
In April 2011, when the ATM was used to match some query terms with a MeSH term, the resulting modified query was different if the query terms matched with the preferred term or an entry term or a UMLS' synonym . If it was the preferred term, the resulting modified query was: q1 = "preferred term"[MeSH term] OR "preferred term"[all fields] OR ("word 1 of preferred term"[All Fields] AND "word 2 of preferred term"[All Fields] AND etc.) (Table 1). If it was an entry term or a UMLS' synonym, the resulting modified query was: q2 = "preferred term"[MeSH term] OR "preferred term"[all fields] OR ("word 1 of preferred term"[All Fields] AND "word 2 of preferred term"[All Fields] AND etc.) OR "entry term"[All Fields] OR ("word 1 of entry term"[All Fields] AND "word 2 of entry term"[All Fields] AND etc.) (Table 1).
However, these queries were not the same compared to Thirion et al.'s strategies , as the word tokenization had only been added recently.
The improvement made by Thirion et al. consisted in limiting noise in MEDLINE and increasing recall in non-indexed PubMed subsets. When a MeSH term was used in a query, this improvement resulted in the retrieval of: 1) citations indexed with this same MeSH term in MEDLINE and 2) non-indexed citation containing any entry term for this MeSH term in its title or abstract. The corresponding query was: q3 = 'preferred term"[MeSH term] OR (("preferred term"[TIAB] OR "entry term 1 "[TIAB] OR "entry term 2 "[TIAB] OR...) NOT Medline[SB]) (Table 1). In contrast to the PubMed ATM, the strategy proposed by Thirion et al. provides the same query, whether or not the query includes preferred terms or entry terms.
In the current study, we propose a new strategy in order to increase recall: adding to the mapped queries all the UMLS synonyms with "ORs": q4 = "preferred term"[MeSH term] OR (("preferred term"[TIAB] OR "entry term 1 "[TIAB] OR "entry term 2 "[TIAB] OR... OR "UMLS synonym 1 "[TIAB] OR "UMLS synonym 2 "[TIAB] OR...) NOT (Medline[SB] OR OldMedline[SB])) (Table 1). The exclusion of OldMedline subset allows this query to focus on Pre-MEDLINE citations ("as supplied by publisher" and "in process" citations), which are not yet manually indexed by NLM curators. Non-indexed citations are not necessarily the latest citations. Nevertheless, according to the NLM customer service , time to index varies greatly between all of the different works that MEDLINE indexes. According to their recent statistical analysis, 25% of the citations are completed within 30 days of receipt, 50% within 60 days, and 75% within 90 days. Furthermore, 82% of Pre-MEDLINE citations that were evaluated in this study were published in 2011 and 11% in 2010. Obviously, when multiple UMLS synonyms contained the same spellings, they were not added in the mapped query. For technical purposes, we limited this list of synonyms to those included in the Health Multi-Terminology Portal http://pts.chu-rouen.fr : SNOMED CT, SNOMED intl, ICD-10, WHO-ART, WHO-ICF, WHO-ICPC2, LOINC, MedDRA, FMA and MEDLINEPlus.
For a quantitative assessment, the number of recent citations retrieved only by the new strategy was computed and compared to the entire number of recent citations retrieved.
To evaluate qualitative changes induced by this modification of mapping, we built Boolean queries based on MeSH terms: q5 = q4 NOT q3 (Table 1). We have selected 20 of the most frequently used MeSH Descriptors (according to the 2011 MEDLINE Baseline Repository data available at http://mbr.nlm.nih.gov/Download/index.shtml#MeSH) from the MeSH Diseases Category (C) where q5 provides citations. The choice of the C (diseases) tree from the MeSH thesaurus was driven by its potential impact on daily health care. Two medical librarians (BT and GK) and two physicians (LR and NG) assessed the relevance of the top 20 answers for each query manually after a careful reading of the title and abstract. Retrieved citations were assessed for relevance according to a three-modality scale used in other standard Information Retrieval test sets : bad, partial or full relevance.
Three factors might have an impact on the number of citation retrieved: (a) the number of sons in MeSH hierarchy (b) number of MeSH synonyms (c) number of UMLS synonyms. These factors were recorded and any association was evaluated using Spearman's correlation.
Evaluators' agreement was measured using kappa statistics (SAS Macro MKAPPA ). Precision was computed at two levels of relevance: using only fully relevant or fully and partially relevant citation. They were then computed for each evaluator and compared using the Friedman test and Chi2 test.
Table 2 summarizes results for the 43 queries we had to perform in order to obtain 20 citations for 20 queries. For the other 23 queries, q5 query did not produce any results: enhancing query using UMLS synonyms did not add any further results. The new strategy led to a heterogeneous 23.7% mean increase in non-indexed citation retrieved (from 0 to 9,876 new citations retrieved). None of the three tested factors (number of sons in MeSH hierarchy, of MeSH synonyms and of UMLS synonyms) were significantly correlated with the number of citations retrieved or the precision.
For the 20 studied MeSH Descriptors, inter-rater agreement was poor: multi-rater's kappa was 0.34. Results of relevance evaluation are summarized in Table 3. The mean precision for fully relevant citation was 48.4% CI95% = [45.8-50.9] but this number does not reflect discrepancies between evaluators: three evaluators (BT, GK and NG) found full relevance around 40% (43.7%, 36.7% and 37.4%, respectively) and one (LR) found 75%. Results are somewhat better for partially relevant citations but have a similar pattern, LR's evaluations were often more relevant than other evaluations: mean partial precision was 59.8% CI95% = [57.3-62.2]. BT, GK and NG found a precision of about 50% (50.1%, 51.7% and 48.2% respectively) whereas LR found 88.1%. Differences between evaluators were significant (p < 0.001, Friedman test). There was also a significant difference of precision depending on the MeSH term (data not shown, p < 0.001, Chi2 test): for 8 MeSH term the full relevance precision was higher than 0.5, for 8 MeSH terms the partial and full relevance precision was less than 0.5.
Enhancing information retrieval is one possible use of UMLS . The new strategy led to a slight increase in non-indexed citation retrieval (23.7%) for a precision very similar to those observed in previous reports studying PubMed performances: Thirion et al.  showed a precision of 54.5%; Lu et al. , for a normal use of PubMed, found a mean rank precision for the 20 top results between 40% and 55%.
Nevertheless, this study has some limitations: First, the absence of a control group to make a comparison led to difficult interpretation of results. However, consistency with literature review suggests that there was no major bias. Second, we used queries based on one MeSH term from the "disease" tree (C). However, would the results be similar for other MeSH tree terms, queries including several MeSH terms or queries including MeSH terms and keywords? Third, there is great variation in the results of the expansion proposed here between queries. The three factors tested were not significantly correlated with precision, but a qualitative assessment of the results was manually performed:
(a) Some UMLS synonyms provide very good results (e.g. "hepatoma" for "liver neoplasms", "Nephropathy" for "Kidney Diseases"), probably because they are very similar to a son of MeSH Descriptor.
(b) Some UMLS synonyms are ambiguous acronyms that generate a lot of noise (e.g. TB for tuberculosis).
(c) Some MeSH descriptors correspond to frequent confounding factors (e.g. "hypertension", "obesity"). Results of retrieved citations are adjusted based on these factors but they are not the real subjects of the citations (mean precision for fully relevant citations: 21.1%, 23% respectively).
We have also tried to explain the number of newly retrieved citations using the q5 query, which varies from 0 to 9,876 (for Diabetes Mellitus; see Table 2).
(a) The number of UMLS synonyms greatly varies from 0 to 38, with a median of 10 (data not shown). This natural source of variation was not confirmed by correlation tests. The difference must be more qualitative:
(b) Some UMLS synonyms do not provide any added value in information retrieval (e.g. all the synonyms finishing by ", NOS" will not provide any citation).
Fourth, the relevance assessment was performed on title and abstract alone, but not with the full text of the article. Although this could have introduced a bias in this study, it seems to us more pragmatic as most end-users select the relevant citation based on title and abstract alone. Lastly, the poor inter-rater agreement measured here (kappa = 0.34) suggests that we do not really know what we are measuring, even if it is common for this type of study . This poor kappa score, and the surprising distribution of results, only highlights differences between users. The improvement proposed here is probably not of interest for some users but may be of interest for others. Based on this study, we have implemented the following three procedures to query MEDLINE via PubMed in the following tool InfoRoute, French Infobutton (URL: inforoute.churouen.fr) :
The classical PubMed ATM
The previous procedure developed by Thirion et al. (semantic expansion with MeSH Entry terms)
The current procedure (semantic expansion with UMLS synonyms)
Different types of users should use these three procedures. Users expecting the most exhaustive results, even at the cost of some noise, should use the latest one. This type of users wants to maximize the recall.
Lu et al.  reviewed 28 different ways to access MEDLINE citations. The search strategy we propose could possibly be the 29th. However, when compared to other teams' strategy to improve PubMed information retrieval, the ones developed by our team modify the ATM and then are applicable in the PubMed interface. In fact, there is no need to integrate and update the MEDLINE bibliographic database in our information system.
Considering the huge number of citations retrieved by each q3 query (frequently more than dozens of thousands, data not shown), the increased number of recent citations retrieved may not lead to an important increase in recall. Nevertheless, the proposed strategy is based on the following assertion: a citation that is not indexed with a MeSH term does not have to be retrieved whatever the semantic expansion was used.
Based on this, the new strategy will only retrieve new citations not belonging to MEDLINE that represent more than ¾ of PubMed citation. We observed a 23.7% increase in recall for the citations aimed by the new strategy, which is not insignificant for the users, especially if they are searching for recent scientific advancements. This improvement mainly concerns new citations (82% of the citations retrieved by q5 have been published less than 4 months earlier). Furthermore, these citations, ranked first by PubMed, may be of great interest for PubMed users who frequently do not read more than the top 20 answers .
In contrast to PubMed, we assumed that when end users search for a disease name in PubMed, they do not add synonyms because of laxity or unawareness. It could be useful to add the son's preferred terms, son's entry terms and son's UMLS synonyms to the query with "ORs". This would eventually lead to an increase in recall and in proportion of queries retrieving additional citation (20 on 43 for this study) and a decrease in precision. However, it would drastically increase query size and resources needs, which are already quite substantial.
The expansion of queries using UMLS' synonyms may not be of interest for all PubMed users, but could be quite useful when seeking for exhaustivity (review, meta-analysis, etc.) as well as when searching for the latest scientific citations. This study highlights the need for specific search tools for each type of user and use-cases.
Automatic term mapping
Foundational model for anatomy
International classification of diseases: tenth revision
International Classification of Functioning: Disability and Health
International Classification of Primary Care: Second edition
Logical Observation Identifiers Names and Codes
Medical Dictionary for Regulatory Activities
Medical subject heading
National library of medicine
- SNOMED CT:
Systematized nomenclature of medicine clinical terms
- SNOMED intl:
Systematized nomenclature of medicine international
Title or abstract
Unified medical language system
World health organization.
Nelson SJ, Johnson WD, Humphreys BL: Relationship in medical subject headings. Relationships in the Organization of Knowledge. Edited by: Bean CA, Green R. 2001, New York: Kluwer Academic Publishers, 171-84.
What's the Difference Between MEDLINE® and PubMed®?. [http://www.nlm.nih.gov/pubs/factsheets/dif_med_pub.html Accessed in 17 February 2012.]
Herskovic JR, Tanaka LY, Hersh W, Bernstam EV: A day in the life of PubMed: Analysis of a typical day's query log. J Am Med Inform Assoc. 2007, 14 (2): 212-20. 10.1197/jamia.M2191.
Hoogendam A, Stalenhoef AF, Robbé PF, Overbeke AJ: Analysis of queries sent to PubMed at the point of care: observation of search behaviour in a medical teaching hospital. BMC Med Inform Decis Mak. 2008, 8: 42-10.1186/1472-6947-8-42.
How PubMed works: Automatic Term Mapping. [http://www.ncbi.nlm.nih.gov/books/NBK3827/#pubmedhelp.How_PubMed_works_aut Accessed in 17 February 2012.]
Lu Z: PubMed and beyond: a survey of web tools for searching biomedical literature. Database. 2011, 2011: baq036-10.1093/database/baq036.
Thirion B, Robu I, Darmoni SJ: Optimization of the PubMed Automatic Term Mapping. Stud Health Technol Inform. 2009, 150: 238-42.
Huang M, Névéol A, Lu Z: Recommending MeSH terms for annotating biomedical articles. J Am Med Inform Assoc. 2011, 18 (5): 660-7. 10.1136/amiajnl-2010-000055. Epub 2011 May 25
Grosjean J, Merabti T, Dahamna B, Kergourlay I, Thirion B, Soualmia LF, Darmoni SJ: Health multi-terminology portal: a semantic added-value for patient safety. Stud Health Technol Inform. 2011, 166: 129-38.
Hersh W, Buckley C, Leone TJ, Hickam D: OHSUmed: An interactive retrieval evaluation and new large test collection for research. Proceedings of the 17th annual international ACM SIGIR conference on Research and development in information retrieval: 1994. 1994, Springer-Verlag New York, 192-201.
Chen B, Zaebst D, Seel L: A Macro to Calculate Kappa Statistics for Categorizations by Multiple Raters. Proceedings of the 30th SAS User Group International conference:. 2005, 155-30. April ; Philadelphia
Hersh W, Price S, Donohoe L: Assessing thesaurus based query expansion using the UMLS metathesaurus. Proceedings of AMIA symposium: November 4-8. 2000, 344-8. ; Los Angeles
Lu Z, Kim W, Wilbur WJ: Evaluation of Query Expansion Using MeSH in PubMed. Inf Retr Boston. 2009, 12 (1): 69-80.
Funk ME, Reid CA: Indexing consistency in MEDLINE. Bulletin of the Medical Library Association. 1983, 2 (71): 176-83.
Darmoni SJ, Pereira S, Névéol A, Massari P, Dahamna B, Letord C, Kedelhué G, Piot J, Derville A, Thirion B: French Infobutton: an academic and... business perspective. AMIA Symp. 2008, IOS Press, 920-
Islamaj Doğan R, Murray GC, Névéol A, Lu Z: Understanding PubMed user search behavior through log analysis. Database. 2009, bap018-
The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1472-6947/12/12/prepub
The authors thank Richard Medeiros, Rouen University Hospital medical educator, for editing this manuscript.
The authors declare that they have no competing interests.
JFG and SJD formulated the idea of this study, design it and participated in writing the draft. NG evaluated citations' relevance, made statistical analysis and wrote the draft. LR, GK and BT have evaluated citation's relevance and have participated in writing. WC has built queries, has retrieved results and has participated in writing the draft. All authors read and approved the final manuscript.
Nicolas Griffon, Wiem Chebil, Laetitia Rollin, Gaetan Kerdelhue, Benoit Thirion, Jean-François Gehanno and Stéfan Jacques Darmoni contributed equally to this work.