RSS-Feed abonnieren

DOI: 10.1055/s-0041-1740358
A Data-Driven Iterative Approach for Semi-automatically Assessing the Correctness of Medication Value Sets: A Proof of Concept Based on Opioids
Autoren
Funding A.G.'s efforts were funded by the National Institute of Mental Health through the “My Data Choices, evaluation of effective consent strategies for patients with behavioral health conditions” (R01 MH108992) grant. A.S.'s efforts were funded by the National Institute on Drug Abuse through the “Mining Social Media Big Data for Toxicovigilance: Automating the Monitoring of Prescription Medication Abuse via Natural Language Processing and Machine Learning Methods” (R01 DA046619) grant.
Abstract
Background Value sets are lists of terms (e.g., opioid medication names) and their corresponding codes from standard clinical vocabularies (e.g., RxNorm) created with the intent of supporting health information exchange and research. Value sets are manually-created and often exhibit errors.
Objectives The aim of the study is to develop a semi-automatic, data-centric natural language processing (NLP) method to assess medication-related value set correctness and evaluate it on a set of opioid medication value sets.
Methods We developed an NLP algorithm that utilizes value sets containing mostly true positives and true negatives to learn lexical patterns associated with the true positives, and then employs these patterns to identify potential errors in unseen value sets. We evaluated the algorithm on a set of opioid medication value sets, using the recall, precision and F1-score metrics. We applied the trained model to assess the correctness of unseen opioid value sets based on recall. To replicate the application of the algorithm in real-world settings, a domain expert manually conducted error analysis to identify potential system and value set errors.
Results Thirty-eight value sets were retrieved from the Value Set Authority Center, and six (two opioid, four non-opioid) were used to develop and evaluate the system. Average precision, recall, and F1-score were 0.932, 0.904, and 0.909, respectively on uncorrected value sets; and 0.958, 0.953, and 0.953, respectively after manual correction of the same value sets. On 20 unseen opioid value sets, the algorithm obtained average recall of 0.89. Error analyses revealed that the main sources of system misclassifications were differences in how opioids were coded in the value sets—while the training value sets had generic names mostly, some of the unseen value sets had new trade names and ingredients.
Conclusion The proposed approach is data-centric, reusable, customizable, and not resource intensive. It may help domain experts to easily validate value sets.
Author Contributions
All authors made substantial contributions to manuscript revisions and approved the final version. A.S. and A.G. contributed to the design of the study. S.L. conducted data analysis under the mentorship of A.S.. A.S. and A.G. supervised the conception, design, and revision of the manuscript. All authors also agree to be accountable for the accuracy and integrity of the work presented here
Publikationsverlauf
Eingereicht: 02. September 2021
Angenommen: 11. Oktober 2021
Artikel online veröffentlicht:
29. Dezember 2021
© 2021. The Author(s). This is an open access article published by Thieme under the terms of the Creative Commons Attribution-NonDerivative-NonCommercial License, permitting copying and reproduction so long as the original work is given appropriate credit. Contents may not be used for commercial purposes, or adapted, remixed, transformed or built upon. (https://creativecommons.org/licenses/by-nc-nd/4.0/)
Georg Thieme Verlag KG
Rüdigerstraße 14, 70469 Stuttgart, Germany
-
References
- 1 Fung KW, Xu J, Gold S. The Use of Inter-terminology Maps for the Creation and Maintenance of Value Sets. Paper presented at: AMIA Annu Symp Proc. Vol 2019. American Medical Informatics Association; Accessed June 8, 2021. /pmc/articles/PMC7153132 2019: 438-447
- 2 Winnenburg R, Bodenreider O. Metrics for assessing the quality of value sets in clinical quality measures. Paper presented at: AMIA Annu Symp Proc. Vol 2013. American Medical Informatics Association; 2013: 1497-1505
- 3 Bodenreider O. Title: Criteria and Metrics for Assessing the Quality of SNOMED CT Value Sets in Clinical Quality Measures. Vol. 2012 ; 2012. Accessed June 8, 2021 at: http://www.ihtsdo.org/fileadmin/user_upload/doc/showcase/show13/
- 4 Winnenburg R, Bodenreider O. Issues in creating and maintaining value sets for clinical quality measures. Paper presented at: AMIA Annu Symp Proc. Vol 2012. American Medical Informatics Association; Accessed June 8, 2021. /pmc/articles/PMC3540585 2012: 988-996
- 5 National Survey on Drug Use and Health (NSDUH). 2019. Results from the 2018 National Survey on Drug Use and Health. Accessed November 21, 2021 at: https://www.samhsa.gov/data/report/
- 6 Ponnapalli A, Grando A, Murcko A, Wertheim P. Systematic Literature Review of Prescription Drug Monitoring Programs. Paper presented at: AMIA Annu Symp Proc. Vol 2018. American Medical Informatics Association; Accessed June 8, 2021. /pmc/articles/PMC6371270 2018: 1478-1487
- 7 Xu QS, Liang YZ. Monte Carlo cross validation. Chemom Intell Lab Syst 2001; 56 (01) 1-11
- 8 American Society of Addiction Medicine (ASAM). Opioids: Brand names, generic names & street names. Published August 28, 2017. Accessed October 10, 2021 at: https://www.asam.org/docs/default-source/education-docs/opioid-names_generic-brand-street_it-matttrs_8-28-17.pdf?sfvrsn=7b0640c2_2
- 9 Zeng Z, Deng Y, Li X, Naumann T, Luo Y. Natural language processing for EHR-based computational phenotyping. IEEE/ACM Trans Comput Biol Bioinformatics 2019; 16 (01) 139-153
- 10 Nesi P, Pantaleo G, Tenti M. Geographical localization of web domains and organization addresses recognition by employing natural language processing, Pattern Matching and clustering. Eng Appl Artif Intell 2016; 51: 202-211
- 11 Sevenster M, Buurman J, Liu P, Peters JF, Chang PJ. Natural language processing techniques for extracting and categorizing finding measurements in narrative radiology reports. Appl Clin Inform 2015; 6 (03) 600-110
- 12 Sarker A, Klein AZ, Mee J, Harik P, Gonzalez-Hernandez G. An interpretable natural language processing system for written medical examination assessment. J Biomed Inform 2019; 98: 103268
- 13 Arnoux-Guenegou A, Girardeau Y, Chen X. et al. The adverse drug reactions from patient reports in social media project: protocol for an evaluation against a gold standard. JMIR Res Protoc 2019; 8 (05) e11448
- 14 Pérez-Pérez M, Pérez-Rodríguez G, Fdez-Riverola F, Lourenço A. Using twitter to understand the human bowel disease community: exploratory analysis of key topics. J Med Internet Res 2019; 21 (08) e12610
- 15 Hostetter J, Wang K, Siegel E, Durack J, Morrison JJ. Using standardized lexicons for report template validation with LexMap, a web-based application. J Digit Imaging 2015; 28 (03) 309-314
- 16 Lafferty J, McCallum A, Pereira F. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data. Dep Pap. Published online June 28, 2001. Accessed May 28, 2020 at: https://repository.upenn.edu/cis_papers/159
- 17 Schuster M, Paliwal KK. Bidirectional recurrent neural networks. IEEE Trans Signal Process 1997; 45 (11) 2673-2681
- 18 Bui DDA, Zeng-Treitler Q. Learning regular expressions for clinical text classification. J Am Med Inform Assoc 2014; 21 (05) 850-857
- 19 Frenz CM. Deafness mutation mining using regular expression based pattern matching. BMC Med Inform Decis Mak 2007; 7: 32
- 20 Chapman WW, Bridewell W, Hanbury P, Cooper GF, Buchanan BG. A simple algorithm for identifying negated findings and diseases in discharge summaries. J Biomed Inform 2001; 34 (05) 301-310
- 21 Brauer F, Rieger R, Mocan A, Barczynski WM. Enabling information extraction by inference of regular expressions from sample entities. Paper presented at: Int Conf Inf Knowl Manag Proc; 2011
- 22 Li Y, Krishnamurthy R, Raghavan S, Vaithyanathan S, Jagadish HV. Regular Expression Learning for Information Extraction. Paper presented at: Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing; 2008: 21-30