Methods Inf Med 2022; 61(03/04): 084-089
DOI: 10.1055/s-0042-1749358
Original Article

Automated Identification of Clinical Procedures in Free-Text Electronic Clinical Records with a Low-Code Named Entity Recognition Workflow

Carmelo Macri
1   Machine Learning Division, Ophthalmic Research Laboratory, University of Adelaide, Adelaide, South Australia, Australia
2   Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia, Australia
3   Discipline of Ophthalmology and Visual Sciences, The University of Adelaide, Adelaide, South Australia, Australia
,
Ian Teoh
1   Machine Learning Division, Ophthalmic Research Laboratory, University of Adelaide, Adelaide, South Australia, Australia
,
Stephen Bacchi
1   Machine Learning Division, Ophthalmic Research Laboratory, University of Adelaide, Adelaide, South Australia, Australia
2   Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia, Australia
,
Michelle Sun
2   Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia, Australia
,
Dinesh Selva
2   Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia, Australia
,
Robert Casson
2   Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia, Australia
,
WengOnn Chan
1   Machine Learning Division, Ophthalmic Research Laboratory, University of Adelaide, Adelaide, South Australia, Australia
2   Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia, Australia
› Institutsangaben

Funding None.

Abstract

Background Clinical procedures are often performed in outpatient clinics without prior scheduling at the administrative level, and documentation of the procedure often occurs solely in free-text clinical electronic notes. Natural language processing (NLP), particularly named entity recognition (NER), may provide a solution to extracting procedure data from free-text electronic notes.

Methods Free-text notes from outpatient ophthalmology visits were collected from the electronic clinical records at a single institution over 3 months. The Prodigy low-code annotation tool was used to create an annotation dataset and train a custom NER model for clinical procedures. Clinical procedures were extracted from the entire set of clinical notes.

Results There were a total of 5,098 clinic notes extracted for the study period; 1,923 clinic notes were used to build the NER model, which included a total of 231 manual annotations. The NER model achieved an F-score of 0.767, a precision of 0.810, and a recall of 0.729. The most common procedures performed included intravitreal injections of therapeutic substances, removal of corneal foreign bodies, and epithelial debridement of corneal ulcers.

Conclusion The use of a low-code annotation software tool allows the rapid creation of a custom annotation dataset to train a NER model to identify clinical procedures stored in free-text electronic clinical notes. This enables clinicians to rapidly gather previously unidentified procedural data for quality improvement and auditing purposes. Low-code annotation tools may reduce time and coding barriers to clinician participation in NLP research.



Publikationsverlauf

Eingereicht: 02. Dezember 2021

Angenommen: 13. April 2022

Artikel online veröffentlicht:
12. September 2022

© 2022. Thieme. All rights reserved.

Georg Thieme Verlag KG
Rüdigerstraße 14, 70469 Stuttgart, Germany