Publication:
Transformers for Clinical Coding in Spanish

Loading...
Thumbnail Image

Date

2021-05-13

Authors

López-García, Guillermo
Jerez, José M.
Ribelles, Nuria
Alba, Emilio
Veredas, Francisco J.

Advisors

Journal Title

Journal ISSN

Volume Title

Publisher

Institute of Electrical and Electronics Engineers (IEEE)
Metrics
Google Scholar
Export

Research Projects

Organizational Units

Journal Issue

Abstract

Automatic clinical coding is an essential task in the process of extracting relevant information from unstructured documents contained in electronic health records (EHRs). However, most research in the development of computer-based methods for clinical coding focuses on texts written in English due to the limited availability of medical linguistic resources in languages other than English. With nearly 500 million native speakers, there is a worldwide interest in processing healthcare texts in Spanish. In this study, we sys tematically analyzed transformer-based models for automatic clinical coding in Spanish. Using a transfer learning-based approach, the three existing transformer architectures that support the Spanish language, namely, multilingual BERT (mBERT), BETO and XLM-RoBERTa (XLM-R), were first pretrained on a corpus of real-world oncology clinical cases with the goal of adapting transformers to the particularities of Spanish medical texts. The resulting models were fine-tuned on three distinct clinical coding tasks, following a multilabel sentence classification strategy. For each analyzed transformer, the domain-specific version out performed the original general domain model across those tasks. Moreover, the combination of the developed strategy with an ensemble approach leveraging the predictive capacities of the three distinct transformers yielded the best obtained results, with MAP scores of 0.662, 0.544 and 0.884 on CodiEsp-D, CodiEsp-P and Cantemist-Coding shared tasks, which remarkably improved the previous state-of-the-art performance by 11.6%, 10.3% and 4.4%, respectively. We publicly release the mBERT, BETO and XLMR transform ers adapted to the Spanish clinical domain at https://github.com/guilopgar/ClinicalCodingTransformerES, providing the clinical natural language processing community with advanced deep learning methods for performing medical coding and other tasks in the Spanish clinical domain.

Description

MeSH Terms

Medical Subject Headings::Information Science::Information Science::Computing Methodologies::Artificial Intelligence::Natural Language Processing
Medical Subject Headings::Analytical, Diagnostic and Therapeutic Techniques and Equipment::Investigative Techniques::Epidemiologic Methods::Data Collection::Records as Topic::Medical Records::Medical Records Systems, Computerized::Electronic Health Records
Medical Subject Headings::Health Care::Health Services Administration::Organization and Administration::Professional Practice::Practice Management::Office Management::Forms and Records Control::Clinical Coding
Medical Subject Headings::Information Science::Information Science::Communication::Language::Linguistics
Medical Subject Headings::Health Care::Health Care Quality, Access, and Evaluation::Delivery of Health Care
Medical Subject Headings::Information Science::Information Science::Computing Methodologies::Computer Systems::Computers
Medical Subject Headings::Psychiatry and Psychology::Behavior and Behavior Mechanisms::Motivation::Goals
Medical Subject Headings::Psychiatry and Psychology::Psychological Phenomena and Processes::Psychology, Applied::Human Engineering::Task Performance and Analysis

DeCS Terms

CIE Terms

Keywords

Clinical coding, Deep learning, Natural language processing, Text classification, Transformers, Codificación clínica, Aprendizaje profundo, Procesamiento de lenguaje natural, Modelos epidemiológicos, Análisis y desempeño de tareas

Citation

López-García G, Jerez JM, Ribelles N, Alba E, Veredas FJ. Transformers for Clinical Coding in Spanish. IEEE Access. 2021;9:72387-72397