Transformers and large language models in healthcare: A review

Artif Intell Med. 2024 Aug:154:102900. doi: 10.1016/j.artmed.2024.102900. Epub 2024 Jun 5.

Abstract

With Artificial Intelligence (AI) increasingly permeating various aspects of society, including healthcare, the adoption of the Transformers neural network architecture is rapidly changing many applications. Transformer is a type of deep learning architecture initially developed to solve general-purpose Natural Language Processing (NLP) tasks and has subsequently been adapted in many fields, including healthcare. In this survey paper, we provide an overview of how this architecture has been adopted to analyze various forms of healthcare data, including clinical NLP, medical imaging, structured Electronic Health Records (EHR), social media, bio-physiological signals, biomolecular sequences. Furthermore, which have also include the articles that used the transformer architecture for generating surgical instructions and predicting adverse outcomes after surgeries under the umbrella of critical care. Under diverse settings, these models have been used for clinical diagnosis, report generation, data reconstruction, and drug/protein synthesis. Finally, we also discuss the benefits and limitations of using transformers in healthcare and examine issues such as computational cost, model interpretability, fairness, alignment with human values, ethical implications, and environmental impact.

Keywords: Electronic Health Records; Healthcare; Large Language Models; Medical Imaging; Natural Language Processing; Transformers.

Publication types

  • Review
  • Research Support, Non-U.S. Gov't

MeSH terms

  • Artificial Intelligence
  • Deep Learning*
  • Delivery of Health Care / organization & administration
  • Electronic Health Records
  • Humans
  • Natural Language Processing*
  • Neural Networks, Computer