Skip to main content
Log in

Leveraging Large Language Models for Clinical Abbreviation Disambiguation

  • Original Paper
  • Published:
Journal of Medical Systems Aims and scope Submit manuscript

Abstract

Clinical abbreviation disambiguation is a crucial task in the biomedical domain, as the accurate identification of the intended meanings or expansions of abbreviations in clinical texts is vital for medical information retrieval and analysis. Existing approaches have shown promising results, but challenges such as limited instances and ambiguous interpretations persist. In this paper, we propose an approach to address these challenges and enhance the performance of clinical abbreviation disambiguation. Our objective is to leverage the power of Large Language Models (LLMs) and employ a Generative Model (GM) to augment the dataset with contextually relevant instances, enabling more accurate disambiguation across diverse clinical contexts. We integrate the contextual understanding of LLMs, represented by BlueBERT and Transformers, with data augmentation using a Generative Model, called Biomedical Generative Pre-trained Transformer (BIOGPT), that is pretrained on an extensive corpus of biomedical literature to capture the intricacies of medical terminology and context. By providing the BIOGPT with relevant medical terms and sense information, we generate diverse instances of clinical text that accurately represent the intended meanings of abbreviations. We evaluate our approach on the widely recognized CASI dataset, carefully partitioned into training, validation, and test sets. The incorporation of data augmentation with the GM improves the model’s performance, particularly for senses with limited instances, effectively addressing dataset imbalance and challenges posed by similar concepts. The results demonstrate the efficacy of our proposed method, showcasing the significance of LLMs and generative techniques in clinical abbreviation disambiguation. Our model achieves a good accuracy on the test set, outperforming previous methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Data Availability

The datasets analyzed during the current study are available in the Clinical Abbreviation Sense Inventory repository, https://conservancy.umn.edu/handle/11299/137703.

References

  1. B. Duganová, “Medical language – a unique linguistic phenomenon,” JAHR, 2019.

  2. C. DesRoches, “Healthcare in the new age of transparency,” Semin Dial, vol. 33, Jul. 2020, doi: https://doi.org/10.1111/sdi.12934

  3. L. Brunetti, J. Santell, and R. Hicks, “The Impact of Abbreviations on Patient Safety,” Joint Commission journal on quality and patient safety / Joint Commission Resources, vol. 33, pp. 576–583, Jul. 2007, doi: https://doi.org/10.1016/S1553-7250(07)33062-6

  4. “Shared Learning-Reported Incidents Involving Hydromorphone,” 2006. [Online]. Available: http://www.ismp-canada.org/download/CMIRPS%20Core%20Data%20Set%20for%20Individu

  5. A. Coghlan, S. Turner, and S. Coverdale, “Danger in discharge summaries: Abbreviations create confusion for both author and recipient,” Intern Med J, vol. 53, Jul. 2021, doi: https://doi.org/10.1111/imj.15582

  6. K. E. Walsh and J. H. Gurwitz, “Medical abbreviations: writing little and communicating less.,” Arch Dis Child, vol. 93, no. 10, pp. 816–817, Oct. 2008, doi: https://doi.org/10.1136/adc.2008.141473

  7. H. Xu, P. D. Stetson, and C. Friedman, “A study of abbreviations in clinical notes.,” AMIA Annu Symp Proc, vol. 2007, pp. 821–825, Oct. 2007.

  8. B. Collard and A. Royal, “The use of abbreviations in surgical note keeping,” Annals of Medicine and Surgery, vol. 4, no. 2, pp. 100–102, 2015, doi: https://doi.org/10.1016/j.amsu.2015.03.008

  9. I. F. Kuhn, “Abbreviations and acronyms in healthcare: when shorter isn’t sweeter.,” Pediatr Nurs, vol. 33, no. 5, pp. 392–398, 2007.

    PubMed  Google Scholar 

  10. T. I. Amosa, L. I. bt Izhar, P. Sebastian, I. B. Ismail, O. Ibrahim, and S. L. Ayinla, “Clinical Errors from Acronym Use in Electronic Health Record: A Review of NLP-based Disambiguation Techniques,” IEEE Access, 2023, doi: https://doi.org/10.1109/ACCESS.2023.3284682

    Article  Google Scholar 

  11. Y. Peng, S. Yan, and Z. lu, “Transfer Learning in Biomedical Natural Language Processing: An Evaluation of BERT and ELMo on Ten Benchmarking Datasets,” Jul. 2019, pp. 58–65. doi: https://doi.org/10.18653/v1/W19-5006

  12. R. Luo et al, “BioGPT: generative pre-trained transformer for biomedical text generation and mining,” Brief Bioinform, vol. 23, no. 6, p. bbac409, Nov. 2022, doi: https://doi.org/10.1093/bib/bbac409

  13. A. Schwartz and M. Hearst, “A Simple Algorithm For Identifying Abbreviation Definitions in Biomedical Text,” Pac Symp Biocomput, vol. 4, pp. 451–462, Feb. 2003, doi: https://doi.org/10.1142/9789812776303_0042

  14. J. Toole, “A Hybrid Approach to the Identification and Expansion of Abbreviations,” May 2000.

  15. V. García, J. Sánchez, and R. Mollineda, “Exploring the Performance of Resampling Strategies for the Class Imbalance Problem,” 2010, pp. 541–549. doi: https://doi.org/10.1007/978-3-642-13022-9_54

  16. R. Navigli, S. Faralli, A. Soroa, O. de Lacalle, and E. Agirre, “Two Birds with One Stone: Learning Semantic Models for Text Categorization and Word Sense Disambiguation,” in Proceedings of the 20th ACM International Conference on Information and Knowledge Management, in CIKM ’11. New York, NY, USA: Association for Computing Machinery, 2011, pp. 2317–2320. doi: https://doi.org/10.1145/2063576.2063955

  17. S. Pakhomov, T. Pedersen, and C. G. Chute, “Abbreviation and acronym disambiguation in clinical discourse.,” AMIA Annu Symp Proc, vol. 2005, pp. 589–593, 2005.

  18. S. Moon, S. Pakhomov, and G. B. Melton, “Automated disambiguation of acronyms and abbreviations in clinical texts: window and training size considerations.,” AMIA Annu Symp Proc, vol. 2012, pp. 1310–1319, 2012.

  19. M. Joshi, S. Pakhomov, T. Pedersen, and C. G. Chute, “A comparative study of supervised learning as applied to acronym expansion in clinical reports.,” AMIA Annu Symp Proc, vol. 2006, pp. 399–403, 2006.

  20. G. P. Finley, S. V. S. Pakhomov, R. McEwan, and G. B. Melton, “Towards Comprehensive Clinical Abbreviation Disambiguation Using Machine-Labeled Training Data.,” AMIA Annu Symp Proc, vol. 2016, pp. 560–569, 2016.

  21. Y. Wu, J. Xu, Y. Zhang, and W. Qi, Clinical Abbreviation Disambiguation Using Neural Word Embeddings. 2015. doi: https://doi.org/10.18653/v1/W15-3822

  22. Y. Wu et al, “A long journey to short abbreviations: developing an open-source framework for clinical abbreviation recognition and disambiguation (CARD).,” J Am Med Inform Assoc, vol. 24, no. e1, pp. e79–e86, Apr. 2017, doi: https://doi.org/10.1093/jamia/ocw109

  23. A. Jaber and P. Mart\’\inez, “Disambiguating Clinical Abbreviations using Pre-trained Word Embeddings,” in Proceedings of the 14th International Joint Conference on Biomedical Engineering Systems and Technologies, {SCITEPRESS} - Science and Technology Publications, 2021. doi: https://doi.org/10.5220/0010256105010508

  24. R. Socher and C. Manning, “Deep Learning for NLP (without Magic).”

  25. V. Kumar, D. Reforgiato Recupero, D. Riboni, and R. Helaoui, “Ensembling Classical Machine Learning and Deep Learning Approaches for Morbidity Identification From Clinical Notes,” IEEE Access, vol. PP, p. 1, Dec. 2020, doi: https://doi.org/10.1109/ACCESS.2020.3043221

    Article  Google Scholar 

  26. V. Joopudi, B. Dandala, and M. Devarakonda, “A convolutional route to abbreviation disambiguation in clinical text,” J Biomed Inform, vol. 86, pp. 71–78, 2018, doi: https://doi.org/10.1016/j.jbi.2018.07.025

    Article  PubMed  Google Scholar 

  27. M. Skreta et al, “Automatically disambiguating medical acronyms with ontology-aware deep learning,” Nat Commun, vol. 12, no. 1, p. 5319, 2021, doi: https://doi.org/10.1038/s41467-021-25578-4

    Article  ADS  PubMed  PubMed Central  CAS  Google Scholar 

  28. C.-X. Zhang, M.-L. Wang, and X.-Y. Gao, “Biomedical Word Sense Disambiguation Based on Graph Attention Networks,” IEEE Access, vol. 10, pp. 123328–123336, 2022, doi: https://doi.org/10.1109/ACCESS.2022.3224802

    Article  Google Scholar 

  29. N. B. Link et al, “Binary acronym disambiguation in clinical notes from electronic health records with an application in computational phenotyping,” Int J Med Inform, vol. 162, p. 104753, 2022, doi: https://doi.org/10.1016/j.ijmedinf.2022.104753

    Article  PubMed  Google Scholar 

  30. A. Rajkomar et al, “Deciphering clinical abbreviations with a privacy protecting machine learning system,” Nat Commun, vol. 13, no. 1, Dec. 2022, doi: https://doi.org/10.1038/s41467-022-35007-9

  31. J. Lee et al, “BioBERT: a pre-trained biomedical language representation model for biomedical text mining,” Bioinformatics, vol. 36, no. 4, pp. 1234–1240, Feb. 2020, doi: https://doi.org/10.1093/bioinformatics/btz682

    Article  MathSciNet  PubMed  CAS  Google Scholar 

  32. J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding,” in NAACL, 2019.

  33. S. Kula, R. Kozik, and M. Choraś, “Implementation of the BERT-derived architectures to tackle disinformation challenges,” Neural Comput Appl, vol. 34, no. 23, pp. 20449–20461, 2022, doi: https://doi.org/10.1007/s00521-021-06276-0

    Article  PubMed  Google Scholar 

  34. S. Hao, P. Zhang, S. Liu, and Y. Wang, “Sentiment recognition and analysis method of official document text based on BERT–SVM model,” Neural Comput Appl, 2023, doi: https://doi.org/10.1007/s00521-023-08226-4

    Article  Google Scholar 

  35. A. Kashyap, H. Burris, C. Callison-Burch, and M. R. Boland, “The CLASSE GATOR (CLinical Acronym SenSE disambiGuATOR): A Method for predicting acronym sense from neonatal clinical notes,” Int J Med Inform, vol. 137, p. 104101, 2020, doi: https://doi.org/10.1016/j.ijmedinf.2020.104101

    Article  PubMed  Google Scholar 

  36. A. Jaber and P. Martínez, “Disambiguating Clinical Abbreviations Using a One-Fits-All Classifier Based on Deep Learning Techniques,” Methods Inf Med, vol. 61, pp. E28–E34, Jun. 2022, doi: https://doi.org/10.1055/s-0042-1742388

    Article  PubMed  PubMed Central  Google Scholar 

  37. P. Kacker, A. Cupallari, A. G. Subramanian, and N. Jain, “ABB-BERT: A BERT model for disambiguating abbreviations and contractions,” Jul. 2022, [Online]. Available: http://arxiv.org/abs/2207.04008

  38. S. Seneviratne, E. Daskalaki, A. Lenskiy, and H. Suominen, “m-Networks: Adapting the Triplet Networks for Acronym Disambiguation,” 2022. [Online]. Available: https://github.com/sandaruSen/m_networks

  39. A. Wagh and M. Khanna, “Clinical Abbreviation Disambiguation Using Clinical Variants of BERT,” in Multi-disciplinary Trends in Artificial Intelligence, R. Morusupalli, T. S. Dandibhotla, V. V. Atluri, D. Windridge, P. Lingras, and V. R. Komati, Eds., Cham: Springer Nature Switzerland, 2023, pp. 214–224.

    Chapter  Google Scholar 

  40. M. Cevik, S. Mohammad Jafari, M. Myers, and S. Yildirim, “Sequence Labeling for Disambiguating Medical Abbreviations,” J Healthc Inform Res, vol. 7, no. 4, pp. 501–526, 2023, doi: https://doi.org/10.1007/s41666-023-00146-1

    Article  PubMed  Google Scholar 

  41. A. Vaswani et al, “Attention Is All You Need,” Jun. 2017.

  42. S. Moon, S. Pakhomov, N. Liu, J. Ryan, and G. Melton, “A sense inventory for clinical abbreviations and acronyms created using clinical notes and medical dictionary resources,” J Am Med Inform Assoc, vol. 21, Jun. 2013, doi: https://doi.org/10.1136/amiajnl-2012-001506

  43. S. P. S. M. G. Moon, “Clinical Abbreviation Sense Inventory. Retrieved from the University of Minnesota Digital Conservancy,” 2012.

  44. J. Kaur and P. Buttar, “A Systematic Review on Stopword Removal Algorithms,” vol. 4, pp. 207–210, Apr. 2018.

  45. Y. Peng, S. Yan, and Z. lu, Transfer Learning in Biomedical Natural Language Processing: An Evaluation of BERT and ELMo on Ten Benchmarking Datasets. 2019. doi: https://doi.org/10.18653/v1/W19-5006

  46. A. E. W. Johnson et al, “MIMIC-III, a freely accessible critical care database,” Sci Data, vol. 3, no. 1, p. 160035, 2016, doi: https://doi.org/10.1038/sdata.2016.35

    Article  MathSciNet  PubMed  PubMed Central  CAS  Google Scholar 

  47. C. Huang, Y. Li, C. C. Loy, and X. Tang, “Learning Deep Representation for Imbalanced Classification,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Jun. 2016, pp. 5375–5384. doi: https://doi.org/10.1109/CVPR.2016.580

  48. M. Khushi et al, “A Comparative Performance Analysis of Data Resampling Methods on Imbalance Medical Data,” IEEE Access, vol. 9, pp. 109960–109975, 2021, doi: https://doi.org/10.1109/ACCESS.2021.3102399

  49. N. Chawla, “Data Mining for Imbalanced Datasets: An Overview,” in Data Mining and Knowledge Discovery Handbook, ISBN 978-0-387-09822-7. Springer Science + Business Media, LLC, 2010, p. 875, vol. 5, 2005, pp. 853–867. doi: https://doi.org/10.1007/0-387-25465-X_40

  50. I. Li et al, “A Neural Topic-Attention Model for Medical Term Abbreviation Disambiguation,” CoRR, vol. abs/1910.1, 2019, [Online]. Available: http://arxiv.org/abs/1910.14076

  51. G. Adams, M. Ketenci, S. Bhave, A. Perotte, and N. Elhadad, “Zero-Shot Clinical Acronym Expansion via Latent Meaning Cells.,” Proc Mach Learn Res, vol. 136, pp. 12–40, Dec. 2020.

  52. M. Agrawal, S. Hegselmann, H. Lang, Y. Kim, and D. Sontag, “Large Language Models are Few-Shot Clinical Information Extractors,” in Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022.

  53. J. Kim et al, “Improved Clinical Abbreviation Expansion via Non-Sense-Based Approaches,” 2020. [Online]. Available: https://github.com/

Download references

Funding

No funds, grants, or other support was received.

Author information

Authors and Affiliations

Authors

Contributions

all authors contributed equally to this work.

Corresponding author

Correspondence to Manda Hosseini.

Ethics declarations

Conflict of Interest

The authors declare that they have no conflict of interest.

Protection of Human and Animal Subjects

Ethical approval was not required for this study.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Electronic Supplementary Material

Below is the link to the electronic supplementary material.

Supplementary Material 1

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Hosseini, M., Hosseini, M. & Javidan, R. Leveraging Large Language Models for Clinical Abbreviation Disambiguation. J Med Syst 48, 27 (2024). https://doi.org/10.1007/s10916-024-02049-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10916-024-02049-z

Keywords

Navigation