Download PDFOpen PDF in browser

Can Language Models Reason about ICD Codes to Guide the Generation of Clinical Notes?

EasyChair Preprint 15731

12 pagesDate: January 18, 2025

Abstract

In the past decade a surge in the amount of electronic health record (EHR) data in the United States, attributed to a favorable policy environment created by the Health Information Technology for Economic and Clinical Health (HITECH) Act of 2009 and the 21st Century Cures Act of 2016. Clinical notes for patients’ assessments, diagnoses, and treatments are captured in these EHRs in free-form text by physicians, who spend a considerable amount of time entering them. Manually writing clinical notes may take considerable amount of time, increasing the patient’s waiting time and could possibly delay diagnoses. Large language models (LLMs), such as GPT-3 possess the ability to generate news articles that closely resemble human-written ones. We investigate the usage of Chain-of-Thought (CoT) prompt engineering to improve the LLM’s response in clinical note generation. In our prompts, we incorporate International Classification of Diseases (ICD) codes and basic patient information along with similar clinical case examples to investigate how LLMs can effectively formulate clinical notes. We tested our CoT prompt technique on six clinical cases from the CodiEsp test dataset using GPT-4 as our LLM and our results show that it outperformed the standard zero-shot prompt.

Keyphrases: CoT, Generative AI, ICD codes, NLP, clinical note gen, info retrieval, large language models

BibTeX entry
BibTeX does not have the right entry for preprints. This is a hack for producing the correct reference:
@booklet{EasyChair:15731,
  author    = {Ivan Makohon and Jian Wu and Bintao Feng and Yaohang Li},
  title     = {Can Language Models Reason about ICD Codes to Guide the Generation of Clinical Notes?},
  howpublished = {EasyChair Preprint 15731},
  year      = {EasyChair, 2025}}
Download PDFOpen PDF in browser