Abstract
This paper investigates the application of Chain of Thought (CoT) reasoning to enhance the performance of smaller language models in medical question-answering tasks. By leveraging CoT prompting strategies, we aim to improve model accuracy and interpretability, especially in resource-constrained settings. Using the PubMedQA dataset, we demonstrate how CoT helps smaller models break down complex medical queries into sequential steps, enabling more structured reasoning. While these models still face challenges in handling highly specialized medical content, CoT significantly improves their viability for healthcare applications. Our findings suggest that further optimization through methods like retrieval-augmented generation could further close the performance gap between smaller and larger models.
| Original language | English |
|---|---|
| Pages (from-to) | 783-787 |
| Number of pages | 5 |
| Journal | Studies in Health Technology and Informatics |
| Volume | 327 |
| DOIs | |
| Publication status | Published - 15 May 2025 |
Keywords
- Chain of Thought
- Large Language Models
- Medical Reasoning