TY - GEN
T1 - Small Language Models in Healthcare: Capabilities, Challenges, and Future Directions
AU - Rao, Sanjeev
AU - Singh, Reyan
AU - Jagya, Kanishk
AU - Sharma, Sahil
PY - 2025/12/31
Y1 - 2025/12/31
N2 - Large Language Models (LLMs) exhibit significant proficiency in various natural language processing tasks. Nonetheless, their direct implementation in healthcare is limited by significant computational expenses, concerns regarding data privacy, and restricted interpretability. In contrast, Small Language Models (SLMs), generally comprising fewer than 8 billion parameters and often less than 3 billion in clinical applications, present a more pragmatic and privacy-aware option. This paper offers an overview of SLMs in healthcare, focusing on architectural innovations, training strategies, and domain-specific adaptations that facilitate their implementation in clinical environments. This work examines healthcare applications in decision support, documentation, and patient interaction, emphasizing technical adaptations including multimodal integration and knowledge distillation. Besides, the paper summarizes risk–benefit considerations to facilitate responsible adoption and proposes an updated taxonomy that classifies models based on deployment readiness and modality support. Finally, the paper concludes by identifying significant challenges and proposing future research directions for the reliable and scalable integration of SLMs in practical healthcare settings.
AB - Large Language Models (LLMs) exhibit significant proficiency in various natural language processing tasks. Nonetheless, their direct implementation in healthcare is limited by significant computational expenses, concerns regarding data privacy, and restricted interpretability. In contrast, Small Language Models (SLMs), generally comprising fewer than 8 billion parameters and often less than 3 billion in clinical applications, present a more pragmatic and privacy-aware option. This paper offers an overview of SLMs in healthcare, focusing on architectural innovations, training strategies, and domain-specific adaptations that facilitate their implementation in clinical environments. This work examines healthcare applications in decision support, documentation, and patient interaction, emphasizing technical adaptations including multimodal integration and knowledge distillation. Besides, the paper summarizes risk–benefit considerations to facilitate responsible adoption and proposes an updated taxonomy that classifies models based on deployment readiness and modality support. Finally, the paper concludes by identifying significant challenges and proposing future research directions for the reliable and scalable integration of SLMs in practical healthcare settings.
KW - Clinical NLP
KW - Edge Deployment
KW - Healthcare AI
KW - Knowledge Distillation
KW - Multimodal Learning
KW - Risk Assessment
KW - Privacy-Preserving AI
KW - Small Language Models (SLMs)
UR - https://pure.ulster.ac.uk/en/publications/a22a54e1-e9dc-4b10-ba66-323d3abf9a82
U2 - 10.1109/ic366947.2025.11290378
DO - 10.1109/ic366947.2025.11290378
M3 - Conference contribution
SN - 979-8-3315-5419-4
T3 - 2025 Seventeenth International Conference on Contemporary Computing (IC3)
SP - 1
EP - 6
BT - 2025 Seventeenth International Conference on Contemporary Computing (IC3)
PB - IEEE Xplore
T2 - 2025 Seventeenth International Conference on Contemporary Computing (IC3)
Y2 - 7 August 2025 through 9 August 2025
ER -