Towards Sustainable Inference of LLMs for Medical Education Through Token Count Minimization

Hyunggu Jung, Jiyoo Min, Yunseo Moon, Jaebeom Son, Mark E. Whipple

Published: 07 Aug 2025, Last Modified: 16 Feb 2026CrossrefEveryoneRevisionsCC BY-SA 4.0
Abstract: The carbon emissions generated during the inference phase of large language models (LLMs) are a growing concern. We propose a method to minimize the number of tokens of input prompts for medical students interacting with LLMs. Using English physical examination course materials, we applied translation and paraphrasing to recommend the most token-efficient prompt. We found that English baseline prompts had fewer tokens than Korean ones, while the paraphrased forms we proposed significantly reduced token counts compared to the baseline prompts.
Loading