Assessing the Macro and Micro Effects of Random Seeds on Fine-Tuning Large Language Models

ACL ARR 2025 February Submission2796 Authors

15 Feb 2025 (modified: 09 May 2025)ACL ARR 2025 February SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: The impact of random seeds in fine-tuning large language models (LLMs) has been largely overlooked despite its potential influence on model performance. In this study, we systematically evaluate the effects of random seeds on LLMs using the GLUE and SuperGLUE benchmarks. We analyze the macro-level impact through traditional metrics like accuracy and F1, calculating their mean and variance to quantify performance fluctuations. To capture the micro-level effects, we introduce a novel metric, consistency, measuring the stability of individual predictions across runs. Our experiments reveal significant variance at both macro and micro levels, underscoring the need for careful consideration of random seeds in fine-tuning and evaluation.
Paper Type: Short
Research Area: Special Theme (conference specific)
Research Area Keywords: evaluation methodologies, evaluation, generalization, metrics, reproducibility, robustness
Languages Studied: English
Submission Number: 2796
Loading