Training-Free Bayesianization for Low-Rank Adapters of Large Language Models

Published: 05 Mar 2025, Last Modified: 22 Mar 2025QUESTION PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Large Language Models, Uncertainty Estimation, Bayesian Inference
TL;DR: We propose a Training-Free Bayesianization approach for LLM adapters that achieves better uncertainty estimation.
Abstract: Estimating the uncertainty of responses of Large Language Models (LLMs) remains a critical challenge. While recent Bayesian methods have demonstrated effectiveness in quantifying uncertainty through low-rank weight updates, they typically require complex fine-tuning or post-training procedures. In this paper, we propose **T**raining-**F**ree **B**ayesianization (**TFB**), a novel framework that efficiently transforms existing off-the-shelf trained low-rank adapters into Bayesian ones without additional training. TFB systematically searches for the maximally acceptable level of variance in the weight posterior, constrained within a family of low-rank isotropic Gaussian distributions. We theoretically demonstrate that under mild conditions, this search process is equivalent to KL-regularized variational optimization, a generalized form of variational inference. Through comprehensive experiments, we show that TFB achieves superior uncertainty estimation and generalization compared to existing methods while eliminating the need for complex training procedures.
Submission Number: 6
Loading