Multi-Metric Adaptive Experimental Design under a Fixed Budget with Validation
TL;DR: We propose a multi-metric adaptive experimental design with validation, introducing a variance-aware sequential halving algorithm that efficiently identifies treatments most likely to pass A/B testing.
Abstract: A/B tests in online experiments face statistical power challenges when testing multiple candidates simultaneously, while adaptive experimental designs (AED) alone fall short in inferring experiment statistics such as the average treatment effect, especially with many metrics (e.g., revenue, safety) and heterogeneous variances. This paper proposes a fixed-budget multi-metric AED framework with a two-phase structure: an adaptive exploration phase to identify the best treatment, and a validation phase with an A/B test to verify the treatment's quality and infer statistics. We propose SHRVar, which generalizes sequential halving (SH) with a novel relative-variance-based sampling and an elimination strategy built on reward $z$ values. It achieves a provable error probability that decreases exponentially, where the exponent $H_3$ generalizes the complexity measure for SH and SHVar with homogeneous and heterogeneous variances, respectively. Numerical experiments demonstrate its performance and robustness.
Submission Number: 927
Loading