Minimizing Human Intervention in Online Classification
TL;DR: We study an online classification problem where the agent can ask a costly expert for guidance.
Abstract: We introduce and study a problem where an agent must sequentially classify user-submitted queries represented by $d$-dimensional embeddings drawn i.i.d. from an unknown distribution. The agent may consult a costly human expert for the
correct label, or guess on her own without receiving feedback. The goal is to minimize regret against an oracle with free expert access. When the time horizon $T$ is at least exponential in the embedding dimension $d$, one can learn the geometry of the class regions: in this regime, we propose the Conservative Hull-based Classifier (CHC), which maintains convex hulls of expert-labeled queries and calls the expert as soon as a query lands outside all known hulls. CHC attains $\mathcal{O}(\log^d T)$ regret in $T$ and is minimax optimal for $d=1$. Otherwise, the geometry cannot be reliably learned without additional distributional assumptions (e.g. a margin condition). We show that when the queries are drawn from a subgaussian mixture, for $T \le e^d$, a simple Explore-Then-Commit (ETC) algorithm achieves regret proportional to $N\log{N}$ where $N$ is the number of classes. To bridge these regimes, we introduce the Generalized Hull-based Classifier (GHC), a practical extension of CHC that allows more aggressive guessing via a tunable threshold parameter. Our approach is validated with experiments, notably on real-world question-answering datasets using embeddings derived from state-of-the-art large language models.
Submission Number: 1309
Loading