Keywords: strategic classification, strategic learning, fairness, societal ML, equilibrium analysis, Stackelberg games, causality, peer learning
TL;DR: We study strategic learning when learners must induce desirable incentives fairly. We show theoretical and empirical trade-offs between accuracy/social welfare and fairness under desirability constraints.
Abstract: \emph{Strategic learning} studies how decision rules interact with agents who may strategically change their inputs/features to achieve better outcomes. In standard settings, models assume that the decision-maker's sole scope is to learn a classifier that maximizes an objective (e.g., accuracy) assuming that agents will best respond. However, real decision-making systems' goals do not always align \emph{exclusively} with producing good predictions. They may need to consider the downstream effects of inducing certain incentives, which translates into certain features being regarded as more \emph{desirable} to change for the decision maker.
Not only that, but the principal may also need to incentivize desirable feature changes equally across heterogeneous agents.
\emph{How much does this constrained optimization (i.e., maximize the principal's objective, while minimizing the disparity in terms of incentivizing desirable effort) cost the principal?} We propose a unified model of principal-agent interaction that captures this trade-off under three additional components: (1) causal dependencies between features, such that changes in one feature affect others; (2) heterogeneous manipulation costs across the agent population; and (3) peer learning, through which agents infer the principal's algorithm. We provide theoretical guarantees on the principal's optimality loss constrained to a particular desirability fairness tolerance for multiple broad classes of fairness measures. Finally, we demonstrate through experiments on real datasets an explicit tradeoff between maximizing accuracy and fairness in desirability effort.
Supplementary Material: zip
Primary Area: alignment, fairness, safety, privacy, and societal considerations
Code Of Ethics: I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics.
Submission Guidelines: I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2026/AuthorGuide.
Anonymous Url: I certify that there is no URL (e.g., github page) that could be used to find authors’ identity.
No Acknowledgement Section: I certify that there is no acknowledgement section in this submission for double blind review.
Submission Number: 21099
Loading