Recognition: unknown
Strategic Classification from Revealed Preferences
read the original abstract
We study an online linear classification problem, in which the data is generated by strategic agents who manipulate their features in an effort to change the classification outcome. In rounds, the learner deploys a classifier, and an adversarially chosen agent arrives, possibly manipulating her features to optimally respond to the learner. The learner has no knowledge of the agents' utility functions or "real" features, which may vary widely across agents. Instead, the learner is only able to observe their "revealed preferences" --- i.e. the actual manipulated feature vectors they provide. For a broad family of agent cost functions, we give a computationally efficient learning algorithm that is able to obtain diminishing "Stackelberg regret" --- a form of policy regret that guarantees that the learner is obtaining loss nearly as small as that of the best classifier in hindsight, even allowing for the fact that agents will best-respond differently to the optimal classifier.
This paper has not been read by Pith yet.
Forward citations
Cited by 1 Pith paper
-
Sequential Strategic Classification with Multi-Stage Selective Classifiers
A new multi-stage sequential model with selective classifiers is proposed to characterize agent actions and design sequences that incentivize genuine improvement rather than gaming in strategic classification.
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.