Learning algorithms for online principal-agent problems (and selling goods online)


Journal Article

In a principal-agent problem, a principal seeks to motivate an agent to take a certain action beneficial to the principal, while spending as little as possible on the reward. This is complicated by the fact that the principal does not know the agent's utility function (or type). We study the online setting where at each round, the principal encounters a new agent, and the principal sets the rewards anew. At the end of each round, the principal only finds out the action that the agent took, but not his type. The principal must learn how to set the rewards optimally. We show that this setting generalizes the setting of selling a digital good online. We study and experimentally compare three main approaches to this problem. First, we show how to apply a standard bandit algorithm to this setting. Second, for the case where the distribution of agent types is fixed (but unknown to the principal), we introduce a new gradient ascent algorithm. Third, for the case where the distribution of agents' types is fixed, and the principal has a prior belief (distribution) over a limited class of type distributions, we study a Bayesian approach.

Duke Authors

Cited Authors

  • Conitzer, V; Garera, N

Published Date

  • October 6, 2006

Published In

  • Icml 2006 Proceedings of the 23rd International Conference on Machine Learning

Volume / Issue

  • 2006 /

Start / End Page

  • 209 - 216

Citation Source

  • Scopus