ABSTRACT
What price should be offered to a worker for a task in an online labor market? How can one enable workers to express the amount they desire to receive for the task completion? Designing optimal pricing policies and determining the right monetary incentives is central to maximizing requester's utility and workers' profits. Yet, current crowdsourcing platforms only offer a limited capability to the requester in designing the pricing policies and often rules of thumb are used to price tasks. This limitation could result in inefficient use of the requester's budget or workers becoming disinterested in the task.
In this paper, we address these questions and present mechanisms using the approach of regret minimization in online learning. We exploit a link between procurement auctions and multi-armed bandits to design mechanisms that are budget feasible, achieve near-optimal utility for the requester, are incentive compatible (truthful) for workers and make minimal assumptions about the distribution of workers' true costs. Our main contribution is a novel, no-regret posted price mechanism, BP-UCB, for budgeted procurement in stochastic online settings. We prove strong theoretical guarantees about our mechanism, and extensively evaluate it in simulations as well as on real data from the Mechanical Turk platform. Compared to the state of the art, our approach leads to a 180% increase in utility.
- Mechanical turk platform. https://www.mturk.com/.Google Scholar
- A. Archer and E. Tardos. Frugal path mechanisms. In SODA, pages 991--999, 2002. Google ScholarDigital Library
- P. Auer, N. Cesa-Bianchi, and P. Fischer. Finite-time analysis of the multiarmed bandit problem. Machine Learning, 47(2-3):235--256, 2002. Google ScholarDigital Library
- P. Auer, R. Ortner, and C. Szepesvari. Improved rates for the stochastic continuum-armed bandit problem. In COLT, pages 454--468, 2007. Google ScholarDigital Library
- M. Babaioff, M. Dinitz, A. Gupta, N. Immorlica, and K. Tal-war. Secretary problems: weights and discounts. In SODA, pages 1245--1254, 2009. Google ScholarDigital Library
- M. Babaioff, S. Dughmi, R. Kleinberg, and A. Slivkins. Dynamic pricing with limited supply. In EC, 2012. Google ScholarDigital Library
- M. Babaioff, N. Immorlica, D. Kempe, and R. Kleinberg. A knapsack secretary problem with applications. In APPROX-RANDOM, pages 16--28, 2007. Google ScholarDigital Library
- A. Badanidiyuru, R. Kleinberg, and Y. Singer. Learning on a budget: Posted price mechanisms for online procurement. In EC, 2012. Google ScholarDigital Library
- Z. Bar-yossef, K. Hildrum, and F. Wu. Incentive-compatible online auctions for digital goods. In SODA, pages 964--970, 2002. Google ScholarDigital Library
- O. Besbes and A. Zeevi. Dynamic pricing without knowing the demand function: Risk bounds and near-optimal algorithms. Operations Research, 2009. Google ScholarDigital Library
- A. Blum and J. Hartline. Near-optimal online auctions. In SODA, 2005. Google ScholarDigital Library
- A. Blum, V. Kumar, A. Rudra, and F. Wu. Online learning in online auctions. In SODA, 2003. Google ScholarDigital Library
- A. Blum and Y. Monsour. Learning, Regret minimization, and Equilibria in Algorithmic Game Theory. Cambridge University Press, 2007.Google Scholar
- M. C. Cary, A. D. Flaxman, J. D. Hartline, and A. R. Karlin. Auctions for structured procurement. In SODA, pages 304--313, 2008. Google ScholarDigital Library
- N. Cesa-Bianchi, Y. Freund, D. P. Helmbold, D. Haussler, R. Schapire, , and M. Warmuth. How to use expert advice. Journal of the ACM, 44(2):427--485, 1997. Google ScholarDigital Library
- S. Chawla, J. D. Hartline, D. L. Malec, and B. Sivan. Multi-parameter mechanism design and sequential posted pricing. In EC, 2010.Google Scholar
- Y. Chen and J. W. Vaughan. A new understanding of prediction markets via no-regret learning. In STOC, 2010.Google ScholarDigital Library
- N. Devanur and J. Hartline. Limited and online supply and the bayesian foundations of prior-free mechanism design. In EC, 2009. Google ScholarDigital Library
- A. Goldberg, J. Hartline, and A. Wright. Competitive auctions and digital goods. In SODA, pages 735--744, 2001. Google ScholarDigital Library
- M. T. Hajiaghayi, R. D. Kleinberg, , and D. C. Parkes. Adaptive limited-supply online auctions. In EC, 2004. Google ScholarDigital Library
- J. J. Horton and L. B. Chilton. The labor economics of paid crowdsourcing. In EC, pages 209--218, 2010. Google ScholarDigital Library
- J. J. Horton and R. J. Zeckhauser. Algorithmic wage negotiations: Applications to paid crowsourcing. In CrowdConf, 2010.Google Scholar
- R. Karlin, D. Kempe, and T. Tamir. Beyond vcg: Frugality of truthful mechanisms. In FOCS, 2005. Google ScholarDigital Library
- R. Kleinberg. Nearly tight bounds for the continuum-armed bandit problem. In NIPS, 2004.Google Scholar
- R. D. .Kleinberg and F. T. Leighton. The value of knowing a demand curve:bounds on regret for online posted-price auctions. In FOCS, 2003. Google ScholarDigital Library
- R. Lavi and N. Nisan. Competitive analysis of incentive compatible on-line auctions. In EC, pages 233--241, 2000. Google ScholarDigital Library
- N. Littlestone and M. K. Warmuth. The weighted majority algorithm. Info and Computation, 70(2):212--261, 1994. Google ScholarDigital Library
- W. Mason and D. J. Watts. Financial incentives and the performance of crowds. In SIGKDD Explor. Newsl., pages 100--108, 2010. Google ScholarDigital Library
- J. J. Shaw, A. D.; Horton and D. L. Chen. Designing incentives for inexpert human raters. In CSCW, pages 275--284, 2011. Google ScholarDigital Library
- Y. Singer. Budget feasible mechanisms. In FOCS, pages 765--774, 2010. Google ScholarDigital Library
- Y. Singer. How to win friends and influence people, truthfully: Influence maximization mechanisms for social networks. In WSDM, 2011. Google ScholarDigital Library
- Y. Singer and M. Mittal. Pricing tasks in online labor markets. In Workshop on Human Computation. ACM, 2011.Google Scholar
- N. Srinivas, A. Krause, S. Kakade, and M. Seeger. Gaussian process optimization in the bandit setting: No regret and experimental design. In ICML, 2010.Google ScholarDigital Library
- Tran-Thanh, L., C. A., J. E. M. de Cote, A. Rogers, and N. R. Jennings. Epsilon-first policies for budget-limited multi-armed bandits. In AAAI, pages 1211--1216, 2010.Google Scholar
- L. Tran-Thanh, A. Chapman, A. Rogers, , and N. R. Jennings. Knapsack based optimal policies for budget-limited multi-armed bandits. http://arxiv.org/abs/1204.1909, 2013. Technical report.Google Scholar
- L. Tran-Thanh, A. Chapman, A. Rogers, and N. R. Jennings. Knapsack based optimal policies for budget-limited multi-armed bandits. In AAAI, 2012.Google ScholarDigital Library
- L. Tran-Thanh, S. Stein, A. Rogers, and N. R. Jennings. Efficient crowdsourcing of unknown experts using multi-armed bandits. In AAAI, 2012.Google Scholar
Index Terms
- Truthful incentives in crowdsourcing tasks using regret minimization mechanisms
Recommendations
Selling to a No-Regret Buyer
EC '18: Proceedings of the 2018 ACM Conference on Economics and ComputationWe consider the problem of a single seller repeatedly selling a single item to a single buyer (specifically, the buyer has a value drawn fresh from known distribution D in every round). Prior work assumes that the buyer is fully rational and will ...
Regret minimization and the price of total anarchy
STOC '08: Proceedings of the fortieth annual ACM symposium on Theory of computingWe propose weakening the assumption made when studying the price of anarchy: Rather than assume that self-interested players will play according to a Nash equilibrium (which may even be computationally hard to find), we assume only that selfish players ...
Characterizing truthful multi-armed bandit mechanisms: extended abstract
EC '09: Proceedings of the 10th ACM conference on Electronic commerceWe consider a multi-round auction setting motivated by pay-per-click auctions for Internet advertising. In each round the auctioneer selects an advertiser and shows her ad, which is then either clicked or not. An advertiser derives value from clicks; ...
Comments