Incentives in Crowdsourcing: A Game-theoretic Approach

Similar documents
Incentivizing High Quality User Contributions: New Arm Generation in Bandit Learning

Incentivizing High Quality User Contributions: New Arm Generation in Bandit Learning

Price of anarchy in auctions & the smoothness framework. Faidra Monachou Algorithmic Game Theory 2016 CoReLab, NTUA

Crowdsourcing Contests

Game Theory & Firms. Jacob LaRiviere & Justin Rao April 20, 2016 Econ 404, Spring 2016

Games, Auctions, Learning, and the Price of Anarchy. Éva Tardos Cornell University

Data Mining. CS57300 Purdue University. March 27, 2018

Incentives and Truthful Reporting in Consensus-centric Crowdsourcing

Copyright (C) 2001 David K. Levine This document is an open textbook; you can redistribute it and/or modify it under the terms of version 1 of the

Efficiency and Robustness of Binary Online Feedback Mechanisms in Trading Environments with Moral Hazard

On Optimal Multidimensional Mechanism Design

Machine learning for Dynamic Social Network Analysis

Practical Peer Prediction for Peer Assessment Victor Shnayder and David C. Parkes Harvard University

Quality Expectation-Variance Tradeoffs in Crowdsourcing Contests

Intro to Algorithmic Economics, Fall 2013 Lecture 1

A Reinforcement Learning Framework for Eliciting High Quality Information

INTERMEDIATE MICROECONOMICS (EC201)

Algorithmic Mechanisms for Internet-based Master-Worker Computing with Untrusted and Selfish Workers

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

Incentives in Crowdsourcing

Mazzeo (RAND 2002) Seim (RAND 2006) Grieco (RAND 2014) Discrete Games. Jonathan Williams 1. 1 UNC - Chapel Hill

MERITOCRACY AS A MECHANISM IN THE CONTEXT OF VOLUNTARY CONTRIBUTION GAMES

Imperfect Price Information and Competition

CMSC 474, Introduction to Game Theory Analyzing Normal-Form Games

arxiv: v2 [cs.gt] 5 May 2017

Oligopoly: How do firms behave when there are only a few competitors? These firms produce all or most of their industry s output.

Online shopping and platform design with ex ante registration requirements

ECN 3103 INDUSTRIAL ORGANISATION

University of California, Davis Date: June 24, PRELIMINARY EXAMINATION FOR THE Ph.D. DEGREE. Answer four questions (out of five)

MOT Seminar. John Musacchio 4/16/09

Newspapers. Joel Sobel. Mullainathan- Shleifer. Newspapers. Gentzkow- Shapiro. Ellman and Germano. Posner. Joel Sobel. UCSD and UAB.

Hidden Actions Analysis in p2p Systems

Game theory (Sections )

Harvard University Department of Economics

Optimal Shill Bidding in the VCG Mechanism

David Easley and Jon Kleinberg November 29, 2010

PERFORMANCE, PROCESS, AND DESIGN STANDARDS IN ENVIRONMENTAL REGULATION

Multi armed Bandits on the Web

Crowdsourcing with All-pay Auctions: a Field Experiment on Taskcn

Buyer Heterogeneity and Dynamic Sorting in Markets for Durable Lemons

VALUE OF SHARING DATA

Sustaining Cooperation in Social Exchange Networks with Incomplete Global Information

CS269I: Incentives in Computer Science Lecture #6: Incentivizing Participation

Matching Markets with Endogenous Information

OPEC CARTEL BEHAVIOR: WHAT IS THEIR OBJECTIVE AND WHAT MAY HAPPEN NOW?

a study on Robust Mechanisms for Social Coordination

Pricing in Dynamic Advance Reservation Games

Estimating Bidders Valuation Distributions in Online Auctions

ECONS 424 STRATEGY AND GAME THEORY MIDTERM EXAM #2

PRICE OF ANARCHY: QUANTIFYING THE INEFFICIENCY OF EQUILIBRIA. Zongxu Mu

Where are we? Knowledge Engineering Semester 2, Basic Considerations. Decision Theory

FIRST FUNDAMENTAL THEOREM OF WELFARE ECONOMICS

Lecture 1:Human Capital Theory. September 27, 2017 Hideo Owan Institute of Social Science

Prizes and Patents: Using Market Signals to Provide Incentives for Innovations

IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, VOL., NO., 1. Dynamic Spectrum Sharing Auction with Time-Evolving Channel Qualities

TOPIC 4. ADVERSE SELECTION, SIGNALING, AND SCREENING

Online shopping and platform design with ex ante registration requirements. Online Appendix

STABILITY AND WELFARE OF MERITOCRATIC GROUP-BASED MECHANISMS

Review from last. Econ 2230: Public Economics. Outline. 1. Dominant strategy implementation. Lecture 9: Mechanism Design lab evidence

Active Learning for Conjoint Analysis

Managing Decentralized Inventory and Transhipment

Part II. Market power

Dynamic Learning in Strategic Pricing Games

Test Codes: QEA and QEB (Both are of 'Descriptive' type) (Junior Research Fellowship in Quantitative Economics)

Reaction Paper Influence Maximization in Social Networks: A Competitive Perspective

Matching: The Theory. Muriel Niederle Stanford and NBER. October 5, 2011

1.. Consider the following multi-stage game. In the first stage an incumbent monopolist

Sponsored Search Markets

The Price of Anarchy in an Exponential Multi-Server

Equilibrium Analysis of Dynamic Bidding in Sponsored Search Auctions

Simple Market Equilibria with Rationally Inattentive Consumers

Does Signaling Solve the Lemons Problem? Timothy Perri * March 31, Abstract

Department of Economics. Harvard University. Spring Honors General Exam. April 6, 2011

Topics in ALGORITHMIC GAME THEORY *

Lecture 4: Will profits in reality be higher or lower than under Cournot? Tom Holden

Inefficiency of Equilibria in Task Allocation by Crowdsourcing

Monitoring and Collusion with Soft Information

The origins of an equilibrium wage distribution. Ofer Cornfeld. June 2013

Distributed Optimization

Approximation in Algorithmic Game Theory

Competitive Franchising

Markets and Values. The Evolution of Intrinsic Motivation. Tim Besley and Maitreesh Ghatak. November 2015

Social Computing Systems

Bilateral and Multilateral Exchanges for Peer-Assisted Content Distribution

Incentive-Compatible Mechanisms for Norm Monitoring in Open Multi-Agent Systems (Extended Abstract)

Incentive-Compatible Mechanisms for Norm Monitoring in Open Multi-Agent Systems (Extended Abstract)

Best-response functions, best-response curves. Strategic substitutes, strategic complements. Quantity competition, price competition

Algoritmi Distribuiti e Reti Complesse (modulo II) Luciano Gualà

Labor Economics, Lectures 9 and 10: Investments in General and Speci c Skills

First-Price Auctions with General Information Structures: A Short Introduction

Homogeneous Platform Competition with Heterogeneous Consumers

On the Reputation of Agent-based Web Services

Wireless Networking with Selfish Agents. Li (Erran) Li Center for Networking Research Bell Labs, Lucent Technologies

Robust Supply Function Bidding in Electricity Markets With Renewables

Economics II - October 27, 2009 Based on H.R.Varian - Intermediate Microeconomics. A Modern Approach

Computational Microeconomics: Game Theory, Social Choice,

ECON 500 Microeconomic Theory MARKET FAILURES. Asymmetric Information Externalities Public Goods

Oligopoly Theory (11) Collusion

Collusion. Sotiris Georganas. February Sotiris Georganas () Collusion February / 31

Intermediate Microeconomics IMPERFECT COMPETITION BEN VAN KAMMEN, PHD PURDUE UNIVERSITY

Transcription:

Incentives in Crowdsourcing: A Game-theoretic Approach ARPITA GHOSH Cornell University NIPS 2013 Workshop on Crowdsourcing: Theory, Algorithms, and Applications Incentives in Crowdsourcing: A Game-theoretic Approach 1 / 26

Users on the Web: Online collective effort Contribution online from the crowds: Incentives in Crowdsourcing: A Game-theoretic Approach 2 / 26

Users on the Web: Online collective effort Contribution online from the crowds: Reviews (Amazon, Yelp), online Q&A sites (Y! Answers, Quora, StackOverflow), discussion forums Wikipedia Social media: Blogs, YouTube,... Incentives in Crowdsourcing: A Game-theoretic Approach 2 / 26

Users on the Web: Online collective effort Contribution online from the crowds: Reviews (Amazon, Yelp), online Q&A sites (Y! Answers, Quora, StackOverflow), discussion forums Wikipedia Social media: Blogs, YouTube,... Crowdsourcing: Paid and unpaid; microtasks and challenges Amazon Mechanical Turk, Citizen Science (GalaxyZoo, FoldIt), Games with a Purpose, contests (Innocentive, Topcoder) Incentives in Crowdsourcing: A Game-theoretic Approach 2 / 26

Users on the Web: Online collective effort Contribution online from the crowds: Reviews (Amazon, Yelp), online Q&A sites (Y! Answers, Quora, StackOverflow), discussion forums Wikipedia Social media: Blogs, YouTube,... Crowdsourcing: Paid and unpaid; microtasks and challenges Amazon Mechanical Turk, Citizen Science (GalaxyZoo, FoldIt), Games with a Purpose, contests (Innocentive, Topcoder) Online education: Peer-learning, peer-grading Incentives in Crowdsourcing: A Game-theoretic Approach 2 / 26

Incentives and collective effort Quality, participation varies widely across systems Incentives in Crowdsourcing: A Game-theoretic Approach 3 / 26

Incentives and collective effort Quality, participation varies widely across systems How to incentivize high participation and effort? Incentives in Crowdsourcing: A Game-theoretic Approach 3 / 26

Incentives and collective effort Quality, participation varies widely across systems How to incentivize high participation and effort? Two components to designing incentives: Social psychology: What constitutes a reward? Incentives in Crowdsourcing: A Game-theoretic Approach 3 / 26

Incentives and collective effort Quality, participation varies widely across systems How to incentivize high participation and effort? Two components to designing incentives: Social psychology: What constitutes a reward? Rewards are limited: How to allocate among self-interested users? A game-theoretic framework for incentive design Incentives in Crowdsourcing: A Game-theoretic Approach 3 / 26

The game-theoretic approach to incentive design System design induces rules specifying allocation of rewards Self-interested users choose actions to maximize own payoff Incentives in Crowdsourcing: A Game-theoretic Approach 4 / 26

The game-theoretic approach to incentive design System design induces rules specifying allocation of rewards Self-interested users choose actions to maximize own payoff Participation ( Endogenous entry ) Incentives in Crowdsourcing: A Game-theoretic Approach 4 / 26

The game-theoretic approach to incentive design System design induces rules specifying allocation of rewards Self-interested users choose actions to maximize own payoff Participation ( Endogenous entry ) Revealing information truthfully (ratings, opinions,... ) Incentives in Crowdsourcing: A Game-theoretic Approach 4 / 26

The game-theoretic approach to incentive design System design induces rules specifying allocation of rewards Self-interested users choose actions to maximize own payoff Participation ( Endogenous entry ) Revealing information truthfully (ratings, opinions,... ) Effort: Quality of content (UGC sites) Output accuracy (crowdsourcing) Quantity: Number of contributions, attemped tasks Speed of response (Q&A forums),... Incentive design: Allocate reward to align agent s incentives with system Incentives in Crowdsourcing: A Game-theoretic Approach 4 / 26

Incentive design for crowdsourcing Reward allocation problem varies across systems: Incentives in Crowdsourcing: A Game-theoretic Approach 5 / 26

Incentive design for crowdsourcing Reward allocation problem varies across systems: Why? Constraints, reward regimes, vary with nature of reward: Monetary; social-psychological (attention, status,... ) Attention rewards: Diverging [GM11, GH11]; subset constraints [GM12] Money-like rewards: Bounded; sum constraints [GM12] Incentives in Crowdsourcing: A Game-theoretic Approach 5 / 26

Incentive design for crowdsourcing Reward allocation problem varies across systems: Why? Constraints, reward regimes, vary with nature of reward: Monetary; social-psychological (attention, status,... ) Attention rewards: Diverging [GM11, GH11]; subset constraints [GM12] Money-like rewards: Bounded; sum constraints [GM12] Observability of (value of) agents output Can only reward what you can see Incentives in Crowdsourcing: A Game-theoretic Approach 5 / 26

Incentive design for crowdsourcing Reward allocation problem varies across systems: Why? Constraints, reward regimes, vary with nature of reward: Monetary; social-psychological (attention, status,... ) Attention rewards: Diverging [GM11, GH11]; subset constraints [GM12] Money-like rewards: Bounded; sum constraints [GM12] Observability of (value of) agents output Can only reward what you can see Perfect rank-ordering: Contests [... ] Imperfect: Noisy votes in UGC [EG13, GH13] Unobservable: Judgement elicitation [DG13] Incentives in Crowdsourcing: A Game-theoretic Approach 5 / 26

Learning & incentives in user-generated content Joint work with Patrick Hummel, ITCS 13 The setting: User-generated content (Reviews, Q&A forums, comments, videos, articles,... ) Quality of contributions varies widely: Incentives in Crowdsourcing: A Game-theoretic Approach 6 / 26

Learning & incentives in user-generated content Joint work with Patrick Hummel, ITCS 13 The setting: User-generated content (Reviews, Q&A forums, comments, videos, articles,... ) Quality of contributions varies widely: Sites want to display best contributions Incentives in Crowdsourcing: A Game-theoretic Approach 6 / 26

Learning & incentives in user-generated content Joint work with Patrick Hummel, ITCS 13 The setting: User-generated content (Reviews, Q&A forums, comments, videos, articles,... ) Quality of contributions varies widely: Sites want to display best contributions But quality is not directly observable: Incentives in Crowdsourcing: A Game-theoretic Approach 6 / 26

Learning & incentives in user-generated content Joint work with Patrick Hummel, ITCS 13 The setting: User-generated content (Reviews, Q&A forums, comments, videos, articles,... ) Quality of contributions varies widely: Sites want to display best contributions But quality is not directly observable: Infer quality from viewer votes Incentives in Crowdsourcing: A Game-theoretic Approach 6 / 26

Learning & incentives in user-generated content Joint work with Patrick Hummel, ITCS 13 The setting: User-generated content (Reviews, Q&A forums, comments, videos, articles,... ) Quality of contributions varies widely: Sites want to display best contributions But quality is not directly observable: Infer quality from viewer votes How to display contributions to optimize overall viewer experience? Incentives in Crowdsourcing: A Game-theoretic Approach 6 / 26

A multi-armed bandit problem Incentives in Crowdsourcing: A Game-theoretic Approach 7 / 26

A multi-armed bandit problem Learning contribution qualities: Multi-armed bandit problem Arms: Contributions Success probability: Contribution s quality Incentives in Crowdsourcing: A Game-theoretic Approach 7 / 26

A multi-armed bandit problem Learning contribution qualities: Multi-armed bandit problem Arms: Contributions Success probability: Contribution s quality Contributors: Agents with cost to quality, benefit from views Incentives in Crowdsourcing: A Game-theoretic Approach 7 / 26

A multi-armed bandit problem Learning contribution qualities: Multi-armed bandit problem Arms: Contributions Success probability: Contribution s quality Contributors: Agents with cost to quality, benefit from views Arms are endogenous! Contributors choose whether to participate, content quality Incentives in Crowdsourcing: A Game-theoretic Approach 7 / 26

A multi-armed bandit problem Learning contribution qualities: Multi-armed bandit problem Arms: Contributions Success probability: Contribution s quality Contributors: Agents with cost to quality, benefit from views Arms are endogenous! Contributors choose whether to participate, content quality What is a good learning algorithm in this setting? Incentives in Crowdsourcing: A Game-theoretic Approach 7 / 26

Overview Strategic contributors: Decide participation, quality Viewers vote on displayed contributions Mechanism: Decides which contribution to display Metric: Equilibrium regret Incentives in Crowdsourcing: A Game-theoretic Approach 8 / 26

Model: Content and feedback Contribution quality q: Probability of viewer upvote Incentives in Crowdsourcing: A Game-theoretic Approach 9 / 26

Model: Content and feedback Contribution quality q: Probability of viewer upvote Stream of T viewers: Each viewer shown, votes on, one contribution Incentives in Crowdsourcing: A Game-theoretic Approach 9 / 26

Model: Content and feedback Contribution quality q: Probability of viewer upvote Stream of T viewers: Each viewer shown, votes on, one contribution Viewers need not vote perfectly : q [0, γ] Incentives in Crowdsourcing: A Game-theoretic Approach 9 / 26

Model: Content and feedback Contribution quality q: Probability of viewer upvote Stream of T viewers: Each viewer shown, votes on, one contribution Viewers need not vote perfectly : q [0, γ] Mechanism should be robust to γ < 1 Incentives in Crowdsourcing: A Game-theoretic Approach 9 / 26

Contributors K = K(T ) potential contributors Incentives in Crowdsourcing: A Game-theoretic Approach 10 / 26

Contributors K = K(T ) potential contributors Contributors are strategic agents Incentives in Crowdsourcing: A Game-theoretic Approach 10 / 26

Contributors K = K(T ) potential contributors Contributors are strategic agents, choosing Whether or not to participate: Incentives in Crowdsourcing: A Game-theoretic Approach 10 / 26

Contributors K = K(T ) potential contributors Contributors are strategic agents, choosing Whether or not to participate: Probability β i Incentives in Crowdsourcing: A Game-theoretic Approach 10 / 26

Contributors K = K(T ) potential contributors Contributors are strategic agents, choosing Whether or not to participate: Probability β i Contribution quality: q i Incentives in Crowdsourcing: A Game-theoretic Approach 10 / 26

Contributors K = K(T ) potential contributors Contributors are strategic agents, choosing Whether or not to participate: Probability β i Contribution quality: q i Actual number of contributions (arms): k(t ) K(T ) Incentives in Crowdsourcing: A Game-theoretic Approach 10 / 26

Contributor utilities Cost: Quality q incurs cost c(q) c(q) increasing, continuously differentiable Incentives in Crowdsourcing: A Game-theoretic Approach 11 / 26

Contributor utilities Cost: Quality q incurs cost c(q) c(q) increasing, continuously differentiable Benefit from views Incentives in Crowdsourcing: A Game-theoretic Approach 11 / 26

Contributor utilities Cost: Quality q incurs cost c(q) c(q) increasing, continuously differentiable Benefit from views Psychological ([Huberman et al 09,... ]) or monetary benefit Incentives in Crowdsourcing: A Game-theoretic Approach 11 / 26

Contributor utilities Cost: Quality q incurs cost c(q) c(q) increasing, continuously differentiable Benefit from views Psychological ([Huberman et al 09,... ]) or monetary benefit ni t : Views allocated to i until period t Total benefit to i: ni T Incentives in Crowdsourcing: A Game-theoretic Approach 11 / 26

Contributor utilities Cost: Quality q incurs cost c(q) c(q) increasing, continuously differentiable Benefit from views Psychological ([Huberman et al 09,... ]) or monetary benefit ni t : Views allocated to i until period t Total benefit to i: ni T Mechanism: Decides which contribution to display at t Incentives in Crowdsourcing: A Game-theoretic Approach 11 / 26

Contributor utilities Cost: Quality q incurs cost c(q) c(q) increasing, continuously differentiable Benefit from views Psychological ([Huberman et al 09,... ]) or monetary benefit ni t : Views allocated to i until period t Total benefit to i: ni T Mechanism: Decides which contribution to display at t Utility: u i = E[n T i (q i, q i, k(t ))] c(q i ) Incentives in Crowdsourcing: A Game-theoretic Approach 11 / 26

Mapping to MAB K(T ) potential contributors, or arms Viewer t: Pull of arm at time t T : Time horizon or total number of viewers Content quality q i : Success probability of arm i Incentives in Crowdsourcing: A Game-theoretic Approach 12 / 26

Mapping to MAB K(T ) potential contributors, or arms Viewer t: Pull of arm at time t T : Time horizon or total number of viewers Content quality q i : Success probability of arm i Actual number of arms k(t ), qualities q i, determined endogenously in response to learning algorithm Incentives in Crowdsourcing: A Game-theoretic Approach 12 / 26

How good is a learning algorithm as a mechanism? Incentives in Crowdsourcing: A Game-theoretic Approach 13 / 26

How good is a learning algorithm as a mechanism? Performance measure: Equilibrium regret Incentives in Crowdsourcing: A Game-theoretic Approach 13 / 26

How good is a learning algorithm as a mechanism? Performance measure: Equilibrium regret Recall contributors choose q [0, γ] Incentives in Crowdsourcing: A Game-theoretic Approach 13 / 26

How good is a learning algorithm as a mechanism? Performance measure: Equilibrium regret Recall contributors choose q [0, γ] Strong regret of mechanism M: Incentives in Crowdsourcing: A Game-theoretic Approach 13 / 26

How good is a learning algorithm as a mechanism? Performance measure: Equilibrium regret Recall contributors choose q [0, γ] Strong regret of mechanism M: Regret wrt γ, Incentives in Crowdsourcing: A Game-theoretic Approach 13 / 26

How good is a learning algorithm as a mechanism? Performance measure: Equilibrium regret Recall contributors choose q [0, γ] Strong regret of mechanism M: Regret wrt γ, in symmetric mixed-strategy Bayes-Nash equilibrium (β, F (q)), Incentives in Crowdsourcing: A Game-theoretic Approach 13 / 26

How good is a learning algorithm as a mechanism? Performance measure: Equilibrium regret Recall contributors choose q [0, γ] Strong regret of mechanism M: Regret wrt γ, in symmetric mixed-strategy Bayes-Nash equilibrium (β, F (q)), T R(T ) = γt E[ q t ] t=1 Incentives in Crowdsourcing: A Game-theoretic Approach 13 / 26

How good is a learning algorithm as a mechanism? Performance measure: Equilibrium regret Recall contributors choose q [0, γ] Strong regret of mechanism M: Regret wrt γ, in symmetric mixed-strategy Bayes-Nash equilibrium (β, F (q)), T R(T ) = γt E[ q t ] t=1 Strong sublinear equilibrium regret: Incentives in Crowdsourcing: A Game-theoretic Approach 13 / 26

How good is a learning algorithm as a mechanism? Performance measure: Equilibrium regret Recall contributors choose q [0, γ] Strong regret of mechanism M: Regret wrt γ, in symmetric mixed-strategy Bayes-Nash equilibrium (β, F (q)), T R(T ) = γt E[ q t ] t=1 Strong sublinear equilibrium regret: lim T R(T ) T = 0 Incentives in Crowdsourcing: A Game-theoretic Approach 13 / 26

How good is a learning algorithm as a mechanism? Performance measure: Equilibrium regret Recall contributors choose q [0, γ] Strong regret of mechanism M: Regret wrt γ, in symmetric mixed-strategy Bayes-Nash equilibrium (β, F (q)), T R(T ) = γt E[ q t ] t=1 R(T ) Strong sublinear equilibrium regret: lim T T every symmetric equilibrium of M = 0 in Incentives in Crowdsourcing: A Game-theoretic Approach 13 / 26

The UCB algorithm, as a mechanism q t i : Estimated quality of i at time t UCB algorithm M UCB : Display all arms once, then Display i = arg max q t i + 2 ln T n t i Incentives in Crowdsourcing: A Game-theoretic Approach 14 / 26

The UCB algorithm, as a mechanism q t i : Estimated quality of i at time t UCB algorithm M UCB : Display all arms once, then Display i = arg max q t i + 2 ln T n t i Theorem: Mechanism M UCB always has a symmetric mixed-strategy equilibrium (β, F (q)) Incentives in Crowdsourcing: A Game-theoretic Approach 14 / 26

UCB as a mechanism: The good news Incentives in Crowdsourcing: A Game-theoretic Approach 15 / 26

UCB as a mechanism: The good news Theorem If K(T ) is such that lim T T K(T ) ln T = : Incentives in Crowdsourcing: A Game-theoretic Approach 15 / 26

UCB as a mechanism: The good news Theorem If K(T ) is such that lim T T K(T ) ln T = : β = 1 in any equilibria of M UCB for sufficiently large T Incentives in Crowdsourcing: A Game-theoretic Approach 15 / 26

UCB as a mechanism: The good news Theorem If K(T ) is such that lim T T K(T ) ln T = : β = 1 in any equilibria of M UCB for sufficiently large T For any fixed q < γ, the probability of choosing quality q q in any equilibrium goes to 0 as T. Incentives in Crowdsourcing: A Game-theoretic Approach 15 / 26

UCB as a mechanism: The good news Theorem If K(T ) is such that lim T T K(T ) ln T = : β = 1 in any equilibria of M UCB for sufficiently large T For any fixed q < γ, the probability of choosing quality q q in any equilibrium goes to 0 as T. M UCB achieves strong sublinear equilibrium regret. Incentives in Crowdsourcing: A Game-theoretic Approach 15 / 26

UCB as a mechanism: The bad news Theorem Suppose lim T T K(T ) = r <. Incentives in Crowdsourcing: A Game-theoretic Approach 16 / 26

UCB as a mechanism: The bad news Theorem T Suppose lim T K(T ) = r <. Then for sufficiently large T, any equilibrium has the property that no agent chooses quality greater than q = cτ 1 (1 + c(0)). Incentives in Crowdsourcing: A Game-theoretic Approach 16 / 26

Improving equilibrium regret: A modified UCB mechanism M UCB MOD : Run UCB on random subset of min{ T, k(t )} arms Exploring random subset: M 1 FAIL [Berry et al 97] M 1 FAIL achieves strong sublinear regret as an algorithm for large K(T ) Incentives in Crowdsourcing: A Game-theoretic Approach 17 / 26

Improving equilibrium regret: A modified UCB mechanism M UCB MOD : Run UCB on random subset of min{ T, k(t )} arms Exploring random subset: M 1 FAIL [Berry et al 97] M 1 FAIL achieves strong sublinear regret as an algorithm for large K(T ), but not as a mechanism Incentives in Crowdsourcing: A Game-theoretic Approach 17 / 26

Improving equilibrium regret: A modified UCB mechanism M UCB MOD : Run UCB on random subset of min{ T, k(t )} arms Exploring random subset: M 1 FAIL [Berry et al 97] M 1 FAIL achieves strong sublinear regret as an algorithm for large K(T ), but not as a mechanism Theorem M UCB MOD achieves strong sublinear equilibrium regret for all γ 1 and cost functions c, for all K(T ) T. Why UCB works. Incentives in Crowdsourcing: A Game-theoretic Approach 17 / 26

Extensions of result M UCB MOD retains strong sublinear equilibrium regret if: Each viewer is shown multiple contributions Explore min{g(t ), k(t )} arms: G(T ), G(T ) = o( T ln T ) Heterogenous types: Cost functions c τ (q) q [δ, γ], δ > 0 Incentives in Crowdsourcing: A Game-theoretic Approach 18 / 26

Open questions Multi-armed bandits with endogenous arms: Strong sublinear equilibrium regret achievable with modified-ucb mechanism Many unanswered questions: Models, mechanisms Probabilistic feedback Sequential contributions Quality-participation tradeoffs with G(T ) Incentives in Crowdsourcing: A Game-theoretic Approach 19 / 26

Open questions Multi-armed bandits with endogenous arms: Strong sublinear equilibrium regret achievable with modified-ucb mechanism Many unanswered questions: Models, mechanisms Probabilistic feedback Sequential contributions Quality-participation tradeoffs with G(T ) What learning algorithms make good mechanisms when arms are endogenous? Incentives in Crowdsourcing: A Game-theoretic Approach 19 / 26

Incentives in crowdsourcing: Unobservable output Crowdsourced evaluation: Replace expert by aggregated evaluation from crowd Image classification & labeling; content rating; abuse detection; MOOCs peer grading,... How to aggregate evaluations from crowd? Workers have different proficiencies; Incentives in Crowdsourcing: A Game-theoretic Approach 20 / 26

Incentives in crowdsourcing: Unobservable output Crowdsourced evaluation: Replace expert by aggregated evaluation from crowd Image classification & labeling; content rating; abuse detection; MOOCs peer grading,... How to aggregate evaluations from crowd? Workers have different proficiencies; possibly unknown to system: Incentives in Crowdsourcing: A Game-theoretic Approach 20 / 26

Incentives in crowdsourcing: Unobservable output Crowdsourced evaluation: Replace expert by aggregated evaluation from crowd Image classification & labeling; content rating; abuse detection; MOOCs peer grading,... How to aggregate evaluations from crowd? Workers have different proficiencies; possibly unknown to system: Learn, weight to maximize accuracy Incentives in Crowdsourcing: A Game-theoretic Approach 20 / 26

Incentives in crowdsourcing: Unobservable output Crowdsourced evaluation: Replace expert by aggregated evaluation from crowd Image classification & labeling; content rating; abuse detection; MOOCs peer grading,... How to aggregate evaluations from crowd? Workers have different proficiencies; possibly unknown to system: Learn, weight to maximize accuracy Input to aggregation problem comes from self-interested agents How to incentivize good evaluations from crowd? Incentives in Crowdsourcing: A Game-theoretic Approach 20 / 26

Incentives in crowdsourced evaluation Incentivizing accurate evaluations, truthful reporting: (i) Unobservable ground truth (ii) Effort-dependent accuracy (Information elicitation with endogenous proficiency) Direct monitoring infeasible: Reward agreement Incentives in Crowdsourcing: A Game-theoretic Approach 21 / 26

Incentives in crowdsourced evaluation Incentivizing accurate evaluations, truthful reporting: (i) Unobservable ground truth (ii) Effort-dependent accuracy (Information elicitation with endogenous proficiency) Direct monitoring infeasible: Reward agreement Problem: Undesirable low-effort/second-guessing equilibria (e.g. always say H ) Incentives in Crowdsourcing: A Game-theoretic Approach 21 / 26

Incentives in crowdsourced evaluation Incentivizing accurate evaluations, truthful reporting: (i) Unobservable ground truth (ii) Effort-dependent accuracy (Information elicitation with endogenous proficiency) Direct monitoring infeasible: Reward agreement Problem: Undesirable low-effort/second-guessing equilibria (e.g. always say H ) Mechanism [Dasgupta-Ghosh, WWW 13]: Maximum effort-truthful reporting is highest-payoff equilibrium! (Assuming no task-specific collusions) Use multiple tasks and ratings: Reward for agreement, Incentives in Crowdsourcing: A Game-theoretic Approach 21 / 26

Incentives in crowdsourced evaluation Incentivizing accurate evaluations, truthful reporting: (i) Unobservable ground truth (ii) Effort-dependent accuracy (Information elicitation with endogenous proficiency) Direct monitoring infeasible: Reward agreement Problem: Undesirable low-effort/second-guessing equilibria (e.g. always say H ) Mechanism [Dasgupta-Ghosh, WWW 13]: Maximum effort-truthful reporting is highest-payoff equilibrium! (Assuming no task-specific collusions) Use multiple tasks and ratings: Reward for agreement, but also identify and penalize blind agreement Incentives in Crowdsourcing: A Game-theoretic Approach 21 / 26

Moving beyond single tasks: Incentivizing overall contribution Problems so far: Incentives for single contribution/task Incentives in Crowdsourcing: A Game-theoretic Approach 22 / 26

Moving beyond single tasks: Incentivizing overall contribution Problems so far: Incentives for single contribution/task Rewarding contributors for overall identity: Badges, leaderboards, reputations,... Virtual rewards for cumulative contribution Gamification rewards valued by agents; contribution to earn reward is costly Badges induce mechanisms! Design affects participation, effort from users Incentives in Crowdsourcing: A Game-theoretic Approach 22 / 26

Moving beyond single tasks: Incentivizing overall contribution Problems so far: Incentives for single contribution/task Rewarding contributors for overall identity: Badges, leaderboards, reputations,... Virtual rewards for cumulative contribution Gamification rewards valued by agents; contribution to earn reward is costly Badges induce mechanisms! Design affects participation, effort from users Incentives in Crowdsourcing: A Game-theoretic Approach 22 / 26

Badges and incentive design Different badge designs online: Absolute milestone badges (StackOverflow, Foursquare), versus competitive top-contributor badges (Y!Answers, Tripadvisor) Information about badge winners (StackOverflow vs Y! Answers) Incentives in Crowdsourcing: A Game-theoretic Approach 23 / 26

Badges and incentive design Different badge designs online: Absolute milestone badges (StackOverflow, Foursquare), versus competitive top-contributor badges (Y!Answers, Tripadvisor) Information about badge winners (StackOverflow vs Y! Answers) What incentives do different badge designs create for participation and effort? Results Game-theoretic analysis of badge design (Easley & Ghosh, ACM EC 13) Absolute or competitive badges? Competitive badges: Fixed number or fraction of participants? Visibility of information: Transparent or not? Incentives in Crowdsourcing: A Game-theoretic Approach 23 / 26

Wrapping up Incentive design for crowdsourcing Data, inputs to algorithms come from self-interested agents Mechanisms rather than algorithms: A game-theoretic framework Incentives in Crowdsourcing: A Game-theoretic Approach 24 / 26

Wrapping up Incentive design for crowdsourcing Data, inputs to algorithms come from self-interested agents Mechanisms rather than algorithms: A game-theoretic framework Lots more to understand! Sequential decision making Eliciting effort with strategic contributors and raters Incentives in Crowdsourcing: A Game-theoretic Approach 24 / 26

Wrapping up Incentive design for crowdsourcing Data, inputs to algorithms come from self-interested agents Mechanisms rather than algorithms: A game-theoretic framework Lots more to understand! Sequential decision making Eliciting effort with strategic contributors and raters Overall contributor rewards; sustaining contribution Learning and incentives: Designing reputations Incentives in Crowdsourcing: A Game-theoretic Approach 24 / 26

Wrapping up Incentive design for crowdsourcing Data, inputs to algorithms come from self-interested agents Mechanisms rather than algorithms: A game-theoretic framework Lots more to understand! Sequential decision making Eliciting effort with strategic contributors and raters Overall contributor rewards; sustaining contribution Learning and incentives: Designing reputations Experimental and empirical: What do agents value, and how? Incentives in Crowdsourcing: A Game-theoretic Approach 24 / 26

Why M UCB MOD works Lemma Any arm with quality q i q max (T ) δ receives Θ(ln T ) attention in expectation for all δ > 0 Theorem q max (T ): Highest-quality explored contribution A purely algorithmic statement; proof by contradiction For any fixed q < γ, the probability that there is some agent explored by M UCB MOD who chooses quality q > q goes to 1 as T in every equilibrium of M UCB MOD. Proof by contradiction: Demonstrate profitable deviation (Involves strategic reasoning, not purely algorithmic) Back to UCB Incentives in Crowdsourcing: A Game-theoretic Approach 25 / 26

Badges and incentive design: An economic framework (Easley & Ghosh, ACM EC 13) Conclusion Design recommendations from analysis: Competitive badges: Reward fixed number, not fraction of competitors Absolute versus competitive badges equivalent if population parameters known With uncertainty, or unknown parameters, competitive badges more robust Sharing information about other users performance: Depends on convexity of value as function of winners Incentives in Crowdsourcing: A Game-theoretic Approach 26 / 26