Peer Prediction Mechanisms and their Connections to Machine Learning Jens Witkowski ∗ ETH Zurich Game Theory Meets Computational Learning Theory Dagstuhl June 22, 2017 ∗ Joint work with David C. Parkes (Harvard) and Rafael Frongillo (Boulder).
Forecasting Peer Prediction Connections to ML Conclusion Proper Scoring Rules Forecaster with belief p ∈ [ 0 , 1 ] reports belief y ∈ [ 0 , 1 ] . 1 � 1 , if event occurs March 1: pay R ( y , ω ) , where ω = 2 if not. 0 , If R ( y , ω ) proper, E p [ R ( y , ω )] maximized by reporting y = p . Example: Quadratic Scoring Rule [Brier, 1950] R q ( y , ω ) = 1 − ( y − ω ) 2 1 / 14
Forecasting Peer Prediction Connections to ML Conclusion “Effective” Scoring Rules What if allowed reports y are restricted? True belief: p = 0 . 7 . Allowed reports: y 1 = 0 . 6 or y 2 = 0 . 79 . p y 1 = 0 . 6 y 2 = 0 . 79 0 1 Quadratic Rule: report y “closest” to p ! Theorem [Friedman, 1983] The Quadratic Rule is effective : if Y ⊆ [ 0 , 1 ] , reporting min y ∈ Y | y − p | maximizes expected score. 2 / 14
Forecasting Peer Prediction Connections to ML Conclusion Motivation Objective: Truthful elicitation of opinions or experiences. 3 / 14
Forecasting Peer Prediction Connections to ML Conclusion Peer Prediction Elicit informative signal (e.g. “high” or “low” experience). Ground truth never observed. Pay agent given her report x i and report of other agent x j . Objective: truthful reporting of signal is a Bayes NE. Key: Agent i ’s signal informative about j ’s signal. S j S i S j = s j S i = s i j j i i 4 / 14
Forecasting Peer Prediction Connections to ML Conclusion Belief Model h 0 . 1 0 . 6 S j = h good 0 . 7 bad 0 . 3 j 0 . 9 0 . 4 l i p ( h ) = 0 . 25 p ( h | h ) = 0 . 46 Belief that S j = h : p ( h | l ) = 0 . 18 5 / 14
Forecasting Peer Prediction Connections to ML Conclusion Simplest Mechanism: Output Agreement � $ 2 if reports agree, Compare two agents’ reports and pay: $ 0 otherwise. Example with S i = h S j = l S j = h = 0 . 54 j = 0 . 46 j i i � � � � payment | x i = h = $ 0 . 92 payment | x i = l = $ 1 . 08 E E Output Agreement not truthful for this belief model! 6 / 14
Forecasting Peer Prediction Connections to ML Conclusion Classical Peer Prediction [Miller et al., 2005] p ( h | l )= 0 . 18 p ( h | h )= 0 . 46 x j x i = h � � R 0 . 46 , x j j i Truthful if p ( h | h ) � = p ( h | l ) ! Intuition Define agent j ’s signal report as event. 1 Restrict possible belief reports to possible posteriors. 2 Crucial: mechanism knows belief model! 7 / 14
Forecasting Peer Prediction Connections to ML Conclusion Shadowing Method [W. and Parkes, 2012] � � Assumption: only y ∈ p ( h | l ) , p ( h | h ) known. δ δ x j x i = h 0 1 y 0 1 y j i � � y + δ, x j R q Truthful: agent prefers “shadow posterior” closer to true belief! Crucial: quadratic scoring rule R q ! 8 / 14
Forecasting Peer Prediction Connections to ML Conclusion Shadowing Method: Key Idea Challenge: no knowledge of posteriors p ( h | h ) or p ( h | l ) ! Compute “Shadow posteriors” y + δ and y − δ : δ δ y p ( h | l ) p ( h | h ) 0 1 Observe: S i = h ⇒ y + δ closer to p ( h | h ) than y − δ . S i = l ⇒ y − δ closer to p ( h | l ) than y + δ . 9 / 14
Forecasting Peer Prediction Connections to ML Conclusion 1. Learning Mechanisms from Reports Can y be learned from data? Idea Sketch � � Known: p ( h ) ∈ p ( h | l ) , p ( h | h ) . Consider other questions/tasks with same prior. Empirical frequency of x = h reports on those will be close to p ( h ) with high probability. Main Result [W. and Parkes, 2013] The Shadowing Method with y = ˆ p i ( h ) is strictly truthful given enough samples where this number depends on a lower bound on the belief change from prior to posterior. 10 / 14
Forecasting Peer Prediction Connections to ML Conclusion 2. Peer Prediction Mechanisms are Loss Functions Classification Loss Peer Prediction Mechanism y i x j + 1 − 1 h l + 1 0 1 h 1 0 sign ( w T x i ) x i − 1 1 0 l 0 1 0/1 Loss Output Agreement y i x j + 1 − 1 h l + 1 0 2 h 2 0 sign ( w T x i ) x i − 1 1 0 l 1 2 Cost-sensitive Loss Shadowing w/ y = 2 3 11 / 14
Forecasting Peer Prediction Connections to ML Conclusion Scoring Rules for Properties Theorem [Frongillo and W., 2017] Peer Prediction Mechanisms are equivalent to scoring rules for properties of probability distributions. Example: Output Agreement c c a a b b arg max x i Pr ( S j = x i | S i = s i ) arg max y i Pr ( Y i = y i | x i , w ) Output Agreement elicits the mode. 12 / 14
Forecasting Peer Prediction Connections to ML Conclusion Conclusion Peer Prediction mechanisms truthfully elicit private opinions or experiences. Connections between PP mechanisms and ML: Learn truthful mechanisms using reports on other items. 1 Mechanisms are equivalent to loss functions eliciting 2 property of conditional label probability. THANK YOU! 13 / 14
References I Brier, G. W. (1950). Verification of Forecasts Expressed in Terms of Probability. Monthly Weather Review, 78(1):1–3. Friedman, D. (1983). Effective Scoring Rules for Probabilistic Forecasts. Management Science, 29(4):447–454. Frongillo, R. and Witkowski, J. (2017). A geometric perspective on minimal peer prediction. ACM Transactions on Economics and Computation (TEAC). Forthcoming. 13 / 14
References II Miller, N., Resnick, P ., and Zeckhauser, R. (2005). Eliciting Informative Feedback: The Peer-Prediction Method. Management Science, 51(9):1359–1373. Witkowski, J. (2014). Robust Peer Prediction Mechanisms. PhD thesis, Department of Computer Science, Albert-Ludwigs-Universität Freiburg. Witkowski, J. and Parkes, D. C. (2012). A Robust Bayesian Truth Serum for Small Populations. In Proceedings of the 26th AAAI Conference on Artificial Intelligence (AAAI’12), pages 1492–1498. 14 / 14
Recommend
More recommend