PRanking with Ranking Koby Crammer Technion – Israel Institute of Technology Based on joint work with Yoram Singer at the Hebrew University of Jerusalem
Problem Machine Prediction User’s Rating Ranking 3 3 0 3 1 Loss
Ranking – Formal Description Problem Setting Online Framework • Instances • Algorithm works in rounds • On each round the ranking • Labels algorithm : – Gets an input instance – Outputs a rank as prediction • Structure 1 2 3 4 5 – Receives the correct rank- value • Ranking rule – Computes loss – Updates the rank-prediction x 1 is preferred over x 2 • Ranking Loss rule
Goal t L t = y i − ˆ y ∑ • Algorithms Loss i i = 1 t L t f y i − f x i ∑ • Loss of a fixed function ( ) = ( ) i = 1 L t − inf f ∈ F L t f • Regret ( ) • No statistical assumptions over data • The algorithm should do well irrespectively of specific sequence of inputs and target labels
Background Binary Classification
The Perceptron Algorithm Rosenblatt, 1958 Hyperplane w • w 1 2
The Perceptron Algorithm Rosenblatt, 1958 Hyperplane w • Get new instance x • Classify x : w • sign( ) (X,1) 1 2
The Perceptron Algorithm Rosenblatt, 1958 Hyperplane w • w Get new instance x • Classify x : w • sign( ) • Update (in case of a mistake) 1 (X,1) 2 1 2
A Function Class for Ranking
Our Approach to Ranking • Project
Our Approach to Ranking • Project • Apply Thresholds < > 4 3 2 1 Rank
Update of a Specific Algorithm it if its not Broken Least change as possible One step at a time
PRank Direction w , • Thresholds w Thresholds 1 2 3 4 5 Rank Levels
PRank Direction w , • Thresholds Rank a new instance x • w 1 2 3 4 5
PRank Direction w , Correct Rank • Thresholds Interval Rank a new instance x • Get the correct rank y • w 1 2 3 4 5
PRank Direction w , • Thresholds Rank a new instance x • Get the correct rank y • Compute Error-Set E • w 1 2 3 4 5
PRank – Update Direction w , • Thresholds Rank a new instance x • Get the correct rank y • Compute Error-Set E • w • Update : –
PRank – Update w Direction w , • Thresholds x Rank a new instance x • Get the correct rank y • x Compute Error-Set E • • Update : w – –
PRank – Summary of Update w Direction w , • Thresholds x Rank a new instance x • x Get the correct rank y • x Compute Error-Set E • • Update : w – –
The PRank Algorithm Get an instance x Maintain Predict : Update Get the true rank y No ? Compute Error set : Yes
Analysis Two Lemmas
Consistency • Can the following happen? w b 4 b 2 b 3 b 2 b 1
Consistency • Can the following happen? No w b b b b b 4 2 3 2 1 • The order of the thresholds is preserved after each round of PRank : .
Regret Bound Given : • Arbitrary input sequence Easy Case: • Assume there exists a model that ranks all the input instances correctly – The total loss the algorithm suffers is bounded Hard Case: • In general L t − inf L ˜ – A “regret” is bounded ( ) t f f ∈ F
Ranking Margin Margin( x ,y ) = min , w 5 3 2 1 4
Ranking Margin Margin( x ,y ) = min , w 5 3 2 1 4
Ranking Margin Margin( x ,y ) = min , w 5 3 2 1 4
Ranking Margin Margin( x ,y ) = min , Margin = min Margin w 5 3 2 1 4
Mistake Bound Given : • Input sequence , • Norm of instances is bounded • Ranked correctly by a normalized ranker with Margin>0 Then : Number of Mistakes PRank Makes
Exploit Structure Loss Range Structure Classification None Under Constraint Regression Metric Over Constraint Ranking Order
Other Approaches • Treat Ranking as Classification or Regression E.g. Basu, Hirsh, Cohen 1998 • Reduce a ranking problem into a classification problem over pair of examples E.g. Freund, Lyer, Schapire, Singer 1998 Herbrich, Graepel, Obermayer 2000 – Not simple to combine preferences predictions over pairs into a singe consistent ordering – No simple adaptation for online settings
Empirical Study
An Illustration • Five concentric ellipses • Training set of 50 points • Three approaches • Pranking • Classification • Regression PRank MC-Perceptron Widrow-Hoff Ranking Classification Regression
Each-Movie database • 74424 registered Users • 1648 listed Movies • Users ranking of movies • 7451 Users saw >100 movies • 1801 Users saw >200 movies
Ranking Loss, 100 Viewers WH Regression MC-Perceptron Classification PRank PRank Over constrained Rank Loss Under constrained Accurately constrained Round
Ranking Loss, 200 Viewers WH Regression MC-Perceptron Classification PRank PRank Rank Loss Round
Demonstration
(2) Movies chosen and ranked by user (1) User choose movies from this list
(3) Press the ‘learn’ key. (5) The system re-ranks a (4) The system re-ranks The systems learns the new fresh set of yet the training set user’s taste unseen movies
(6) Press the ‘flip’ button (7) The flipped list to see what movies you should not view
• Many alternatives to formulate ranking • Choose one that models best your problem • Exploit and Incorporate structure • Specifically: – Online algorithm for ranking problems via projections and conservative update of the projection’s direction and the threshold values – Experiments on a synthetic dataset and on Each- Movie data set indicate that the PRank algorithm performs better then algorithms for classification and regression
Recommend
More recommend