for optimizing the partial auc
play

for Optimizing the Partial AUC Harikrishna Narasimhan (Joint work - PowerPoint PPT Presentation

A Structural SVM Based Approach for Optimizing the Partial AUC Harikrishna Narasimhan (Joint work with Shivani Agarwal) A paper on this work has been accepted in ICML 2013 Learning with Binary Supervision Learning with Binary Supervision


  1. A Structural SVM Based Approach for Optimizing the Partial AUC Harikrishna Narasimhan (Joint work with Shivani Agarwal) A paper on this work has been accepted in ICML 2013

  2. Learning with Binary Supervision

  3. Learning with Binary Supervision

  4. Learning with Binary Supervision http://www.google.com/imghp

  5. Learning with Binary Supervision http://www.google.com/imghp

  6. Learning with Binary Supervision http://www.google.com/imghp

  7. Learning with Binary Supervision Good evaluation metric? http://www.google.com/imghp

  8. Learning with Binary Supervision Good evaluation metric? http://www.google.com/imghp

  9. Learning with Binary Supervision …….. x 1 + x 2 + x 3 + x m + Positive Instances Training …….. Set x 1 - x 2 - x 3 - x n - Negative Instances

  10. Learning with Binary Supervision …….. x 1 + x 2 + x 3 + x m + Positive Instances Training …….. Set x 1 - x 2 - x 3 - x n - Negative Instances GOAL? Learn a scoring function

  11. Learning with Binary Supervision …….. x 1 + x 2 + x 3 + x m + Positive Instances Training …….. Set x 1 - x 2 - x 3 - x n - Negative Instances GOAL? Learn a scoring function Rank objects x 5 + x 3 + x 1 - x 6 + …. x n -

  12. Learning with Binary Supervision …….. x 1 + x 2 + x 3 + x m + Positive Instances Training …….. Set x 1 - x 2 - x 3 - x n - Negative Instances GOAL? Learn a scoring function Rank objects Build a classifier x 5 + x 5 + x 3 + x 3 + Threshold or x 1 - x 1 - x 6 + x 6 + …. …. x n - x n -

  13. Learning with Binary Supervision …….. x 1 + x 2 + x 3 + x m + Positive Instances Training …….. Set x 1 - x 2 - x 3 - x n - Negative Instances GOAL? Learn a scoring function Rank objects Build a classifier Quality of score function? x 5 + x 5 + x 3 + x 3 + Threshold or x 1 - x 1 - x 6 + x 6 + …. …. x n - x n -

  14. Learning with Binary Supervision …….. x 1 + x 2 + x 3 + x m + Positive Instances Training …….. Set x 1 - x 2 - x 3 - x n - Negative Instances GOAL? Learn a scoring function Rank objects Build a classifier Quality of score function? x 5 + x 5 + x 3 + x 3 + Threshold or x 1 - x 1 - Threshold Assignment x 6 + x 6 + …. …. x n - x n -

  15. Receiver Operating Characteristic Curve Captures how well a prediction model discriminates between positive and negative examples

  16. Receiver Operating Characteristic Curve Captures how well a prediction model discriminates between positive and negative examples Full AUC

  17. Receiver Operating Characteristic Curve Captures how well a prediction model discriminates between positive and negative examples Vs Full AUC Partial AUC

  18. Ranking http://www.google.com/

  19. Ranking http://www.google.com/

  20. Medical Diagnosis http://www.google.com/imghp

  21. Medical Diagnosis http://www.google.com/imghp

  22. Bioinformatics ― Drug Discovery ― Gene Prioritization ― Protein Interaction Prediction ― …… http://www.google.com/imghp

  23. Bioinformatics ― Drug Discovery ― Gene Prioritization ― Protein Interaction Prediction ― …… http://www.google.com/imghp

  24. Partial Area Under the ROC Curve is critical to many applications

  25. Partial AUC Optimization • Many existing approaches are either heuristic or solve special cases of the problem. Partial Area Under the ROC Curve is critical to many applications

  26. Partial AUC Optimization • Many existing approaches are either heuristic or solve special cases of the problem. • Our contribution : A new support vector method for optimizing the general partial AUC measure. Partial Area Under the ROC Curve is critical to many applications

  27. Partial AUC Optimization • Many existing approaches are either heuristic or solve special cases of the problem. • Our contribution : A new support vector method for optimizing the general partial AUC measure. • Based on Joachims’ Structural SVM approach for optimizing full AUC, but leads to a trickier inner combinatorial optimization problem. Partial Area Under the ROC Curve is critical to many applications

  28. Partial AUC Optimization • Many existing approaches are either heuristic or solve special cases of the problem. • Our contribution : A new support vector method for optimizing the general partial AUC measure. • Based on Joachims’ Structural SVM approach for optimizing full AUC, but leads to a trickier inner combinatorial optimization problem. • Improvements over baselines on several real-world applications Partial Area Under the ROC Curve is critical to many applications

  29. ROC Curve Receiver Operating Characteristic Curve 20 15 Scores 14 assigned 13 by f 11 9 8 6 5 3 2 0

  30. ROC Curve Receiver Operating Characteristic Curve 20 15 14 13 11 9 8 6 5 3 2 0

  31. ROC Curve Receiver Operating Characteristic Curve 20 15 14 13 11 9 8 6 5 3 2 0

  32. ROC Curve Receiver Operating Characteristic Curve 20 15 14 13 11 9 8 6 5 3 2 0

  33. Partial AUC Optimization

  34. Partial AUC Optimization Minimize:

  35. Partial AUC Optimization Discrete and Minimize: Non-differentiable

  36. Partial AUC Optimization Discrete and Minimize: Non-differentiable Convex Upper Bound on “ ”

  37. Partial AUC Optimization Discrete and Minimize: Non-differentiable Convex Upper Bound on “ ” + Regularizer

  38. Partial AUC Optimization Discrete and Minimize: Non-differentiable Convex Upper Bound on “ ” + Regularizer Structural SVM

  39. Partial AUC Optimization Discrete and Minimize: Non-differentiable Convex Upper Bound on “ ” + Regularizer Structural SVM • Extends Joachims’ approach for full AUC optimization, but leads to a trickier combinatorial optimization step. T. Joachims, “A Support Vector Method for Multivariate Performance Measures”, ICML 2005.

  40. Partial AUC Optimization Discrete and Minimize: Non-differentiable Convex Upper Bound on “ ” + Regularizer Structural SVM • Extends Joachims’ approach for full AUC optimization, but leads to a trickier combinatorial optimization step. • Efficient solver with the same time complexity as that for full AUC. T. Joachims, “A Support Vector Method for Multivariate Performance Measures”, ICML 2005.

  41. Structural SVM Based Approach

  42. Structural SVM Based Approach Ordering of {x 1 , x 2 , …, x s } n +1 +1 +1 +1 +1 -1 -1 +1 +1 +1 m -1 -1 +1 +1 -1 -1 -1 +1 +1 -1

  43. Structural SVM Based Approach Ordering of {x 1 , x 2 , …, x s } n +1 +1 +1 +1 +1 +1 +1 +1 +1 +1 compared -1 -1 +1 +1 +1 +1 +1 +1 +1 +1 IDEAL m with -1 -1 +1 +1 -1 +1 +1 +1 +1 +1 -1 -1 +1 +1 -1 +1 +1 +1 +1 +1

  44. Structural SVM Based Approach Ordering of {x 1 , x 2 , …, x s } n +1 +1 +1 +1 +1 +1 +1 +1 +1 +1 compared -1 -1 +1 +1 +1 +1 +1 +1 +1 +1 IDEAL m with -1 -1 +1 +1 -1 +1 +1 +1 +1 +1 -1 -1 +1 +1 -1 +1 +1 +1 +1 +1

  45. Structural SVM Based Approach Ordering of {x 1 , x 2 , …, x s } n +1 +1 +1 +1 +1 +1 +1 +1 +1 +1 compared -1 -1 +1 +1 +1 +1 +1 +1 +1 +1 IDEAL m with -1 -1 +1 +1 -1 +1 +1 +1 +1 +1 -1 -1 +1 +1 -1 +1 +1 +1 +1 +1 Upper Bound on (1 – pAUC)

  46. Structural SVM Based Approach Ordering of {x 1 , x 2 , …, x s } n +1 +1 +1 +1 +1 +1 +1 +1 +1 +1 compared -1 -1 +1 +1 +1 +1 +1 +1 +1 +1 IDEAL m with -1 -1 +1 +1 -1 +1 +1 +1 +1 +1 -1 -1 +1 +1 -1 +1 +1 +1 +1 +1 Upper Bound on (1 – pAUC) Regularizer

  47. Structural SVM Based Approach Ordering of {x 1 , x 2 , …, x s } n +1 +1 +1 +1 +1 +1 +1 +1 +1 +1 compared -1 -1 +1 +1 +1 +1 +1 +1 +1 +1 IDEAL m with -1 -1 +1 +1 -1 +1 +1 +1 +1 +1 -1 -1 +1 +1 -1 +1 +1 +1 +1 +1 Upper Bound on (1 – pAUC) Regularizer pAUC Loss

  48. Structural SVM Based Approach Ordering of {x 1 , x 2 , …, x s } n +1 +1 +1 +1 +1 +1 +1 +1 +1 +1 compared -1 -1 +1 +1 +1 +1 +1 +1 +1 +1 IDEAL m with -1 -1 +1 +1 -1 +1 +1 +1 +1 +1 -1 -1 +1 +1 -1 +1 +1 +1 +1 +1 Upper Bound on (1 – pAUC) Regularizer Exponential Number of Output pAUC Loss Matrices!!

  49. Optimization Solver

  50. Optimization Solver Repeat: 1. Solve OP for a subset of constraints.

  51. Optimization Solver Repeat: 1. Solve OP for a subset of constraints. 2. Add the most violated constraint.

  52. Converges in Optimization Solver constant number of iterations Repeat: 1. Solve OP for a subset of constraints. 2. Add the most violated constraint. T. Joachims, “Training linear SVMs in linear time”, KDD 2006.

  53. Converges in Optimization Solver constant number of iterations Repeat: 1. Solve OP for a subset of constraints. 2. Add the most violated constraint. T. Joachims, “Training linear SVMs in linear time”, KDD 2006.

Recommend


More recommend