lecture 20 regression
play

Lecture 20: Regression Dr. Chengjiang Long Computer Vision - PowerPoint PPT Presentation

Lecture 20: Regression Dr. Chengjiang Long Computer Vision Researcher at Kitware Inc. Adjunct Professor at RPI. Email: longc3@rpi.edu Recap Previous Lecture 2 C. Long Lecture 20 April 16, 2018 Outline Regression Overview Linear


  1. Lecture 20: Regression Dr. Chengjiang Long Computer Vision Researcher at Kitware Inc. Adjunct Professor at RPI. Email: longc3@rpi.edu

  2. Recap Previous Lecture 2 C. Long Lecture 20 April 16, 2018

  3. Outline Regression Overview • Linear Regression • Support Vector Regression • Logistic Regression • Deep Neural Network for Regression • 3 C. Long Lecture 20 April 16, 2018

  4. Outline Regression Overview • Linear Regression • Support Vector Regression • Logistic Regression • Deep Neural Network for Regression • 4 C. Long Lecture 20 April 16, 2018

  5. Regression Overview Hierarchy clustering • Logistic Regression Gaussian Mixture Model • Neural Networks 5 C. Long Lecture 20 April 16, 2018

  6. One Example ...... 6 C. Long Lecture 20 April 16, 2018

  7. Evaluation Metrics Root mean-square error (RMSE) • RMSE is a popular formula to measure the error rate of a • regression model. However, it can only be compared between models whose errors are measured in the same units. Mean absolute error (MAE) • MAE has the same unit as the original data, and it can only be • compared between models whose errors are measured in the same units. It is usually similar in magnitude to RMSE, but slightly smaller. 7 C. Long Lecture 20 April 16, 2018

  8. Evaluation Metrics Relative Squared Error (RSE) • Unlike RMSE, the relative squared error (RSE) can be compared • between models whose errors are measured in the different units. mean value Relative Absolute Error (RAE) • Like RSE , the relative absolute error (RAE) can be compared • between models whose errors are measured in the different units.. 8 C. Long Lecture 20 April 16, 2018

  9. Outline Regression Overview • Linear Regression • Support Vector Regression • Logistic Regression • Deep Neural Network for Regression • 9 C. Long Lecture 20 April 16, 2018

  10. Linear Regression Given data with n dimensional variables and 1 target- • variable (real number) where The objective: Find a function f that returns the best fit. • Assume that the relationship between X and y is • approximately linear. The model can be represented as (w represents coefficients and b is an intercept) 10 C. Long Lecture 20 April 16, 2018

  11. Linear Regression To find the best fit , we minimize the sum of squared • errors -> Least square estimation The solution can be found by solving • (By taking the derivative of the above objective function w.r.t. ) In MATLAB, the back-slash operator computes a least • square solution. 11 C. Long Lecture 20 April 16, 2018

  12. Linear Regression To avoid over-fitting, a regularization term can be introduced (minimize a magnitude of w) 12 C. Long Lecture 20 April 16, 2018

  13. Outline Regression Overview • Linear Regression • Support Vector Regression • Logistic Regression • Deep Neural Network for Regression • 13 C. Long Lecture 20 April 16, 2018

  14. Support Vector Regression Find a function , f ( x ), with at most ε- deviation from • the target y 14 C. Long Lecture 20 April 16, 2018

  15. Support Vector Regression 15 C. Long Lecture 20 April 16, 2018

  16. Soft margin 16 C. Long Lecture 20 April 16, 2018

  17. How about a non-linear case? 17 C. Long Lecture 20 April 16, 2018

  18. Linear versus Non-linear SVR 18 C. Long Lecture 20 April 16, 2018

  19. Dual problem 19 C. Long Lecture 20 April 16, 2018

  20. Kernel trick 20 C. Long Lecture 20 April 16, 2018

  21. Dual problem for non-linear case 21 C. Long Lecture 20 April 16, 2018

  22. Architecture of a regression machine [Alex J. Smola et al. A tutorial on support vector regression, 2004.] URL: https://alex.smola.org/papers/2004/SmoSch04.pdf 22 C. Long Lecture 20 April 16, 2018

  23. Outline Regression Overview • Linear Regression • Support Vector Regression • Logistic Regression • Deep Neural Network for Regression • 23 C. Long Lecture 20 April 16, 2018

  24. Logistic Regression Takes a probabilistic approach to learning • discriminative functions (i.e., a classifier) Logistic regression model: • 24 C. Long Lecture 20 April 16, 2018

  25. Interpretation of Hypothesis Output 25 C. Long Lecture 20 April 16, 2018

  26. Another Interpretation Equivalently , logistic regression assumes that • In other words , logistic regression assumes that the log • odds is a linear function of x 26 C. Long Lecture 20 April 16, 2018

  27. Logistic Regression Assume a threshold and ... • 27 C. Long Lecture 20 April 16, 2018

  28. Non-Linear DecisionBoundary Can apply basis function expansion to features , same • as with linear regression ? 28 C. Long Lecture 20 April 16, 2018

  29. Logistic Regression 29 C. Long Lecture 20 April 16, 2018

  30. Logistic Regression Objective Function Can’t just use squared loss as in linear regression • – Using the logistic regression model results in a non-convex optimization 30 C. Long Lecture 20 April 16, 2018

  31. Deriving the Cost Function via Maximum Likelihood Estimation 31 C. Long Lecture 20 April 16, 2018

  32. Deriving the Cost Function via Maximum Likelihood Estimation 32 C. Long Lecture 20 April 16, 2018

  33. Intuition Behind the Objective 33 C. Long Lecture 20 April 16, 2018

  34. Intuition Behind the Objective 34 C. Long Lecture 20 April 16, 2018

  35. Intuition Behind the Objective 35 C. Long Lecture 20 April 16, 2018

  36. Regularized Logistic Regression We can regularize logistic regression exactly as before • 36 C. Long Lecture 20 April 16, 2018

  37. Gradient Descent for Logistic Regression 37 C. Long Lecture 20 April 16, 2018

  38. Gradient Descent for Logistic Regression 38 C. Long Lecture 20 April 16, 2018

  39. Gradient Descent for Logistic Regression 39 C. Long Lecture 20 April 16, 2018

  40. Multi-Class Classification 40 C. Long Lecture 20 April 16, 2018

  41. Multi-Class Logistic Regression 41 C. Long Lecture 20 April 16, 2018

  42. Multi-Class Logistic Regression Split into One vs Rest : • Train a logistic regression classifier for each class • i to predict the probability that y = i with 42 C. Long Lecture 20 April 16, 2018

  43. Implementing Multi-Class Logistic Regression 43 C. Long Lecture 20 April 16, 2018

  44. Outline Regression Overview • Linear Regression • Support Vector Regression • Logistic Regression • Deep Neural Network for Regression • 44 C. Long Lecture 20 April 16, 2018

  45. DNN Regression For a two-layer MLP: • The network weights are adjusted to minimize an • output cost function 45 C. Long Lecture 20 April 16, 2018

  46. Computing the Partial Derivatives for Regression We use SSE and for a two layer network the linear final outputs • can be written : We can then use the chain rules for derivatives , as for the Single • Layer Perceptron , to give the derivatives with respect to the two sets of weights 46 C. Long Lecture 20 April 16, 2018

  47. Deriving the Back Propagation Algorithm for Regression All we now have to do is substitute our derivatives into the weight • update equations Then if the transfer function f(x) is a Sigmoid we can use f′(x) = • f(x).(1 – f(x)) to give These equations constitute the Back-Propagation Learning • Algorithm for Regression. 47 C. Long Lecture 20 April 16, 2018

  48. Classification + Localization: Task 48 C. Long Lecture 20 April 16, 2018

  49. Idea #1: Localization as Regression 49 C. Long Lecture 20 April 16, 2018

  50. Simple Recipe for Classification + Localization Step 1: Train ( or download ) a classification model • ( AlexNet , VGG , GoogLeNet ) 50 C. Long Lecture 20 April 16, 2018

  51. Simple Recipe for Classification + Localization Step 2: Attach new fully-connected “regression head” • to the network 51 C. Long Lecture 20 April 16, 2018

  52. Simple Recipe for Classification + Localization Step 3: Train the regression head only with SGD and • L2 loss 52 C. Long Lecture 20 April 16, 2018

  53. Simple Recipe for Classification + Localization Step 4: At test time use both heads • 53 C. Long Lecture 20 April 16, 2018

  54. Per-class vs class agnostic regression 54 C. Long Lecture 20 April 16, 2018

  55. Where to attach the regression head? 55 C. Long Lecture 20 April 16, 2018

  56. Aside: Localizing multiple objects 56 C. Long Lecture 20 April 16, 2018

  57. Aside: Human Pose Estimation 57 C. Long Lecture 20 April 16, 2018

  58. DNN Regression Applications Great results in: • Computer Vision • ○ Object Localization / Detection as DNN Regression ○ Self-driving Steering Command Prediction ○ Human Pose Regression Finance • ○ Currency Exchange Rate ○ Stock Price Prediction ○ Forecasting Financial Time Series ○ Crude Oil Price Prediction 58 C. Long Lecture 20 April 16, 2018

Recommend


More recommend