introduction to machine learning
play

Introduction to Machine Learning Multilayer Perceptron Barnabs - PowerPoint PPT Presentation

Introduction to Machine Learning Multilayer Perceptron Barnabs Pczos The Multilayer Perceptron 2 Multilayer Perceptron 3 ALVINN: AN AUTONOMOUS LAND VEHICLE IN A NEURAL NETWORK Dean A. Pomerleau, Carnegie Mellon University, 1989


  1. Introduction to Machine Learning Multilayer Perceptron Barnabás Póczos

  2. The Multilayer Perceptron 2

  3. Multilayer Perceptron 3

  4. ALVINN: AN AUTONOMOUS LAND VEHICLE IN A NEURAL NETWORK Dean A. Pomerleau, Carnegie Mellon University, 1989 Training: using simulated road generator 4

  5. Gradient Descent We want to solve: 5

  6. Starting Point 6

  7. Starting Point 7

  8. Fixed step size can be too big 8

  9. Fixed step size can be too small 9

  10. 10

  11. 11

  12. Character Recognition with MLP Matlab: appcr1 12

  13. The network Noise-free input: 26 different letters of size 7x5 13

  14. Noisy inputs 14

  15. Matlab MLP Training % Create MLP hiddenlayers=[10, 25]; net1 = feedforwardnet(hiddenlayers); net1 = configure(net1,X,T); %View view(net1); %Train net1 = train(net1,X,T); %Test Y1 = net1(Xtest); 15

  16. Prediction errors ▪ Network 1 was trained on clean images 16 ▪ Network 2 was trained on noisy images. 30 noisy copies of each letter are created

  17. The Backpropagation Algorithm 17

  18. Multilayer Perceptron 18

  19. The gradient of the error 19

  20. Notation 20

  21. Some observations 21

  22. The backpropagated error 22

  23. The backpropagated error Lemma 23

  24. The backpropagated error Therefore, 24

  25. The backpropagation algorithm 25

  26. 26

  27. 27

  28. 28

  29. 29

  30. 30

  31. 31

  32. 32

  33. What functions can multilayer perceptrons represent? 33

  34. Perceptrons cannot represent the XOR function f(0,0)=1, f(1,1)=1, f(0,1)=0, f(1,0)=0 What functions can multilayer perceptrons represent? 34

  35. Hilbert’s 13 th Problem 1902: 23 “most important” problems in mathematics The 13 th Problem: “Solve 7-th degree equation using continuous functions of two parameters .” Conjecture : It can’t be solved… Related conjecture: Let f be a function of 3 arguments such that Prove that f cannot be rewritten as a composition of finitely many functions of two arguments. Another rewritten form: Prove that there is a nonlinear continuous system of three variables that cannot be decomposed with finitely many functions of two variables. 35

  36. Function decompositions f(x,y,z) =Φ 1 (ψ 1 (x), ψ 2 (y ))+Φ 2 (c 1 ψ 3 (y)+c 2 ψ 4 (z),x) ψ 1 Φ 1 x ψ 2 f(x,y,z) Σ y c 1 ψ 3 Φ 2 Σ z c 2 ψ 4 36

  37. Function decompositions 1957, Arnold disproves Hilbert ’s conjecture. 37

  38. Function decompositions Corollary: Issues: This statement is not constructive. 38

  39. Universal Approximators Kur Hornik, Maxwell Stinchcombe and Halber White : “Multilayer feedforward networks are universal approximators ”, Neural Networks, Vol:2(3), 359-366, 1989 Definition: Σ N (g) neural network with 1 hidden layer: Definition: Theorem: 39

  40. Universal Approximators Definition: Theorem: ( Blum & Li, 1991) Formal statement: 40

  41. Proof GOAL: Integral approximation in 1-dim: Integral approximation in 2-dim: x i x i x i          i i j 41 i

  42. Proof GOAL: x i x i x i The indicator function of X i polygon can be learned by this neural network: 1 if x is in X i -1 otherwise The weighted linear combination of these indicator functions will be a good approximation of the original function f 42

  43. Proof This linear equation can also be solved. 43

  44. Thanks for your attention! 44

Recommend


More recommend