deciphering the face deciphering the face
play

Deciphering the Face Deciphering the Face Aleix M. Martinez - PowerPoint PPT Presentation

Deciphering the Face Deciphering the Face Aleix M. Martinez Computational Biology Computational Biology and Cognitive Science Lab aleix@ece.osu.edu l i @ d Human-Computer Interaction Politics Human Human face face face face Art


  1. Deciphering the Face Deciphering the Face Aleix M. Martinez Computational Biology Computational Biology and Cognitive Science Lab aleix@ece.osu.edu l i @ d

  2. Human-Computer Interaction Politics Human Human face face face face Art Sign Language Language Cognitive Cognitive Computer Vision Science

  3. Models of Face Perception • Features: Shape vs. texture. … … • 2D vs. 3D • Form of the computational space: p p Continuous vs. Categorical

  4. What we are going to show • What is the form of the computational space in human face perception? Hybrid approach : in human face perception? Hybrid approach : Linear combination of continuous representations of categories representations of categories. c 1 + c 2 + … + c n 1 2 n • What are the dimensions? Mostly configural . • In computer vision we need precise detailed detection of faces and facial features. detect o o aces a d ac a eatu es.

  5. Same or different? Identity

  6. Same or different? Identity

  7. Same or different? Identity, expression, gender, etc. Identity

  8. Dimensions of the Face Space Same or different? Configural processing

  9. Form of the Computational Face Space Computational Face Space Exemplar-based model Exemplar based model Exemplar … cells cells Norm-based model Mid-level vision Low-level vision

  10. Facial Expressions of Emotion

  11. Muscle Positions Model

  12. Muscle Positions Model • Global shape (bone structure) determines identity – configural. y g • But ONLY muscles are responsible for expression interaction for expression, interaction …

  13. Angry Emotion perception in Configural Processing emotionless faces f l Sad Neth & Martinez, JOV, 2009. Neutral Neutral

  14. 100% Stimuli 75% Neth & Martinez, JOV, 2009. 50% 25%

  15. Less, same, more. Experiment Experiment

  16. Configural Processing Sad 90 * * * 80 * * * * 70 60 Less 50 * Same * * 40 * More 30 * * * * 20 * 10 0 -100% -75% -50% -25% 0% 25% 50% 75% 100% Neth & Martinez, JOV, 2009.

  17. Configural Processing Angry 90 * * * 80 * * 70 * * 60 Less 50 Same * * 40 More * * 30 * * * 20 * * * 10 0 -100% -75% -50% -25% 0% 25% 50% 75% 100% Neth & Martinez, JOV, 2009.

  18. Norm-based Face Space Sadness 100% Multidimensional 75% 75% Face Space S 50% - density + density + density 25% Easier MEAN - density + density density 100% More difficult Anger Neth & Martinez, JOV, 2009.

  19. Configural Processing Neth & Martinez, JOV, 2009.

  20. Computational Space Neth & Martinez, Vision Research, 2010

  21. Thinner face Thinner face Computational Space Neth & Martinez, Vision Research, 2010 Wider face

  22. American Gothic Illusion Neth & Martinez, Vision Research, 2010

  23. Why Configural Features?

  24. 15 x 10 pixels

  25. Why Configural cues? sad neutral angry Neth & Martinez, Vision Research, 2010; Du & Martinez, 2011

  26. Proposed Hybrid Model: Recognizing other emotion labels Recognizing other emotion labels c 1 c + c 2 + c + + … + c n + c Happily Angrily g y surprised surprised Martinez, CVPR, 2011

  27. Configural Processing = Precise detection of facial features detection of facial features 4 2 pixels 4.2 pixels 3,930 error images (1.5%) (1.5%) Ding & Martinez, PAMI, 2010

  28. Face Detection

  29. Features VS context Observation : Most detections are near the correct location – they are not incorrect, they are imprecise . location they are not incorrect, they are imprecise . Key idea : Use context information to train where not t d t to detect faces and facial features. t f d f i l f t Ding & Martinez, CVPR, 2008; PAMI, 2010

  30. Features VS context Observation : Most detections are near the correct location – they are not incorrect, they are imprecise . location they are not incorrect, they are imprecise . Key idea : Use context information to train where not t d t to detect faces and facial features. t f d f i l f t Ding & Martinez, CVPR, 2008; PAMI, 2010

  31. Features VS context Observation : Most detections are near the correct location – they are not incorrect, they are imprecise . location they are not incorrect, they are imprecise . Key idea : Use context information to train where not t d t to detect faces and facial features. t f d f i l f t Ding & Martinez, CVPR, 2008; PAMI, 2010

  32. Subclass Discriminant Analysis y Between subclass Between-subclass scatter matrix: ( ( ) ( ) ( ) ) H C ∑∑ ∑∑ i = μ − μ μ − μ T T Σ Σ p . B ij ij ij = = i 1 j 1 Basis vectors: Basis vectors: = Λ Σ V Σ V . B X How many subclasses (H): Minimize the conflict, K . Zhu & Martinez, PAMI, 2006

  33. Precise Detailed Detection Error: 6.2 pixels (2%) vs Manual: 4.2 (1.5%) E 6 2 i l (2%) M l 4 2 (1 5%) Ding & Martinez, CVPR, 2008; PAMI, 2010

  34. 36 Detection + non-rigid SfM Gotardo & Martinez, PAMI, 2011; Gotardo & Martinez, CVPR, 2011.

  35. Take Home Messages • What is the form of the computational space in human face perception? Linear combination of known categories. c 1 + c 2 + … + c n 1 2 n • What are the dimensions? Mostly configural . fi l Wh t th di i ? M tl • Precise detection of facial features .

  36. CBCSL Paulo Gotardo, Shichuan Du, Don Neth, Liya Ding, Onur Paulo Gotardo, Shichuan Du, Don Neth, Liya Ding, Onur Hamsici, Samuel Rivera, Fabian Benitez, Hongjun Jia, Di You. National Institutes of Health National Institutes of Health National Science Foundation

Recommend


More recommend