Joint Probabilistic Matching Using m -Best Solutions S. Hamid Rezatofighi Anton Milan Zhen Zhang Qinfeng Shi Antony Dick Ian Reid 1
Introduction One-to-One Graph Matching in Computer Vision • Action Recognition • Feature Point Matching • Multi-Target Tracking ⋮ ⋮ • Person Re-Identification 2
Introduction Most existing works focus on • Feature and/or metric learning [Zhao et al ., CVPR 2014 , Liu et al ., ECCV 2010] • Developing better solvers [Cho et al ., ECCV 2010 , Zhou & De la Torre, CVPR 2013] The optimal solution does not necessarily yield the correct matching assignment To improving the matching results, we propose • to consider more feasible solutions • a principle approach to combine the solutions 3
One-to-One Graph Matching Formulating it as a constrained binary program ⋮ ⋮ 4
One-to-One Graph Matching Formulating it as a constrained binary program 0 𝑦 1 1 𝑦 1 ⋮ ⋮ ⋮ 𝑂 𝑦 𝑁 5
One-to-One Graph Matching Formulating it as a constrained binary program 0 𝑦 1 𝑘 = {0,1} 1 𝑦 𝑗 𝑦 1 𝑂 𝑈 ⊆ 𝑁×(𝑂+1) 𝑘 , … , 𝑦 𝑁 0 , 𝑦 1 1 , … , 𝑦 𝑗 𝑌 = 𝑦 1 ⋮ ⋮ ⋮ 𝑂 𝑦 𝑁 6
One-to-One Graph Matching Formulating it as a constrained binary program 0 𝑌 ∗ = argmin 𝑔 𝑌 𝑦 1 1 𝑌 ∈ 𝒴 𝑦 1 Or 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 ⋮ where 𝑘 = 0,1 , 𝑘 𝒴 = ቄ𝑌 = 𝑦 𝑗 ∀𝑗,𝑘 | 𝑦 𝑗 𝑘 ≤ 1, ∀𝑘: ∑ 𝑦 𝑗 𝑘 = 1 ∀𝑗: ∑ 𝑦 𝑗 ቅ ⋮ ⋮ 𝑂 𝑦 𝑁 7
One-to-One Graph Matching Formulating it as a constrained binary program 0 𝑌 ∗ = argmin 𝑔 𝑌 𝑦 1 1 𝑌 ∈ 𝒴 𝑦 1 Or 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 ⋮ where 𝑘 = 0,1 , 𝑘 𝒴 = ቄ𝑌 = 𝑦 𝑗 ∀𝑗,𝑘 | 𝑦 𝑗 𝑘 ≤ 1, ∀𝑘: ∑ 𝑦 𝑗 𝑘 = 1 ∀𝑗: ∑ 𝑦 𝑗 ቅ ⋮ ⋮ 𝑂 𝑦 𝑁 8
One-to-One Graph Matching Formulating it as a constrained binary program 0 𝑌 ∗ = argmin 𝑔 𝑌 𝑦 1 1 𝑌 ∈ 𝒴 𝑦 1 Or 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 ⋮ where 𝑘 = 0,1 , 𝑘 𝒴 = ቄ𝑌 = 𝑦 𝑗 ∀𝑗,𝑘 | 𝑦 𝑗 𝑘 ≤ 1, ∀𝑘: ∑ 𝑦 𝑗 𝑘 = 1 ∀𝑗: ∑ 𝑦 𝑗 ቅ ⋮ ⋮ 𝑂 𝑦 𝑁 9
One-to-One Graph Matching Formulating it as a constrained binary program 0 𝑌 ∗ = argmin 𝑔 𝑌 𝑦 1 1 𝑌 ∈ 𝒴 𝑦 1 Or 𝐵𝑌 ≤ 𝐶 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 ⋮ where 𝑘 = 0,1 , 𝑘 𝒴 = ቄ𝑌 = 𝑦 𝑗 ∀𝑗,𝑘 | 𝑦 𝑗 𝑘 ≤ 1, ∀𝑘: ∑ 𝑦 𝑗 𝑘 = 1 ∀𝑗: ∑ 𝑦 𝑗 ቅ ⋮ ⋮ 𝑂 𝑦 𝑁 10
One-to-One Graph Matching Formulating it as a constrained binary program 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∈ 𝒴 Or 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 where 𝑘 = 0,1 , 𝑘 𝒴 = ቄ𝑌 = 𝑦 𝑗 ∀𝑗,𝑘 | 𝑦 𝑗 𝑘 ≤ 1, ∀𝑘: ∑ 𝑦 𝑗 𝑘 = 1 ∀𝑗: ∑ 𝑦 𝑗 ቅ ⋮ ⋮ 11
One-to-One Graph Matching Examples of joint matching distribution 𝑞 𝑌 and cost 𝑔 𝑌 in different applications • Multi-target tracking [Zheng et al. , CVPR 2008] and person re-identification [Das et al. , ECCV 2014 ] 𝑘 𝑘 𝑦 𝑗 𝑔 𝑌 = 𝐷 𝑈 𝑌 or equivalently 𝑞 𝑌 ∝ ς 𝑞 𝑦 𝑗 • Feature point matching [Leordeanu et al. , IJCV 2011] 𝑔 𝑌 = 𝑌 𝑈 𝑅 𝑌 • Stereo matching [Meltzer et al. , ICCV 2005] and iterative closest point [Zheng, IJCV 1994] higher-order constraints in addition to one-to-one constraints 12
Marginalization VS MAP Estimates In general, globally optimal solution may or may not be easily achieved. 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 𝑌 ∈ 𝒴 Even the optimal solution does not necessarily yield the correct matching assignment 13
Marginalization VS MAP Estimates In general, globally optimal solution may or may not be easily achieved. 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 𝑌 ∈ 𝒴 Even the optimal solution does not necessarily yield the correct matching assignment • Visual similarity • Other ambiguities in the matching space 14
Marginalization VS MAP Estimates In general, globally optimal solution may or may not be easily achieved. 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 𝑌 ∈ 𝒴 Even the optimal solution does not necessarily yield the correct matching assignment • Visual similarity • Other ambiguities in the matching space 15
Marginalization VS MAP Estimates In general, globally optimal solution may or may not be easily achieved. 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 𝑌 ∈ 𝒴 Even the optimal solution does not necessarily yield the correct matching assignment 16
Marginalization VS MAP Estimates In general, globally optimal solution may or may not be easily achieved. 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 𝑌 ∈ 𝒴 Even the optimal solution does not necessarily yield the correct matching assignment 17
Marginalization VS MAP Estimates Motivation to use marginalization Encoding the entire distribution to untangle potential ambiguities MAP only considers one single value of that distribution Improving matching ranking due to averaging / smoothing property Exact marginalization is NP-hard Requiring all feasible permutations to built the joint distribution Solution Approximation using m -Best solutions 18
Marginalization Using m -Best Solutions Marginalization by considering a fraction of the matching space Using m -highest joint probabilities 𝑞 𝑌 / m -lowest values for 𝑔 𝑌 19
Marginalization Using m -Best Solutions Marginalization by considering a fraction of the matching space Using m -highest joint probabilities 𝑞 𝑌 / m -lowest values for 𝑔 𝑌 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 ∗ is 𝑌 1 1-st optimal solution 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∈ 𝒴 20
Marginalization Using m -Best Solutions Marginalization by considering a fraction of the matching space Using m -highest joint probabilities 𝑞 𝑌 / m -lowest values for 𝑔 𝑌 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 ∗ is 𝑌 2 2-nd optimal solution 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∈ 𝒴 21
Marginalization Using m -Best Solutions Marginalization by considering a fraction of the matching space Using m -highest joint probabilities 𝑞 𝑌 / m -lowest values for 𝑔 𝑌 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 ∗ is 𝑌 3 3-rd optimal solution 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∈ 𝒴 22
Marginalization Using m -Best Solutions Marginalization by considering a fraction of the matching space Using m -highest joint probabilities 𝑞 𝑌 / m -lowest values for 𝑔 𝑌 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 ∗ is 𝑌 𝑙 k- th optimal solution 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∈ 𝒴 23
Marginalization Using m -Best Solutions Marginalization by considering a fraction of the matching space Using m -highest joint probabilities 𝑞 𝑌 / m -lowest values for 𝑔 𝑌 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 ∗ is 𝑌 𝑙 k -th optimal solution 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∈ 𝒴 24
Marginalization Using m -Best Solutions Marginalization by considering a fraction of the matching space Using m -highest joint probabilities 𝑞 𝑌 / m -lowest values for 𝑔 𝑌 𝑌 ∗ = argmax 𝑞 𝑌 𝑌 ∈ 𝒴 ∗ is 𝑌 𝑙 k -th optimal solution 𝑌 ∗ = argmin 𝑔 𝑌 𝑌 ∈ 𝒴 Approximation error bound decreases exponentially by increasing number of solutions [Rezatofighi et al. , ICCV 2015] 25
Computing the m -Best Solutions Naïve exclusion strategy ∗ = argmin 𝑔 𝑌 𝑌 1 𝐵𝑌 ≤ 𝐶 26
Computing the m -Best Solutions Naïve exclusion strategy ∗ = argmin 𝑔 𝑌 𝑌 2 𝐵𝑌 ≤ 𝐶 ∗ ≤ ∗ 𝑌, 𝑌 1 𝑌 1 1 − 1 27
Computing the m -Best Solutions Naïve exclusion strategy ∗ = argmin 𝑔 𝑌 𝑌 3 𝐵𝑌 ≤ 𝐶 ∗ ≤ ∗ 𝑌, 𝑌 1 𝑌 1 1 − 1 ∗ ≤ ∗ 𝑌, 𝑌 2 𝑌 2 1 − 1 28
Computing the m -Best Solutions Naïve exclusion strategy ∗ = argmin 𝑔 𝑌 𝑌 𝑙 𝐵𝑌 ≤ 𝐶 ∗ ≤ ∗ 𝑌, 𝑌 1 𝑌 1 1 − 1 ∗ ≤ ∗ 𝑌, 𝑌 2 𝑌 2 1 − 1 ⋮ ∗ ∗ 𝑌, 𝑌 𝑙−1 ≤ 𝑌 𝑙−1 1 − 1 29
Computing the m -Best Solutions Naïve exclusion strategy ∗ = argmin 𝑔 𝑌 General approach 𝑌 𝑙 𝐵𝑌 ≤ 𝐶 Impractical for large values of m 𝐵𝑌 ≤ ሖ ሖ 𝐶 30
Computing the m -Best Solutions Naïve exclusion strategy ∗ = argmin 𝑔 𝑌 General approach 𝑌 𝑙 𝐵𝑌 ≤ 𝐶 Impractical for large values of m 𝐵𝑌 ≤ ሖ ሖ 𝐶 Binary Tree Partitioning Partitioning the space into a set of disjoint subspaces [Rezatofighi et al., ICCV 2015 ] Efficient approach Not a good strategy for weak solvers 31
Experimental Results Query images Gallery images 0 𝑑 1 Person Re-Identification None of them 1 𝑑 1 𝑂 𝑑 𝑁 32
Recommend
More recommend