POSTER #124 Differentially Private Learning of Geometric Concepts Uri Stemmer Ben-Gurion University joint work with Haim Kaplan, Yishay Mansour, and Yossi Matias
Privately Learning Union of Polygons POSTER #124 π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error
Privately Learning Union of Polygons POSTER #124 π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data:
Privately Learning Union of Polygons POSTER #124 π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data: οΌ Every labeled example represents the (private) information of one individual οΌ Goal: the output hypothesis does not reveal information that is specific to any single individual
Privately Learning Union of Polygons POSTER #124 π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data: οΌ Every labeled example represents the (private) information of one individual οΌ Goal: the output hypothesis does not reveal information that is specific to any single individual οΌ Requirement: the output distribution is insensitive to any arbitrarily change of a single input example (an algorithm satisfying this requirement is differentially private )
Privately Learning Union of Polygons POSTER #124 π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data: οΌ Every labeled example represents the (private) information of one individual οΌ Goal: the output hypothesis does not reveal information that is specific to any single individual οΌ Requirement: the output distribution is insensitive to any arbitrarily change of a single input example (an algorithm satisfying this requirement is differentially private ) Why y is tha hat t a go good od pr privacy vacy definiti nition? on? Even if an observer knows all other data point but mine, and now she sees the outcome of the computation, then she still cannot learn βanythingβ on my data point
Privately Learning Union of Polygons POSTER #124 π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data
Privately Learning Union of Polygons POSTER #124 π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data Motivation: Analyzing Usersβ Location Reports β’ Analyzing GPS navigation data β’ Learning the shape of a flood or a fire based on reports β’ Identifying regions with poor cellular reception based on reports
Privately Learning Union of Polygons POSTER #124 π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data
Privately Learning Union of Polygons POSTER #124 π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data Differential Privacy and Discretization β’ Impossibility results for differential privacy show that this problem (and even much simpler problems) cannot be solved over infinite domains
Privately Learning Union of Polygons POSTER #124 π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data Differential Privacy and Discretization β’ Impossibility results for differential privacy show that this problem (and even much simpler problems) cannot be solved over infinite domains β’ We assume that input points come from π π = π, π, β¦ , π Γ π, π, β¦ , π for a discretization parameter π
Privately Learning Union of Polygons POSTER #124 π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data Differential Privacy and Discretization β’ Impossibility results for differential privacy show that this problem (and even much simpler problems) cannot be solved over infinite domains β’ We assume that input points come from π π = π, π, β¦ , π Γ π, π, β¦ , π for a discretization parameter π β’ Furthermore, the sample complexity must grow with the size of the discretization
Privately Learning Union of Polygons POSTER #124 π π π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data
Privately Learning Union of Polygons POSTER #124 π π π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data Previous Result Private learner with sample complexity π π β π¦π©π‘ π and runtime β π π (using a generic tool of MTβ 07)
Privately Learning Union of Polygons POSTER #124 π π π Given: π points in β π with binary labels: π π , π π π=π Assume: β collection of polygons πΈ π , β¦ , πΈ π with a total of al most π edges s.t. βπ β π : π π β πΈ π βΊ π π = π π Find: Hypothesis π: β π β π, π with small error, while providing differential privacy for the training data Previous Result New Result Private learner with sample complexity Private learner with sample complexity π π β π¦π©π‘ π and runtime β π π π β π¦π©π‘ π and runtime πͺπ©π¦π³ π, π¦π©π‘ π π (using a generic tool of MTβ 07)
Recommend
More recommend