probabilistic and logistic circuits
play

Probabilistic and Logistic Circuits: A New Synthesis of Logic and - PowerPoint PPT Presentation

Probabilistic and Logistic Circuits: A New Synthesis of Logic and Machine Learning Guy Van den Broeck Stanford Nov 14, 2018 Foundation: Logical Circuit Languages Negation Normal Form Circuits = (sun rain rainbow) [Darwiche 2002]


  1. Probabilistic and Logistic Circuits: A New Synthesis of Logic and Machine Learning Guy Van den Broeck Stanford Nov 14, 2018

  2. Foundation: Logical Circuit Languages

  3. Negation Normal Form Circuits Δ = (sun ∧ rain ⇒ rainbow) [Darwiche 2002]

  4. Decomposable Circuits Decomposable [Darwiche 2002]

  5. Tractable for Logical Inference • Is there a solution? (SAT) ✓ – SAT( 𝛽 ∨ 𝛾 ) iff SAT( 𝛽 ) or SAT( 𝛾 ) ( always ) – SAT( 𝛽 ∧ 𝛾 ) iff SAT( 𝛽 ) and SAT( 𝛾 ) ( decomposable ) • How many solutions are there? (#SAT) • Complexity linear in circuit size 

  6. Deterministic Circuits Deterministic [Darwiche 2002]

  7. How many solutions are there? (#SAT)

  8. How many solutions are there? (#SAT) Arithmetic Circuit

  9. Tractable for Logical Inference • Is there a solution? (SAT) ✓ ✓ • How many solutions are there? (#SAT) • Stricter languages (e.g., BDD, SDD): ✓ – Equivalence checking ✓ – Conjoin/disjoint/negate circuits • Complexity linear in circuit size  • Compilation into circuit language by either – ↓ exhaustive SAT solver – ↑ conjoin/disjoin/negate

  10. Learning with Logical Constraints

  11. Motivation: Video [Lu, W. L., Ting, J. A., Little, J. J., & Murphy, K. P. (2013). Learning to track and identify players from broadcast sports videos.]

  12. Motivation: Robotics [Wong, L. L., Kaelbling, L. P., & Lozano-Perez, T., Collision-free state estimation. ICRA 2012]

  13. Motivation: Language • Non-local dependencies: At least one verb in each sentence • Sentence compression If a modifier is kept, its subject is also kept • Information extraction • Semantic role labeling … and many more! [Chang, M., Ratinov, L., & Roth, D. (2008). Constraints as prior knowledge],…, [ Chang, M. W., Ratinov, L., & Roth, D. (2012). Structured learning with constrained conditional models.], [https://en.wikipedia.org/wiki/Constrained_conditional_model]

  14. Motivation: Deep Learning [Graves, A., Wayne, G., Reynolds, M., Harley, T., Danihelka, I., Grabska- Barwińska , A., et al.. (2016). Hybrid computing using a neural network with dynamic external memory. Nature , 538 (7626), 471-476.]

  15. Running Example Courses: Data • Logic (L) • Knowledge Representation (K) • Probability (P) • Artificial Intelligence (A) Constraints • Must take at least one of Probability or Logic. • Probability is a prerequisite for AI. • The prerequisites for KR is either AI or Logic.

  16. Structured Space unstructured structured L K P A L K P A 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 1 • Must take at least one of 0 0 1 0 0 0 1 0 Probability ( P ) or Logic ( L ). 0 0 1 1 0 0 1 1 • Probability is a prerequisite 0 1 0 0 0 1 0 0 for AI ( A ). 0 1 0 1 0 1 0 1 • 0 1 1 0 The prerequisites for KR ( K ) is 0 1 1 0 0 1 1 1 0 1 1 1 either AI or Logic. 1 0 0 0 1 0 0 0 1 0 0 1 1 0 0 1 1 0 1 0 1 0 1 0 7 out of 16 instantiations 1 0 1 1 1 0 1 1 are impossible 1 1 0 0 1 1 0 0 1 1 0 1 1 1 0 1 1 1 1 0 1 1 1 0 1 1 1 1 1 1 1 1

  17. Boolean Constraints unstructured structured L K P A L K P A 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 1 0 0 1 0 0 0 1 0 0 0 1 1 0 0 1 1 0 1 0 0 0 1 0 0 0 1 0 1 0 1 0 1 0 1 1 0 0 1 1 0 0 1 1 1 0 1 1 1 1 0 0 0 1 0 0 0 1 0 0 1 1 0 0 1 1 0 1 0 1 0 1 0 7 out of 16 instantiations 1 0 1 1 1 0 1 1 are impossible 1 1 0 0 1 1 0 0 1 1 0 1 1 1 0 1 1 1 1 0 1 1 1 0 1 1 1 1 1 1 1 1

  18. Learning in Structured Spaces + Data Constraints (Background Knowledge) (Physics) Learn ML Model Today‟s machine learning tools don‟t take knowledge as input! 

  19. Deep Learning with Logical Constraints

  20. Deep Learning + Data Constraints with Deep Neural Learn Logical Knowledge Network Neural Network Logical Constraint Output Input Output is probability vector p , not Boolean logic!

  21. Semantic Loss Q: How close is output p to satisfying constraint? Answer: Semantic loss function L( α , p ) • Axioms, for example: – If p is Boolean then L( p,p ) = 0 – If α implies β then L( α , p ) ≥ L(β , p ) ( α more strict ) • Properties: SEMANTIC – If α is equivalent to β then L( α , p ) = L( β , p ) Loss! – If p is Boolean and satisfies α then L( α , p ) = 0

  22. Semantic Loss: Definition Theorem: Axioms imply unique semantic loss: Probability of getting x after flipping coins with prob. p Probability of satisfying α after flipping coins with prob. p

  23. Example: Exactly-One • Data must have some label We agree this must be one of the 10 digits: • Exactly-one constraint 𝒚 𝟐 ∨ 𝒚 𝟑 ∨ 𝒚 𝟒 ¬𝒚 𝟐 ∨ ¬𝒚 𝟑 → For 3 classes: ¬𝒚 𝟑 ∨ ¬𝒚 𝟒 • Semantic loss: ¬𝒚 𝟐 ∨ ¬𝒚 𝟒 Only 𝒚 𝒋 = 𝟐 after flipping coins Exactly one true 𝒚 after flipping coins

  24. Semi-Supervised Learning • Intuition: Unlabeled data must have some label Cf. entropy constraints, manifold learning • Minimize exactly-one semantic loss on unlabeled data Train with 𝑓𝑦𝑗𝑡𝑢𝑗𝑜𝑕 𝑚𝑝𝑡𝑡 + 𝑥 ∙ 𝑡𝑓𝑛𝑏𝑜𝑢𝑗𝑑 𝑚𝑝𝑡𝑡

  25. MNIST Experiment Competitive with state of the art in semi-supervised deep learning

  26. FASHION Experiment Same conclusion on CIFAR10 Outperforms Ladder Nets!

  27. What about real constraints? Paths cf. Nature paper Good variable assignment Bad variable assignment (does not represent route) (represents route) 184 16,777,032 Unstructured probability space: 184+16,777,032 = 2 24 Space easily encoded in logical constraints  [Nishino et al.]

  28. How to Compute Semantic Loss? • In general: #P-hard  • With a logical circuit for α : Linear! • Example: exactly-one constraint: L( α , p ) = L( , p ) = - log( ) • Why? Decomposability and determinism!

  29. Predict Shortest Paths Add semantic loss for path constraint Is output Is prediction Are individual a path? the shortest path? edge predictions This is the real task! correct? (same conclusion for predicting sushi preferences, see paper)

  30. Probabilistic Circuits

  31. Logical Circuits  L K L  P A  P   L   P  A  L  K L   P  L P P K  K A  A A  A Can we represent a distribution over the solutions to the constraint?

  32. Recall: Decomposability  L K L  P A  P   L   P  A  L  K L   P  L P P K  K A  A A  A AND gates have disjoint input circuits

  33. Recall: Determinism L ⊥ ¬ P ⊥ ¬ L ⊥ L ⊥ ¬ P ⊥ ¬ L K P A L ¬ P ¬ A P ¬ L ¬ K P K ¬ K A ¬ A A ¬ A Input: L, K, P, A are true and ¬L, ¬K, ¬P, ¬A are false Property: OR gates have at most one true input wire

  34. PSDD: Probabilistic SDD 0.1 0.6 0.3 1 0 1 0 1 0 0.6 0.4 1 0 1 0 L ⊥ ¬ P ⊥ L ⊥ ¬ P ⊥ ¬ L ⊥ ¬ L K P A L ¬ P ¬ A P ¬ L ¬ K P 0.8 0.2 0.25 0.75 0.9 0.1 K ¬ K A ¬ A A ¬ A Syntax: assign a normalized probability to each OR gate input

  35. PSDD: Probabilistic SDD 0.1 0.6 0.3 0.6 1 0 1 0 1 0 0.4 1 0 1 0 L ⊥ ¬ P ⊥ ¬ L ⊥ L ⊥ ¬ P ⊥ ¬ L K P A L ¬ P ¬ A P ¬ L ¬ K P Input: L, K, P, A 0.8 0.2 0.75 0.25 0.9 0.1 are true A ¬ A A ¬ A K ¬ K Pr( L,K,P,A ) = 0.3 x 1 x 0.8 x 0.4 x 0.25 = 0.024

  36. Each node represents a normalized distribution! 0.1 0.6 0.3 1 0 1 0 1 0 0.6 0.4 1 0 1 0  L K L  P A  P   L  K L   P   L   P  A L P P 0.8 0.2 0.25 0.75 0.9 0.1 A  A A  A A  A Can read probabilistic independences off the circuit structure

  37. Tractable for Probabilistic Inference • MAP inference : Find most-likely assignment to x given y (otherwise NP-hard) • Computing conditional probabilities Pr(x|y) (otherwise #P-hard) • Sample from Pr(x|y) • Algorithms linear in circuit size  (pass up, pass down, similar to backprop)

  38. Parameters are Interpretable 0.1 0.6 0.3 Probability of course P given L 1 0 1 0 1 0 0.6 0.4 1 0 1 0  L  K L   P   L K L  P A  P   L   P  A L P P 0.8 0.2 0.25 0.75 0.9 0.1 K  K A  A A  A Student takes course P Student takes course L Explainable AI DARPA Program

  39. Learning Probabilistic Circuit Parameters

  40. Learning Algorithms • Closed form max likelihood from complete data • One pass over data to estimate Pr(x|y) Not a lot to say: very easy!  • Where does the structure come from? For now: simply compiled from constraint…

  41. Combinatorial Objects: Rankings rank sushi rank sushi 1 fatty tuna 1 shrimp 10 items : 2 sea urchin 2 sea urchin 3,628,800 3 salmon roe 3 salmon roe rankings 4 shrimp 4 fatty tuna 5 tuna 5 tuna 20 items : 6 squid 6 squid 2,432,902,008,176,640,000 7 tuna roll 7 tuna roll rankings 8 see eel 8 see eel 9 egg 9 egg cucumber cucumber 10 10 roll roll

Recommend


More recommend