2QBF workshop paper report: Graph Neural Network in the 2QBF - - PowerPoint PPT Presentation

2qbf workshop paper report graph neural network in the
SMART_READER_LITE
LIVE PREVIEW

2QBF workshop paper report: Graph Neural Network in the 2QBF - - PowerPoint PPT Presentation

2QBF workshop paper report: Graph Neural Network in the 2QBF Zhanfu Yang Content What is SAT u What is QBF/2QBF u GNN in SAT u GNN in 2QBF u What we find out u GNN-based heuristics in CEGAR-based solvers u Current work u SAT


slide-1
SLIDE 1

2QBF workshop paper report: Graph Neural Network in the 2QBF

Zhanfu Yang

slide-2
SLIDE 2

Content

u

What is SAT

u

What is QBF/2QBF

u

GNN in SAT

u

GNN in 2QBF

u

What we find out

u

GNN-based heuristics in CEGAR-based solvers

u

Current work

slide-3
SLIDE 3

SAT Problem

u

Built from variables, operators AND (conjunction, also denoted by ∧), OR (disjunction, ∨), NOT (negation, ¬), and parentheses.

u

A formula is said to be satisfiable if it can be made TRUE by assigning appropriate logical values (i.e. TRUE, FALSE) to its variables.

u

Given a formula, to check whether it is satisfiable.

[1] https://en.wikipedia.org/wiki/Boolean_satisfiability_problem

slide-4
SLIDE 4

QBF Problem

u

QBF is an extension of propositional formula, which allows quantifiers (∀ and ∃) over the boolean variables

u

QBF problems are PSPACE-complete, lies between NP-completeness of SAT problems and the semi-decidability of predicate logic problems

[1] Hans Kleine Büning and Uwe Bubeck. Theory of quantified boolean formulas. In Handbook of Satisfiability, 2009.

slide-5
SLIDE 5

2QBF Formulae

u

Furthermore, 2-QBF (QBF with only two alternative quantifiers) is ∑"

# $%&'()*)v

u

∀x1∃x2 .sat. (x1 ∨ ¬x2) ∧ (¬x1 ∨ x2)

[1] Hans Kleine Büning and Uwe Bubeck. Theory of quantified boolean formulas. In Handbook of Satisfiability, 2009.

slide-6
SLIDE 6

Why we choose 2QBF?

u

The subclass 2QBF is worthy of study in its own right, away from the more general context of QBF .

u

2QBF is a gentler generalization of SAT than general QBF , techniques that are useful in SAT algorithms sometimes adapt more easily and more usefully to 2QBF than they do to QBF .

[1] Darsh Ranjan, Daijue Tang, and Sharad Malik. A comparative study of 2qbf algorithms, 2014

slide-7
SLIDE 7

Graph neural Network

u

GNNs broadly refer to neural architectures that convey message flows over graphs, which are devised to learn the embedding of nodes and graphs.

u

denotes the hidden state of node v at the k th layer.

u

denotes the neighbors of node v.

slide-8
SLIDE 8

GNN in SAT

u

Embedding for SAT . one kind of nodes represent all literals (boolean variables and their negations, denoted as L ) , and the other kind of nodes represent clauses (denoted as C).

u

Graph Representation of (x1 ∨ ¬x2) ∧ (¬x1 ∨ x2):

u

[1] Daniel Selsam, Matthew Lamm, Benedikt Bünz, Percy Liang, Leonardo de Moura, and David L. Dill. Learning. a SAT solver from single-bit

  • supervision. ICLR 2019
slide-9
SLIDE 9

GNN Embedding in SAT

u

Emb L and Emb C denotes embedding matrices of literals and clauses respectively,

slide-10
SLIDE 10

GNN in 2QBF

u

(x1 ∨ ¬x2) ∧ (¬x1 ∨ x2)

slide-11
SLIDE 11

GNN Embedding in 2QBF

u

Accordingly, in GNN architectures, the separated ∀-literals and ∃-literals are embedded via different machine learning modules.

slide-12
SLIDE 12

Experiment

u

Evaluate:

u

SAT:

u

UNSAT:

u

We trained our GNNs with different amount of training data (40 pairs, 80 pairs, and 160 pairs of satisfiable/unsatisfiable formulas) and different numbers of message-passing iterations (8 iters, 16 iters, and 32 iters), and then tested the converged models on 600 pairs of new instances.

slide-13
SLIDE 13

Why Custom GNN failed?

u

Difficulty in Proving UNSAT for Propositional Logic (1) GNN-bases SAT solvers had trouble predicting unsatisfiability with high confidence (2)message-passing in GNN is rather similar to some incomplete SAT solvers

u

GNN-based 2QBF Solver is Conjecturally Infeasible

[1]

slide-14
SLIDE 14

Learn GNN-based HeuristicsI for 2QBF

u

Traditional decision procedures (such as CEGAR-based solvers (Rabe et al., 2018)) has a way to incrementally construct proof.

u

However, it’s not being ranking.

slide-15
SLIDE 15

Experiment

u

Ranking

u

The evaluations are done on 4 separate datasets TrainU: 1000 unsatisfiable formulas used for training TrainS: 1000 satisfiable formulas used for training TestU: 600 unsatisfiable formulas used for testing TestS: 600 satisfiable formulas used for testing’

u

with 2 baselines:

  • -: vanilla CEGAR without ranking
  • MaxSAT
slide-16
SLIDE 16

Result-1 Ranking CANDIDATES/COunterexample

slide-17
SLIDE 17

Result-2 Combination

slide-18
SLIDE 18

Current work

u

Larger Dataset(Experiment)

u

Difficulty in Proving UNSAT for Propositional Logic(proof)

u

Powerful of GNN Embedding in the 2QBF problem(proof)

slide-19
SLIDE 19

Thanks