machine learning matched action parameters
play

Machine Learning Matched Action Parameters Phiala Shanahan - PowerPoint PPT Presentation

Machine Learning Matched Action Parameters Phiala Shanahan Motivation: ML for LQCD First-principles nuclear physics beyond A=4 How finely tuned is the emergence of nuclear structure in nature? Interpretation of intensity-frontier


  1. Machine Learning 
 Matched Action Parameters Phiala Shanahan

  2. Motivation: ML for LQCD First-principles nuclear physics beyond A=4 How finely tuned is the emergence of nuclear structure in nature? Interpretation of intensity-frontier experiments Scalar matrix elements in A=131 
 XENON1T dark matter direct detection search Axial form factors of Argon A=40 
 DUNE long-baseline neutrino expt. Double-beta decay rates of Calcium A=48 Exponentially harder 
 Need exponentially 
 problems improved algorithms

  3. Machine learning for LQCD APPROACH 
 Machine learning as ancillary tool for lattice QCD Will need to } Accelerate gauge-field 
 accelerate all stages generation of lattice QCD Optimise extraction of physics 
 workflow to achieve from gauge field ensemble physics goals ONLY apply where quantum field theory can be rigorously preserved

  4. Accelerating HMC: action matching QCD gauge field configurations sampled via Hamiltonian dynamics + Markov Chain Monte Carlo Updates diffusive Lattice spacing 0 Number of updates to change ∞ fixed physical length scale “Critical slowing-down” 
 of generation of uncorrelated samples

  5. Multi-scale HMC updates Given coarsening and refinement procedures… coarsen refine Endres et al., PRD 92, 114516 (2015)

  6. Multi-scale HMC updates Perform HMC updates at coarse level Fine ensemble 
 rethermalise 
 with fine action 
 coarsen to make exact HMC 
 Multiple layers of coarsening Significantly cheaper approach to continuum limit … Endres et al., PRD 92, 114516 (2015)

  7. Multi-scale HMC updates Perform HMC updates at coarse level encode same long-distance physics Map a subset of physics parameters MUST KNOW 
 in the coarse space and match to parameters of coarse coarsened ensemble 
 QCD action that OR reproduce ALL physics Solve regression problem directly: 
 parameters of fine “Given a coarse ensemble, what simulation parameters generated it?”

  8. Machine learning LQCD Neural networks excel on problems where Combination of units 
 Basic data unit is meaningful has little meaning Image recognition Pixel Image Label “Colliding 
 Neural 
 black holes” network

  9. Machine learning LQCD Neural networks excel on problems where Combination of units 
 Basic data unit is meaningful has little meaning Parameter identification Element of a colour Ensemble of lattice QCD Label matrix at one discrete gauge field configurations space-time point Parameters 
 Neural 
 0 6 7 3 5 of action network 8 4 2 1 6

  10. Machine learning LQCD Ensemble of lattice QCD CIFAR benchmark image set for machine learning gauge fields 64 3 x128 x 4 x N c2 x 2 
 32 x 32 pixels x 3 cols 
 ≃ 10 9 numbers ≃ 3000 numbers ~1000 samples 60000 samples Ensemble of gauge fields has Each image has meaning meaning Long-distance correlations Local structures are important are important Gauge and translation- Translation-invariance invariant with periodic within frame boundaries

  11. 
 Regression by neural network Lattice QCD 
 Parameters of 
 gauge field 
 lattice action Few real 
 ~10 7- 10 9 real 
 numbers numbers NEURAL NETWORK Complete: not restricted to affordable subset of physics parameters Instant: once trained over a parameter range

  12. Naive neural network Simplest approach Ignore physics symmetries Train simple neural network (state-of-the-art ~10 9 ) on regression task Fully-connected structure Far more degrees of freedom than number of training samples available “Inverted data Recipe for hierarchy” overfitting!

  13. Naive neural network Training and validation 
 datasets Quark mass parameter Parameters of training and validation datasets O(10,000) independent configurations generated at each point Validation configurations randomly selected from generated streams Parameter related 
 Spacing in evolution stream >> to lattice spacing correlation time of physics observables

  14. Naive neural network Neural net predictions 
 SUCCESS? on validation data sets Quark mass parameter - 0.7 No sign of overfitting * * * * Training and validation loss equal - 0.8 Accurate predictions for * * * * validation data - 0.9 * * * * BUT fails to generalise to - 1.0 Ensembles at other parameters * * * New streams at same - 1.1 1.75 1.80 1.85 1.90 parameters Parameter related to lattice spacing NOT POSSIBLE IF CONFIGS 
 ARE UNCORRELATED True parameter values Confidence interval from ensemble of gauge fields

  15. Naive neural network Stream of generated gauge fields at given parameters … Training/validation data selected from configurations spaced to be decorrelated (by physics observables) Network succeeds for validation configs Network has identified from same stream as training configs feature with a longer correlation length than any Network fails for configs from new known physics observable stream at same parameters

  16. Naive neural network Naive neural network that does not respect symmetries fails at parameter regression task BUT Identifies unknown feature of gauge fields with a longer correlation length than any known physics observable τ max τ int = 1 1 X 2 + lim ρ ( τ ) , Network feature autocorrelation ρ (0) τ max →∞ τ =0 40 Network-identified feature Autocorrelation in evolution 1.0 autocorrelation time time using identification of 30 0.8 parameters of configurations at the end of a training stream 0.6 20 0.4 Max physics observable autocorrelation time 10 0.2 0.0 0 0 50 100 150 200 0 50 100 150 200

  17. 
 Regression by neural network Lattice QCD 
 Parameters of 
 gauge field 
 lattice action Few real 
 ~10 7- 10 9 real 
 numbers numbers NEURAL NETWORK Complete: not restricted to affordable subset of physics parameters Instant: once trained over a parameter range

  18. 
 Regression by neural network Lattice QCD 
 Parameters of 
 gauge field 
 lattice action Custom network structures Few real 
 ~10 7- 10 9 real 
 numbers numbers Respects gauge-invariance, translation-invariance, boundary conditions NEURAL NETWORK Emphasises QCD-scale physics Range of neural network structures find same minimum Complete: not restricted to affordable subset of physics parameters Instant: once trained over a parameter range

  19. Symmetry-preserving network Network based on symmetry-invariant features Closed Wilson loops 
 Loops (gauge-invariant) Correlated products of loops at various W 3 × 2 ( y ) length scales y Volume-averaged and ˆ ν U µ ( x ) ˆ µ x + ˆ rotation-averaged x µ

  20. Symmetry-preserving network Network based on symmetry-invariant features Fully-connected network structure First layer samples from set of possible symmetry- invariant features Number of degrees of freedom of network comparable to size of training dataset

  21. Gauge field parameter regression Neural net predictions 
 Predictions on 
 on validation data sets new datasets 1.05 1.05 Quark mass parameter * * * * * * * * * * * * * * 1.00 1.00 ** * 0.95 0.95 * * * * * * * * * * * 0.90 0.90 0.85 0.85 * * * * * * * * * * 0.80 0.80 0.75 0.75 1.75 1.80 1.85 1.90 1.95 2.00 1.75 1.80 1.85 1.90 1.95 2.00 Parameter related 
 True parameter values to lattice spacing Confidence interval from 
 ensemble of gauge fields

  22. Gauge field parameter regression Neural net predictions 
 Predictions on 
 on validation data sets new datasets 1.05 1.05 Quark mass parameter * * * * * * * * * * * * * * 1.00 1.00 ** * 0.95 0.95 SUCCESS! 
 * * * * * * * * * * * 0.90 0.90 Accurate parameter regression 0.85 0.85 and successful generalisation * * * * * * * * * * 0.80 0.80 0.75 0.75 1.75 1.80 1.85 1.90 1.95 2.00 1.75 1.80 1.85 1.90 1.95 2.00 Parameter related 
 True parameter values to lattice spacing Confidence interval from 
 ensemble of gauge fields

  23. Gauge field parameter regression PROOF OF PRINCIPLE 
 Step towards fine lattice generation 
 at reduced cost 1. Generate one fine configuration 2. Find matching coarse action 3. HMC updates in coarse space 4. Refine and rethermalise Accurate matching minimises cost of updates in fine space Guarantees 
 correctness Shanahan, Trewartha, Detmold, PRD (2018) [1801.05784]

  24. Tests of network success How does neural network regression perform compared with other approaches? Consider very closely-spaced validation ensembles at new parameters Sets along lines of constant 1x1 Wilson loop (most precise feature allowed by Set A network) Much closer spacing than separation of Set B training ensembles

  25. Tests of network success How does neural network regression perform compared with other approaches? Consider very closely-spaced validation ensembles at new parameters: not distinguishable to principal component analysis in loop space Histograms of dominant eigenvectors Eigenvalues 150 150 Set A 3 Set B 100 100 50 50 ◦ ◦ ◦ 2 ◦ ◦ ◦ ◦ 0 0 ◦ ◦ ◦ ◦ ◦ ◦ - 2.00 - 1.95 - 1.90 - 1.85 1.50 1.55 1.60 1.65 ◦ ◦ ◦ ◦ 1 ◦ ◦ ◦ ◦ ◦ ◦ 150 150 ◦ ◦ ◦ ◦ ◦ ◦ ◦ ◦ ◦ ◦ ◦ 0 100 100 50 50 - 1 0 5 10 15 0 0 1.18 1.20 1.22 1.24 1.26 0.10 0.11 0.12 0.13 0.14

Recommend


More recommend