learning noise in quantum information processors
play

Learning Noise in Quantum Information Processors Travis L Scholten - PowerPoint PPT Presentation

Learning Noise in Quantum Information Processors Travis L Scholten @Travis_Sch Center for Quantum Information and Control, Center for Computing Research, Sandia National Laboratories, Albuquerque, USA University of New Mexico, Albuquerque, USA


  1. Learning Noise in Quantum Information Processors Travis L Scholten @Travis_Sch Center for Quantum Information and Control, Center for Computing Research, Sandia National Laboratories, Albuquerque, USA University of New Mexico, Albuquerque, USA QTML 2017 CCR Sandia National Laboratories is a multimission laboratory managed and operated by National Technology and Engineering Solutions of Sandia LLC, a wholly owned subsidiary of Honeywell International Inc. for the U.S. Department of Energy’s National Nuclear Security Administration under contract DE-NA0003525. Center for Computing Research

  2. There are lots of applications at the intersection of QI/QC and ML… quantum machine quantum information machine learning learning processing annealing simulated annealing quantum annealing quantum gibbs sampling quantum topological markov chain monte-carlo quantum BM algorithms feed forward neural net quantum PCA quantum SVM quantum quantum NN classification perceptron neural nets Quantum ODE solvers quantum clustering quantum data fitting quantum rejection sampling / HHL control and metrology quantum control phase estimation reinforcement learning tomography hamiltonian learning Biamonte, et. al, arXiv: 1611.09347 26

  3. …I want to focus on how ML can improve characterization of quantum hardware . quantum machine quantum information machine learning learning processing annealing simulated annealing quantum annealing quantum gibbs sampling quantum topological markov chain monte-carlo quantum BM algorithms feed forward neural net quantum PCA quantum SVM quantum quantum NN classification perceptron neural nets Quantum ODE solvers quantum clustering quantum data fitting quantum rejection sampling / HHL control and metrology quantum control phase estimation reinforcement learning tomography hamiltonian learning Biamonte, et. al, arXiv: 1611.09347 25

  4. Quantum device characterization (QCVV) techniques arranged by amount learned and time required. Full Amount learned Slow Fast Limited Speed of learning 24

  5. Tomography is very informative , but time-consuming ! Gate set Full tomography Process Amount learned tomography State tomography Slow Fast Limited Speed of learning 23

  6. Randomized benchmarking is fast , but yields limited information . Gate set Full tomography Process Amount learned tomography State tomography Slow Fast Randomized Benchmarking (Several variants, leads to different Limited kinds of information learned.) Speed of learning 22

  7. Depending on how much we want to learn, and how quickly, machine learning could be useful. Gate set Full tomography Machine Learning Process Amount learned tomography Caveat: “Speed” doesn’t include *training time* State tomography Slow Fast Randomized Benchmarking Limited Speed of learning 21

  8. Depending on how much we want to learn, and how quickly, machine learning could be useful. Gate set Full tomography Machine Learning Process Amount learned tomography Can machine learning State extract information about noise tomography Slow Fast affecting near and medium-term quantum hardware? Randomized Benchmarking Limited Speed of learning 20

  9. Noise affects the outcome probabilities of quantum circuits. How can we learn about noise using the data we get from running quantum circuits ? 19

  10. Noise in quantum hardware affects the outcome probabilities of circuits. Example: over-rotation error of a single-qubit gate Y π / 2 | 0 i Pr(0) = Tr( | 0 ih 0 |E ( | 0 ih 0 | )) = 1 2 (1 � sin θ ) (The circuit we write down) (Noise affects outcome probability) 18

  11. Gate set tomography (GST) provides a set of structured circuits we can use for learning. GST assumes the device is a black box, described by M ρ a gate set. G 2 . . . G 1 GST prescribes certain circuits to run that collectively amplify all types of noise . Y π / 2 | 0 i X π / 2 Y π / 2 Z π / 2 | 0 i Y π / 2 Y π / 2 | 0 i Standard use: Outcome probabilities are analyzed Blume-Kohout, et. al, by pyGSTi software to arXiv 1605.07674 estimate the noisy gates 17

  12. Gate set tomography (GST) provides a set of structured circuits we can use for learning. GST prescribes certain circuits to run that collectively amplify all types of noise . l = 1 Y π / 2 Y π / 2 Y π / 2 Y π / 2 Y π / 2 | 0 i | 0 i Y π / 2 Y π / 2 | 0 i l = 2 l = 4 Circuits have varying length, up to some maximum length L. l = 1 l = 2 l = 1 , 2 , 4 , · · · , L Why? Longer circuits are more sensitive to noise. 16

  13. To do machine learning on GST data sets, embed them in a feature space. (GST data set) ## Columns = minus count, plus count   {} 100 0 0 Gx 44 56 . 56   f = ( f 1 , f 2 , · · · ) ∈ R d   Gy 45 55 . 55     GxGx 9 91 . 91     GxGxGx 68 32 . 32   GyGyGy 70 30 . 3 The dimension of the feature space grows with L because more circuits are added. 15

  14. Noise changes some components of the feature vectors. How can we identify the “signature” of a noise process using GST feature vectors? 14

  15. Principal component analysis (PCA) is a useful tool for understanding the structure of GST feature vectors. PCA finds a low-dimensional representation of data by looking for directions of maximum variance . Compute covariance matrix & diagonalize C = P K j =1 σ j σ j σ T j σ 1 ≥ σ 2 · · · ≥ σ K Defines a map: f → P K j =1 ( f · σ j ) σ j ( d = K ) 13

  16. Projection onto a 2-dimensional PCA subspace reveals a structure to GST feature vectors . 1% depolarizing 1% coherent Different noise types and noise strengths tend to cluster ! (PCA performed on entire dataset, then individual feature vectors transformed.) 4.5% depolarizing 4.5% coherent 12

  17. Adding longer circuits makes the clusters more distinguishable. Longer GST circuits amplify noise, making the clusters more distinguishable. We can use this structure to do classification ! (An independent PCA was done for each L.) 11

  18. Classification is possible because the data sets cluster based on noise type and strength! Project feature vectors Label feature vectors Train a soft- based on PCA based on noise margin , linear support vector machine ( SVM ) 10

  19. Classification is possible because the data sets cluster based on noise type and strength! 96% accuracy?? Cross-validation required! Project feature vectors Label feature vectors Train a soft- based on PCA based on noise margin , linear support vector machine ( SVM ) 9

  20. Under cross-validation, the SVM has reasonably low inaccuracy. SVM is fairly accurate - largest inaccuracy ~2% 20-fold shuffle-split cross-validation (25% withheld for testing) 8

  21. The accuracy of the SVM is affected by the number of components and maximum sequence length. 20-fold shuffle-split cross-validation scheme used, with 25% of the data withheld for testing on each split. A “one-versus-one” multi-class classification scheme was used. 7

  22. Can a classifier learn the difference between arbitrary stochastic and arbitrary coherent noise? Coherent Noise Ideal Stochastic Noise ρ = − i [ H 0 , ρ ] ˙ ρ = − i [ H 0 , ρ ] ˙ ρ = − i [ H 0 , ρ ] ˙ + A ρ A † − 1 2 { A † A, ρ } − i [ e, ρ ] E = Λ � G 0 E = V � G 0 E = G 0 ΛΛ T 6 = I V V T = I 6

  23. Classification in a 2-dimensional subspace is harder, due to structure of PCA-projected feature vectors. “Radio dish” type structure Linear classifier infeasible with only 2 PCA components 5

  24. Preliminary results indicate a linear, soft-margin SVM can classify these two noise types in higher dimensions. Gap goes away if noise <= 10**-2 removed from data For each L: - 10 values of noise strength in [10**-4, 10**-1] - 260 random instances 20-fold shuffle-split cross-validation scheme used, with 25% of the data withheld for testing on each split. A “one-verus-one” multi-class classification scheme was used. 4

  25. Specific machine learning tools can analyze GST circuits and learn about noise. Gate set Full tomography Machine Learning Process Amount learned tomography State tomography Slow Fast SVMs & PCA + GST Circuits Randomized Benchmarking Limited Speed of learning 3

  26. Specific machine learning tools can analyze GST circuits and learn about noise. Gate set Full tomography What else can we learn?? Machine Learning Process What circuits do we need?? Amount learned tomography State tomography Slow Fast SVMs & PCA + GST Circuits Randomized Benchmarking Limited Speed of learning 2

  27. quantum machine quantum information machine learning learning processing annealing There are lots of problems simulated annealing quantum annealing quantum gibbs sampling at the intersection of quantum topological device characterization markov chain monte-carlo quantum BM algorithms and machine learning ! feed forward neural net quantum PCA quantum SVM quantum quantum NN classification perceptron neural nets Quantum ODE solvers quantum clustering quantum data fitting quantum rejection sampling / HHL control and metrology quantum control phase estimation reinforcement learning tomography hamiltonian learning Biamonte, et. al, arXiv: 1611.09347 1

  28. Thank you! @Travis_Sch 0

Recommend


More recommend