Neural-network architectures have been increasingly used to represent quantum many-body wave functions. These networks require a large number of variational parameters and are challenging to optimize using traditional methods, as gradient descent. Stochastic reconfiguration (SR) has been effective with a limited number of parameters, but becomes impractical beyond a few thousand parameters. Here, we leverage a simple linear algebra identity to show that SR can be employed even in the deep learning scenario. We demonstrate the effectiveness of our method by optimizing a Deep Transformer architecture with 3 x 105 parameters, achieving state-of-the-art ground-state energy in the J1-J2 Heisenberg model at J2/J1 = 0.5 on the 10 x 10 square lattice, a challenging benchmark in highly-frustrated magnetism. This work marks a significant step forward in the scalability and efficiency of SR for neural-network quantum states, making them a promising method to investigate unknown quantum phases of matter, where other methods struggle.Stochastic reconfiguration (SR) is a method to enhance the optimization of variational functions in quantum many-body systems but becomes impractical beyond a few thousand parameters. Here, the authors combine an alternative SR formulation with a Deep Transformer architecture with 105 parameters, achieving the best variational energy for the J1-J2 Heisenberg model on the 10 x 10 square lattice
A simple linear algebra identity to optimize large-scale neural network quantum states / Rende, Riccardo; Viteritti, Luciano Loris; Bardone, Lorenzo; Becca, Federico; Goldt, Sebastian. - In: COMMUNICATIONS PHYSICS. - ISSN 2399-3650. - 7:1(2024), pp. 1-8. [10.1038/s42005-024-01732-4]
A simple linear algebra identity to optimize large-scale neural network quantum states
Rende, Riccardo;Bardone, Lorenzo;Becca, Federico;Goldt, Sebastian
2024-01-01
Abstract
Neural-network architectures have been increasingly used to represent quantum many-body wave functions. These networks require a large number of variational parameters and are challenging to optimize using traditional methods, as gradient descent. Stochastic reconfiguration (SR) has been effective with a limited number of parameters, but becomes impractical beyond a few thousand parameters. Here, we leverage a simple linear algebra identity to show that SR can be employed even in the deep learning scenario. We demonstrate the effectiveness of our method by optimizing a Deep Transformer architecture with 3 x 105 parameters, achieving state-of-the-art ground-state energy in the J1-J2 Heisenberg model at J2/J1 = 0.5 on the 10 x 10 square lattice, a challenging benchmark in highly-frustrated magnetism. This work marks a significant step forward in the scalability and efficiency of SR for neural-network quantum states, making them a promising method to investigate unknown quantum phases of matter, where other methods struggle.Stochastic reconfiguration (SR) is a method to enhance the optimization of variational functions in quantum many-body systems but becomes impractical beyond a few thousand parameters. Here, the authors combine an alternative SR formulation with a Deep Transformer architecture with 105 parameters, achieving the best variational energy for the J1-J2 Heisenberg model on the 10 x 10 square latticeFile | Dimensione | Formato | |
---|---|---|---|
s42005-024-01732-4.pdf
accesso aperto
Licenza:
Creative commons
Dimensione
846.98 kB
Formato
Adobe PDF
|
846.98 kB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.