Neural-network architectures have been increasingly used to represent quantum many-body wave functions. These networks require a large number of variational parameters and are challenging to optimize using traditional methods, as gradient descent. Stochastic reconfiguration (SR) has been effective with a limited number of parameters, but becomes impractical beyond a few thousand parameters. Here, we leverage a simple linear algebra identity to show that SR can be employed even in the deep learning scenario. We demonstrate the effectiveness of our method by optimizing a Deep Transformer architecture with 3 x 105 parameters, achieving state-of-the-art ground-state energy in the J1-J2 Heisenberg model at J2/J1 = 0.5 on the 10 x 10 square lattice, a challenging benchmark in highly-frustrated magnetism. This work marks a significant step forward in the scalability and efficiency of SR for neural-network quantum states, making them a promising method to investigate unknown quantum phases of matter, where other methods struggle.Stochastic reconfiguration (SR) is a method to enhance the optimization of variational functions in quantum many-body systems but becomes impractical beyond a few thousand parameters. Here, the authors combine an alternative SR formulation with a Deep Transformer architecture with 105 parameters, achieving the best variational energy for the J1-J2 Heisenberg model on the 10 x 10 square lattice

A simple linear algebra identity to optimize large-scale neural network quantum states / Rende, Riccardo; Viteritti, Luciano Loris; Bardone, Lorenzo; Becca, Federico; Goldt, Sebastian. - In: COMMUNICATIONS PHYSICS. - ISSN 2399-3650. - 7:1(2024), pp. 1-8. [10.1038/s42005-024-01732-4]

A simple linear algebra identity to optimize large-scale neural network quantum states

Rende, Riccardo;Bardone, Lorenzo;Becca, Federico;Goldt, Sebastian
2024-01-01

Abstract

Neural-network architectures have been increasingly used to represent quantum many-body wave functions. These networks require a large number of variational parameters and are challenging to optimize using traditional methods, as gradient descent. Stochastic reconfiguration (SR) has been effective with a limited number of parameters, but becomes impractical beyond a few thousand parameters. Here, we leverage a simple linear algebra identity to show that SR can be employed even in the deep learning scenario. We demonstrate the effectiveness of our method by optimizing a Deep Transformer architecture with 3 x 105 parameters, achieving state-of-the-art ground-state energy in the J1-J2 Heisenberg model at J2/J1 = 0.5 on the 10 x 10 square lattice, a challenging benchmark in highly-frustrated magnetism. This work marks a significant step forward in the scalability and efficiency of SR for neural-network quantum states, making them a promising method to investigate unknown quantum phases of matter, where other methods struggle.Stochastic reconfiguration (SR) is a method to enhance the optimization of variational functions in quantum many-body systems but becomes impractical beyond a few thousand parameters. Here, the authors combine an alternative SR formulation with a Deep Transformer architecture with 105 parameters, achieving the best variational energy for the J1-J2 Heisenberg model on the 10 x 10 square lattice
2024
7
1
1
8
260
https://doi.org/10.1038/s42005-024-01732-4
https://arxiv.org/abs/2310.05715
Rende, Riccardo; Viteritti, Luciano Loris; Bardone, Lorenzo; Becca, Federico; Goldt, Sebastian
File in questo prodotto:
File Dimensione Formato  
s42005-024-01732-4.pdf

accesso aperto

Licenza: Creative commons
Dimensione 846.98 kB
Formato Adobe PDF
846.98 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.11767/143210
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? 3
social impact