Transferable neural wavefunctions for solids
File(s)s43588-025-00872-z.pdf (1.62 MB)
Published online version
Author(s)
Gerard, L
Scherbela, M
Sutterud, H
Foulkes, WMC
Grohs, P
Type
Journal Article
Abstract
Deep-learning-based variational Monte Carlo has emerged as a highly
accurate method for solving the many-electron Schrödinger
equation. Despite favorable scaling with the number of electrons, 𝒪𝒪𝒪nel
4),
the practical value of deep-learning-based variational Monte Carlo is limited
by the high cost of optimizing the neural network weights for every system
studied. Recent research has proposed optimizing a single neural network
across multiple systems, reducing the cost per system. Here we extend this
approach to solids, which require numerous calculations across different
geometries, boundary conditions and supercell sizes. We demonstrate that
optimization of a single ansatz across these variations significantly reduces
optimization steps. Furthermore, we successfully transfer a network trained
on 2 × 2 × 2 supercells of LiH, to 3 × 3 × 3 supercells, reducing the number of
optimization steps required to simulate the large system by a factor of 50
compared with previous work.
accurate method for solving the many-electron Schrödinger
equation. Despite favorable scaling with the number of electrons, 𝒪𝒪𝒪nel
4),
the practical value of deep-learning-based variational Monte Carlo is limited
by the high cost of optimizing the neural network weights for every system
studied. Recent research has proposed optimizing a single neural network
across multiple systems, reducing the cost per system. Here we extend this
approach to solids, which require numerous calculations across different
geometries, boundary conditions and supercell sizes. We demonstrate that
optimization of a single ansatz across these variations significantly reduces
optimization steps. Furthermore, we successfully transfer a network trained
on 2 × 2 × 2 supercells of LiH, to 3 × 3 × 3 supercells, reducing the number of
optimization steps required to simulate the large system by a factor of 50
compared with previous work.
Date Issued
2025-10-22
Date Acceptance
2025-08-20
Citation
Nature Computational Science, 2025
ISSN
2662-8457
Publisher
Nature Research
Journal / Book Title
Nature Computational Science
Copyright Statement
© The Author(s) 2025 Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
License URL
Publication Status
Published online
Date Publish Online
2025-10-22