NNQS-Transformer: An Efficient and Scalable Neural Network Quantum States Approach for Ab Initio Quantum Chemistry
Author: Yangjun Wu; Chu Guo; Yi Fan; Pengyu Zhou; Honghui Shang
@inproceedings{SC23_nnqs,
abstract = {Neural network quantum state (NNQS) has emerged as a promising candidate for quantum many-body problems, but its practical applications are often hindered by the high cost of sampling and local energy calculation. We develop a high-performance NNQS method for ab initio electronic structure calculations. The major innovations include: (1) A transformer based architecture as the quantum wave function ansatz; (2) A data-centric parallelization scheme for the variational Monte Carlo (VMC) algorithm which preserves data locality and well adapts for different computing architectures; (3) A parallel batch sampling strategy which reduces the sampling cost and achieves good load balance; (4) A parallel local energy evaluation scheme which is both memory and computationally efficient; (5) Study of real chemical systems demonstrates both the superior accuracy of our method compared to state-of-the-art and the strong and weak scalability for large molecular systems with up to 120 spin orbitals.},
address = {New York, NY, USA},
articleno = {42},
author = {Wu, Yangjun and Guo, Chu and Fan, Yi and Zhou, Pengyu and Shang, Honghui},
booktitle = {Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis},
doi = {10.1145/3581784.3607061},
isbn = {9798400701092},
keywords = {many-body schrödinger equation, neural network quantum state, quantum chemistry, autoregressive sampling, transformer based architecture},
location = {Denver, CO, USA},
numpages = {13},
publisher = {Association for Computing Machinery},
series = {SC '23},
title = {NNQS-Transformer: An Efficient and Scalable Neural Network Quantum States Approach for Ab Initio Quantum Chemistry},
url = {https://doi.org/10.1145/3581784.3607061},
year = {2023}
}