Core Concepts
Retentive networks (RetNets) offer a computationally more efficient alternative to transformers as ansatz networks for neural quantum states (NQS) in ab initio quantum chemistry calculations, achieving comparable accuracy while significantly reducing time complexity, especially for larger systems.
Stats
The authors provide a threshold ratio of problem-to-model size (n_seq > 1.75 * d_model) past which the RetNet's inference time complexity surpasses that of the transformer.
The paper presents ground state energy calculations (in Hartree) for various molecules (H2O, N2, O2, H2S, PH3, LiCl, Li2O) using RetNets, transformers, MADE, CCSD, and FCI, demonstrating comparable accuracy between the neural network-based methods and established techniques.
Quotes
"Unlike transformers, RetNets overcome this time complexity bottleneck by processing data in parallel during training, and recurrently during inference."
"Our findings support the RetNet as a means of improving the time complexity of NQS without sacrificing accuracy."