Hi all,
I'm a PhD student in #MachineLearning at the technical university of Munich #TUM. I'm currently working on machine learning on graphs and machine learning-driven computional chemistry.
#ml #GraphNeuralNetworks #GNNs #compchem
Hi all,
I'm a PhD student in #MachineLearning at the technical university of Munich #TUM. I'm currently working on machine learning on graphs and machine learning-driven computional chemistry.
#ml #GraphNeuralNetworks #GNNs #compchem
@rastinza In our work, we found that by training a neural wave function on multiple geometries we do not sacrifice any performance but only have to train a single model reducing training times significantly.
While there is still work to be done to scale neural wave functions, we believe this is an important step in reducing their computational cost.
@rastinza we in fact perform highly accurate ab-initio calculations. In many small system, such ML-driven ab-initio methods report the lowest variational results in the literature. What we mean with "not sacrificing performance" is that compared to neural wave function-based baselines we don't lose any accuracy despite solving many Schrödinger equations simultaneously.
The plot below shows our Potential Energy Surface Network (PESNet) in comparison to other neural wave function-based methods.
Recently, it has been shown that neural networks not only approximate the ground-state wave functions of a single molecular system well but can also generalize to multiple geometries. While such generalization significantly speeds up training, each energy evaluation still requires Monte Carlo integration which limits the evaluation to a few geometries. In this work, we address the inference shortcomings by proposing the Potential learning from ab-initio Networks (PlaNet) framework, in which we simultaneously train a surrogate model in addition to the neural wave function. At inference time, the surrogate avoids expensive Monte-Carlo integration by directly estimating the energy, accelerating the process from hours to milliseconds. In this way, we can accurately model high-resolution multi-dimensional energy surfaces for larger systems that previously were unobtainable via neural wave functions. Finally, we explore an additional inductive bias by introducing physically-motivated restricted neural wave function models. We implement such a function with several additional improvements in the new PESNet++ model. In our experimental evaluation, PlaNet accelerates inference by 7 orders of magnitude for larger molecules like ethanol while preserving accuracy. Compared to previous energy surface networks, PESNet++ reduces energy errors by up to 74%.