Beyond training data: how elemental features enhance ML-based formation energy predictions

Abstract

Quantum mechanics (QM) based modeling allows for accurate prediction of molecular and atomic interactions, enabling simulations of many materials and chemical properties. However, the high computational cost of QM models leads to a need for faster computational methods to study atomic-scale interactions. Graph Neural Networks fit to QM calculations have been used as a computationally efficient alternative to QM. Still, generalization to diverse unseen compounds is challenging due to the many possible chemistries and structures. In this work, we demonstrate the effectiveness of utilizing element features in facilitating generalization to compounds containing completely new elements in the dataset. Our findings show that we can even randomly exclude up to ten percent of the elements from the dataset without significantly compromising the model's performance.

Graphical abstract: Beyond training data: how elemental features enhance ML-based formation energy predictions

Supplementary files

Article information

Article type
Paper
Submitted
05 May 2025
Accepted
16 Aug 2025
First published
04 Sep 2025
This article is Open Access
Creative Commons BY license

Digital Discovery, 2025, Advance Article

Beyond training data: how elemental features enhance ML-based formation energy predictions

H. Mahdavi, V. Honavar and D. Morgan, Digital Discovery, 2025, Advance Article , DOI: 10.1039/D5DD00182J

This article is licensed under a Creative Commons Attribution 3.0 Unported Licence. You can use material from this article in other publications without requesting further permissions from the RSC, provided that the correct acknowledgement is given.

Read more about how to correctly acknowledge RSC content.

Social activity

Spotlight

Advertisements