Predefined attention-focused mechanism using center-environment features: A machine learning study of alloying effects on stability of Nb5Si3 alloys
Abstract
Digital encoding of material structures using graph-based features combined with deep neural networks often lacks local specificity. Additionally, incorporating a self-attention mechanism increases architectural complexity and demands extensive data. To overcome these challenges, we developed a Center-Environment (CE) feature representation—a less data-intensive, physics-informed predefined attention mechanism. The pre-attention mechanism underlying the CE model shifts attention from complex black-box machine learning (ML) algorithms to explicit feature models with physical meaning, reducing data requirements while enhancing the transparency and interpretability of ML models. This CE-based ML approach was employed to investigate the alloying effects on the structural stability of Nb5Si3, with the objective of guiding data-driven compositional design for ultra-high-temperature NbSi superalloys. The CE features leveraged the Atomic Environment Type (AET) method to characterize the local low-symmetry physical environments of atoms. The optimized CEAET models reasonably predicted double-site substitution energies in α-Nb5Si3, achieving a mean absolute error (MAE) of 329.43 meV/cell. The robust transferability of the CEAET models was demonstrated by their successful prediction of untrained β-Nb5Si3 structures. Site occupancy preferences were identified for B, Si, and Al at Si sites, and for Ti, Hf, and Zr at Nb sites within β-Nb5Si3. This CE-based ML approach represents a broadly applicable and intelligent computational design method, capable of handling complex crystal structures with strong transferability, even when working with small datasets.
- This article is part of the themed collection: 2023 and 2024 Accelerate Conferences