Open Access Article
Daniel
Schwalbe-Koda
*abd,
Nitish
Govindarajan
*acd and
Joel B.
Varley
ad
aMaterials Science Division, Lawrence Livermore National Laboratory, Livermore, CA 94550, USA
bDepartment of Materials Science and Engineering, University of California, Los Angeles, Los Angeles, CA 90095, USA. E-mail: dskoda@ucla.edu
cSchool of Chemistry, Chemical Engineering, and Biotechnology, Nanyang Technological University, 21 Nanyang Link, Singapore 637371, Singapore. E-mail: nitish.govindarajan@ntu.edu.sg
dLaboratory for Energy Applications for the Future (LEAF), Lawrence Livermore National Laboratory, Livermore, CA 94550, USA
First published on 9th December 2024
Sampling high-coverage configurations and predicting adsorbate–adsorbate interactions on surfaces are highly relevant to understand realistic interfaces in heterogeneous catalysis. However, the combinatorial explosion in the number of adsorbate configurations among diverse site environments presents a considerable challenge in accurately estimating these interactions. Here, we propose a strategy combining high-throughput simulation pipelines and a neural network-based model with the MACE architecture to increase sampling efficiency and speed. By training the models on unrelaxed structures and energies, which can be quickly obtained from single-point DFT calculations, we achieve excellent performance for both in-domain and out-of-domain predictions, including generalization to different facets, coverage regimes and low-energy configurations. From this systematic understanding of model robustness, we exhaustively sample the configuration phase space of catalytic systems without active learning. In particular, by predicting binding energies for over 14 million structures within the neural network model and the simulated annealing method, we predict coverage-dependent adsorption energies for CO adsorption on six Cu facets (111, 100, 211, 331, 410 and 711) and the co-adsorption of CO and CHOH on Rh(111). When validated by targeted post-sampling relaxations, our results for CO on Cu correctly reproduce experimental interaction energies reported in the literature, and provide atomistic insights on the site occupancy of steps and terraces for the six facets at all coverage regimes. Additionally, the arrangement of CO on the Rh(111) surface is demonstrated to substantially impact the activation barriers for the CHOH bond scission, illustrating the importance of comprehensive sampling on reaction kinetics. Our findings demonstrate that simplified data generation routines and evaluating generalization of neural networks can be deployed at scale to understand lateral interactions on surfaces, paving the way towards realistic modeling of heterogeneous catalytic processes.
Given the multitude of site types and the combinatorial increase in the number of configurations with adsorbate coverage, exhaustive sampling and estimation of coverage-dependent adsorption energies are intractable using DFT simulations. This is especially true when larger supercell sizes are used to quantify coverage effects at longer length scales, and remain challenging even for small adsorbates with few degrees of freedom, such as *H or *CO. As a result, surrogate models bootstrapped from DFT simulations have been used to explore the large configurational space to estimate coverage-dependent adsorption energies.12 Examples of such surrogate models include: (1) analytical relationships based on first or higher order lateral interaction models,13–15 (2) cluster expansion based approaches that include a sum of on-site, two- and higher-body interactions,8,16–19 and (3) machine learning (ML) based approaches, including neural networks (NNs), that predict total adsorption energies from atom-centered descriptors.20–23 Each of these strategies has advantages and limitations. For instance, analytical expressions that are often used to account for lateral interactions in mean-field microkinetic models are computationally efficient, but rely on sampling of limited configurations at discrete coverages and use interpolated functional forms (e.g., piecewise-linear)13 to estimate coverage-dependent adsorption energies. As spatial correlations are not accounted for in such models, they can lead to large errors in the prediction of surface coverages and catalytic activity.24 Cluster expansion approaches explicitly account for spatial correlations between adsorbates and have been extensively used in lattice based kinetic Monte Carlo (kMC) simulations.25 However, cluster expansion requires selecting (small) discrete clusters whose complexity increases exponentially with number of adsorbates and active sites. This also hinders the transferability of these models beyond a single (low symmetry) facet type, particularly to higher index facets that have more complex active site environments. Finally, ML-based approaches are data-intensive and require careful analysis to avoid generalization failures when used beyond their training datasets. As such, these methods rely on active learning to augment training sets whenever novel coverage configurations are found,21 or on incremental model training by stepwise generation of coverage configurations.22 When using relaxed configurations obtained from DFT-based geometry optimizations, ML approaches require larger computational resources to accurately sample the entire configuration space of interest, including high adsorbate coverages, co-adsorption, and low symmetry surfaces with diverse site environments.
In this work, we introduce a fast and scalable data pipeline for quantifying coverage effects on catalyst surfaces using high-throughput workflows and ML, with an emphasis on generalization and transferability. In particular, we use the MACE architecture,26 based on many-body message passing neural networks exhibiting good generalization behavior,26,27 to predict coverage dependent adsorption energies with high accuracy for both in- and out-of-domain tasks. Using binding energies for unrelaxed structures as training data, MACE models exhibit high accuracy across different facets, coverage regimes, and thoroughly sample combinatorial spaces of coverage configurations without retraining the model. By combining the NN model with workflow management implemented in mkite,28 we sampled over 14 million configurations of CO adsorbed on six Cu facets (111, 100, 211, 331, 410 and 711) with varying coordination environments, which are systems of relevance for electrochemical CO2 reduction.29 Binding energies of the most stable structures, as predicted with the NN model, are computed with single-point DFT calculations and demonstrated to agree nearly perfectly with the predictions. Our approach is further validated using over a thousand structural relaxations, from which differential binding curves and trends are obtained for coverage-dependent *CO adsorption on the six Cu facets. Finally, we demonstrate that our pipeline is also applicable to co-adsorption systems using the *CHOH intermediate on CO-covered Rh(111), an active catalyst for thermal CO hydrogenation, as a case study.15 This combination of simpler data generation and understanding of NN generalization demonstrates how new ML pipelines can be developed and employed to model increasingly complex interfaces in heterogeneous catalysis at scale.
To efficiently generate data for combinatorial adsorption spaces while also avoiding active learning loops, we propose to: (1) train a model on unrelaxed energies (i.e., single-point DFT calculations) from randomly sampled configurations, and (2) extend beyond this limited space by assessing the generalization behavior of the ML models. Because sampling the space of relaxed structures with ML models is much more expensive and less reliable than that of unrelaxed structures, step (1) reduces the computational cost associated with performing structural relaxations when creating training data. As a consequence, more data can be generated to train the ML models at a reduced computational expense. A second benefit of our approach is bypassing the need for active learning approaches, simplifying the simulation workflows. Fig. 1b summarizes the data generation strategy for this work, and the complete relationship between workflow objects implemented in mkite is shown on Fig. S3.† Instead of enumerating configurations, we randomly sampled configurations on a per-facet, per-coverage basis (see the Methods section for details). In particular, we generated up to 100 unique configurations for each facet and coverage where nCO < 10 (low coverage regime) and 50 unique configurations per facet and coverage where 10 ≤ nCO ≤ 18 (high coverage regime). These configurations were generated by randomly sampling adsorption sites that avoid distances smaller than 2.0 Å (1.7 Å) between adsorbates in the low (high) coverage regime, then deduplicating symmetrically equivalent structures (see Methods). This resulted in 6793 configurations for the different CO coverages and Cu facets considered in this work (Table S3†). Then, single-point DFT simulations are used to compute unrelaxed binding energies of *CO adsorbed on the Cu facets of interest. These randomly sampled distributions span a wide range of energies, as exemplified in Fig. 1c for low and high *CO coverages on Cu(111), and detailed in Fig. S4 and S5.† Already from the randomly generated data, known trends from the sampling outcomes can be observed. For example, at higher coverages, average binding energies tend to shift towards higher energies in all facets (Fig. S4†), although with different magnitudes in each facet due to different binding sites.
Because randomly sampled datasets are often not representative of the lowest-energy structures, quantitative analysis of the physical phenomena cannot be performed until the configuration space is thoroughly sampled. To evaluate how models generalize beyond these datasets, we proposed five tests to assess the performance of NN models beyond their training domain (Fig. 1d). In addition to in-domain train-test splits, additional errors metrics can be obtained by: training the model on data from one facet and testing it on data from another facet; training the model on lower coverages of a single facet and testing it on higher coverages of the same facet; or omitting certain coverages from the training set. These tests can provide additional information to assess generalization behavior, thus helping decide if a model can reliably obtain low-energy structures despite being trained on a higher-energy dataset.
To explain the facet dependence of the aforementioned errors, we first evaluated whether models are indeed performing in out-of-distribution assumptions. Given that binding sites in higher-index facets can be represented using binding sites from low-index facets,40 it is relevant to quantify the extent of their similarity from an ML standpoint, thus beyond the microfacet notation. To perform this analysis independently of the learned features, we represented the adsorption sites of all facets using a fixed descriptor, namely the Smooth Overlap of Atomic Positions (SOAP)41 (see Methods). Then, by performing all pairwise comparisons between adsorption sites across all facets, we obtained a measure of similarity between the sets of unique adsorption sites. Because the resulting similarity between two facets corresponds to a matrix, we compute the Hausdorff distance between sites, which is a measure of how far is the furthest point of the test set compared to all points in the train set. Finally, the similarities are then compared based on their percentile within the distribution of all similarities. These results are shown in Fig. 2b. Although a qualitative visualization can help visualize the richness of the binding space (Fig. S7†), the quantitative analysis in Fig. 2b further confirms that adsorption sites from low-index facets are indeed contained in all stepped surfaces from an ML perspective, as shown by the higher percentiles in the columns of Cu(100) and Cu(111). Interestingly, however, adsorption sites from Cu(100) are more similar to those in Cu(111) than the other way around, explaining the asymmetry of generalization errors between Cu(100) and Cu(111). Similar patterns can be observed across Cu(211), Cu(331), and Cu(711). In fact, when the RMSE in Fig. 2a is visualized against the similarity in Fig. 2b, reasonable correlations are found between the results, as shown in Fig. 2c. The Pearson correlation coefficient between the two quantities ranges from −0.61 (for Cu(100) as the test set) to −0.95 (for Cu(331) as the test set). Errors scale similarly across facets, with the exception of Cu(410). Its RMSE values remain consistently lower than its counterparts when used as a test set, which could be an artifact of lower coverages when the number of *CO adsorbates is normalized by the number of surface sites. Indeed, the distribution of binding energies for Cu(410) is closer to those from the (100, 111, 211) facets than to those observed in the lower-symmetry counterparts (331, 711) (Fig. S5†). This suggests that, in addition to similarity in binding sites, the performance of the models is also connected to the coverage and density of adsorbates. Importantly, it also emphasizes that, from a machine learning perspective, the results do not correspond plainly to an ill-defined interpolation. Rather, the test data is far away from the entire training set and thus is not guaranteed to be within reasonable ranges of interpolation. While these absolute thresholds are challenging to determine in a universal way, the analysis shows that the model performance is strongly related to the surrogate metric of “extrapolation” shown in Fig. 2.
To test whether generalization across number of adsorbates plays a substantial role in the models, we compared models trained with different coverage regimes (Fig. 1d). The results of the comparison are shown in Fig. 3a. Errors of models trained and tested on all coverages (dark blue in 3a) are equivalent to those in Fig. 2a, and serve as a baseline for this test. First, we trained per-facet models on subsets of the data by adding all even coverages to the training set and all odd coverages in the test set. The results, depicted in Fig. 3a as “different covs.” (cyan), show that slightly worse performances are found despite half of the coverages being discarded. Fig. S8† further shows that, beyond the error values, correlation coefficients also remain nearly perfect for the test set predictions. This observation implies not only that the model can properly interpolate within the space of coverages, but also that future dataset construction can be optimized by removing some intermediate coverages. This approach could drastically reduce the total number of single-point DFT calculations to be performed when generating the initial training set. Secondly, we verified whether models trained on low-coverage structures (nCO < 10) could accurately predict binding energies from high-coverage structures (nCO ≥ 10). In a previous work,22 this approach was performed in a step-wise manner to ensure the reliability of predictions, but coupling step-by-step exploration and model retraining can be time-consuming. Ideally, a model with high generalization capacity could bypass the need for sequential data generation and predict directly the binding energies for higher coverage regimes. Fig. 3a shows that, although the RMSE degrades when models are tested on substantially higher coverages, errors remain small compared to the baseline (red bars). Even in the worst case in Fig. 3a, obtained with models trained on low-coverage structures for Cu(711), correlation coefficients for the predictions remain nearly perfect (Fig. S9†). This remarkable performance suggests that accurate models of lateral interactions are being learned within the MACE models, and offer reliable predictions outside of their training domain. Moreover, the lower error of Cu(410) compared to other stepped facets further confirms the previous hypothesis that lateral interaction models between *CO adsorbates are either easier to fit within Cu(410) compared to Cu(211), Cu(331), and Cu(711), or more limited in scope given the larger lateral sizes of the Cu(410) supercells under study (Table S1†).
Given the success of the MACE models in predicting binding energies beyond their training domain, we evaluated whether a single model could be created for all facets and all coverages at once without loss of accuracy. This would suggest that the MACE models are able to capture all binding energies in a single model, facilitating and increasing the convenience of sampling new structures with a single model. Fig. 3c shows a breakdown of the test RMSE values for the MACE model trained on the entire dataset (all facets and coverages) at once. Errors for individual facets are equivalent to the in-domain RMSE values shown in Fig. 2a, with deviations on the order of 1 meV per CO. Across all facets and coverages, the RMSE is found to be uniform and generally smaller than 20 meV per CO. This further confirms the ability of the models to represent the entire configuration space of *CO adsorbates on Cu within different facets and coverages.
To better understand whether this performance could have been achieved without the MACE models, we systematically decreased the complexity of the architectures and models themselves. Because of the diversity in adsorption sites for Cu(711) (see Fig. 2b), we selected this facet as the training set for this study. First, we reduced the number of body-order correlations (ν) in MACE to understand whether complex many-body interactions can substantially affect the errors. Fig. 3c shows that small test error differences are found for in-domain data when 3-body (ν = 2) and 4-body messages (ν = 3) are used, but the performance degrades more substantially when the model is restricted to use 2-body (ν = 1) messages. These observations are similar to those in ref. 26. The trends in Fig. 3c also remain constant across smaller dataset sizes, and illustrate that the representation capacity of the models are likely limited by the interaction order. To further test this result, we verified whether simpler models can obtain comparable performance to MACE (Fig. 3d). For instance, we trained a NN on the SOAP fingerprint centered on the adsorbates, as well as a plain SchNet model,42 which is similar to the recently proposed ACE-GCN,22 to predict binding energies of all Cu facets after being trained on Cu(711) configurations. Although the in-domain performance of both models were reasonable, with 41 and 29 meV per CO RMSE when the models were tested on held-out Cu(711) data, the average RMSE for out-of-domain test sets in both systems was much higher, at 201 and 174 meV per CO, compared to an average RMSE of 112 meV per CO for the MACE model (see also Fig. S11†). Following the clue from Fig. 3c on the role of many-body interactions, we tested a simple model that explicitly accounts for lateral interactions when computing the binding energy. The simplified model, which decomposes the binding energy into site-centered contributions and lateral interactions (see ESI Text and Fig. S10†), exhibits a substantial improvement over SchNet in terms of generalization behavior (Fig. 3d and S11†) despite using a fixed descriptor. Remarkably, the simplified NN model shows a smaller error when generalizing from Cu(711) to Cu(211) or Cu(331) than MACE (Fig. S11†), but a larger error for the other facets. We hypothesize that this effect reflects the larger similarity between the SOAP vectors of adsorption sites from Cu(711) in the training set and the ones from Cu(211) and Cu(331) in the test set, as shown in Fig. 2b. On the other hand, this observation also indicates that while learnable representations in MACE effectively capture environments beyond those in the Cu(711) training set, it slightly compromises the ability to understand other specific facets despite their similarity (Fig. 2b). Furthermore, this behavior may not be exclusive to the MACE architecture and can be likely achieved by other models demonstrating near state-of-the-art performance in recent benchmarks.43,44 Nevertheless, this example further demonstrates that improved generalization behavior in coverage-dependent interactions can be improved by the priors in the NN architecture, and thus are not entirely explained by correlations in the dataset of *CO on Cu facets.
To extensively sample the configuration space of nCO on the six Cu facets, we used a Markov chain Monte Carlo (MCMC) sampling strategy within the Metropolis–Hastings45,46 and simulated annealing47 algorithms (see Methods). An example of the temperature and energy profiles from the MCMC sampling for Cu(410) with nCO = 10 is shown in Fig. 4a. By taking advantage of batched evaluation of the NN model, we sampled 1000 replicas in parallel for each nCO and facet, resulting in over 14.5 million energy evaluations to sample the lowest energy configurations (see Fig. S12 and S13†). The sampling method systematically sampled low energy configurations despite starting from random structures (Fig. S14†), demonstrating the success of the sampling approach. To validate whether the model predictions were accurate and the low energy structures are not generalization failures, we ranked the structures by energy and selected the top-3 unique configurations per facet and per coverage for DFT evaluation. After performing 312 single-point calculations for these systems, we compared their unrelaxed binding energies against the original dataset. Fig. 4b shows that the DFT-calculated binding energies of MCMC-sampled structures (red) are substantially lower in energy than the original dataset (dark blue), as expected. This shift towards lower DFT energies is also visualized on the total, per-facet distributions of energies (Fig. 4c). Therefore, the MACE model trained on all facets was effective in sampling low-energy configurations without active learning loops and despite being trained on a dataset of randomly sampled structures. When the performance of the model is assessed by comparing the predicted binding energies for the top sampled structures and the resulting DFT values, a near-perfect correlation is found (Fig. 4d). Although the RMSE of predictions (91 meV per CO) is higher than the baseline test errors (<15 meV per CO), the Spearman's correlation coefficient is equal to 0.999. This indicates that the model predicts shifted energies with respect to the ground truth due to a skewed training set, but reproduces all correlations in the underlying data. The strategy also enabled sampling higher coverage configurations (nCO > 18) for the Cu(410) case which were initially absent from the training set (see Fig. 4b), further supporting the model's ability to generalize towards higher coverages in a production analysis.
To evaluate correlations between relaxed and unrelaxed energies, we performed DFT relaxations for 675 configurations from the original (thus randomly sampled) dataset across all facets and coverage regimes. These data points were obtained by attempting to relax a subset of nearly 2000 configurations whose unrelaxed energies were lower than its counterparts on a per-facet, per-coverage basis. However, not all structural relaxations led to systems where all *CO remained adsorbed on the Cu facet. Most of the relaxation calculations for the original dataset had to be stopped due to desorption of *CO from the facet, or large wall times expected for job completion on the high-performance computing cluster. Therefore, the final 675 relaxed structures and energies only represent the subset of successful relaxations. Although this is an inherent bias in the data—e.g., systems for which there is desorption have a single-point energy in the training set, but not relaxed energies—it still represents the subspace of interest, which is that of structures with faster DFT relaxations and closer to the local optima. In this sense, the absence of high-energy structures (including desorbed ones) is exactly the phase space of interest and does not affect the conclusions regarding the predicted desorption energies. Aiming to sample this space allows us to focus on the effect of low-energy configurations of *CO on Cu facets in their bound states.
First, we verified if relaxed binding energies correlate with their unrelaxed counterparts. If the energies are completely uncorrelated, this would suggest that sampling unrelaxed structures is not a good strategy for exploring the space of relaxed configurations. Fig. 5a illustrates this correlation for Cu(410) (see Fig. S15† for all facets). The Spearman's correlation coefficient (ρS) between relaxed and unrelaxed binding energies of *CO on Cu(410) is 0.58, suggesting that lower unrelaxed energies usually lead to systems with lower relaxed energies. Similar trends are found among other facets, with most values of ρS higher than 0.5. The only exception is Cu(711), for which fewer data points are available, with a correlation coefficient of 0.350. On the other hand, the best correlation coefficient found for Cu(211) (ρS = 0.739). To quantify whether relaxing the lowest unrelaxed energies often leads to the lowest relaxed binding energies, we computed the recall of the top-N relaxed energies given the top-N unrelaxed ones. Fig. 5b exemplifies this recall for Cu(410) for three different cases and compares it to the expected recall for a baseline, random sampling method. The recall is always better than the baseline when the top-5, top-10, and top-15 (un)relaxed structures are considered, exhibiting a higher area under the recall curve and quick rise with low percentiles of unrelaxed energies. Similar trends are observed for other facets (Fig. S16†), with particularly successful results for Cu(100), Cu(211), Cu(331), and Cu(711). The case of Cu(111) shows a mismatch between the structures with lowest unrelaxed energies and relaxed ones, as also seen in Fig. S15.† For some coverages, the best configurations of *CO on Cu do not have the top unrelaxed energies, but intermediate values that require further sampling of the unrelaxed configurations to obtain the lowest energy configurations of the original dataset.
Given the reasonable agreement between unrelaxed and relaxed energies from the recall curves, we verified whether relaxed energies could be inferred directly from unrelaxed ones on a per-facet, per-coverage basis. To do so, we analyzed the distribution of relaxed binding energies across all systems under study. Fig. S17 and S18† show that the range of relaxed binding energies is much smaller than the range of unrelaxed binding energies in the original dataset (compare with Fig. S4†). In fact, even when systems with different unrelaxed energies are selected, they often collapse to very similar relaxed energies, as shown by the small standard deviation from Fig. S4.† Because structural relaxations may substantially rearrange the position of the adsorbates on the surface, structures with different unrelaxed energies may relax towards similar structures, with the difference between unrelaxed energies compensated by sufficient structural reconstruction. To verify this hypothesis, we compared the relaxation energy difference (ΔEb = E(relax)b − E(unrelax)b) with the unrelaxed binding energies on a per-facet manner. Fig. S19† shows a nearly-linear relationship between ΔEb and the unrelaxed binding energies, suggesting that even across different coverages and facets, improvements in energy due to relaxation are related to the initial unrelaxed binding affinities. This suggests that systems with higher unrelaxed energies may still undergo relaxation to lower energies given enough reconstruction and provided that structural rearrangement is achieved within DFT relaxation trajectories. To further verify this claim, we quantified the structural reconstruction of each trajectory by comparing the initial and final configuration using the Pointwise Distance Distribution (PDD).48 This structural invariant is continuous, guaranteed to recover identical structures, and has been proven useful to compare a range of materials.48,49 Furthermore, it bypasses the need for root mean square deviations (RMSD) along the relaxation trajectory, which can be deceptive depending on the atom assignment, the periodic boundary conditions, and the number of atoms over which the deviations were averaged against. Fig. 5c and S20† further confirm that these differences in binding energy are due to rearrangement of atom positions on the surface, with strong correlations between ΔEb and the PDD distance. As expected, configurations with higher nCO undergo larger restructuring during geometry relaxations (higher PDD distance). Taken together, these results suggest that sampling configurations with low unrelaxed energies can lead to smaller ΔEb and, consequently, may require fewer relaxation steps to converge. Although recalling the ground state configuration cannot be guaranteed within any method, our approach provides an efficient strategy to minimize the significant computational cost associated with structural relaxations.
The previous results show that relaxing structures with low unrelaxed binding energies on a per-facet, per-coverage basis is an efficient way to estimate low-energy relaxed structures without having to exhaustively sample the space of relaxed structures. To evaluate the properties of interest, we performed geometry optimizations for all 312 low-energy structures obtained from the MCMC sampling using DFT calculations (see Methods). When the resulting energies from the sampled dataset are compared with the energies from the original dataset, we found that the sampled structures relax to lower energies than their counterparts (Fig. S21†), further supporting the previous results. The only exception is Cu(111), as expected from the shifted recall curves in Fig. S16.† Using the final low-energy (relaxed) structures allows us to investigate the coverage of *CO on specific adsorption sites for the different facets. For example, the coverage-dependent occupancy of *CO adsorbates on the sites of Cu(410) is shown in Fig. 5d. To avoid labeling the surface atoms with discrete coordination numbers after reconstruction upon geometry optimization, we computed an average coordination number (aCN) for each copper atom on the surface using a sum over neighbors defined from a smooth cutoff (see Methods). Then, for each carbon atom, we compute the average coordination of the binding site by taking the average of aCN of all the neighboring copper atoms. Fig. 5d shows with a red bar the aCN of each CO for the lowest energy structure associated with each coverage. At low coverage, *CO adsorbs preferentially on undercoordinated sites that have lower aCN values, specifically the on-top sites on steps (see also Fig. 5e and S22† for a visual guide). After step sites are occupied, lateral repulsion between adsorbates forces the occupancy of higher coordination sites, with a larger preference for terrace sites (aCN < 11) for coverages above 0.25 monolayer (nCO = 6 in our slab model). Then, as the coverage continues to increase, the occupancy of terrace and overcoordinated sites increases, with high preference for on-top sites, until all sites are occupied. This behavior agrees very well with *CO desorption experiments on Cu(410). For example, Makino and Okada50 verified with infrared reflection-adsorption spectroscopy (IRAS) and temperature-programmed desorption (TPD) that the preferential adsorption sites for *CO at low coverages is the on-top site of steps, with their saturation point at ≈0.25 monolayer. Similar behavior is found for all stepped facets considered in this work, where *CO preferentially adsorbed on the undercoordinated sites at low coverages, and structural relaxations tilted the *CO adsorbates away from each other due to lateral interactions. On the other hand, for the flat Cu(100) and Cu(111) surfaces, the relaxation led to geometrical deviation of adsorbates from their initial configuration only at higher coverages (Fig. S23–S28†). Our results show that Cu(100) is preferred over Cu(111) at higher CO coverages, as also proposed from other experimental and computational investigations.35 Additionally, we find that the on-top sites are occupied at low coverages, with the occupancy of the multi-coordinated sites (hollow, bridge) increasing at higher coverages (Fig. S23 and S24†).
To further illustrate how the calculated results can predict the experimental behavior of these systems, we computed the differential binding energy curves of *CO on all facets. By considering all relaxations, we computed the Boltzmann average of relaxed binding energies on a per-coverage, per-facet basis. Then, we fitted a piecewise continuous and differentiable linear-cubic polynomial to the integral binding energy (see Methods). The data points and fitted integral binding energy curve are shown in Fig. S29.† Then, by differentiating the integral binding energy curve with respect to the coverage, we obtained the differential binding energy curves in Fig. 6. A qualitative analysis of Fig. 6 highlights the similar behaviors of *CO adsorption on Cu(100) and Cu(111), given the similar adsorption modes on the flat surfaces (also see Fig. S23 and S24†). However, Cu(100) allows up to 0.7 monolayer of *CO coverage, with lateral repulsions starting at 0.66 monolayer, in contrast to lower thresholds for Cu(111). This finding is in excellent agreement with experimental observations,51 where lateral repulsions (i.e., appearance of a shoulder peak in the TPD spectra) is observed to dominate the adsorption at 0.6 monolayer. The behavior of Cu(711) also shows a large saturation point for *CO, with increasing differential binding energies near 0.5 monolayer. However, due to the nature of the stepped surface (Fig. S28†), *CO adsorption energies are more favorable compared to (100) and (111), shifting the differential binding energies down. In the case of the other stepped surfaces, Cu(211) and Cu(331) have similar adsorption energies at the low coverage regime, but their smaller curvature reflects the less favorable energies of terrace sites occupied at higher coverages compared to undercoordinated steps (Fig. S25 and S26†). Indeed, for Cu(211), it was observed that CO prefers to bind upright on the top sites of the step edge based on DFT calculations and scanning tunneling microscopy (STM) imaging experiments.52 At coverages greater than 0.5ML, the authors note that obtaining the most favorable adsorption geometry was non-trivial, with both top–top and top-bridge configurations being reported in experiments. Both these observations are consistent with the lowest energy configurations we obtain for Cu(211) for different coverages as shown in Fig. S25.† Furthermore, when the differential binding energy of these facets are compared against experimental results, good agreement is found despite the limitations of the RPBE functional in predicting the low-coverage binding preferences and *CO adsorption energies on Cu compared to experiments.53 Cu(100) and Cu(111) have experimental desorption energies at low coverages equal to 530 ± 15 and 490 ± 15 meV per CO,50 respectively, while our computed values are 432 and 440 meV per CO. Cu(211) exhibits an experimental *CO desorption energy of 605 ± 15 meV per CO (attributed to the step-edge),50 similar to our calculated desorption energy of 619 meV per CO at low coverage. Finally, Cu(410) exhibits an experimental desorption energy of 725 ± 31 meV per CO (attributed to the step-edge),50 in excellent agreement with the calculated value of 728 meV per CO from the binding energy curve at low coverage where *CO exclusively occupies the step-edges (Fig. S27†). Taken together, these results further validate the combination of fast data pipelines, ML-accelerated sampling, and selective relaxation when predicting coverage-dependent binding energies for *CO on Cu. The overall agreement of our results with experimental observations offer a roadmap on the extension of these methods towards other catalyst systems—for instance, to accelerate the sampling of potential- and coverage-dependent adsorption energies and properties relevant in electrocatalysis.54
To further exemplify the importance of configurational sampling in modeling reactions where co-adsorption plays a role, we computed the activation barriers for *CH–OH bond scission on Rh(111) in the presence of three different CO configurations with the same coverage of 0.55 monolayer. The barriers were calculated using the nudged elastic band (NEB) method and DFT simulations using CO configurations with initial unrelaxed energies of 0, 150, and 250 meV relative to the lowest energy structure. The results of the NEB calculations are summarized in Fig. 7c (see Fig. S33† for detailed NEB trajectories). Because of the structural relaxations prior to NEB calculations, the energy differences between the three initial systems changed slightly to 0, 72, and 286 meV, but still conserved the energy ranking. All systems are observed to undergo not only the *CH–OH bond scission, but also different restructuring of the adsorbate configurations during the reaction pathway. Because sterical hindrance from *CO does not allow the formation of separate *CH and *OH products, the reaction progresses by first reorganizing the CO into a higher energy state. The energy barrier for this restructuring depends on the initial configuration, with configurations 1 and 3 in Fig. 7c showing barriers of 0.46 and 0.38 eV for this restructuring, leading to intermediate structures less stable than the initial ones. Whereas in system 1 the restructuring is due to CO rearrangement on the surface, in system 3 this increase in energy is associated with the inversion the OH configuration instead (Fig. S34†). After this restructuring, both systems proceed with the CH–OH bond scission on the Rh(111) surface with minor rearrangements of CO adsorbates on the surface. However, even though the energy of restructured configuration 1 is similar to the energy of initial configuration 3 (see also Table S4†), configurations 1 and 3 exhibit energy barriers of 0.85 and 1.98 eV with respect to the intermediate state, respectively (Fig. 7d), demonstrating that the initial arrangement of CO strongly influences the reaction kinetics. As another example, configuration 2 does not show a single barrier with the formation of a more stable intermediate, but the pathway only increases in energy relative to the initial state until the CH–OH bond scission. In our simulations, this led to a concerted CH–OH bond breaking and CO rearrangement, with a single barrier of 1.31 eV along the reaction pathway (see also Fig. S32 and S33†). Because these three systems vary only by the initial *CO arrangement, but exhibit substantially different energetics (both kinetic and thermodynamic) and reaction pathways, this example demonstrates the importance of sampling low-energy configurations and how leveraging generalization in ML models can facilitate this sampling method. Importantly, because exploring combinatorial spaces of co-adsorption can be overly expensive, we showed how data-efficient pipelines can enable a more realistic representation of these and other complex interfaces.
| Eb = EnCO+surf − Esurf − nECO(g), | (1) |
m) was initially relaxed using DFT calculations, and a lattice parameter of 2.60 Å was obtained. Then, surfaces for all copper facets (111, 100, 211, 331, 410 and 711) were generated using the Atomic Simulation Environment,64 and constrained to have at least 6 layers and lateral size of 9 Å. The atomic positions of the two topmost layers were then relaxed using DFT, as described above. A similar procedure was adopted for Rh, but only the Rh(111) facet was considered in our study of co-adsorption. A supercell containing 6 layers and 9 surface sites was created for Rh(111).
recipe in the
package (v 0.1.1). The generator takes in a surface and a single adsorbate as inputs. Then, it identifies all distinct high-symmetry adsorption sites on the surface using the AdsorbateSiteFinder class65 implemented in pymatgen.66 Because fully enumerating all combinations of adsorption sites with n adsorbates is intractable (see Table S2† for rough estimates), we instead randomly sampled n different adsorption sites and checked whether the distances between sampled sites is larger than a given threshold, ensuring there is no overlap between adsorbate positions. For lower coverages (n < 10), this threshold was set to 2.0 Å, while for higher coverages (n ≥ 10) this threshold was set to 1.7 Å. The sampling and checking procedure is repeated until a desired number of configurations is achieved or if a maximum number of attempts is reached. In our case, we sampled up to 100 valid configurations for lower coverages (n < 10) and 50 valid configurations for higher coverages (n ≥ 10) for all facets, with a maximum of 106 attempts. For all these systems, adsorbates were placed 2.0 Å away from the surface, including on the step sites. This distance was selected to be close to equilibrium distances between CO and Cu. Whereas this is a hyperparameter, it is not expected to strongly influence the overall workflow if it remained within reasonable ranges (e.g., 1.8 or 2.2 Å), as all energy shifts would likely retain the major trends. For the cases of co-adsorption, this procedure was first performed with *CHOH as the only adsorbate. As discussed in the main text, the four most favorable structures created from *CHOH adsorption were used as initial structures for sampling the CO coverage.
(v. 2.1.0).67 The cosine distance was used to obtain the distance between two environments.
![]() | (2) |
package in Python (v. 0.5.3). The 2D UMAP plot was produced by using the cosine distance between binding sites, and using 10 neighbors as parameter.
package (https://github.com/dwiddo/average-minimum-distance, v. 1.4.1).48 A total of k = 50 neighbors were used when computing the PDD matrices. The final distance between the two structures is the earth mover's distance between the PDD matrices.
![]() | (3) |
![]() | (4) |
For each CO adsorbed on the copper surfaces, the coordination number reported in Fig. 5d is the coordination number of the closest copper atom. In the case of near-degeneracies of distances, defined when the k-nearest neighbors have distances within 0.05 Å of the smallest distance, the reported aCN is the average of aCNs of all k neighboring atoms.
:
20
:
20 for train/validation/test. This ensures that there is no data imbalance between splits, which could bias the results. One exception is the case where only one or two adsorbates are on the surface. Because of their importance for on-site energy predictions and their scarcity, all configurations with nCO ≤ 2 were added to the training set only, and none to validation and tests sets. The dataset splits were then aggregated on a per-facet basis to train individual models per facet, and concatenated together for the all-facet model. As figures of merit such as root mean squared error depend on test splits, all comparisons between models (such as Fig. 3) were performed using the exact same train–validation–test splits.
. The number of radial basis functions was set to 8, with a cutoff of 5.0 Å.
(v. 2.1.0).67 A simple feedforward neural network was implemented in PyTorch to predict atom-centered contributions which are later summed to obtain the total binding energy of the system. The NN model had 3 hidden layers with 600 neurons each, mish activation function, and an input size of 539, corresponding to the length of the SOAP vector. A mean squared error loss was used to train the models. NN parameters were updated using the AdamW optimizer71 using a learning rate of 10−3, weight decay of 10−2, default β values of β1 = 0.9, β2 = 0.999, and ε = 10−8. The learning rate was reduced by a factor of 0.5 when the loss plateaus for 20 consecutive epochs, and the model was trained for 900 epochs in four NVIDIA V100 GPUs in the Lassen supercomputer, with a batch size of 50. A stochastic weight averaging (SWA) callback was also used after epoch 700, with a starting learning rate of 3 × 10−4 and cosine annealing function every 10 epochs. Training and evaluation routines were implemented using PyTorch Lightning.
The sampling procedure follows the Metropolis–Hastings algorithm within the simulated annealing method. The cooling profile is chosen to be a quadratic decay, following the equation
![]() | (5) |
![]() | (6) |
Although more sophisticated sampling strategies have shown to improve the acceptance ratios,21 the simpler, parallelized implementation of the MCMC algorithm demonstrated enough success in sampling the configuration space of the cases studied in this work (see Fig. S12 and S13†).
| E(int)b = θEb/n, | (7) |
![]() | (8) |
To avoid computing numerical derivatives of the models, we fitted a piecewise continuous and differentiable polynomial for the integral binding energy curves,
![]() | (9) |
The equation above can be simplified when the function and its first and second derivatives are constrained to be continuous at θ = θ0, leading to
To fit the three parameters (a, b, θ0) to a relevant binding energy curve, relaxed energies were averaged using a Boltzmann average at 298 K of energies on a per-facet, per-coverage basis. Then, the parameters for the integral binding energy (eqn (7)) were obtained using the non-linear least squares method implemented in SciPy.
Footnote |
| † Electronic supplementary information (ESI) available. See DOI: https://doi.org/10.1039/d4dd00328d |
| This journal is © The Royal Society of Chemistry 2025 |