Yuehui Xian,
Yunfan Wang,
Pengfei Dang,
Xinquan Wan,
Yumei Zhou*,
Xiangdong Ding*,
Jun Sun and
Dezhen Xue*
State Key Laboratory for Mechanical Behavior of Materials, Xi'an Jiaotong University, Xi'an 710049, China. E-mail: zhouyumei@xjtu.edu.cn; dingxd@mail.xjtu.edu.cn; xuedezhen@xjtu.edu.cn
First published on 2nd July 2025
Bayesian optimization (BO) has been widely employed for alloy composition design, but faces unique challenges in this domain when maximizing the acquisition function (AF), which is a critical step for selecting the best candidate. While various optimization methods exist for maximizing AF, material composition design presents difficulties that include the need to translate compositions into material features, rapid polynomially expanding design spaces as component numbers increase, and compositional constraints (e.g., sum to 100%). To address this issue, we propose a strategy that leverages numerical feature gradient for efficient AF maximization in material composition design. By establishing a differentiable pipeline from alloy compositions, through material features and model predictions, to AF values, our strategy enables efficient navigation from initial compositional guesses to optimal solutions. This approach reduces the complexity of the inner optimization problem from rapid polynomial (i.e., in the case of full enumeration) to empirically observed linear scale with respect to the number of components, making it efficient for medium-scaled design spaces (up to 10 components) while showing potential for scaling to larger compositional spaces. Additionally, initiating the process with randomly generated compositions promotes more diverse solutions, as evidenced by a slower decay of compositional state entropy compared to traditional enumeration-based approaches. Furthermore, the flexibility of our method allows for tailoring the optimization process by adjusting key settings, such as the number of initial compositions, the choice of AFs, surrogate models, and the formulas used to calculate material features. We envision this strategy as a scalable and modular methodology for advancing materials design, particularly in the composition design of high-entropy alloys, ceramics, and perovskites, where elemental compositions can be adjusted as continuous variables.
However, maximizing AF values (often referred as inner loop optimization,28,30 Section 2 in the ESI†) is often challenging. While BO traditionally assumes continuous input variables, material composition design often involves discrete variables (e.g., discretized component percentages) and transforming raw compositions into material features. These characteristics complicate the maximization process, rendering standard gradient-based approaches difficult to apply directly. Consequently, many existing compositional design works rely on evaluating the AF values across the entire design space to identify the optimal candidate.2,18,31–33 This approach, known as exhaustive enumeration (brute force), becomes increasingly intractable as the size of the compositional space shows a rapid polynomial growth with the number of components for material composition design34 (Section 1 in the ESI†). Consequently, this limitation has constrained the search space in many studies to fewer than 107 possible compositions,18,35–37 which is far smaller than the full potential compositional space of high-entropy materials.5,34,35,38,39 Moreover, this exhaustive enumeration frequently leads to the repeated selection of similar compositions across consecutive experimental iterations, concentrating the search around narrow compositional regions and resulting in compositions with marginally improved material properties.1,2
Various approaches have been proposed to address the challenge of AF maximization, including local search,40 probabilistic reparameterization,41 and applying continuous relaxation and then discretize (round) for discrete domains.42 Among these approaches, gradient-based methods have demonstrated promising performance in maximizing AF values.30,43,44 In composition design, obtaining gradient from surrogate models like Gaussian Process Regression (GPR) with respect to raw compositions might seem straightforward. However, a common practice in materials informatics involves transforming raw compositions into material features,37,45,46 where the composition values often have constraints (e.g., ensuring components sum to 100%). These features are defined using various mathematical formulas related to elemental properties and mole fractions, such as (weighted) minimum/maximum operations, along with mole averages of atomic radius, valence electron number, and electronegativity.46–48 While these transformations enhance the performance of machine learning surrogate models,17,35,37,49,50 improving the efficiency of composition design,35,37,50 they also complicate both the definition and propagation of gradients. This complexity arises because feature transformations may involve non-differentiable operations (such as min/max), making closed-form gradient calculation through features challenging. As a result, computing and leveraging gradients for AF maximization becomes a challenging task within this workflow. This difficulty has led many works to adopt alternative strategies, such as exhaustive enumeration32,33 or heuristic methods.51
To address this inner loop optimization challenge in material composition design, we proposed a strategy that leverages numerically computed feature gradient for efficient AF maximization (Section 3 in the ESI†). By establishing an end-to-end piecewise differentiable pipeline from alloy compositions, through material features and model predictions, to the AF values, our strategy enables efficient navigation from initial compositional guesses to optimal solutions. This approach reduces the complexity of the inner optimization problem from a rapid polynomial to a linear scale with respect to the number of components, making it feasible for medium-scaled design spaces (up to 10 components) with the potential for scaling to larger compositional spaces. Additionally, initiating the process with randomly generated compositions promotes more diverse solutions, as indicated by a slower decay of compositional state entropy compared to traditional enumeration-based approach. This broader exploration mitigates the issue of compositional concentration. Even though our results indicate possible incomplete maximization of AF values (i.e., the algorithm may not always find the global maximum of the AF, but rather identify the largest local maximum from a subset of all possible local maxima), the search spans a wider range of promising compositional regions, allowing for stable performance with smaller deviations across parallel BO runs. This proposed method, leveraging gradients propagated through feature transformations, is applicable to the composition design of a wide array of materials, including ceramics, metallic glasses, and high-entropy perovskites, where elemental compositions can be adjusted as continuous variables.
The remainder of this article is structured as follows: We first present our feature gradient methodology in Section 2. Next, we describe the benchmark environments used to evaluate our approach in Section 3. We then demonstrate the efficacy of our approach in Section 4. This is followed by a discussion part regarding various key settings in Section 5. Finally, we summarize our contributions and outline promising directions for future research.
The complete feature set is formed by the combination of elemental properties and feature transformation functions, totaling 240 distinct features. The elemental property set includes 30 metrics, spanning from fundamental atomic characteristics to complex physicochemical properties (Table S5†). The feature transformation functions comprise 8 formulations (Table S1†), including weighted averages, max/min operations, etc. These transformations enable the integration of compositions with elemental properties to create material features that effectively capture composition–property relationships through enhanced information sharing among constituent elements.
The mapping from compositions, through material features and GPR predictions, to AF values that we considered in this work, including Expected Improvement (EI), Probability of Improvement (POI) and Upper Confidence Bound (UCB), can be implemented as a differentiable computational graph. This allows for end-to-end gradient computation through a chain, linking the alloy compositions, material features, GPR predictions, and AF values, as expressed by ∇cα(g(ε(c))) (Fig. 1a(1)). Although deriving an analytical form for this chain is challenging, we utilize numerical differentiation to obtain gradient values, leveraging PyTorch's differentiation tools.52,53 The details, including the mean and uncertainty predictions of GPR, the definitions of considered AFs (EI, POI and UCB), the computation of feature transformations (ε(c)), and the numerical evaluation of gradients through features and GPR predictions to AF values, are provided in Section 3 of the ESI.†
Using this numerical gradient ∇cα(g(ε(c))), AF maximization is achieved efficiently, as demonstrated in Fig. 1a(2). Optimization methods, such as Sequential Least Squares Programming (SLSQP), guide the search toward local optima. Starting from random initial composition guesses, this gradient-based approach rapidly identifies compositions that locally maximize the AFs. The SLSQP method incorporates linear compositional constraints (e.g., positive concentrations) into the optimization objective through the Lagrange Multiplier method, essentially constructing an equivalent Lagrangian formulation of the constrained optimization of scalar AFs. This incorporation of constraints ensures that compositions proposed by the gradient-guided search remain realizable throughout the optimization process. In composition design, constraints such as element solubility and the requirement that total mole fractions sum to 100% often lead to a rugged compositional space. To address this, we include a rejection sampling54 step to generate feasible initial compositions. This sampling ensures that the initial compositions are well-distributed within the feasible design space, improving the effectiveness of the optimization process.
Random initial sampling of compositions also helps to prevent the optimization from converging prematurely to specific regions, promoting broader exploration of the design space. Consequently, this strategy yields a diverse set of optima compared to traditional enumeration-based approaches (detailed in Table S3†), as schematically illustrated in Fig. 1b.
Finally, Table 1 presents pseudo-code for our BO method, which incorporates feature gradient for efficient AF maximization. The code is available at https://github.com/wsxyh107165243/FeatureGradientBO. By optimizing the AF with feature gradient, our method significantly accelerates the BO process, allowing for efficient exploration of large compositional spaces with greatly reduced computational effort.
The SMAs environment focuses on phase change alloys used in thermal management devices, targeting key properties such as high enthalpy change (ΔH) for greater energy density, low thermal hysteresis (ΔT) for rapid response, and the working temperature Tw (defined as the average of martensite and austenite transformation peak temperatures, Mp and Ap) near the operating temperature. A figure of merit (FOMsma) is defined as: , where ΔHN, ΔTN and TwN are normalization factors for enthalpy change, hysteresis, and the deviation from the target working temperature, respectively. ΔTwN is the difference between the objective working temperature and the predicted working temperature.
The Ti environment evaluates mechanical properties, specifically yield strength (σY), ultimate strength (σU), and Vickers hardness (ν). These properties are optimized through a figure of merit (FOMti) defined as: , where σYN, σUN and νN are normalization factors for yield strength, ultimate strength, and Vickers hardness, respectively.
The HEAs environment focuses on yield strength (σY), ultimate strength (σU), and elongation (ε), with the goal of balancing strength and ductility. The figure of merit (FOMhea) for this environment is defined as: , where εN is the normalization factor for elongation. These figures of merit (FOMs) combine multiple properties into single objectives for the BO algorithms to optimize, reflecting practical material development priorities and providing a complex, multi-objective optimization landscape suitable for testing our composition optimization algorithms. It should be noted that our current implementation uses an equally weighted sum of multiple properties as the figure of merit, which serves as optimization objective for method validation but does not directly address multi-objective optimization, which is a limitation of the current approach.
The neural networks used in these environments were implemented with an architecture inspired by the approach outlined in ref. 55. Each neural network begins with a convolutional section followed by a residual connection, after which process conditions are concatenated with the outputs before entering the fully connected section. The convolutional section consists of two layers with a kernel size of 1 × Nfeature × Nfeature, and batch normalization is applied after each layer to enhance model robustness. The resulting output is flattened and passed through two fully connected layers, configured as (Nelem × Nfeature) × 128 and 128 × 1, incorporating Exponential Linear Units (ELUs) for nonlinear activation and a dropout layer to improve stability. The networks were trained using the Adam optimizer with a batch size of 16.
For training, we used 501 composition-property pairs for HEAs, 603 pairs for Ti alloys, and 295 pairs for SMAs, ensuring reliable model performance. These data were collected from both our laboratory and published literature, with details provided in the Section 6 of the ESI.† The nine neural networks (three per alloy type) were trained as detailed in the Methods section and in Section 6 of the ESI.† Before training, each dataset was randomly split into training (70%), validation (15%), and test (15%) subsets. As shown in left panels of Fig. 2, training was performed for up to 1000 epochs with a learning rate of 5 × 10−4. The training loss and validation R2 scores demonstrated significant improvements during the initial 500 epochs, with diminishing returns in later epochs. The R2 score on the test set at the 500th epoch closely matched that of the validation set, confirming stable training. Consequently, 500 epochs were selected as the optimal training duration for building the neural networks. The performance of these networks, as shown in right panels of Fig. 2, was further evaluated using 10-fold cross-validation, yielding R2 values between 0.443 and 0.917 and MAE values between 0.187 and 0.492. It is worth noting that the neural network performance shows lower accuracy for Ti alloy tensile strength prediction (R2 = 0.443). This reduced performance likely stems from the variability in tensile strength measurements, which are highly sensitive to specimen preparation, testing conditions, and structural variations. These neural networks are then integrated to form FOMs specific to each alloy class, which will be optimized using BO algorithms.
For alloys with 3–5 components, the total number of possible compositions remains within 1010, allowing for a direct comparison of the performance of both optimization methods. In these scenarios, our feature gradient strategy for inner optimization achieved results comparable to the enumeration approach, as shown in Fig. 3.
As the number of components increased, however, enumeration became increasingly computationally impractical. Therefore, for compositions with higher numbers of components, only the gradient-based inner optimization trajectories are presented. Our results demonstrate that the proposed feature gradient strategy is scalable and efficient for composition design involving up to 10 components. This scalability is achieved by empirically scaling the number of initial compositional guesses linearly with the number of components, suggesting potential applicability for even larger systems. Expanding the search space often increases the likelihood of discovering compositions with enhanced properties, especially in the case of Ti alloys shown in Fig. 3. The scalability underscores the advantages of using gradient-based optimization in complex compositional spaces, where direct enumeration becomes computationally prohibitive.
For multi-component alloys, the number of possible compositions shows rapid polynomial growth with the number of components (see Section 1 in the ESI†), making inner loop optimization increasingly challenging. Comparing the computation time for inner loop optimization between the traditional enumeration approach and our feature gradient method provides valuable insights. Based on our hardware, the estimated time to calculate the AF value for a single composition (t0) is approximately 6 × 10−7 seconds. In the enumeration approach, the number of evaluated compositions (N) corresponds to the entire design search space, resulting in a time consumptions (t) of approximately t0 × N. As shown in Fig. 4, both N and t increase with the number of components (Nelem).
For our feature gradient strategy, as the gradient calculation must be included, the time to evaluate a single gradient tgrad is ∼1.5 × 10−3 seconds using the same hardware. However, by employing gradient-based optimization methods like SLSQP, the number of evaluated compositions (N) is significantly reduced. In our approach, N depends on the number of initial compositions, and to our experience, scales linearly with the number of components (Nelem), at approximately 20 × Nelem. For example, with a training set of 40 compositions, we observed that the SLSQP algorithm typically requires ∼10 gradient calculations to find the near zero-gradient composition from a random initial composition point. Therefore, with a 4-component system (Nelem = 4), the total computation time would be approximately 10 × 20 × Nelem × tgrad ≈ 10 × 20 × 4 × (1.5 × 10−3) = 1.2 seconds per inner loop in BO framework. Overall, the time consumption for our feature gradient method remains far lower than that of the enumeration method.
Fig. 4 illustrates the number of evaluated compositions and corresponding time consumption as a function of component count for both strategies. Our results demonstrate that the proposed method achieves substantially higher efficiency and lower time consumption in the inner loop, particularly as the number of components increases. The computational complexity of our method is primarily determined by the number of initial random guesses in the inner loop, exhibiting only a weak, linear correlation with the number of components. For up to six components, the enumeration method remains feasible, with a computational time of around one hour. However, based on the intersection of the two curves in Fig. 4b, when the target component number exceeds four, our feature gradient strategy becomes the preferred approach.
By leveraging the high efficiency of our feature gradient strategy for AF maximization, we can initiate the search from multiple randomly generated compositions, enabling several parallel search routes. Each route progresses along the gradient ascent direction until the gradient value, ∇cα(g(ε(c))) approaches zero. With a sufficiently large number of starting compositions (e.g., 20 × Nelem), the optimized solutions are more likely to approximate the global optimum. This multi-starting point approach efficiently explores diverse regions of the AF landscape in each inner loop, which subsequently leads to solutions that are both well-distributed across the design space and possess desirable material properties through the outer loop process. This approach addresses a key limitation of traditional enumeration-based BO, which often selects similar compositions repeatedly across consecutive iterations, thereby concentrating the search on narrow compositional regions.1,2 In contrast, our gradient-based strategy promotes broader exploration, yielding a more comprehensive coverage of the entire composition design space.
To quantitatively evaluate the broader exploration behavior of our strategy, we introduced state entropy as a metric, following the definition from:56,57
The state entropy analysis from multiple parallel optimization trajectories strongly supports our findings, as shown in Fig. 5a–c. Notably, in all three test environments (SMAs, Ti alloys, and HEAs), our strategy consistently maintains higher state entropy values, demonstrating a broader exploration of the compositional space. The higher state entropy values observed in our feature gradient strategy arise from its fundamental operational characteristics. While enumeration-based methods deterministically identify the single global maximum of the AF, our approach identifies multiple local maxima dependent on randomly selected initial compositions. For each inner loop, we perform multiple gradient-based ascent runs (20 × Nelem), potentially discovering diverse local maxima across the compositional space. These identified local maxima constitute a subset from the power set of all local maxima within the whole compositional design space, and may vary across parallel inner loops. This introduces an inherent stochasticity as initial compositions are randomly selected and the GPR landscape is often multimodal. Consequently, the feature gradient method explores diverse promising regions, manifesting as higher state entropy values even though it may not guarantee finding the exact global maximum in every iteration.
To further visualize this broader exploration, we performed t-SNE (t-Distributed Stochastic Neighbor Embedding) analysis on two representative optimization trajectories from both methods in the 4-component SMAs environment (Fig. 5d). t-SNE preserves short Euclidean distances and represents higher compositional similarity in a 2-dimensional space, allowing high-dimensional compositional data to be effectively visualized. The broader exploration achieved by our strategy is evident in the more scattered compositions within the light pink region of Fig. 5e. Additional pairwise distribution analysis of elemental compositions (Fig. S2, ESI†) further corroborates these findings, showing that the enumeration method produces more peaked compositional distributions concentrated in limited regions, while the gradient optimization approach maintains broader, more uniform exploration across the compositional space.
A key observation is that our strategy promotes broader exploration while also achieving more stable performance across parallel optimization tests, even though complete global maximization is not guaranteed. This trade-off suggests that sacrificing the guarantee of finding the exact argmax significantly reduces computational time in the inner loop, with only negligible losses in material performance. Such a balance is particularly valuable in scenarios where computational efficiency takes precedence over marginal improvements in solutions. Furthermore, we demonstrate that the increased exploration enabled by our strategy enhances the robustness of BO, making it less sensitive to the selection of initial data. This robustness is further supported by the consistently higher state entropy values achieved by our method.
Our results show that the proposed strategy offers a computational advantage in finding near-optimal AF values, outperforming the enumeration-based method by significantly reducing computational time through the use of feature gradient and gradient ascent, which require evaluating only a small subset of compositions. Additionally, systematic state entropy analysis confirms that our approach achieves broader exploration of the design space, reducing the risk of premature convergence and enhancing the robustness of BO. This improved efficiency and reduced dependency on initial data highlight the potential of our method to advance computational materials design.
Our strategy involves hyperparameters that need to be determined for effective deployment. One key hyperparameter is the number of randomly initialized starting compositions for inner loop optimization. In the inner loop, gradient ascent is performed on these starting compositions, with a larger number of starting points enabling a more thorough exploration of the composition design space. While increasing the number of starting compositions can lead to better acquisition function values, it also results in higher computational costs. Experimental results across different numbers of starting compositions (10, 20, 40, and 80) demonstrate consistent improvements in the FOM with more starting points, as shown in Fig. 7. As this budget increases, the gradient-based results increasingly approximate towards the enumeration-based results. This convergence occurs because with more initial composition points distributed throughout the compositional space, gradient ascent can identify more of the local optima that would otherwise only be found through exhaustive enumeration. When the budget increases to a sufficient level where gradient ascent identifies all local maxima across the compositional space, the optimization result becomes effectively consistent with the enumeration-based method. The inset in Fig. 7 illustrates this convergence trend, showing how the performance increases as the number of initial compositions increases from 10 to 80, with diminishing marginal returns at higher values, indicating a saturation effect. To balance the trade-off between achieving higher FOM values and managing computational costs, we empirically propose that the number of starting compositions should scale linearly with the number of components, following the formula 20 × Nelem. This scaling provides a practical balance between BO performance and computational efficiency.
The differentiable pipeline proposed in this work connects alloy compositions, material features, surrogate models, and AFs. While we demonstrated the effectiveness of our strategy using EI as the AF, it is important to note that this methodology is not limited to EI. Other AFs that are analytically differentiable, such as Upper Confidence Bound (UCB, with trade-off parameter κ = 1.96) and Probability of Improvement (POI), can also be integrated into our feature gradient strategy. This mathematical property broadens the applicability of our feature gradient strategy to a range of AFs, each offering unique exploration–exploitation trade-offs. To explore this broader applicability, we extended our experiments to include UCB and POI as AFs. Using the same experimental procedures as for EI, we conducted comparative tests on the SMA dataset with four components. The results, shown in Fig. 8a, indicate consistent optimization performance across all tested AFs when using the feature gradient strategy. All three AFs exhibit similar trends: an initial phase of steep improvement during the first 10 experimental iterations, followed by stable convergence. Among the tested AFs, UCB demonstrated marginally faster initial convergence, while POI showed slightly higher variance in the early stages. However, all three AFs converged to similar final FOM values (approximately 0.7), with overlapping confidence intervals during the stable phase. These results highlight the robustness of our feature gradient strategy and its consistent effectiveness regardless of the choice of AF.
For material features, various mathematical formulas beyond the traditional weighted averaging approach can be employed. This broader class of functions (Table S1†) expands the feature space beyond linear combinations of elemental properties, potentially capturing non-linear relationships and richer physical information. However, this generalization complicates the calculation from initial compositions to final AF values. To evaluate the impact of this generalization on the optimization outcomes, we examined its influence on the FOM values obtained through BO. We utilized eight mathematical formulas commonly used in HEAs research59 and combined them with 30 elemental properties, generating a pool of 240 elemental property–formula combinations. Genetic algorithms were applied for feature subset selection, optimizing the 10-fold cross-validation (R2) to identify the most promising combinations. Comparative experiments between traditional weighted averaging and expanded mathematical functionals revealed notable differences in optimization performance (Fig. 8b). The approach employing a broader range of formulas exhibited superior performance during the initial stages of optimization, with a significant positive performance difference observed within the first three BO iterations (inset of Fig. 8b). This result underscores the potential benefits of incorporating more sophisticated mathematical formulas into feature calculations. However, the increased complexity of these functionals significantly amplifies the computational burden in traditional enumeration-based inner loop optimization, potentially limiting its practical implementation. This observation highlights the strong advantage of the feature gradient strategy, which efficiently handles these complex formulas while maintaining computational tractability. By enabling the practical use of advanced mathematical formulas, the gradient-based approach not only enhances BO performance but also ensures computational efficiency, providing a robust framework for advancing materials design.
Apart from the AF, another key factor for successful BO is the choice of surrogate model, which directly impacts the effectiveness of the property prediction and optimization process. While Gaussian Process Regression (GPR) remains the most widely used surrogate model, alternative approaches, such as Random Forest (RF), have gained attention for their robust predictive performance and reduced sensitivity to hyperparameter tuning. However, the non-differentiability of RF models poses a significant challenge for inner loop optimization, particularly when employing efficient gradient-based strategies. This limitation highlights the need for differentiable alternatives, such as Deep Gaussian Processes (DGPs),60,61 which not only preserve differentiability but also excel in capturing complex composition–property relationships. Additionally, DGPs offer enhanced computational efficiency when handling larger datasets, making them a promising option for high-dimensional materials design problems. Choosing an appropriate surrogate model therefore requires striking a careful balance between predictive accuracy, computational efficiency, and differentiability to enable effective inner loop optimization. This balance is especially critical in materials design, where the intricate complexity of composition–property relationships necessitates both precise modeling and computationally tractable optimization methods.
In the supplementary, we further briefly discuss various design choices, including the use of 20 × Nelem and random sampling, as well as potential issues with gradient failure. We note that for EI, a common challenge is that AF values gradually approach 0 as experimental iterations proceed. This occurs because the probability of finding improvements diminishes with better solutions found, leading to numerical difficulties with vanishing EI values and gradients.62 Additionally, determining which elemental properties are most relevant for specific performance metrics presents a significant challenge. The potential elemental feature space is large,63,64 encompassing hundreds of candidates, many of which may have minimal correlation with the target performance. Feature selection, which involves identifying the optimal subset of elemental properties that meaningfully influence performance, remains an active research area in materials informatics. Our selection of 30 candidate elemental properties represents a balance between domain knowledge, empirical experience, and computational tractability, though more sophisticated features could potentially enhance performance further. Notably, some acquisition functions, such as entropy-based functions, may not be differentiable, limiting our method's applicability in such cases. Currently, our research focuses on pure numerical experiments, with the validity of recommendations ensured by BO itself. We do not conduct further material synthesis or characterization experiments, and it is important to note that a better AF score does not necessarily lead to improved experimental results. This limitation is partly related to our current approach of encoding multiple material properties into a single FOM with equal weights. While this formulation serves our purpose of testing optimization strategies on complex landscapes, it may not reflect specific application preferences.
A limitation of our method is its reliance on a single aggregated FOM, which may not effectively capture trade-offs between competing material properties for specific applications. Addressing this limitation would expand the applicability of our approach to more complex materials design scenarios where such trade-offs are essential. Further improvements to this approach could include several promising directions. First, extending the framework to handle multi-objective optimization problems, allowing for the exploration of Pareto-optimal solutions and preference-based optimization rather than using naive weighted property combinations. Second, enhancing the sampling strategies in the inner loop. Advanced methods for generating diverse initial compositions as starting points for gradient ascent could further improve the optimization result. Such strategies might include diversity-promoting sampling techniques that better cover the compositional space. Additionally, developing adaptive sampling approaches that dynamically adjust the number of initial points based on the complexity of the acquisition function landscape, for example by increasing sampling density with more evaluated compositions during sequential BO iterations, would enhance the method's efficacy. These enhancements would collectively address both the balance between competing objectives and the effectiveness of inner optimization, potentially leading to even more effective materials discovery.
Footnote |
† Electronic supplementary information (ESI) available. See DOI: https://doi.org/10.1039/d5dd00080g |
This journal is © The Royal Society of Chemistry 2025 |