Open Access Article
Nila Mandal
,
James Maniscalco,
Mark Aindow
and
Qian Yang
*
University of Connecticut, Storrs, CT, USA. E-mail: qyang@uconn.edu
First published on 2nd March 2026
High-throughput screening enabled by structure–property prediction models is a powerful approach for accelerating materials discovery. However, while machine learning of structure–property models have become widespread, its application to mixtures remains limited due to increased complexity and the scarcity of available data. Machine learning methods for high-throughput screening of eutectic mixtures have been proposed in recent years, but there remain challenges due to the lack of diverse, open-access datasets and the need for feature engineering based on chemical knowledge. To overcome these limitations, we propose a method using Siamese graph neural networks trained solely on structural information, without requiring any prior chemical descriptors, to predict eutectic melting temperatures. We demonstrate on a dataset of molten salt eutectics that this approach can reach similar performance to chemistry-based models that require significantly more prior knowledge. We show that lower-order mixtures may be used to augment data on higher-order mixtures. Interestingly, our model trained on inorganic molten salts seems to learn information about the ideal mixture model. We also evaluate the efficacy of using our inorganic molten salt model for transfer learning with a variety of organic eutectic mixtures.
These important applications are hindered by drawbacks of existing computational and experimental methods for determining mixture properties, which are often slow and resource-intensive. A machine learning approach for high-throughput screening of eutectic melting temperatures is thus highly desirable. However, such approaches are often hindered by the lack of diverse, publicly available experimental datasets and the computational cost of simulation methods for generating data. As a result, most recent studies rely on classical machine learning algorithms with highly informative but expensive engineered features that can be effectively trained on smaller datasets. In this work, we demonstrate that a deep learning-based approach leveraging only structural information is sufficient to train effective models for predicting eutectic melting temperatures of binary molten salt mixtures, using a dataset of 2244 data points. Our approach utilizes Siamese graph neural networks4 and incorporates ideas from Janossy pooling5 to effectively handle mixtures. We also demonstrate that individual components' melting point data can be used to augment the mixture datasets, and produce models that can extrapolate from inorganic to organic materials (Fig. 1).
Our architecture includes several novel contributions, including the ability to learn melting points of binary eutectic mixtures from structural data alone, and the ability to learn a model correlated with the ideal thermodynamic model without requiring single component melting points or enthalpy values, or eutectic compositions xe values. Although single component melting points are not required, we also demonstrate that by optionally using them as additional datapoints for data augmentation rather than features, we can further improve our model's predictive performance as well as achieve good predictions for previously unknown single component melting temperatures.
Much work has been done on melting point prediction for ionic liquids, including comparisons of methods such as k-nearest neighbor regression, gradient boosting, random forests, support vector machines, and graph neural networks.7,8 Structural descriptors such as molecule fingerprints, Coulomb matrices, or other engineered features can be used to improve prediction accuracy. Fingerprints which encode functional groups have been found to be particularly effective.9 However, Acar et al.10 argues that all of these fingerprint methods are expensive to compute, and not practical for either large datasets or datasets which include large molecules. Instead, they test a very simple fully connected neural network with one dropout layer, and compare the correlation between melting point and the various engineered features they use. Although these engineered features may be less computationally complex than the fingerprint methods evaluated by Low et al.,9 expertise in chemistry is necessary in order to understand and generate these descriptors for any dataset.
![]() | ||
| Fig. 2 From ref. 11, a phase diagram illustrating the relationships between Ta, Tb, xe, and Te. Eutectic mixtures are mixtures with a melting point lower than that of the individual components of the mixture. The mixture proportion with the lowest possible melting temperature is referred to as xe, and that melting temperature is referred to as Te.† | ||
Determining eutectic melting points is challenging for materials screening because it is infeasible to manually test every possible mixture of candidate components. Research into data-driven approaches to Te prediction has been significantly hindered by the lack of large, publicly available datasets. Although published eutectic datasets exist, they are often small (100 examples or less) and focused on a highly specific family of materials (for example, fatty acids or organic explosives). Due to these limited dataset sizes, machine learning approaches to eutectic melting point prediction typically rely on highly informative chemical descriptors such as individual melting points, enthalpies of melting, and xe values, which are themselves difficult to obtain or compute.
If all of those values are available, Te can be computed using the ideal thermodynamic model (ITM). However, if any of these values are unavailable in existing literature for a given material, it is impossible for the ITM to compute Te for any mixture which includes that material. Even when all necessary values are known, the predicted Te values from the ITM are not perfect. For example, Ravichandran et al.1 found that for their data, the ITM predictions had a root mean squared error (RMSE) of 86.6 K. The correlation between Te and the individual melting points of the mixture components has been used to estimate how much melting point depression the eutectic mixture will have;12 however, even this extremely rough estimation requires knowledge of the melting points of each individual component, which is itself challenging to obtain for new materials exploration.
Ravichandran et al.1 approach eutectic Te prediction for molten salts by using an ensemble model composed of the ideal thermodynamic model, a gradient boosting model, and a Roost model. Of these, the Roost model is the only part of the ensemble which does not require expensive engineered features. The ensemble model, referred to in the text as the “mean model,” takes the average of the three models' predictions to determine its final output. They consider mixtures of up to six eutectic components, and their data is a subset of the experimentally-determined molten salt eutectic mixtures, proportions, and Te values compiled by Janz et al.13 Our work considers a different subset of mixtures from the same compilation.
There are several smaller datasets we are aware of for different families of organic eutectic mixtures. Kahwaji et al.11 published a set of fatty acid eutectics as well as a computational method that they designed to support eutectic phase change materials exploration. Guendouzi et al.14 published a set of melting points for individual fatty acids, which overlap with Kahwaji's individual components. Luu et al.15 and Lavrinenko et al.16 both published sets of deep eutectic solvents drawn from literature, with substantial overlap between the mixtures they compiled. More recently, databases for binary and ternary DES and corresponding ML models have also been developed.17
AttentiveFP,20 which is used in this work, is one such message passing graph neural network architecture, with the addition of an attention mechanism to help represent intramolecular interactions beyond each node's one-hop neighborhood. The architecture has a series of layers for atom embedding, followed by a series of layers for molecule embedding. For each graph, AttentiveFP generates a “state node” which is connected to every node in the graph; this is used to aggregate information from the whole graph to compute a learned representation of the entire molecule.
A significant challenge in GNNs is identifying an appropriate pooling method. Graph pooling should generally be order-invariant, because the nodes in a graph typically are not meant to relate to each other in any sequential ordering. However, simple order-invariant methods like mean or max pooling fail to capture important structural information that is not explicitly represented by node features.21 Learned pooling methods, such as hierarchical pooling or Janossy pooling5 may produce better pooled representations.
In order to handle mixtures, we incorporate ideas from Janossy pooling,5 which refers to building permutation-invariant functions by taking the average, or an approximation of the average, of a permutation-sensitive function's output over every possible permutation. In the context of graph neural networks, this refers to permutations of the nodes in the graph. In the general case, an approximation method is necessary for tractability; as such, Murphy et al.5 present three categories of approximation methods. However, in problem domains where the number of possible permutations per input is small, as in this work, it may be reasonable to compute this naively.
Siamese neural networks4 were originally developed to measure the similarity or difference between pairs of input data. Since determination of binary eutectic Te values is an inherently pairwise problem, and Te and xe are correlated with the difference between the melting points of the mixture components,12 Siamese neural networks are well-suited to our problem.
Siamese neural networks have been applied to many problems in chemistry and materials science due to their efficacy in one- or few-shot learning and similarity comparison. In particular, they have been widely used to screen molecules for drug discovery, predict molecule solubility or toxicity, and predict drug response similarity.4,22–24 Siamese graph neural networks have also been used with engineered graph level features for drug-responsiveness prediction on cancer treatments.25
We further use a set of fatty acids,11 a set of explosives,12 and a set of quinones28 which form eutectic mixtures as small test sets for evaluating zero-shot transfer from inorganic to organic eutectics (Table 2).
| Dataset | Tasks | Number of data points | Avg. atoms per molecule |
|---|---|---|---|
| Molten salts A,B data | Training from scratch | 2244 | 5 |
| Molten salts A,A data | Training from scratch | 149 | 5 |
| DES A,B data | Transfer learning | 239 | 24 |
| DES A,A data | Transfer learning | 88 | 24 |
| Fatty acids | Inference only | 102 | 43 |
| Explosives | Inference only | 74 | 21 |
| Quinones | Inference only | 26 | 18 |
For the purposes of this work, we focus on the minimum Te value for each unique mixture pair represented, because we are interested in predicting the lowest melting point achievable by any given pair. All temperature values presented here are in Kelvin unless otherwise stated.
After filtering duplicates, we take a stratified sample of mixtures (which we refer to as A,B pairs) and individual components (A,A pairs) into the training and test set (90% and 10%, respectively). We then divide the molten salt training set into 10 stratified folds for cross-validation.
The interactions between the two molecules in the mixture influence its melting temperature, so rather than computing a similarity or difference metric between the two, we concatenate the two molecule representations so that the following layers can learn a function of the molecules' interactions. By concatenating A and B, it is possible for the fully connected layers in the Siamese branches to learn something about the interactions between the two molecules, as opposed to only learning their difference.
However, concatenating them implies an ordering, and because we are not incorporating information about individual melting temperatures or mixture proportions, we cannot enforce a meaningful ordering between the two. We address this by utilizing the central idea from Janossy pooling to enforce order invariance by concatenating the representations in every possible ordering, i.e. (A, B) and (B, A). One might extend to higher-order mixtures by implementing a permutation-sampling approximation of Janossy pooling, selecting a value n as the maximum number of permutations to be sampled per mixture.
These concatenated molecule representations are passed through Siamese branches consisting of fully connected layers, with dropout layers in between. The output of these branches are mean-pooled to generate a final order-invariant representation of the mixture pair, which is then passed through the output layer to predict Te. A diagram of the architecture is depicted in Fig. 6, and all selected hyperparameters and optimization methods are detailed in Table 3.
| Hyperparameter | Selected value |
|---|---|
| Batch size | 256 |
| Learning rate | 0.001 |
| LR Scheduler start factor | 0.350 |
| LR Scheduler iterations | 481 |
| Embedding size | 128 |
| Hidden channels | 256 |
| AttentiveFP layers | 6 |
| AttentiveFP timesteps | 4 |
| Dropout 1 | 0.383 |
| Fully connected layer 1 size | 8 |
| Dropout 2 | 0.002 |
| Fully connected layer 2 size | 8 |
We train some models using only the molten salt eutectic mixtures, which we refer to as A,B pairs. We also train some models using both the molten salt eutectic mixtures and the individual components for which we have melting point values, in order to test whether augmenting the dataset with these individual components may improve model performance. The individual components are represented as pairs in which both components are identical, which we call A,A pairs. We compare the performance of models trained in these two ways (Fig. 7), and evaluate whether the individual component data is useful for data augmentation.
We emphasize that these single component melting points are used as data points, not features. When we compare these results to the exact same architecture and feature representation trained on A,B pairs only, we see that the gain in performance is approximately 4.5° on the overall set, and 12.5° in the low temperature subset. Table 4 shows that the model trained on A,A and A,B pairs have improved performance on all test sets compared to the model trained only on A,B pairs.
| Dataset | Model | RMSE | MAE | MAPE | R2 | Std. dev. | Baseline MAE | Baseline MAPE |
|---|---|---|---|---|---|---|---|---|
| Full dataset from ref. 1* | ITM | 86.6 | — | — | — | — | — | — |
| Test set from ref. 1* | Mean model | 65.4 | — | — | — | — | — | — |
| Random partition from ref. 1 | A,B only, trained on1 | 95.45 ± 13.76 | 68.90 ± 10.29 | 0.09 | 0.92 | 309.86 | 220.23 | 0.85 |
| A,B test set | A,A + A,B | 103.08 ± 20.05 | 73.04 ± 9.09 | 0.10 | 0.93 | 369.9 | 267.5 | 0.46 |
| A,B test set | A,B only | 107.55 ± 11.79 | 77.74 ± 9.42 | 0.11 | 0.92 | 369.9 | 267.5 | 0.46 |
| A,B where Te < 500 K | A,A + A,B | 59.53 ± 10.33 | 47.73 ± 9.27 | 0.13 | 0.60 | 78.41 | 63.11 | 0.19 |
| A,B where Te < 500 K | A,B only | 72.04 ± 20.03 | 49.76 ± 15.03 | 0.15 | 0.48 | 78.41 | 63.11 | 0.19 |
| A,A test set | A,A + A,B | 231.61 ± 86.17 | 189.56 ± 69.85 | 0.44 | 0.89 | 629.34 | 443.35 | 0.81 |
| A,A test set | A,B only | 372.26 ± 169.92 | 275.45 ± 129.39 | 0.67 | 0.64 | 629.34 | 443.35 | 0.81 |
As shown in Fig. 8, our model trained on A,B and A,A pairs also achieves prediction on individual components' melting points with a R2 of 0.89, and a RMSE of 231.61 K. In the context of the standard deviation of the A,A dataset, this prediction error is one-third of the RMSE of a baseline trivial model (the best possible constant model), indicating that despite the relatively large numerical value of the RMSE, the model has learned significant predictive information. This suggests that by including only small quantities of A,A data in training, the final model can learn to estimate individual component melting points while also achieving improved prediction performance on A,B mixtures.
A recent work by Ravichandran et al.1 also used subsets of Janz et al.13 to train machine learning models for eutectic melting temperature prediction; their selection of binary mixtures is a subset of those used in our work. While their experiments resulted in lower RMSE values than ours, their proposed methods require all individual component melting points and enthalpies (ITM method), manually curated engineered features (GBM method), or xe values (Roost method), none of which are required by our method. For comparison, took a random training and test partition of the binary mixtures used in their dataset, and trained our architecture on that partition using only structural information. The results, as shown in Table 4, indicate that our model can achieve performance close to the ITM without requiring individual component melting points and enthalpies.
We note that differences in performance may also be in part due to the difference in our datasets; they considered a narrower selection of binary mixtures than ours, and included other n-ary mixtures which we did not.
First, we select the molten salts model which performs best on our DES training set. Organic training data including both A,A and A,B pairs were input into our architecture, and then the molecule features output by the AttentiveFP layers (F(A) and F(B)) were extracted. We scale the features and then use (PCA) to reduce them from 256 features per molecule to 5 features per molecule. We then concatenate each molecule pair in both possible orders, and fit a kernel ridge regression model to the data. We use a grid search to determine the optimal kernel ridge regression hyperparameters. We evaluate the best model on the DES test set, as well as evaluate zero-shot predictions on other smaller organic datasets.
![]() | ||
| Fig. 11 Prediction results on fatty acids. (a) presents 0-shot predictions from the model trained on molten salts, while (b) presents predictions after extracting DES features from the A,A + A,B model, fitting a kernel ridge regression model, and then testing on fatty acids. Models from our experiments frequently made predictions on fatty acid eutectic mixtures which had a strong linear correlation with the target values. Mixtures including oleic acid consistently “branched” off from the rest of the group. This data was computed using an ideal thermodynamic model by Kahwaji et al.11 This suggests our architecture is able to learn a function correlated with the ideal thermodynamic model without having information about individual components' melting points or enthalpies of melting. (a) Predictions from molten salts model. (b) Predictions from transfer learning KRR with DES. | ||
Some models trained on molten salts are able to assign organic materials to the correct average melting temperatures of their corresponding molecule family (e.g. fatty acids) despite having never seen an organic molecule in training. These models are also able to assign families of closely-related molecules to similar values; for example, all quinone mixtures receive similar predictions, and fatty acid mixtures' predictions form distinct patterns. This indicates that our model is learning similar feature embeddings for mixtures that we expect to be similar to one another. Therefore, the molecule representation learned by the GNN layers of our architecture from inorganic data can be used for transfer learning to organic DES data to achieve improved predictions on DES mixtures. However, our KRR model transfers less well to both the explosives and quinone datasets.
Mixtures from the same family of materials have predictions that are close together; fatty acids, in particular, seem to follow a distinct trend. This suggests that mixtures that belong to the same family are being assigned similar learned feature embeddings; these are the molecule graph features that we extract from our GNN layers. Then we scale, apply PCA, and fit the kernel ridge regression model.
Through this process, we are able to predict DES Te values with an RMSE of 37.57 K. We are also able to predict melting temperatures of the individual components from the DES dataset with an RMSE of 67.30 K (Table 5).
| Dataset | RMSE | MAE | MAPE | R2 | Std. dev. | Baseline MAE | Baseline MAPE |
|---|---|---|---|---|---|---|---|
| 0-shot prediction from best molten salt model | |||||||
| DES test set | 90.26 ± 16.20 | 71.64 ± 16.33 | 0.25 | 0.05 | 69.96 | 51.7 | 0.12 |
| Fatty acids test set | 43.97 ± 5.23 | 36.73 ± 5.02 | 0.12 | 0.29 | 20.15 | 17.25 | 0.06 |
| Explosives12 | 63.19 ± 13.31 | 51.32 ± 10.31 | 0.14 | 0.04 | 35.32 | 25.73 | 0.07 |
| Quinones28 | 75.31 ± 13.22 | 67.21 ± 13.34 | 0.21 | 0.33 | 19.62 | 12.95 | 0.04 |
![]() |
|||||||
| 0-shot prediction from molten salt model chosen for transfer learning | |||||||
| DES test set | 55.79 ± 13.25 | 40.58 ± 10.84 | 0.13 | 0.0 | 69.96 | 51.7 | 0.12 |
| Fatty acids test set | 20.73 ± 2.87 | 16.56 ± 2.45 | 0.06 | 0.38 | 20.15 | 17.25 | 0.06 |
| Explosives12 | 66.69 ± 12.41 | 56.37 ± 8.89 | 0.16 | 0.03 | 35.32 | 25.73 | 0.07 |
| Quinones28 | 31.30 ± 10.02 | 24.97 ± 8.16 | 0.07 | 0.01 | 19.62 | 12.95 | 0.04 |
![]() |
|||||||
| KRR transfer learning predictions | |||||||
| DES test set | 37.57 ± 4.55 | 31.21 ± 4.29 | 0.11 | 0.41 | 69.96 | 51.7 | 0.12 |
| DES AA test set | 67.30 ± 24.77 | 47.17 ± 20.09 | 0.12 | 0.62 | 77.05 | 61.32 | 0.17 |
| 0-shot prediction from KRR model | |||||||
| Fatty acids test set | 18.11 ± 2.64 | 13.34 ± 1.73 | 0.05 | 0.39 | 20.15 | 17.25 | 0.06 |
| Explosives12 | 75.53 ± 7.80 | 62.91 ± 7.47 | 0.17 | 0.02 | 35.32 | 25.73 | 0.07 |
| Quinones28 | 48.60 ± 5.01 | 44.08 ± 5.36 | 0.14 | 0.0 | 19.62 | 12.95 | 0.04 |
![]() |
|||||||
| Other works for comparison | |||||||
| Odegova et al.17 | 41.00 | — | — | 0.78 | 77 | — | — |
The most relevant work for comparison is Odegova et al.,17 from which we obtain the majority of our DES dataset. Their work achieves an RMSE of 41 K, but requires much more expensive features than our work, including the individual component melting points and corresponding xe values. However, their overall goal is to predict the melting points at any given mixture ratio, which differs from our work which is trying to predict the minimum reachable eutectic melting point regardless of ratio.
The KRR model trained on DES data has slightly better zero-shot performance on fatty acids than zero-shot predictions directly from the molten salts model. The molten salt A,A + A,B model zero-shot predictions for fatty acids follow a trend roughly aligned with the x = y line with performance approximately equal to trivial baselines, while after transfer learning and KRR, all error values are slightly better than the trivial baselines. This result is meaningful: achieving baseline-level zero-shot performance indicates that the trained model correctly identifies the approximate average eutectic melting temperature of fatty acids, which exhibits a standard deviation of only about 20 K across the test set. Given that the molten salt training data have a standard deviation roughly an order of magnitude larger, this represents a significant achievement.
Both the zero-shot predictions from the molten salts model and the KRR model that was trained with deep eutectic solvent data are able to perform well on the fatty acids. The fatty acids are the only computationally generated dataset we examined in this work, as Kahwaji et al.11 used the ideal thermodynamic model to compute these. Furthermore, fatty acids are themselves a type of deep eutectic solvent, although we ensured that the fatty acids in this set were not present in the DES training or test sets. Both of these factors could potentially make this set an easier extrapolation problem than the explosives or quinone sets.
We also include test results on explosive eutectics12 and quinones.28 Our predictive performance on these datasets is poor in the context of their respective baselines. However, the neural network and KRR predictions are informative with regards to our models' ability to recognize which mixtures belong to similar families.
Feature extraction and KRR, using DES training data, results in improved predictions on the DES test set, as expected. However, this model still appears unable to extrapolate to other families of organic mixtures, such as explosives and quinones. The same model performs well on fatty acids, which are a type of deep eutectic solvent not seen during model training. Therefore it is possible that the difficulty of extrapolating to explosives and quinones is because these families of materials are not similar enough to deep eutectic solvents.
Our results on the computationally-generated fatty acids dataset suggest that our model is learning a function correlated with the ideal thermodynamic model despite not requiring any individual component melting point or enthalpy data.
Our architecture is able to learn eutectic melting points from only the graph structure and element composition of the molecules in the mixture, and to learn models that are correlated with the ideal thermodynamic model without requiring the xe value, individual components' melting points, or enthalpies. As a result, this architecture can be used to screen a much wider variety of materials than comparable methods, drastically reducing the barriers to high-throughput screening for eutectic mixtures. This has the potential to speed up materials design for sustainable energy applications such as development of battery electrolytes, thermal energy storage materials, and molten salt nuclear reactors.
This method is also able to use individual components' melting temperature data to augment the eutectic mixture data, leading to both improvements in the prediction of melting temperatures for mixtures, as well as the ability to predict additional individual components' melting temperatures.
We also publish a curated dataset of 2244 molten salt eutectic pairs, including their structural information and eutectic melting points, in order to promote further research by the community.
Code repository is also available at: https://github.com/nilamandal/SGNN_Eutectic_Mixtures/.
Footnote |
| † Reprinted from Thermochimica Acta, 660, S. Kahwaji and M. A. White, Prediction of the properties of eutectic fatty acid phase change materials, 94–100, 2018, with permission from Elsevier. |
| This journal is © The Royal Society of Chemistry 2026 |