Open Access Article
Mohammad
Soleymanibrojeni
,
Celso Ricardo
Caldeira Rego
,
Meysam
Esmaeilpour
and
Wolfgang
Wenzel
*
Institute of Nanotechnology (INT), Karlsruhe Institute of Technology (KIT), Hermann-von-Helmholtz-Platz 1, 76344 Eggenstein-Leopoldshafen, Germany. E-mail: wolfgang.wenzel@kit.edu
First published on 19th December 2023
Li-ion batteries store electrical energy by electrochemically reducing Li ions from a liquid electrolyte in a graphitic electrode. During these reactions, electrolytic species in contact with the electrode particles form a solid-electrolyte interphase (SEI), a layer between the electrode and electrolyte. This interphase allows the exchange of Li ions between the electrode and electrolyte while blocking electron transfer, affecting the performance and life of the battery. A network of reactions in a small region determines the final structure of this interphase. This complex problem has been studied using different multi-scale computational approaches. However, it is challenging to obtain a comprehensive characterization of these models in connection with the effects of model parameters on the output, due to the computational costs. In this work, we propose an active learning workflow coupled with a kinetic Monte Carlo (kMC) model for formation of a SEI as a function of reaction barriers including electrochemical, diffusion, and aggregation reactions. This workflow begins by receiving an initial database from a design-of-experiment approach to train an initial Gaussian process classification model. By iterative training of this model in the proposed workflow, we gradually extended the model's validity over a larger subset of reaction barriers. In this workflow, we took advantage of statistical tools to reduce the uncertainty of the model. The trained model is used to study the features of the reaction barriers in the formation of a SEI, which allows us to obtain a new and unique perspective on the reactions that control the formation of a SEI.
000 trajectories for different combinations of reaction rates, of which more than half resulted in the growth of some appreciable volume fraction of inorganic and organic SEIs. Given that it is difficult to transfer the rates from first-principles calculations directly and considering that these calculations also require additional approximations, for instance, regarding the homogeneity of the environment for a specific reaction, we aim to investigate which of these reactions is the most important one in observing a particular outcome. We used this initial trajectory to create an initial training dataset for training a Gaussian process classification model. The model learns the relation between reaction barriers and the output of the kMC model. The active learning workflow is implemented to identify uncertainty regions within the reaction barrier space. This approach significantly reduces the effort compared with an exploration of parameter space where each set of barriers is independently sampled. This is accomplished by testing the model with a new sample dataset at the end of each training cycle, and determining the uncertainty of the model about these sampled points. New calculations with the kMC model are performed in these regions and these model outputs were added to the training dataset in the next cycle. By adding new results to the previous dataset, we trained a new model with updated certainty and uncertainty until we reached a representative model with a reliable understanding of the relationship between the reaction barriers and the output of the kMC model. The representative model enabled us to determine which model parameters have more deterministic effects on formation of a SEI. Moreover, this information allows us to use the kMC model in the framework of the multiscale modeling approach, which requires embedding of the lower scale model in the larger scale models, taking into account different considerations such as finding the regions of validity, calibration, and scale matching of the parameters of the two models.19,20 In the following Section 2, we present our methodology including preprocessing of data, training of the model, and performing the active learning cycles, followed by results and discussion in Section 3, which includes the discussions on the model error in active learning cycles. We conclude with the discussion on the effects of reaction barriers on the formation of a SEI based on the kMC model.
| No. | Reactants/event | Product(s) | Barrier range [eV] | Rep. color |
|---|---|---|---|---|
| 1 | Electrode surface + Li+/EC− | Li+·oEC− | 0.28–0.50 | Green |
| 2 | Electrode surface + Li+·oEC− + Li+ | Li2CO3 + C2H4 ↑ | 0.27–0.47 | Red |
| 3 | Li2CO3 surface + Li+/EC− | Li+·oEC− | 0.31–0.55 | Green |
| 4 | Li+·oEC− + Li+·oEC− | Li2EDC | 0.30–0.53 | Orange |
| 5 | Li2EDC + Li2EDC | (Li2EDC)2 | 0.55–0.97 | Blue |
| 6 | (Li2EDC)2 + Li2EDC | Organic SEI | 0.47–0.83 | Magenta |
| 7 | Li2CO3 surface + Li+·oEC− + Li+ | Li2CO3 | 0.49–0.88 | Red |
| 8 | Li2EDC + organic SEI | Organic SEI | 0.46–0.81 | Magenta |
| 9 | (Li2EDC)2 + organic SEI | Organic SEI | 0.46–0.81 | Magenta |
| 10 | (Li2EDC)2 + (Li2EDC)2 | Organic SEI | 0.46–0.81 | Magenta |
| 11 | Organic SEI + organic SEI | Organic SEI | 0.46–0.82 | Magenta |
| 12 | Diffusion of (Li2EDC)2 | — | 0.40–0.70 | — |
| 13 | Diffusion of Li2EDC | — | 0.35–0.61 | — |
| 14 | Diffusion of C2H4OCOOLi | — | 0.35–0.62 | — |
| 15 | Diffusion of an organic SEI | — | 0.38–0.68 | — |
| 16 | Li+·oEC− going out of the box | — | 0.01 | — |
| 17 | (Li2EDC)2 going out of the box | — | 0.01 | — |
| 18 | Organic SEI going out of the box | — | 0.01 | — |
| 19 | Li2EDC going out of the box | — | 0.01 | — |
000 two-dimensional kMC calculations of size 50 × 50 nm2, with the position and the type of species, such that the coordinates of each pixel represent 1 nm2. Each kMC simulation receives 15 reaction barriers as its parameters, as shown in Table 1, and returns an output with the species type and their coordinates. A descriptor transforms the model output. The descriptor used for this purpose bins the radial distances from the center of the electrode at coordinate (0, 25) to a radius of 25 nm at 0.25 nm steps (100 indices) and counts the SEI species (Li+·oEC−, Li2CO3, and Li2EDC, and dimerizations of Li2EDC) at each bin and normalizes this number. An independent component analysis (ICA)21 is performed on this set of 100-dimensional vectors to reduce the dimensionality to 10. The ICA is performed using the FastICA algorithm.22 The ICA deals with the process of decomposing an observed variable Xn into a matrix product of As, where An×m is the unknown mixing matrix and sm is the independent component of a non-Gaussian distribution with a reduced dimension to m. The output of this section is a dataset of size (50
000 × 10) and a trained FastICA model for the transformation of the kMC outputs in the following steps.
, and assigning each data point i to a cluster k using the indicator function
.![]() | (1) |
The clustering starts with a 1000 initial random guess for the location of centroids, and the final result is the lowest squared sum of the distances of the points from their assigned cluster. The output of this step is a set of 50
000 labeled data, cluster centroids, and a K-mean model for labeling the future kMC outputs.
061, and 12
561. The class empty is the smallest class with 561 members, and it is not trimmed for datasets larger than 2000.
![]() | (2) |
The term p(f|X,ỹ) is defined as inference over f. There are approximate methods to handle the determination of posteriors that are analytically intractable or expensive due to the involvement of non-Gaussian likelihoods or the inversion of large matrices. Variational inference is a method of approximate Bayesian inference. In this method, the inducing random variable u and inducing points Z, u = f(Z) are introduced, and the inference over f is approximated by using a distribution q(u).24 The distribution of the inducing points is optimized during the training of the model so that the approximate posterior, q(u), becomes close to the true posterior, p(f|X,ỹ). This closeness is determined by using Kullback–Leibler divergence.25 In the optimization of this divergence, the evidence p(ỹ|X) finds a lower bound. The prediction using approximate variational inference is defined according to eqn (3), and the actual class label obtained from
.26
![]() | (3) |
The probability of the observation of the corresponding distribution of classes is defined in eqn (4).
![]() | (4) |
In this work, we used the VariationsStrategy class from the gpytorch package for creating and training the model and an anisotropic radial basis function kernel with a diagonal length scale matrix of Λ, eqn (5), which finds the relationships between the 15 input features (reaction barriers), x and x′, for each of the four classes, a total of 60 length scales.
![]() | (5) |
The DirichletClassificationLikelihood with heteroscedastic noise and α∈ = 10−2, loss function by the VariationalELBO class from gpytorch, and the Adam optimizer of the pytorch package with a learning rate of 10−3 were used. At each cycle, the dataset was divided into 30% test and 70% training sets. The inducing points were equally distributed (by class) and randomly selected from 5% of the training set. The location of the inducing points is also optimized. The optimization was carried out in 200
000 steps. The output of this section is a trained model based on the training dataset and inducing points corresponding to the 5% of that training dataset.
| ∣B-CM∣ < 1.2 A | (6) |
| kpca(X,X) = (γXX⊤ + C0)d | (7) |
, and the maximum is 1. To consider a prediction uncertain, we have set a threshold of 0.40 for this measure. The higher the threshold, the more data must be sent for direct query with kMC. The output of this section is an uncertain dataset of variable size to be queried directly with the kMC program.
The SEI-Init WaNo reads a Pickle-file containing a collection of 50
000 kMC outputs after applying the descriptor (2.1), and a Json-file of their relevant reaction barriers. After the data processing step described in Sections 2.1, 2.2, and 2.3, the Tdata.json file is sent to the SEI-AL Wano. In the SEI-AL WaNo, the active learning cycle, including training the model, sampling the reaction barrier space, and selecting the informative data points, as described in Sections 2.4, 2.6, and 2.7 is performed. This WaNo generates the models and the data (PF.json) and is also responsible for creating a list of data points for a direct query with the kMC program in the KMC WaNo, 2.8.
The DB-Generator collects all the processed data from the previous WaNos into a single YAML file. It also triggers the Update-PF to update the dataset after omitting the outliers as described in Sections 2.5 and 2.9, for the next active learning cycle. This framework allows monitoring multiple batches of calculations for independent parameters during different active learning cycles. Additionally, a database is generated after each active learning cycle, which we store in a designated GitHub repository.
A Colab notebook is used for analyzing, visualizing, and processing the database. All the workflow nodes, notebooks, dependencies, and documentation are available on the following repository: https://github.com/KIT-Workflows/SEI-Model-Active-Learning.
000 kMC model output is classified into the four classes. In the classification problems, it is necessary to classify the dataset into classes with independent and distinguishable features. This requires an efficient description-conversion-classification schema. This schema effectively maps a continuous model output space into a discrete class label space. The number of classes was determined based on our expectations of the present classification problem and its distinguishable features. The classes inorganic SEI and organic SEI are the output classes for SEI products. In the class unfinished, intermediate products are mainly formed, but in the class empty, the reactions end without producing any SEI components. It is important to study the unfinished class because it helps identify the reactions that control the decomposition of the electrolyte species, or the loss of active lithium, without forming a SEI. The class empty carries information about the possibility of not producing any SEI even though the system has an active electrode and excess reactants. In the training of the model on the initial dataset, we aimed to find the smallest dataset size with the lowest error to improve the computational efficiency of the workflow. The initial dataset not only plays a role in the training of the model, but it also determines the sampling of the space, as shown in Section 2.6. We created smaller datasets, as explained in Section 2.3, called trimmed datasets, and evaluated the error of the model trained with them. The error of the classification model is the ratio of misclassified test points to the size of the test set. In the classification problem, it is not only important to be able to predict the correct class given the input parameters, but also to avoid misclassification in other class labels. This makes the expression and handling of errors in classification problems different from regression problems.
A general representation of the four classes is given in Fig. 2. The procedure for labeling the kMC outputs is presented in Fig. 3 (also ESI Fig. S1–S3†). A more detailed view of the features of classes is provided by presenting the amount of organic and inorganic SEI produced in each class. Fig. 4 shows the corresponding main component of the SEI in the inorganic and organic classes. In Fig. 4a, an inorganic SEI is produced with a maximum of 4 layers (per electrode width of 50 nm), and this maximum corresponds to the length of electron tunneling that was set in our model to be 4 nm. In this class, a minimal amount of organic SEI is produced. In Fig. 4b, the trend is reversed. We see a minimal amount of inorganic SEI (typically one layer) that is the product of the rapid reduction of Li+·EC on the electrode surface. The plot shows a wide range of the amount of produced organic SEI, up to a maximum of 14 layers and most frequently between 2 and 4 layers (per electrode width of 50 nm). Fig. 4c and d show negligible amounts of either organic or inorganic SEI in their classes. The analysis of the error of the model trained with the trimmed datasets in Fig. 6 shows that the lowest error was for the dataset of size 9561. The range of error was between 0.03 and 0.05. We can see that increasing the dataset size did not necessarily lower the error of the model. The dataset with a size of 9561 is selected as the initial training dataset. Fig. 5b illustratively shows the trimmed dataset and its corresponding SEI classes. The results show that our labeled dataset worked well to train an initial classification model by using the provided description-conversion-classification schema and the method to find the smallest efficient trimmed dataset. The workflow continues with a much more efficient dataset and a lower computational cost for training the model at each cycle.
In contrast to the creation of the initial dataset through a design-of-experiment approach, the present active learning approach uses the uncertainty of the model to dynamically choose the next set of training data points to expand the domain of validity of the model. The benefit of this approach is the curation of a training dataset in a fast and efficient way. The strategy of selection of new points, estimation of uncertainty, decision of the size of new data points and the computational cost of labeling, and monitoring the classification error after each cycle are essential in order to successfully train a model through an active learning workflow. The selection of the most informative data points in active learning reduces the need for labeling large amounts of data, making the training process more efficient. Additionally, by including various representative data points the model can generalize better to unseen data. The probability that a test point belongs to one of the four classes is used as a metric to measure the uncertainty of the classification model, as shown in Section 2.7. We determined the class labels of the uncertain test points through a direct query with the kMC program, followed by a description, conversion, and classification of the kMC outputs. The uncertainty of the model about a test point is determined based on a predefined threshold. This threshold also controls the computational costs of direct query and labeling, as well as the growth of the dataset and the computational cost of training the model in the next cycle. The uncertain test points and their determined class labels are added to the dataset for training the model in the next cycle. In each cycle, we analyzed the classification error of the trained model, described as the ratio of misclassified test points to the size of the test dataset, and the confusion matrix, which shows the misclassifications for each class and the instances in which other classes were confused with a given class. Thus, a set of workflow hyperparameters controls the training of a model through active learning. The workflow design in this work guarantees the reproducibility of the results and facilitates control over these hyperparameters.
The model was trained through 15 cycles of active learning. The error of the model with the training step is shown in Fig. 7a. Fig. 7b shows the classification error of the model for each cycle. The figure shows that up to the 8th cycle, there are fluctuations in the error, and after that, the changes in error are reduced. Fig. 7c and d show the confusion matrices after the 14th and 15th cycles. The confusion matrices can be analyzed for each row and each column. A row in the confusion matrix shows the misclassification of the given class. A column, on the other hand, shows the misclassifications of other classes as the corresponding class of the column. For example, in Fig. 7d, the row related to class empty shows that of the total test points belonging to class empty, 0.03 of them were classified as class inorganic SEI, 0.096 as class organic SEI, and 0.015 as class unfinished. The column related to class empty shows that 0.02 of test points of class inorganic SEI, 0.01 of class organic SEI, and 0.004 of class unfinished were classified as class empty.
The overall error at the 14th cycle is 0.070. In the 15th cycle, we see an overall error of 0.071, but with different details. In Fig. 7d, the result shows an increase in confusion between class empty and class organic SEI compared to the 14th cycle. But this confusion is not symmetric, which means there is no confusion between class organic SEI and class empty. In fact, the confusion with class empty is overall minimal. This means it is not likely that the regions of reaction barrier space corresponding to the classes inorganic SEI, organic SEI, or unfinished will be confused with class empty.
We used the confusion matrix at the 15th cycle to calculate the probability of actually observing the organic SEI class using the trained model, using eqn (8). This equation includes the confusion and errors from other classes and determines the reliability of a prediction for the class organic SEI.
![]() | (8) |
In this equation, AO and PO are the actual class organic SEI and predicted class organic SEI, respectively, and similarly, AE, AI, and AU, are actual class empty, inorganic SEI and unfinished, respectively (ESI eqn (3)†). This means that when the model predicts a data point in the test set as belonging to class organic SEI, the probability of it actually belonging to class organic SEI is 0.918. Similar analyses for other classes are as follows: P(AI|PI) = 0.940, P(AU|PU) = 0.952, and P(AE|PE) = 0.842. These are a comprehensive measure of the accuracy of the model in the prediction of a given class in connection with the probability of observing that class, and confusion of the model about features of classes. The results express that the probability of actual occurrence of class unfinished is greater than that of class inorganic SEI, followed by that of class organic SEI. It is also less likely to observe class empty. In this regard, it should be noted that the class empty corresponds to a condition where the kMC model produces no reaction products despite the electrolyte being in direct contact with the active electrolyte material. This particular condition explains the smaller size of class empty.
One reason for the higher chance of misclassification for class organic SEI compared to class inorganic SEI is due to the broadness of features in this class, as shown in Fig. 4b. This class has an overlap with the features of other classes, which makes other classes an extreme example of class organic SEI.
At the end of each cycle, 50 direct queries with the kMC program were performed. After performing the outlier detection method, as shown in Section 2.9, the final number of new entries to the dataset can be lower than 50. This number is a workflow hyperparameter that controls the computational cost of labeling the new model outputs and training the next model. Fig. 8 shows the number of new entries added to the dataset from each class at the end of each active learning cycle. The lowest number belonged to the class inorganic SEI, and the class empty came in second. The total number of points added to the dataset was as follows: for class inorganic SEI 36, class empty 100, class unfinished 127, and class organic SEI 164. The model inquired about more data points to capture the general features of class organic SEI. On the other hand, the model needed comparatively fewer additional data points about class inorganic SEI and class empty.
In this active learning workflow, a total of 427 new data points were added to the dataset, the training of the model after each cycle converged, and the error of the model evolved and reached a certain level after 15 cycles without significantly increasing the dataset size. The addition of new data points to the dataset during each cycle allowed the model to continuously improve its performance. The convergence of the training after each cycle indicates that the model was able to effectively learn from the newly added data. This shows the hyperparameters of the workflow and the description-conversion-classification schema explained in the previous section integrated well in the active learning cycles.
The behavior of the model after each training cycle was studied using the prediction score, as described in Section 2.7. A sample dataset was created as described in Section 2.6, and each sample point received a prediction score from the trained model at the end of each cycle. Then, for each of the four classes, we obtained a distribution of prediction scores that has an upper quartile, median, and lower quartile. These quartiles allow us to study the performance of the model for each class after each cycle. In addition, we can examine the prediction score of the model for the higher, median, and lower ranges separately through active learning cycles. By analyzing the distribution of prediction scores, we can identify any patterns or trends in the behavior of the model after each cycle. This information allows us to assess the consistency and stability of the model during the active learning cycles, enabling us to make informed decisions regarding necessary adjustments in the performance of the workflow and the number of active learning cycles. In addition, each test point had a relative change in its prediction score with each cycle. We used the relative prediction probability score (RPPS), eqn (9), to discuss this change for the sample dataset. The RPPS provides a quantitative measure of the change in the prediction score for each sample point using the trained model at any cycle, compared to the initial trained model. For each of the four classes, we obtained a distribution of RPPS that has an upper quartile, median, and lower quartile. This information is used to study the relative change in the prediction score for different classes with different prediction scores.
![]() | (9) |
Fig. 9 shows the range of prediction scores between 0.25 and 1. The figure shows a reduction in the prediction score at the end of the 15th cycle compared to the initial prediction score for the sample points with high prediction scores. At the same time, the trend is an increase in prediction score for the samples that initially had a low score, which was defined by our framework hyperparameter of 0.4, as described in Section 2.7. The first quartile of RPPS for classes inorganic SEI, organic SEI, unfinished, and empty is −0.27, −0.20, −0.21, and −0.36, respectively. The sample points that received high prediction scores using the initial model maintained their high scores after active learning cycles. For example, a 0.2 reduction in the prediction score in the range of 0.8 to 0.9 is still in the range of 0.64 to 0.72. The result shows that for the sample points with prediction scores very close to 1, the prediction score does not change after 15 cycles, and RPPS is zero. Through the progression of the active learning cycles, the model readjusts its understanding of the features compared to its initial comprehension. Considering the kernel of the Gaussian process model, eqn (5), the similarity of the features is captured via two metrics: one is their Euclidean distance, and the other is the kernel length scale. An additional learned noise is also added to the model during the training. With the addition of more informative points to the model, the kernel length scale for each feature of each class is adjusted so that all the points within a class are close together (ESI Fig. S4†). Within this optimization of the model hyperparameters, an optimal point is reached where all previous and newly added data points share a maximum likelihood of membership in a class, and this comes with a possible reduction in prediction scores for some data points that initially had a high prediction score. In Fig. 9, the horizontal axis shows the prediction score in the 15th cycle for different classes. The quartiles of the sample dataset having different prediction scores, which are shown as dented color bars on the top, are used for providing more details in the next figure.
Fig. 10 shows the changes in the first (Q1), second (Q2), and third (Q3) quartiles of the prediction scores for the sample dataset vs. the active learning cycles. The Q3 for all classes is consistently high, except for class empty, which shows an increasing trend after the first cycle. The Q2 of the prediction score of the test set for the classes inorganic SEI, unfinished, and empty also shows an increasing trend. The results show a stable value of this quartile of the prediction score of around 0.8 for all classes, except class inorganic SEI, which has reached above 0.9 to 0.93. In contrast to other classes, we see that after the 4th cycle, the Q2 of the class organic SEI reduces from 0.88 to below 0.85 and finally to 0.82, which is closer to that of the classes unfinished and empty. As shown in Fig. 10b and d, the last two cycles show opposite trends. The results show that where the Q2 for class empty increases, it decreases for class organic SEI. This behavior was also visible in the confusion matrix in Fig. 7c and d. The results show that the model's status at the 15th cycle is more representative than at the 14th cycle. The Q1 of the prediction score of the classes also shows similar trends to their own Q2. Except for class inorganic, which reached above 0.7, other classes stabilized between 0.5 and 0.6.
The effectiveness of this active learning workflow in training the classification model is described using a confusion matrix for the different classes at the last cycle, the difference in the number of new entries added to the dataset from each class, and the details of the prediction score using the trained model at each cycle. These results show that the model had different training performances for each class. For classes inorganic SEI and empty, the training through active learning cycles showed a distinct improvement in prediction scores. This can be due to the easier detection of features in these classes compared to the more complex classes of organic SEI and unfinished. The confusion matrix of class unfinished showed minimal confusion with other classes. On the other hand, the model added more new entries about class unfinished compared to class inorganic SEI and class empty. The model added the most new entries related to class organic SEI. The progression of the prediction scores during the active learning cycles shows that with the addition of more entries related to class organic SEI, the prediction scores changed and reached a level of stability. This result also demonstrates the effectiveness of the presented sampling procedure in identifying the most informative data points from the reaction barrier space to improve the ability of the model to find the features of this class. This active learning workflow ensures that the training dataset is representative, resulting in a more accurate training and generalized model.
We created a sample test, as described in Section 2.6. The model classifies the test set. The minimum prediction score threshold of 0.6 was also set to exclude less certain predictions. We obtained a distribution of reaction barriers for each of the 15 reactions in each of the four classes. These distributions provide valuable insights into the behavior of the features across different classes, allowing us to identify patterns and draw conclusions. We measured the upper quartile, median, and lower quartile from each distribution. Quartiles are used to understand a distribution's spread and the presence of outliers. They are particularly useful for the characterization of a distribution that is skewed or has extreme values. We defined a parameter Δ as eqn (10),
Δi = Q3,i − Q1,i, i∈{1, 2, …, 15} | (10) |
which is the interquartile range between the third and the first quartile of the generated parameters for each SEI class. We used this metric as a measure of the expansion of the parameter distribution for each SEI class. The location of the second quartile, or median, is also used as the representative value of the feature for each class. In this case, the reaction barrier with the unit of eV, which determines the rate of the reaction, is the median, and the observed spread of this reaction barrier is Δ2, with the unit of [eV]2. The collection of results is given in Tables 2 and 3 (also ESI Fig. S5–S7†). These tables provide a comprehensive overview of the reaction barriers and their spread for each reaction and each class. The data presented allows a clear comparison between the reactions of each class. It also helps with understanding the factors that contribute to the observation of each class. We discuss the table once for each class and once for the reactions. The results for each class can be interpreted using Δ2 [eV]2 and Q2 [eV]. In these tables, the reactions are divided into three categories: electrochemical (1–4), aggregation (5–11), and diffusion (12–15), in order of their appearance on the table. The electrochemical reactions are divided into two categories: those that happen directly in contact with the electrode and those that happen through electron tunneling. It should be noted that the reactions in this table have been rearranged compared to Table 1 for more contextual clarity. The output of the model is more sensitive to the parameters with a lower Δ2, compared to the parameters with a higher Δ2. The median is used for comparing the reaction barriers between different classes. We also used the median to test these generated parameters to see if they would lead to the correct outputs. For class inorganic SEI, Table 2 shows a small value for Δ2 for the 4th reaction, which controls the electron tunneling from the electrode. The second lowest Δ2 is for the 6th reaction, which controls the aggregation of Li2EDC species. For other reactions, Δ2 is in the same order, and the largest values belong to the 14th and 15th reactions that control diffusion of the aggregated organic SEI.
| Reaction | Class inorganic SEI | Class organic SEI | ||
|---|---|---|---|---|
| Δ 2 | Q 2 [eV] | Δ 2 | Q 2 [eV] | |
| (1) Electrode: Li+/EC− → Li+·oEC− | 2.66 × 10−2 | 0.349 | 2.96 × 10−2 | 0.347 |
| (2) Electrode: Li+·oEC− + Li+ → Li2CO3 + C2H4 | 2.53 × 10−2 | 0.344 | 1.96 × 10−2 | 0.292 |
| (3) Tunneling: Li+/EC− → Li+·oEC− | 2.69 × 10−2 | 0.371 | 2.66 × 10−2 | 0.374 |
| (4) Tunneling: Li+·oEC− + Li+ → Li2CO3 + C2H4 | 6.40 × 10−5 | 0.484 | 2.13 × 10−2 | 0.628 |
| (5) Li+·oEC− + Li+·oEC− → Li2EDC | 2.66 × 10−2 | 0.380 | 2.59 × 10−2 | 0.364 |
| (6) Li2EDC + Li2EDC → (Li2EDC)2 | 1.74 × 10−2 | 0.801 | 1.35 × 10−2 | 0.636 |
| (7) (Li2EDC)2 + Li2EDC → organic·SEI | 2.34 × 10−2 | 0.592 | 2.50 × 10−2 | 0.554 |
| (8) (Li2EDC)2 + (Li2EDC)2 → organic·SEI | 2.99 × 10−2 | 0.562 | 2.76 × 10−2 | 0.547 |
| (9) Organic·SEI + Li2EDC → organic·SEI | 2.62 × 10−2 | 0.557 | 2.31 × 10−2 | 0.544 |
| (10) Organic·SEI + (Li2EDC)2 → organic·SEI | 2.59 × 10−2 | 0.556 | 2.22 × 10−2 | 0.558 |
| (11) Organic·SEI + organic·SEI → organic·SEI | 2.76 × 10−2 | 0.555 | 2.40 × 10−2 | 0.557 |
| (12) Li+·oEC− + Li+/EC− → Li+/EC− + Li+·oEC− | 2.79 × 10−2 | 0.452 | 2.64 × 10−2 | 0.407 |
| (13) Li2EDC + Li+/EC− → Li+/EC− + Li2EDC | 2.10 × 10−2 | 0.471 | 9.60 × 10−3 | 0.476 |
| (14) (Li2EDC)2 + Li+/EC− → Li+/EC− + (Li2EDC)2 | 3.17 × 10−2 | 0.469 | 3.31 × 10−2 | 0.480 |
| (15) Organic·SEI + Li+/EC− → Li+/EC− + organic·SEI | 3.24 × 10−2 | 0.465 | 2.92 × 10−2 | 0.468 |
| Reaction | Class unfinished | Class empty | ||
|---|---|---|---|---|
| Δ 2 | Q 2 [eV] | Δ 2 | Q 2 [eV] | |
| (1) Electrode: Li+/EC− → Li+·oEC− | 2.96 × 10−2 | 0.350 | 2.19 × 10−2 | 0.362 |
| (2) Electrode: Li+·oEC− + Li+ → Li2CO3 + C2H4 | 1.23 × 10−2 | 0.286 | 3.72 × 10−3 | 0.456 |
| (3) Tunneling: Li+/EC− → Li+·oEC− | 2.66 × 10−2 | 0.372 | 2.16 × 10−2 | 0.376 |
| (4) Tunneling: Li+·oEC− + Li+ → Li2CO3 + C2H4 | 1.56 × 10−2 | 0.646 | 1.23 × 10−2 | 0.679 |
| (5) Li+·oEC− + Li+·oEC− → Li2EDC | 2.76 × 10−2 | 0.365 | 2.31 × 10−2 | 0.371 |
| (6) Li2EDC + Li2EDC → (Li2EDC)2 | 2.22 × 10−2 | 0.813 | 3.76 × 10−2 | 0.705 |
| (7) (Li2EDC)2 + Li2EDC → organic·SEI | 3.20 × 10−2 | 0.592 | 2.79 × 10−2 | 0.587 |
| (8) (Li2EDC)2 + (Li2EDC)2 → organic·SEI | 2.72 × 10−2 | 0.557 | 2.56 × 10−2 | 0.568 |
| (9) Organic·SEI + Li2EDC → organic·SEI | 2.79 × 10−2 | 0.574 | 2.10 × 10−2 | 0.523 |
| (10) Organic·SEI + (Li2EDC)2 → organic·SEI | 2.76 × 10−2 | 0.555 | 2.19 × 10−2 | 0.570 |
| (11) Organic·SEI + organic·SEI → organic·SEI | 2.40 × 10−2 | 0.560 | 2.56 × 10−2 | 0.570 |
| (12) Li+·oEC− + Li+/EC− → Li+/EC− + Li+·oEC− | 2.69 × 10−2 | 0.411 | 2.76 × 10−2 | 0.413 |
| (13) Li2EDC + Li+/EC− → Li+/EC− + Li2EDC | 7.92 × 10−3 | 0.371 | 4.36 × 10−3 | 0.341 |
| (14) (Li2EDC)2 + Li+/EC− → Li+/EC− + (Li2EDC)2 | 3.03 × 10−2 | 0.472 | 3.20 × 10−2 | 0.476 |
| (15) Organic·SEI + Li+/EC− → Li+/EC− + organic·SEI | 3.13 × 10−2 | 0.458 | 3.88 × 10−2 | 0.463 |
For class organic SEI, as shown in Table 2, the lowest Δ2 belongs to the 13th reaction, which controls the diffusion of Li2EDC, followed by the 6th reaction which controls the aggregation of Li2EDC species, followed by the 2nd and 4th reactions, which control the electrochemical reduction of Li+·oEC− on the surface of the electrode and through electron tunneling, respectively.
For class unfinished in Table 3, the lowest Δ2 belongs to the 13th reaction, which controls the diffusion of Li2EDC, followed by the 2nd and 4th reactions, which control the electrochemical reduction of Li+·oEC− on the surface of the electrode and through electron tunneling, respectively.
For class empty, as shown in Table 3, the lowest Δ2 belongs to the 2nd reaction which controls the electrochemical reduction of Li+·oEC− on the surface of the electrode, followed by the 13th reaction which controls the diffusion of Li2EDC, followed by the 4th reaction which controls the electrochemical reduction of Li+·oEC− through electron tunneling.
The results show a repeating pattern of the lowest Δ2 for the 2nd, 4th, 6th, and 13th reactions. Tables 2 and 3 show the median (Q2) of the reaction barriers for each class. For class inorganic, the reduction of Li+·oEC− through electron tunneling should happen with a barrier of 0.484 eV and the aggregation of Li2EDC species should happen with a barrier of 0.801 eV. This high barrier for aggregation of Li2EDC species also appears for class unfinished and empty, but it is lower for class organic SEI, with a value of 0.636 eV. This difference indicates that aggregation of Li2EDC species is more important for the formation of an organic SEI than other aggregation reactions. For class organic SEI, the results show that the diffusion of Li2EDC happens with a higher barrier compared to class unfinished and class empty. This barrier is 0.476 eV while for class unfinished and empty, it is 0.371 and 0.341 eV, respectively. These results indicate that faster diffusion and slower aggregation of Li2EDC lead to the delay or even prevention of the formation of an organic SEI.
These results can be summarized in the following order:
• For an inorganic SEI, Li+·oEC− should be reduced to Li2CO3 with the reaction barrier a lot lower than the reaction for its aggregation into Li2EDC, as they are competing reactions.
• For an organic SEI, slower diffusion of Li2EDC and faster aggregation of Li2EDC into (Li2EDC)2 are essential to obtain an organic SEI.
• Slow reduction of Li+·oEC−, along with sluggish aggregation of Li2EDC into (Li2EDC)2, combined with fast diffusion of Li2EDC, cause delay in the formation of a SEI.
• Slow formation of Li+·oEC− and fast diffusion of their aggregate Li2EDC, lead to absence of either an organic or inorganic SEI.
Fig. 11 shows the results of kMC calculations based on the parameters generated by the model in Tables 2 and 3 for the different classes (also ESI Fig. S8†). Each calculation was carried out for 1800 s. Fig. 11a shows the concentration (the number of pixels occupied by the species divided by the system size) of different reactants and products of the kMC model for the inorganic SEI sample. The plot shows that the concentration of Li2EDC product reaches a maximum of 7.0 × 10−2 after 3.0 × 10−4 s and with the same trend, the inorganic SEI, Li2CO3 is produced. Fig. 11b shows that the starting time for formation of (Li2EDC)2 is around 7.0 × 10−5 s. After 6.06 × 10−4 s the concentration of Li2EDC reaches its maximum of 7.0 × 10−2, which subsequently causes the organic SEI to reach its maximum concentration.
Fig. 11c shows that Li2EDC is continuously produced with an increase in its concentration, while the production of Li+·oEC− shows a steady and stable trend with a concentration of around 1.0 × 10−2, without formation of any further organic or inorganic SEI. Fig. 11d shows that both Li2EDC and Li+·oEC− were produced at a limited concentration of maximum 4.0 × 10−3 followed by a downfall trend to zero, without production of any inorganic or organic SEI. The event of diffusion out of the box occurred for inorganic SEI 137, organic SEI 8, unfinished 584, and empty 29 times (ESI Fig. S9†). Delay in the formation of a SEI causes the reduced species to diffuse out of the simulation box without the formation of SEI products, which means exhaustion of Li resources. This result shows that the model learned the features of the parameter space corresponding to each class of SEI.
Footnote |
| † Electronic supplementary information (ESI) available. See DOI: https://doi.org/10.1039/d3ta06054c |
| This journal is © The Royal Society of Chemistry 2024 |