Open Access Article
This Open Access Article is licensed under a
Creative Commons Attribution 3.0 Unported Licence

Machine learning for flow batteries: opportunities and challenges

Tianyu Li , Changkun Zhang * and Xianfeng Li *
Division of Energy Storage, Dalian National Laboratory for Clean Energy (DNL), Dalian Institute of Chemical Physics, Chinese Academy of Sciences, Zhongshan Road 457, Dalian 116023, China. E-mail: zhangchk17@dicp.ac.cn; lixianfeng@dicp.ac.cn

Received 17th January 2022 , Accepted 6th April 2022

First published on 7th April 2022


Abstract

With increased computational ability of modern computers, the rapid development of mathematical algorithms and the continuous establishment of material databases, artificial intelligence (AI) has shown tremendous potential in chemistry. Machine learning (ML), as one of the most important branches of AI, plays an important role in accelerating the discovery and design of key materials for flow batteries (FBs), and the optimization of FB systems. In this perspective, we first provide a fundamental understanding of the workflow of ML in FBs. Moreover, recent progress on applications of the state-of-art ML in both organic FBs and vanadium FBs are discussed. Finally, the challenges and future directions of ML research in FBs are proposed.


image file: d2sc00291d-p1.tif

Tianyu Li

Tianyu Li is currently a postdoctoral research associate at Dalian National Laboratory for Clean Energy, Dalian Institute of Chemical Physics, Chinese Academy of Science. He received his B.S. degree from Dalian University of Technology in 2014 and Ph.D. degree from Dalian University of Technology in 2019. His current research focuses on the R&D of key components for electrochemical energy storage by machine learning and theoretical calculations.

image file: d2sc00291d-p2.tif

Changkun Zhang

Changkun Zhang is a full professor at Dalian Institute of Chemical Physics, Chinese Academy of Science. He was a postdoctoral researcher at University of Texas at Austin from 2016 to 2020. He received his B.S. in Chemical Engineering and Technology from Harbin Engineering University and Ph.D. in Chemical Engineering from Dalian Institute of Chemical Physics, Chinese Academy of Science. His work focuses on the novel key materials for next-generation electrochemical energy storage, including redox-active molecules and electrolyte microstructures.

image file: d2sc00291d-p3.tif

Xianfeng Li

Xianfeng Li received his Ph.D. degree in Polymer Chemistry and Physics from Jilin University. He was appointed as a full professor at DICP, CAS, in 2012. He currently serves as the head of the energy storage division at DICP. His research interests include key materials and core technologies of flow batteries (vanadium flow battery, zinc-based flow batteries, and novel flow battery systems), innovation battery technologies (lead carbon battery, supercapacitor and lithium/sodium/potassium/zinc-based battery), the structure design and simulation of batteries, and battery systems, including their testing and evaluation, industrial development and application demonstration.


Introduction

Worldwide economic growth has boosted the demand for energy, while the massive use of fossil fuels continues to cause environmental issues. To reduce greenhouse gas emission and meet the growing demand for energy consumption, more attention has been paid to renewable energy sources such as solar and wind power. To integrate the intermittent and unstable renewable energies into the grid, there is an urgent need for a safe, economic and environmental-friendly large scale energy-storage system to balance the renewable energy supply and electricity demand.1–3 Several energy-storage technologies, such as physical storage methods (e.g., pumped storage hydro,4 flywheels,5 and compressed-air6), electrochemical methods (e.g., Li-ion,7 lead-acid,8 and FBs9,10), and chemical methods (e.g., hydrogen energy storage11), are available for electricity storage. Owing to the merits of decoupled energy and power, high safety, high efficiency, and long cycle life, FBs are well suited for the integration of renewable energy into the grid up to 100 MW for long term energy storage of 4 hours or longer.10

In a FB system, energy is typically stored in electrolyte solutions, which normally consist of redox-active couples (i.e., catholytes and anolytes) and are separated on the opposite side of an ion conductive membrane. A schematic diagram of a vanadium flow battery (VFB) is provided in Fig. 1. Electrochemical redox reactions occur on the electrode surfaces in the electrochemical cell to convert chemical energy into electricity during discharge to supply the power, and the reversible electrochemical redox reactions do the opposite to store electricity during charge. The modern concept of FBs was proposed by the National Aeronautics and Space Administration (NASA) in 1970s,12 in which Fe3+/Fe2+ and Cr3+/Cr2+ redox-active couples were employed as the catholyte and anolyte, respectively.13 Since then, many new FBs have been designed, including metal- or inorganic-based redox species (e.g., VFBs,14,15 Zn–Fe,16,17 Zn–Br,18,19 Zn–I,20,21 Zn–Mn22) and organic redox-active molecules (e.g., quinones,23,24 TEMPO,25,26 viologen,27 and phenazine25). VFBs, which employ vanadium with different valences as catholytes (V3+/V2+) and anolytes (VO2+/VO2+), are regarded as one of the most promising large-scale energy storage technologies among the mental-based FBs, and have already been commercially implemented in recent years.2 Zinc-based FBs are another promising large-scale energy storage technology, owing to the low redox potential of −0.76 V (vs. the standard hydrogen electrode SHE) of the zinc deposition/stripping reaction, high theoretical capacity (820 mA h g−1, 5855 mA h cm−3), and low cost of zinc. However, the stability of zinc-based FBs needs to be further improved. Since the first organic FB system was reported,23 more research studies have focused on the discovery and design of novel organic redox-active species (ORASs). ORASs are normally composed of carbon, hydrogen, oxygen, nitrogen, and sulphur, which are earth-abundant elements.9,28,29 Thus, they could potentially offer low-cost electrolytes. Furthermore, the properties, including solubility, redox potential, kinetics, and stability, can be precisely tuned by molecular engineering.30 Therefore, organic flow batteries (OFBs) are regarded as a potential option for large-scale energy storage.


image file: d2sc00291d-f1.tif
Fig. 1 A schematic diagram of a VFB.

However, it is time-consuming to use traditional trial-and-error methods from the lab design to commercialization for a new material. Compared with traditional trial-and-error methods, computational chemistry has made great contributions to provide useful information for the research and development (R&D) of new materials, especially the density functional theory (DFT). Moreover, with the help of some modern chemical-simulation toolkits, the high-throughput calculation screening can speed up the discovery of new materials. However, the large-scale screening of new materials still takes too much time due to the high computational cost of high-precision DFT calculations.31

With the rapid development of computer ability and data science, AI (defined as the simulation of human intelligence processes by computer systems) has opened another door to modern research and attracted worldwide attention. In recent years, AI has been employed in the fields of natural language processing,32 image recognition,33 and autonomous driving.34 It shows great potential to surpass the existing cognitive level of human beings in some fields.35 ML is one of the most important independent disciplines of AI, which has been rapidly developed and widely applied recently.36–38 ML algorithms can automatically mine implicit relationships hidden behind the data from a large amount of data. With the development of chemoinformatics, it has vast applications in the fields of chemistry and material science, such as quantum chemistry,39,40 drug discovery,41 and molecular design,42 reaction prediction and reverse synthesis,43 and automated synthesis.44

In this perspective, we introduce the basic workflow of ML and discuss the R&D of ML in FBs, highlighting the successful applications of state-of-art ML in OFBs and VFB systems. Moreover, the prediction of the physical and electrochemical properties of organic redox-active molecules for FBs based on high-throughput computational simulations is included. Finally, we provide a perspective on the main limitations and future research directions of ML for FBs, and present a realistic outlook.

The workflow of ML

The general application workflow of ML is shown in Fig. 2. ML is an interdisciplinary multi-field, involving probability theory, statistics, approximation theory, convex analysis, algorithm and others. It applies algorithms to establish the hidden relationships between massive data and specific properties. Thus, the first step is dataset construction, which collects sufficient data samples for ML application. The second step is feature engineering, which creates new features based on the raw data by mathematical representation. This mathematical representation should refine the correlation between features, resulting in a few new features to reflect the sample information. Followed by randomly dividing the data into training and test datasets, a ML algorithm is employed to establish a model between features and target function with training dataset. The parameters of the algorithm are adapted to achieve the goal that the accuracy and computation cost of the model are acceptable. Then, the accuracy (performance) of the model is evaluated by a test dataset. Finally, the validated model can be employed to predict the properties of unknown data.
image file: d2sc00291d-f2.tif
Fig. 2 The general application workflow of ML.

High-throughput calculations and database

A database is the basis of ML. Generally, the database can be constructed by the following methods.
(1) Doing high-throughput experiments or calculations. Severson et al. generated a database with 124 commercial lithium iron phosphate/graphite cells cycled under fast-charging conditions by experiment, and then applied ML to predict the cycle life of the cell by the discharge voltage curves from the early cycle capacity degradation.45
(2) Extracting and summarizing the data from publications and patents. However, most of the data reported in the literature are the best results under optimal conditions from successful experiments. The failed data from experiments are also generally deliberately hidden, which would cause serious data bias.
(3) Taking advantage of the existing open chemical or material databases. Such as GDB-13,46 GDB-17,47 ZINC,48 NREL materials database,49 OQMD50 and others. Some notable material databases are shown in Table 1. The computation method plays an important role in generating material properties. By employing open databases, we can avoid the time-consuming data preparation and generation process. Another advantage of databases obtained by high-throughput calculations is that they provide an open source and user-friendly Materials Application Programming Interfaces (API) for users.51 If the database does not include the properties we need, high-throughput calculations can be applied to obtain them. Some existing open-source automated interfaces, such as Materials Project,52 Python Materials Genomics (pymatgen),53 the open quantum materials database (OQMD)50 and others, can provide modules to perform high-throughput calculations including classical molecular dynamics and quantum mechanics calculations (e.g., ab initio and DFT calculations).
Table 1 Some open source materials databases
Database name URL Descriptions
The Materials Project https://materialsproject.org Computed information on known and predicted materials including inorganic compounds, organic molecules, nanoporous materials
OMDB https://omdb.mathub.io An open access electronic structure database for 3-dimensional organic crystals
NRELMatDB https://materials.nrel.gov A computational materials database focus on materials for renewable energy applications
OQMD https://oqmd.org DFT calculated thermodynamic and structural properties of 815[thin space (1/6-em)]654 materials
GDB-13 https://www.cbligand.org/gdb13 Databases of 970 million hypothetical small organic molecule
GDB-17 https://gdb.unibe.ch/downloads Databases of 166 billion hypothetical small organic molecules
PubChem https://pubchem.ncbi.nlm.nih.gov Include freely accessible chemical information for small organic molecules
ZNIC https://zinc.docking.org A database for purchasable compounds
NIST Chemistry WebBook https://webbook.nist.gov/chemistry Thermochemical data for over 7000 organic and small inorganic compounds, reaction thermochemistry data for over 8000 reactions, IR spectra for over 16[thin space (1/6-em)]000 compounds, mass spectra for over 33[thin space (1/6-em)]000 compounds and so on
CCDC https://ccdc.cam.ac.uk A database for crystal structure data
COD https://www.crystallography.net/cod A database for crystal structures of organic, inorganic, metal–organics compounds and minerals, excluding biopolymers
ChemSpider https://www.chemspider.com Chemical information based on chemical structures, including physical and chemical properties of compounds


Feature engineering

Feature engineering normally includes data pre-processing, feature extraction, selection and construction, which employs relevant knowledge in the data science to create features that enable ML algorithms to achieve the best performance. After the original dataset is constructed, feature engineering is a key step for applying ML algorithms. For an organic energy storage material, how molecules are represented (which is called a descriptor) and whether the descriptor contains key information will directly determine the performance upper limit of the model. The commonly used molecular descriptors include physicochemistry properties (e.g., log[thin space (1/6-em)]P, pKa, molecular weight, and properties from DFT or semi-empirical calculations), molecular fingerprints (e.g., MACCS keys,54 PubChem fingerprints,55 MolPrint2D,56,57 Morgan fingerprint,58 and others), molecular abbreviation (e.g., SMILES,59 InChI,60 and SMARTS61), molecular graph (e.g., Coulomb matrix39), and grid representation. After transforming a molecule into a computer-recognized mathematical representation, a feature selection method is employed to eliminate irrelevant or redundant features, so as to reduce the number of features, improve the model accuracy, and reduce the training time of the model. Filter feature selection, wrapper feature selection, and embedded feature selection are three commonly used methods. In short, feature engineering is a complex but extraordinarily crucial process for the application of ML, which directly affects the performance upper limit of the model.

ML algorithms and models

Algorithms are the key to ML, and it can be generally divided into classical ML algorithms based on statistics and neural network. The classical ML algorithms normally include Bayesian, Decision Tree (DT), Support Vector Machine (SVM), Cluster analysis and Random Forest (RF). The Scikit-learn package62 in Python has integrated most of the classical ML algorithms, which can be easily accessed. The popular artificial neural network mainly includes fully connected neural network (FCNN), convolutional neural network (CNN), recurrent neural network (RNN), auto encoder (AE), and generative adversarial network (GAN). Many open source ML frameworks, such as TensorFlow,63 PyTorch,64 and Theano,65 can be accessed to establish the neural network.

Supervised learning algorithms are commonly applied in properties prediction of organic redox-active molecules. It requires labels (e.g., physicochemical properties to be predicted) of the sample in the dataset. A higher accuracy of the learning model can be acquired with higher accuracy of the label and better representation of the sample. In specific applications, the most suitable ML algorithm can often be determined by screening and verifying existing algorithms. It should be noted that the optimal model based on the training dataset and the validation dataset may not be the most suitable model. This is due to the over-fitting or poor generalization ability of the algorithm. Therefore, it needs to be further tested and verified by an external test dataset to determine the optimal prediction model.

Linear model. Linear model is the simplest ML algorithm. Assuming that there are n samples, each sample includes a input vector Xi = (x1i, x2i, …, xdi) and a corresponding value yi. In linear mode, the predicted value, f(Xi) = ω1x1i + ω2x2i + … + ωdxdi + b or f(Xi) = ωTxi + b, where ω = (ω1, ω1,…, ωd) is the regression coefficients vector. The error between the difference of the predicted value f(Xi) and the true value yi is called loss. A loss function is the sum of each error for n samples. The mean squared error (MSE) is commonly used to avoid the positive and negative differences between the predicted value f(Xi) and the true value yi. If we use MSE as the loss function, it can be calculated as follows,
 
image file: d2sc00291d-t1.tif(2.1)

The object of the learning algorithm is to find optimal regression coefficients vector ω and intercept b to minimize the loss function. To prevent overfitting, regularization is normally used. Lasso regularization66 and ridge regularization67 are commonly used methods to prevent overfitting.

SVM. SVM is a supervised ML algorithm that can be used for both classification and regression. In SVM, each data item is plotted as a point in n-dimensional space (where n is a number of features) with the value of each feature being the value of a particular coordinate. The aim of the algorithm is to find the hyper-plane that differentiates the classes very well. For a regression problem, Support Vector Regression (SVR) is applied. Different from the aim of linear regression, which is to minimize the loss function, SVR gives the flexibility to define how much error is acceptable in the model and will find an appropriate line (or hyperplane in higher dimensions) to fit the data. There are five commonly used kernels functions, including linear kernel, polynomial kernel, Laplacian kernel, Gaussian kernel, and sigmoid kernel. SVM can solve high-dimensional problems with small samples, and can handle the interaction of nonlinear features without relying on the entire data set. However, the efficiency becomes poor if the samples are too large. The interpretation of high-dimensional features is not very clear, and it is sensitive to missing data and the kernel.
DT. DT is a supervised learning algorithm. A DT is a flowchart-like structure in which each internal node represents a “test” on an attribute, each branch represents the outcome of the test, and each leaf node represents a class label. There are three commonly used approaches for DT, which are ID3, C4.5, and CART algorithm. Random forest (RF) algorithm, which is a type of ensemble learning containing multiple DT models, can increase the robustness of a tree-based algorithm.
NN. Inspired by the action mechanism of neurons in the brain, a NN is composed of many nodes (called artificial neurons), containing an input layer, one or more hidden layers, and an output layer. Each node connects to another, and has an associated weight and threshold. If the output of any individual node is above the specified threshold value, that node is activated, sending data to the next layer of the network. Otherwise, no data is passed along to the next layer of the network. Based on different purposes, NNs can be classified into different types, including perceptron, deep neural network (DNN), convolutional neural network (CNN), recurrent neural network (RNN) and others. The activation function is one of the important parts in NNs. The input is calculated by the active function in the neural network, and the result is passed to the next neuron or output as the result of the NN model. Optimization algorithms, such as stochastic gradient descent (SGD), RMSprop, Adagrad, Adadelta and Adam, are commonly used to calculate and optimize the loss of the model.

Model evaluation

Evaluating the model is a core part of building an effective ML model. The ML model learned by the algorithm cannot cover all situations, which will lead to a difference between the actual predicted output and the true value of the sample. The error of the ML model on the training and test set is called the training error and test error, respectively, and the error on the new sample is called the generalization error. The aim of the ML model is to have a small generalization error. However, it is hard to calculate the generalization error. Thus, the commonly used method is to divide the dataset into a training set, a validation set and a test set, and they are applied to train the model, adjust the parameters, and calculate the test error, respectively. The test error is regarded as the approximate evaluation of the generalization error to evaluate the accuracy of the ML model.
Bias and variance. The generalization error is the sum of bias, variance and noise, which is determined by the learning algorithm, the sufficiency of the data, and the difficulty of the learning task, respectively. If the bias is small but the variance is high, the ML model is overfitting, which indicates the model is too complex. If the bias and variance are both very high, the ML model is underfitting, which indicates the model is too simple. The process to balance the overfitting and underfitting is to adjust the internal variables (hyperparameters) of the model. Cross validation can be applied to evaluate the generalization performance of the model, which is commonly used in ML.

Application of ML in FBs

In this section, the application of ML for the design and development of organic redox-active species (ORASs) and VFBs is summarized. Database, which was built through high-throughput virtual screen (HTVS) for ORASs is first introduced. The application of ML mainly focuses on redox potential, solubility and stability prediction. Thus, we discuss the establishment of the model for each specific problem and the performance of the ML model. Finally, some ML applications in VFBs, such as electrode structure, membrane optimization and system cost and performance prediction, are reviewed.

Database construction for ORASs

To build the structure–property relationship for ORASs, a database is needed. HTVS have been found to be an effective technology for database generation, especially for organic molecules, owing to its low cost and high efficiency. Er et al. has applied a HTVS approach to study potential candidate organic molecules for quinone derivatives using DFT calculations.68 The virtual library includes 17 core structures of quinones, which were decorated with 18 substituents (–N(CH3)2, –NH2, –OCH3, –OH, –SH, –CH3, –SiH3, –F, –Cl, –C2H3, –CHO, –COOCH3, –CF3, –CN, –COOH, –PO3H2, –SO3H, and–NO2), as shown in Fig. 3. The single and full substitutions were taken into consideration for the redox potential and solvation-free energy calculation. DFT calculations were employed to calculate the energy of the quinone molecules. The total database includes 1710 quinone (Q)/hydroquinone (QH2) molecular couples. Results showed that it should be possible to adjust the standard redox potential from as high as 0.6 VNHE to −1.5 VNHE. Ultimately, 408 Q/QH2 couples were screened with E0 < 0.2 V vs. SHE and E0 > 0.9 V vs. SHE. Tabor et al. extended the research of the redox potential and solvation-free energy to stability for Q/QH2 couples.69 They combined DFT and semi-empirical calculations to study the decomposition or instability mechanisms from the virtual screening of more than 140[thin space (1/6-em)]000 quinone pairs. Their research indicated that HTVS is an effective method to construct databases for ORASs.
image file: d2sc00291d-f3.tif
Fig. 3 A schematic representation of the molecular screening library. The parent BQ, NQ, and AQ isomers are shown on the left (white). These quinone isomers are functionalized with 18 different R-groups singly (gray) and fully (green) to generate a total of 1710 quinone molecules. Reproduced with permission from ref. 68. Copyright 2015 Royal Society of Chemistry.

Carbonyl reductions to alcohols and amines are two of the most common carbon redox transformations in biology. The database of redox potentials of carbonyl reductions to alcohols and amines can also be used for ORASs. A constructed database with standard potentials of more than 315[thin space (1/6-em)]000 redox reactions involving approximately 70[thin space (1/6-em)]000 compounds by PM7 calculation was calibrated with Gaussian process (GP) regression, which can be fully assessed at https://github.com/aspuru-guzik-group/gp_redox_rxn.70 The construction of the molecular structure–property database using the HTVS strategy for ORASs mainly includes three stages: molecule library generation, molecule structure and property generation, and database creation.71 The process for the development of RedDB is shown in Fig. 4.


image file: d2sc00291d-f4.tif
Fig. 4 A schematic overview of the various tasks that have been undertaken for the development of RedDB. Reproduced with permission from ref. 71. Copyright 2021 ChemRxiv.

The main weakness of HTVS is the failure to address the error between the calculated value and the experiment. Wedege et al. systematically studied the redox potential, solubility, and stability of 28 quinone-based compounds by experiment, and found that the position of the substituents plays a key role in determining the redox potential and solubility, while the number of substitutions only has a minor influence.72 Their experimental data showed that the standard redox potential of sulfonated AQs [electron-withdrawing groups (EWGs)] is only 200–300 mV higher than that of hydroxylated ones [electron-donating groups (EDGs)], which was not fully consistent with the EWGs increasing the redox potential of the anthraquinone (AQ) derivatives, while EDGs decrease the redox potential of AQs.68 The reliability of each sample in the database is the basis of ML. Therefore, the accuracy of theory calculation for HTVS needs to be verified in the first place. Another weakness of HTVS in building a database is that the calculation level employed by researchers varies from one to another, which builds a barrier for other researchers when combining these databases for further ML applications. Thus, more comprehensive, standard data samples should be established to share between databases.

Applications of ML in ORASs

ML is widely applied in the field of property predictions. For ORASs, the solubility, redox potential and the stability are most frequently involved. Ideal ORASs should have high solubility, rational redox potential and high stability to have high energy density and long cycle life. However, obtaining these properties by experiments and high-precise computation is expensive and time-consuming. Therefore, the ML approach with high efficiency attracted much attention for these property predictions. Since the accuracy of the ML model is highly related to the amount and reliability of the data, the prediction of the properties of ORASs is often combined with high-throughput quantum mechanical calculations.

The estimation of aqueous solubility of organic compound has been widely studied for many years.73–82 The prediction methods can normally fall into three categories: the first type is based on group contribution method.73,83 The second employs regression from molecular parameters, which is also regarded as quantitative structure–activity relationships (QSAR).84,85 The third applies a data-driven approach (ML algorithms), including SVM,75,77 RF,86 NN,80,87 and others,81 to establish a correlation between the molecular structure and solubility. In 2019, Sorkun et al. built a new database AqSolDB, which consisted of 9982 unique organic compounds, for the prediction of aqueous solubility.88 In 2021, Francoeur et al. proposed a ML tool, SolTranNet (the general architecture is shown in Fig. 5) which contained 3393 parameters, for the fast aqueous solubility prediction based on the AqSolDB dataset.82 A specific ML model called multiple descriptor multiple kernel (MultiDK) method was developed to discover ORASs for aqueous organic flow batteries (AOFBs).89 By combination of binary descriptors (e.g., Morgan fingerprint and the MACCS keys) and nonbinary descriptors (e.g., the physicochemical molecular property), the prediction accuracy of aqueous solubility improved significantly. The combination of the Tanimoto similarity kernel for binary descriptors and linear kernel for nonbinary descriptors showed better solubility prediction performance than LR. As a result, MultiDK can predict pH-dependent solubility for quinone and its derivatives. Although many solubility models have been developed and validated by researchers, the overall performance of these models for a blind dataset was not as encouraging as that reported in the published papers.90 The main reason is the lack of high quality experimental data and structural diversity of the training dataset. Thus, to construct more reliable solubility prediction models, more attention was needed to obtain high quality experimental data and a diverse molecular structure dataset construction.


image file: d2sc00291d-f5.tif
Fig. 5 General architecture of SolTranNet. Each item in a blue box is a tuned hyper-parameter. Reproduced with permission from ref. 82. Copyright 2021 American Chemical Society.

Redox potential is another important property for ORASs. The redox potential of an organic redox-active couple can be calculated from the thermodynamic cycle (Fig. 6) by first-principles computational methods,91e.g., DFT calculation.92 Recently, some structure-redox potential prediction models were established by ML to predict the redox potential of ORASs.70,93–96 For instance, Doan et al. applied the Gaussian process regression (GPR) to forecast the oxidation potential of homobenzylic ethers (HBEs) molecules for application in nonaqueous flow batteries.97 The dataset was constructed by DFT calculations, and contained 1400 HBEs. The GPR model contained a Matérn kernel, a generalization of the radial basis function, to construct the covariance matrices. A total of 9% of HBE molecules were identified with the desirable Eox ∈ [1.40 V, 1.70 V, vs. NHE]. An active learning framework based on Bayesian Optimization (BO) was then applied to find materials with desirable oxidation potentials more efficiently. Their results showed that the BO were more than 5-fold improvement in computational efficiency compared to the random selection.


image file: d2sc00291d-f6.tif
Fig. 6 Thermodynamic cycle to calculate the equilibrium redox potential in the solution. Reproduced with permission from ref. 93. Copyright 2020 Elsevier Ltd. All rights reserved.

Although a practical application example of ML to discover novel ORASs directly and then synthesize accordingly has been rare up to now, ML still provides a meaningful future vision for accelerating the discovery of new battery materials. Most recently, by combining ML with high-precision theoretical calculations and experiments, Zhang et al. found that indigo trisulfonate [Indigo-3(SO3H)] showed higher solubility, capacity retention, and coulombic efficiency than AQDS and its predecessors.96 Allam et al. applied high-throughput screening methods based on the DFT-ML framework to design novel organic electrode materials of Li-ion batteries.93,98 Their dataset includes various derivatives of functionalized graphene flakes, ketones, quinones, corannulenes, and coronenes, which can be found from their previous work.99–103 DFT was employed to calculate the primary feature (electronic properties) for the input of ML, including electron affinity, highest occupied molecular orbital (HOMO) and lowest unoccupied molecular orbital (LUMO), and the HOMO–LUMO gap. Three pipelines were used for each of the three algorithms, including ANN, gradient boosting regression (GBR), and Kernel ridge regression (KRR), to train the models, which are summarized in Fig. 7. The results showed that the model trained by KRR in Pipeline3 has the best performance (with an MSE of 0.025) for the prediction of the redox potential. The trained KRR model was tested by an unknown dataset, which includes 17 sumanene derivatives. An average error and a Pearson correlation of 3.94% and ∼97%, respectively, were obtained between the DFT- and KRR-predicted redox potentials.


image file: d2sc00291d-f7.tif
Fig. 7 Overall breakdown of the three pipelines for all three learning models. Pipeline 1 represents the base protocol, in which the models were trained directly using the 10 primary features. Pipeline 2 depicts the placement of a Pearson correlation filter, in addition to a relative contribution analysis (RCA) and recursive feature elimination (RFE). Lastly, pipeline 3 depicts the addition of composite features and feature elimination using LASSO. Reproduced with permission from ref. 93. Copyright 2020 Elsevier Ltd.

Recently, ML was successfully employed in the prediction of the redox potentials of phenazine derivatives.96 The workflow is shown in Fig. 8. The dataset includes 185 molecules. 2D, 3D and molecular fingerprints were generated as the input features. Twenty linear or non-linear ML algorithms (LR, Ridge Regression, Lasso, Elastic-Net, LARS Lasso, Orthogonal Matching Pursuit, Bayesian Ridge Regression, Automatic Relevance Determination Regression, Passive Aggressive, Huber Regression, Kernel ridge Regression, SVM, GPR, Decision Trees, Bagging meta-estimator, Random Forest, AdaBoost, Gradient Boosting Regression, Artificial Neural Network, Nearest Neighbors Regression) were employed to build the redox potential prediction model, and high accuracy was obtained for all of the above models for both training and test datasets (i.e., R2 > 0.98, MSE < 0.008 V and MAE < 0.07 V). Their results indicated that the linear model can obtain high performance when the features are properly selected. The trained best-performance ML model was applied to predict the redox potential of previously reported promising ORASs, including tetra-amino-phenazine (TAPZ), hexa-amino-phenazine (HAPZ), and octa-amino-phenazine (OAPZ), and the predicted redox potential was less than 0.07 V (<3%).


image file: d2sc00291d-f8.tif
Fig. 8 The pipeline of applying ML to predict the redox potentials of phenazine derivatives. Reproduced with permission from ref. 96. Copyright 2021 ChemRxiv.

Implementation of ML in FBs

Apart from predicting the properties of ORASs, ML can also be utilized in electrode design,38,104 membrane design105 and system optimization for FBs.106 As the place in which electrochemical reaction occurs, the pore structure and specific surface area of the electrode will affect the efficiencies of the FB. Wan et al. combined a data generation method with ML algorithms to design porous electrodes with large specific surface area and high hydraulic permeability for FBs.104 Stochastic reconstruction method, morphological algorithm and lattice Boltzmann method were adopted to construct the dataset, which contains 2275 fibrous structures (shown in Fig. 9). LR, ANN, and RF algorithms were employed to construct the model for the specific surface area and hydraulic permeability prediction of porous electrodes. ANN achieved the best accuracy with a test error of 1.91% and 11.48% for specific surface area and hydraulic permeability, respectively. More than 700 promising porous electrode materials were screened by combination of a genetic algorithm with ANN. Graphite felt electrodes with an 80% increase for the specific surface area and 50% increase for the hydraulic permeability compared to the commercial one can be found, which indicates that ML has an attractive potential in the design of porous electrode materials for FBs.
image file: d2sc00291d-f9.tif
Fig. 9 Computational methods used in the dataset generation. (a) A simple example illustrating the calculation method of the specific surface area. The area of voxel facets belonging to both the solid phase and pore phase is regarded as the effective area (colored in red). (b) Streamlined plot of the simulated velocity field within a three-dimensional fibrous structure. The insert is the streamlined plot of the slice z = 60 μm. (c) Comparison between the simulated specific surface area and the empirical equation (filament analogue model). (d) Comparison between the simulated hydraulic permeability. (e) Illustration of the two examples stored in our dataset. Each case has the four input variables and the two output variables. Reproduced with permission from ref. 104. Copyright 2021 Elsevier Ltd.

The membrane is one of the key components in VFBs. The performance of VFB is directly affected by membranes. Recently, LR and ANN algorithms were applied to predict the performance of a PBI porous membrane treated by various solvents (Fig. 10). Nine solvent properties and five experimental parameters were used as input.105 The mean absolute percentage error (MAPE) of the above models can be achieved within 1% for both voltage efficiency (VE) and energy efficiency (EE). The reliability of this model was further demonstrated by experiment. This model was further applied to screen for the proper solvent for the solvent treatment of the PBI porous membrane, and alcohols were regarded as the most suitable solvent to regulate the porous structure.


image file: d2sc00291d-f10.tif
Fig. 10 A schematic workflow of applying ML to screen suitable solvents for the solvent treatment of a PBI porous membrane. Reproduced with permission from ref. 105. Copyright 2021 Royal Society of Chemistry.

ML can be applied to optimize and design the microstructure of electrodes combined with the computational fluid dynamics simulations, and to guide the surface functionalization combined with DFT calculations. For membranes, ML can be applied to optimize the fabrication conditions and to understand the polymer (or porous) structure–property (e.g., ion selectivity and ion conductivity) relationship. For electrolytes, ML shows enormous potential in ORASs design and the optimization of electrolyte interactions.

Currently, VFBs are at a commercial demonstration stage. However, the relatively high cost restricts their further commercialization. The performance and cost of a VFB system is highly related to the stack and electrolyte. Thus, it is extremely important for the optimization of VFB stacks and systems in a more efficient way. LR models were thus established to predict the power cost, energy cost, voltage efficiency (VE), energy efficiency (EE), utilization ratio of the electrolyte (UE) of VFB systems based on the operating current density, materials and structure parameters.106 The MAPEs of these models can achieve a precision within 1% for VE and EE, and within 5.2% for UE. The coefficients of the models demonstrated that the future development of materials for the VFB stack should focus on reducing the electrochemical polarization and ohmic polarization at high current densities, and the design of the flow field should monitor the enhancement of mass transfer to decrease the concentration polarization of FB stacks.

Summary and prospects

ML, especially deep learning technology, has already been applied in the development of FBs, from key materials design to system performance–cost relationship optimization. However, in spite of these meaningful advances that have been achieved in the past decade, ML in FBs is still in its infancy. Thus, enormous work should be paid to establish effective ML models for FBs (Fig. 11).
image file: d2sc00291d-f11.tif
Fig. 11 Prospects for the future research of ML for FBs.

First, a primary focus can be put on the construction and sharing of the relative database and algorithms. A database with the sufficient amount of data and reliable data is the first key step in the wide application of ML in FBs. How to share the data more efficiently is another challenge. Currently, there are only some open sources of chemoinformatics databases,71,107 and the validity and amount are limited. In addition, innovative ML algorithms, including clustering, principal component analysis (PCA), autoencoder (AE), generative adversarial network (GAN) and meta-learning models like neural Turing machines108 and imitation learning algorithms,109 are promising solutions for key materials design in FBs. Finally, the closed-loop design of key materials in FBs has not been formed yet. Future research should combine simulation, ML, and experiments to truly realize the effective guidance and application in material design.

Improving the performance of the stack and reducing the cost of the system are critical for further application of FBs. ML has been already applied to connect the stack performance and system cost of VFB.98 The performance-cost models can also be built for other FBs, such as Zn-based and organic FB systems. Moreover, more attention can be paid to the combination of the time series prediction methods, such as the ML algorithm with computational fluid dynamics (CFD) simulations, to establish the concentration gradient, electric field gradient, and pressure gradient models. This can further guide the design of the flow field structure and improve the performance of the stack.

Monitoring system operating parameters (such as voltage, current, state of charge (SOC), and temperature) and the further prediction of these parameters are critical to the stable and safe operation of FB systems. ML can also be applied to predict the above operating parameters based on the huge amount of operating data generated by the FB systems, and further guide the operation of the systems. Moreover, ML can be applied to optimize the overall cost of a FB system by establishing a model between the operation parameters and performance. With the establishment of more commercial FB systems, a large amount of data (e.g., efficiencies, capacity, charge–discharge curve of each cycle) will be generated during the operation of those systems in the future. These data can be used by ML to establish the operational performance models of the FB systems (e.g., battery cycle life45,110), which can be involved in the intelligent control system. Considering the high cost to build a FB system, the abovementioned research requires the cooperation between researchers and enterprises. If enough FB systems data are collected, ML can be applied to predict the cycle life and lower the system cost.

Interpretation of ML can provide inspiration and supplemental information for the understanding of mechanisms and laws for the original design, discovery of energy materials, and guide the optimization of stack and systems. Therefore, how to establish a model with a clear mechanism, which can be understood from a human perspective, will be an important research direction in the future and will mutually promote the novel key materials design of FBs.

Data availability

Data sharing is not applicable to this article as no new data were created or analyzed in this study.

Author contributions

Tianyu Li: investigation, visualization, funding acquisition and writing – original draft. Changkun Zhang: conceptualization, supervision and writing – review & editing. Xianfeng Li: conceptualization, funding acquisition, supervision and writing – review & editing.

Conflicts of interest

There are no conflicts to declare.

Acknowledgements

This work was supported by the National Natural Science Foundation of China (Grant No. U1808209), Strategic Priority Research Program of the CAS (XDA21070100), and China Postdoctoral Science Foundation (2021T140651).

Notes and references

  1. A. Z. Weber, M. M. Mench, J. P. Meyers, P. N. Ross, J. T. Gostick and Q. Liu, J. Appl. Electrochem., 2011, 41, 1137,  DOI:10.1007/s10800-011-0348-2.
  2. W. Lu, X. Li and H. Zhang, Phys. Chem. Chem. Phys., 2018, 20, 23–35,  10.1039/C7CP07456E.
  3. B. Li and J. Liu, Natl. Sci. Rev., 2017, 4, 91–105,  DOI:10.1093/nsr/nww098.
  4. S. Rehman, L. M. Al-Hadhrami and M. M. Alam, Renewable Sustainable Energy Rev., 2015, 44, 586–598,  DOI:10.1016/j.rser.2014.12.040.
  5. M. E. Amiryar and K. R. Pullen, Appl. Sci., 2017, 7, 286,  DOI:10.3390/app7030286.
  6. A. G. Olabi, T. Wilberforce, M. Ramadan, M. A. Abdelkareem and A. H. Alami, J. Energy Storage, 2021, 34, 102000,  DOI:10.1016/j.est.2020.102000.
  7. D. Di Lecce, R. Verrelli and J. Hassoun, Green Chem., 2017, 19, 3442–3467,  10.1039/C7GC01328K.
  8. D. A. J. Rand and P. T. Moseley, in Electrochemical Energy Storage for Renewable Sources and Grid Balancing, ed. P. T. Moseley and J. Garche, Elsevier, Amsterdam, 2015, pp. 201–222,  DOI:10.1016/B978-0-444-62616-5.00013-9.
  9. C. Zhang, L. Zhang, Y. Ding, S. Peng, X. Guo, Y. Zhao, G. He and G. Yu, Energy Storage Mater., 2018, 15, 324–350,  DOI:10.1016/j.ensm.2018.06.008.
  10. W. Liu, W. Lu, H. Zhang and X. Li, Chem.–Eur. J., 2019, 25, 1649–1664,  DOI:10.1002/chem.201802798.
  11. P. Colbertaldo, S. B. Agustin, S. Campanari and J. Brouwer, Int. J. Hydrogen Energy, 2019, 44, 9558–9576,  DOI:10.1016/j.ijhydene.2018.11.062.
  12. H. Thaller Lawrence and S. Ohio, Electrically rechargeable redox flow cell, US Pat., US3996064A, 1976 Search PubMed.
  13. M. Lopez-Atalaya, G. Codina, J. R. Perez, J. L. Vazquez and A. Aldaz, J. Power Sources, 1992, 39, 147–154,  DOI:10.1016/0378-7753(92)80133-V.
  14. G. Kear, A. A. Shah and F. C. Walsh, Int. J. Energy Res., 2012, 36, 1105–1120,  DOI:10.1002/er.1863.
  15. C. Ding, H. Zhang, X. Li, T. Liu and F. Xing, J. Phys. Chem. Lett., 2013, 4, 1281–1294,  DOI:10.1021/acsenergylett.8b01828.
  16. Z. Kang, C. Wu, L. Dong, W. Liu, J. Mou, J. Zhang, Z. Chang, B. Jiang, G. Wang, F. Kang and C. Xu, ACS Sustainable Chem. Eng., 2019, 7, 3364–3371,  DOI:10.1021/acssuschemeng.8b05568.
  17. C. Xie, Y. Duan, W. Xu, H. Zhang and X. Li, Angew. Chem., Int. Ed., 2017, 56, 14953–14957,  DOI:10.1002/anie.201708664.
  18. P. Singh and B. Jonshagen, J. Power Sources, 1991, 35, 405–410,  DOI:10.1016/0378-7753(91)80059-7.
  19. C. Wang, Q. Lai, P. Xu, D. Zheng, X. Li and H. Zhang, Adv. Mater., 2017, 29, 1605815,  DOI:10.1002/adma.201605815.
  20. C. Xie, H. Zhang, W. Xu, W. Wang and X. Li, Angew. Chem., Int. Ed., 2018, 57, 11171–11176,  DOI:10.1002/anie.201803122.
  21. C. Xie, Y. Liu, W. Lu, H. Zhang and X. Li, Energy Environ. Sci., 2019, 12, 1834–1839,  10.1039/C8EE02825G.
  22. C. Xie, T. Li, C. Deng, Y. Song, H. Zhang and X. Li, Energy Environ. Sci., 2020, 13, 135–143,  10.1039/C9EE03702K.
  23. B. Yang, L. Hoober-Burkhardt, F. Wang, G. K. Surya Prakash and S. R. Narayanan, J. Electrochem. Soc., 2014, 161, A1371–A1380,  DOI:10.1149/2.0161807jes.
  24. B. Huskinson, M. P. Marshak, C. Suh, S. Er, M. R. Gerhardt, C. J. Galvin, X. Chen, A. Aspuru-Guzik, R. G. Gordon and M. J. Aziz, Nature, 2014, 505, 195–198,  DOI:10.1038/nature12909.
  25. J. Winsberg, C. Stolze, S. Muench, F. Liedl, M. D. Hager and U. S. Schubert, ACS Energy Lett., 2016, 1, 976–980,  DOI:10.1021/acsenergylett.6b00413.
  26. Y. Liang, Z. Tao and J. Chen, Adv. Energy Mater., 2012, 2, 742–769,  DOI:10.1002/aenm.201100795.
  27. J. Noack, N. Roznyatovskaya, T. Herr and P. Fischer, Angew. Chem., Int. Ed., 2015, 54, 9776–9809,  DOI:10.1002/anie.201410823.
  28. M. Park, J. Ryu, W. Wang and J. Cho, Nat. Rev. Mater., 2017, 2, 16080–16097,  DOI:10.1038/natrevmats.2016.80.
  29. Y. Ding, C. Zhang, L. Zhang, Y. Zhou and G. Yu, Chem. Soc. Rev., 2018, 47, 69–103,  10.1039/C7CS00569E.
  30. D. G. Kwabi, Y. Ji and M. J. Aziz, Chem. Rev., 2020, 120, 6467–6489,  DOI:10.1021/acs.chemrev.9b00599.
  31. L. Cheng, R. S. Assary, X. Qu, A. Jain, S. P. Ong, N. N. Rajput, K. Persson and L. A. Curtiss, J. Phys. Chem. Lett., 2015, 6, 283–291,  DOI:10.1021/jz502319n.
  32. G. G. Chowdhury, Annu. Rev. Inf. Sci. Technol., 2003, 37, 51–89,  DOI:10.1002/aris.1440370103.
  33. K. He, X. Zhang, S. Ren and J. Sun, Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2016, 770–778,  DOI:10.1109/CVPR.1997.609286.
  34. J. Levinson, J. Askeland, J. Becker, J. Dolson, D. Held, S. Kammel, J. Z. Kolter, D. Langer, O. Pink, V. Pratt, M. Sokolsky, G. Stanek, D. Stavens, A. Teichman, M. Werling and S. Thrun, 2011 IEEE Intelligent Vehicles Symposium (IV), 2011, pp. 163-168,  DOI:10.1109/IVS.2011.5940562.
  35. P. Raccuglia, K. C. Elbert, P. D. F. Adler, C. Falk, M. B. Wenny, A. Mollo, M. Zeller, S. A. Friedler, J. Schrier and A. J. Norquist, Nature, 2016, 533, 73–76,  DOI:10.1038/nature17439.
  36. B. Sanchez-Lengeling and A. Aspuru-Guzik, Science, 2018, 361, 360–365,  DOI:10.1126/science.aat2663.
  37. Y. Liu, B. Guo, X. Zou, Y. Li and S. Shi, Energy Storage Mater., 2020, 31, 434–450,  DOI:10.1016/j.ensm.2020.06.033.
  38. J. Bao, V. Murugesan, C. J. Kamp, Y. Shao, L. Yan and W. Wang, Adv. Theory Simul., 2020, 3, 1900167,  DOI:10.1002/adts.201900167.
  39. M. Rupp, A. Tkatchenko, K.-R. Müller and O. A. von Lilienfeld, Phys. Rev. Lett., 2012, 108, 058301,  DOI:10.1103/PhysRevLett.108.058301.
  40. K. Hansen, G. Montavon, F. Biegler, S. Fazli, M. Rupp, M. Scheffler, O. A. von Lilienfeld, A. Tkatchenko and K.-R. Müller, J. Chem. Theory Comput., 2013, 9, 3404–3419,  DOI:10.1021/ct400195d.
  41. J. M. Stokes, K. Yang, K. Swanson, W. Jin, A. Cubillos-Ruiz, N. M. Donghia, C. R. MacNair, S. French, L. A. Carfrae, Z. Bloom-Ackermann, V. M. Tran, A. Chiappino-Pepe, A. H. Badran, I. W. Andrews, E. J. Chory, G. M. Church, E. D. Brown, T. S. Jaakkola, R. Barzilay and J. J. Collins, Cell, 2020, 180, 688–702,  DOI:10.1016/j.cell.2020.01.021.
  42. R. Gómez-Bombarelli, J. N. Wei, D. Duvenaud, J. M. Hernández-Lobato, B. Sánchez-Lengeling, D. Sheberla, J. Aguilera-Iparraguirre, T. D. Hirzel, R. P. Adams and A. Aspuru-Guzik, ACS Cent. Sci., 2018, 4, 268–276,  DOI:10.1021/acscentsci.7b00572.
  43. M. H. S. Segler, M. Preuss and M. P. Waller, Nature, 2018, 555, 604–610,  DOI:10.1038/nature25978.
  44. J. M. Granda, L. Donina, V. Dragone, D.-L. Long and L. Cronin, Nature, 2018, 559, 377–381,  DOI:10.1038/s41586-018-0307-8.
  45. K. A. Severson, P. M. Attia, N. Jin, N. Perkins, B. Jiang, Z. Yang, M. H. Chen, M. Aykol, P. K. Herring, D. Fraggedakis, M. Z. Bazant, S. J. Harris, W. C. Chueh and R. D. Braatz, Nat. Energy, 2019, 4, 383–391,  DOI:10.1038/s41560-019-0356-8.
  46. L. C. Blum and J.-L. Reymond, J. Am. Chem. Soc., 2009, 131, 8732–8733,  DOI:10.1021/ja902302h.
  47. L. Ruddigkeit, R. van Deursen, L. C. Blum and J.-L. Reymond, J. Chem. Inf. Model., 2012, 52, 2864–2875,  DOI:10.1021/ci300415d.
  48. T. Sterling and J. J. Irwin, J. Chem. Inf. Model., 2015, 55, 2324–2337,  DOI:10.1021/acs.jcim.5b00559.
  49. V. Stevanović, S. Lany, X. Zhang and A. Zunger, Phys. Rev. B: Condens. Matter Mater. Phys., 2012, 85, 115104,  DOI:10.1103/PhysRevB.85.115104.
  50. J. E. Saal, S. Kirklin, M. Aykol, B. Meredig and C. Wolverton, JOM, 2013, 65, 1501–1509,  DOI:10.1007/s11837-013-0755-4.
  51. S. P. Ong, S. Cholia, A. Jain, M. Brafman, D. Gunter, G. Ceder and K. A. Persson, Comput. Mater. Sci., 2015, 97, 209–215,  DOI:10.1016/j.commatsci.2014.10.037.
  52. A. Jain, S. P. Ong, G. Hautier, W. Chen, W. D. Richards, S. Dacek, S. Cholia, D. Gunter, D. Skinner, G. Ceder and K. A. Persson, APL Mater., 2013, 1, 011002,  DOI:10.1063/1.4812323.
  53. S. P. Ong, W. D. Richards, A. Jain, G. Hautier, M. Kocher, S. Cholia, D. Gunter, V. L. Chevrier, K. A. Persson and G. Ceder, Comput. Mater. Sci., 2013, 68, 314–319,  DOI:10.1016/j.commatsci.2012.10.028.
  54. J. L. Durant, B. A. Leland, D. R. Henry and J. G. Nourse, J. Chem. Inf. Comput. Sci., 2002, 42, 1273–1280,  DOI:10.1021/ci010132r.
  55. A. Cereto-Massagué, M. J. Ojeda, C. Valls, M. Mulero, S. Garcia-Vallvé and G. Pujadas, Methods, 2015, 71, 58–63,  DOI:10.1016/j.ymeth.2014.08.005.
  56. A. Bender, H. Y. Mussa, R. C. Glen and S. Reiling, J. Chem. Inf. Comput. Sci., 2004, 44, 170–178,  DOI:10.1021/ci034207y.
  57. A. Bender, H. Y. Mussa, R. C. Glen and S. Reiling, J. Chem. Inf. Comput. Sci., 2004, 44, 1708–1718,  DOI:10.1021/ci0498719.
  58. H. L. Morgan, J. Chem. Doc., 1965, 5, 107–113,  DOI:10.1021/c160017a018.
  59. D. Weininger, J. Chem. Inf. Comput. Sci., 1988, 28, 31–36,  DOI:10.1021/ci00057a005.
  60. S. Heller, A. McNaught, S. Stein, D. Tchekhovskoi and I. Pletnev, J. Cheminf., 2013, 5, 7,  DOI:10.1186/1758-2946-5-7.
  61. N. Jeliazkova and N. Kochev, Mol. Inf., 2011, 30, 707–720,  DOI:10.1186/s13321-017-0203-5.
  62. F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss and V. Dubourg, J. Mach. Learn. Res., 2011, 12, 2825–2830 Search PubMed.
  63. M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis, J. Dean, M. Devin, S. Ghemawat, G. Irving and M. Isard, 12th USENIX symposium on operating systems design and implementation, 2016, pp. 265–283 Search PubMed.
  64. A. Paszke, S. Gross, S. Chintala, G. Chanan, E. Yang, Z. DeVito, Z. Lin, A. Desmaison, L. Antiga and A. Lerer, Automatic differentiation in PyTorch, 2017 Search PubMed.
  65. J. Bergstra, O. Breuleux, F. Bastien, P. Lamblin, R. Pascanu, G. Desjardins, J. Turian, D. Warde-Farley and Y. Bengio, Proceedings of the Python for scientific computing conference (SciPy), 2010, pp. 1–7 Search PubMed.
  66. R. Tibshirani, J. Roy. Stat. Soc. B Stat. Methodol., 1996, 58, 267–288,  DOI:10.1111/j.2517-6161.1996.tb02080.x.
  67. A. E. Hoerl and R. W. Kennard, Technometrics, 1970, 12, 55–67,  DOI:10.2307/1271436.
  68. S. Er, C. Suh, M. P. Marshak and A. Aspuru-Guzik, Chem. Sci., 2015, 6, 885–893,  10.1039/C4SC03030C.
  69. D. P. Tabor, R. Gómez-Bombarelli, L. Tong, R. G. Gordon, M. J. Aziz and A. Aspuru-Guzik, J. Mater. Chem. A, 2019, 7, 12833–12841,  10.1039/C9TA03219C.
  70. A. Jinich, B. Sanchez-Lengeling, H. Ren, R. Harman and A. Aspuru-Guzik, ACS Cent. Sci., 2019, 5, 1199–1210,  DOI:10.1021/acscentsci.9b00297.
  71. E. Sorkun, Q. Zhang, A. Khetan and S. Er, ChemRxiv, Cambridge Open Engage, Cambridge, 2021,  DOI:10.26434/chemrxiv.14398067.v1.
  72. K. Wedege, E. Dražević, D. Konya and A. Bentien, Sci. Rep., 2016, 6, 39101,  DOI:10.1038/srep39101.
  73. G. Klopman, S. Wang and D. M. Balthasar, J. Chem. Inf. Comput. Sci., 1992, 32, 474–482,  DOI:10.1021/ci00009a013.
  74. T. Suzuki, J. Comput.-Aided Mol. Des., 1991, 5, 149–166,  DOI:10.1007/BF00129753.
  75. P. Lind and T. Maltseva, J. Chem. Inf. Comput. Sci., 2003, 43, 1855–1859,  DOI:10.1021/ci034107s.
  76. J. Wang, G. Krudy, T. Hou, W. Zhang, G. Holland and X. Xu, J. Chem. Inf. Model., 2007, 47, 1395–1404,  DOI:10.1021/ci700096r.
  77. T. Cheng, Q. Li, Y. Wang and S. H. Bryant, J. Chem. Inf. Model., 2011, 51, 229–236,  DOI:10.1021/ci100364a.
  78. H.-s. Kim, K.-J. Lee, Y.-K. Han, J. H. Ryu and S. M. Oh, J. Power Sources, 2017, 348, 264–269,  DOI:10.1016/j.jpowsour.2017.03.019.
  79. L. E. VanGelder, B. E. Petel, O. Nachtigall, G. Martinez, W. W. Brennessel and E. M. Matson, ChemSusChem, 2018, 11, 4139–4149,  DOI:10.1002/cssc.201802029.
  80. S. Boobier, D. R. J. Hose, A. J. Blacker and B. N. Nguyen, Nat. Commun., 2020, 11, 5753,  DOI:10.1038/s41467-020-19594-z.
  81. G. Panapitiya, M. Girard, A. Hollas, V. Murugesan, W. Wang and E. Saldanha, arXiv preprint arXiv:2105, 12638, 2021,  DOI:10.48550/arXiv.2105.12638.
  82. P. G. Francoeur and D. R. Koes, J. Chem. Inf. Model., 2021, 61, 2530–2536,  DOI:10.1021/acs.jcim.1c00331.
  83. G. Klopman and H. Zhu, J. Chem. Inf. Comput. Sci., 2001, 41, 439–445,  DOI:10.1021/ci000152d.
  84. D. Zhou, Y. Alelyunas and R. Liu, J. Chem. Inf. Model., 2008, 48, 981–987,  DOI:10.1021/ci800024c.
  85. T. S. Schroeter, A. Schwaighofer, S. Mika, A. Ter Laak, D. Suelzle, U. Ganzer, N. Heinrich and K.-R. Müller, J. Comput.-Aided Mol. Des., 2007, 21, 485–498,  DOI:10.1007/s10822-007-9125-z.
  86. D. S. Palmer, N. M. O’Boyle, R. C. Glen and J. B. O. Mitchell, J. Chem. Inf. Model., 2007, 47, 150–158,  DOI:10.1021/ci060164k.
  87. J. Huuskonen, M. Salo and J. Taskinen, J. Chem. Inf. Comput. Sci., 1998, 38, 450–456,  DOI:10.1021/ci970100x.
  88. M. C. Sorkun, A. Khetan and S. Er, Sci. Data, 2019, 6, 143,  DOI:10.1038/s41597-019-0151-1.
  89. S. Kim, A. Jinich and A. Aspuru-Guzik, J. Chem. Inf. Model., 2017, 57, 657–668,  DOI:10.1021/acs.jcim.6b00332.
  90. A. J. Hopfinger, E. X. Esposito, A. Llinàs, R. C. Glen and J. M. Goodman, J. Chem. Inf. Model., 2009, 49, 1–5,  DOI:10.1021/ci800436c.
  91. R. S. Assary, F. R. Brushett and L. A. Curtiss, RSC Adv., 2014, 4, 57442–57451,  10.1039/C4RA08563A.
  92. J. F. Kucharyson, L. Cheng, S. O. Tung, L. A. Curtiss and L. T. Thompson, J. Mater. Chem. A, 2017, 5, 13700–13709,  10.1039/C7TA01285C.
  93. O. Allam, R. Kuramshin, Z. Stoichev, B. W. Cho, S. W. Lee and S. S. Jang, Mater. Today Energy, 2020, 17, 100482,  DOI:10.1016/j.mtener.2020.100482.
  94. J. C. Ortiz-Rodríguez, J. A. Santana and D. D. Méndez-Hernández, J. Mol. Model., 2020, 26, 70,  DOI:10.1007/s00894-020-4331-x.
  95. S. Ghule, S. R. Dash, S. Bagchi, K. Joshi and K. Vanka, ACS Omega, 2022, 7(14), 11742–11755,  DOI:10.1021/acsomega.1c06856.
  96. Q. Zhang, A. Khetan, E. Sorkun, F. Niu, A. Loss, I. Pucher and S. Er, Energy Storage Mater., 2022, 47, 167–177,  DOI:10.1016/j.ensm.2022.02.013.
  97. H. A. Doan, G. Agarwal, H. Qian, M. J. Counihan, J. Rodríguez-López, J. S. Moore and R. S. Assary, Chem. Mater., 2020, 32, 6338–6346,  DOI:10.1021/acs.chemmater.0c00768.
  98. O. Allam, B. W. Cho, K. C. Kim and S. S. Jang, RSC Adv., 2018, 8, 39414–39420,  10.1039/C8RA07112H.
  99. Y. Zhu, K. C. Kim and S. S. Jang, J. Mater. Chem. A, 2018, 6, 10111–10120,  10.1039/C8TA01671B.
  100. J. Kang, K. C. Kim and S. S. Jang, J. Phys. Chem. C, 2018, 122, 10675–10681,  DOI:10.1021/acs.jpcc.8b00827.
  101. P. Sood, K. C. Kim and S. S. Jang, J. Energy Chem., 2018, 27, 528–534,  DOI:10.1016/j.jechem.2017.11.009.
  102. K. C. Kim, T. Liu, S. W. Lee and S. S. Jang, J. Am. Chem. Soc., 2016, 138, 2374–2382,  DOI:10.1021/jacs.5b13279.
  103. T. Liu, K. C. Kim, B. Lee, Z. Chen, S. Noda, S. S. Jang and S. W. Lee, Energy Environ. Sci., 2017, 10, 205–215,  10.1039/C6EE02641A.
  104. S. Wan, X. Liang, H. Jiang, J. Sun, N. Djilali and T. Zhao, Appl. Energy, 2021, 298, 117177,  DOI:10.1016/j.apenergy.2021.117177.
  105. T. Li, W. Lu, Z. Yuan, H. Zhang and X. Li, J. Mater. Chem. A, 2021, 9, 14545–14552,  10.1039/D1TA02421C.
  106. T. Li, F. Xing, T. Liu, J. Sun, D. Shi, H. Zhang and X. Li, Energy Environ. Sci., 2020, 13, 4353–4361,  10.1039/d0ee02543g.
  107. J. Sun, N. Jeliazkova, V. Chupakhin, J.-F. Golib-Dzib, O. Engkvist, L. Carlsson, J. Wegner, H. Ceulemans, I. Georgiev, V. Jeliazkov, N. Kochev, T. J. Ashby and H. Chen, J. Cheminf., 2017, 9, 17,  DOI:10.1186/s13321-017-0203-5.
  108. A. Zylberberg, S. Dehaene, P. R. Roelfsema and M. Sigman, Trends Cognit. Sci., 2011, 15, 293–300,  DOI:10.1016/j.tics.2011.05.007.
  109. Y. Duan, M. Andrychowicz, B. C. Stadie, J. Ho, J. Schneider, I. Sutskever, P. Abbeel and W. Zaremba, in Advances in Neural Information Processing Systems, ed. I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan and R. Garnett, Curran Associates, Inc., 2017, vol. 30,  DOI:10.48550/arXiv.1703.07326.
  110. P. M. Attia, A. Grover, N. Jin, K. A. Severson, T. M. Markov, Y.-H. Liao, M. H. Chen, B. Cheong, N. Perkins, Z. Yang, P. K. Herring, M. Aykol, S. J. Harris, R. D. Braatz, S. Ermon and W. C. Chueh, Nature, 2020, 578, 397–402,  DOI:10.1038/s41586-020-1994-5.

This journal is © The Royal Society of Chemistry 2022
Click here to see how this site uses Cookies. View our privacy policy here.