Juhwan
Noh
,
Geun Ho
Gu
,
Sungwon
Kim
and
Yousung
Jung
*
Department of Chemical and Biomolecular Engineering, Korea Advanced Institute of Science and Technology (KAIST), 291, Daehak-ro, Yuseong-gu, Daejeon 34141, Republic of Korea. E-mail: ysjn@kaist.ac.kr
First published on 15th April 2020
Developing high-performance advanced materials requires a deeper insight and search into the chemical space. Until recently, exploration of materials space using chemical intuitions built upon existing materials has been the general strategy, but this direct design approach is often time and resource consuming and poses a significant bottleneck to solve the materials challenges of future sustainability in a timely manner. To accelerate this conventional design process, inverse design, which outputs materials with pre-defined target properties, has emerged as a significant materials informatics platform in recent years by leveraging hidden knowledge obtained from materials data. Here, we summarize the latest progress in machine-enabled inverse materials design categorized into three strategies: high-throughput virtual screening, global optimization, and generative models. We analyze challenges for each approach and discuss gaps to be bridged for further accelerated and rational data-driven materials design.
There are two mapping directions (i.e. forward and inverse) in materials informatics. In a forward mapping, one essentially aims to predict the properties of materials using materials structures as input, encoded in various ways such as simple attributes of constituent atoms, compositions, structures in graph forms, etc. In an inverse mapping, by contrast, one defines the desired properties first and attempts to find materials with such properties in an inverse manner using mathematical algorithms and automations. While forward mapping mainly deals with property prediction given structures, inverse mapping focuses on the “design” aspect of materials informatics towards target properties. For effective inverse design, therefore, one needs (1) efficient methods to explore the vast chemical space towards the target region (“exploration”), and (2) fast and accurate methods to predict the properties of a candidate material along with chemical space exploration (“evaluation”).
The purpose of this mini-review is to survey exciting new developments of methods to perform inverse design by “exploring” the chemical space effectively towards the target region. We will particularly highlight the design of inorganic solid-state materials since there are excellent recent review articles in the literature for the molecular version of inverse design.3,4 To structure this review, we categorize the inverse design strategies of inorganic crystals as summarized in Fig. 1, namely, high-throughput virtual screening (HTVS), global optimization (GO), and generative ML models (GM), largely borrowing the classification of Sanchez-Lengeling and Aspuru-Guzik3 and Butler et al.5 Among them, HTVS may be regarded as an extended version of the direct approach since it goes through the library and evaluates its function one by one, but the data-driven nature of the automated, extensive, and accelerated search in the functional space makes it potentially included in the inverse design strategy.6
One of the drawbacks of HTVS, however, is that, the search is limited by the user-selected library (either the experimental database or substituted computational database) and experts' (sometimes biased) intuitions are still involved in selecting the database, and thus potentially high-performing materials that are not in the library can be missed out. Also, since the screening is run over the database blindly without any preferred directions to search, the efficiency can be low in HTVS. One way to expedite the brute-force search toward the optimal material is to perform global optimization (GO) in the chemical space. In evolutionary algorithms (EAs), one form of GO, for example, mutations and crossover allow effective visits of various local minima by leveraging the previous histories of configurational visits, and therefore can generally be more efficient and also go beyond the chemical space defined by known materials and their structural motifs unlike HTVS.7
The data-driven GM is another promising inverse design strategy.3 The GM is a probabilistic ML model that can generate new data from the continuous vector space learned from the prior knowledge on dataset distribution.3,21 The key advantage of GMs is their ability to generate unseen materials with target properties in the gap between the existing materials by learning their distribution in the continuous space. While both the EA and GM can generate completely new materials not in the existing database, they differ by the way each approach utilizes data. The EA learns the geometric landscape of the functionality manifold (energy and properties) implicitly as the iteration evolves, while the GM learns the distribution of the whole target functional space during training in an implicit (i.e. adversarial learning) or explicit (i.e. variational inference) manner.
Below we summarize the current status and successful examples of these three main strategies (HTVS, GO, and GM) of the data-driven inorganic inverse design approach. We also discuss several challenges for the practical application of accelerated inverse materials design and also offer some promising future directions.
Using the computational HTVS going through the existing database, Reed and co-workers9 discovered 21 new Li-solid electrolyte materials by screening 12831 Li-containing materials in the materials project (MP).10 Singh and co-workers11 newly identified 43 photocatalysts for CO2 conversion through the theory/experiment combined screening framework for 68860 materials available in MP. However, as discussed above, moving beyond the known materials is critical, and to address it, a new functional photoanode material has been discovered by enumerating hypothetical materials by substituting elements to the existing crystals.12 Recently, data mining-13 and deep learning-based14 algorithms for elemental substitution are proposed to effectively search through the existing crystal templates, and Sun et al.15 discovered a large number of metal nitrides using the data-mined elemental substitution algorithm which accelerated the experimental discovery of nitrides by a factor of 2 compared to the average rate of discovery listed on the inorganic crystal structure database, ICSD.16,17
Despite those successful results, the large computational cost for property evaluation using DFT calculations is still a main bottleneck in the computational HTVS, and to overcome the latter challenge, ML-aided property prediction has begun to be implemented (see Table 1 and ref. 18 and 19 for an extensive review on ML used in property predictions). Herein, we mainly focus on ML models predicting the stability of crystal structures since the stability represented by the formation energy is a widely used quantity, though crude, to approximate synthesizability in many materials designs.
Representation | Invertibility | Invariance | Model | Application |
Supervised learning (property prediction in HTVS) | ||||
Atomic properties56,57 | No | Yes | SVR | Predicting melting temperature, bulk and shear modulus, bandgap |
Crystal site-based representation20 | Yes | Yes | KRR | Predicting formation energy of ABC2D6 elpasolite structures |
Average atomic properties22 | No | Yes | Ensembles of decision trees | Predicting the formation energy of inorganic crystal structures |
Voronoi-tessellation-based representation58 | No | Yes | Random forest | Predicting the formation energy of quaternary Heusler compounds |
Crystal graph24 | No | Yes | GCNN | Predicting formation enthalpy of inorganic compounds |
Unsupervised learning (GM) | ||||
3D atomic density59,79 | Yes | No | VAE | Generation of inorganic crystals |
3D atomic density and energy grid shape60 | Yes | No | GAN | Generation of porous materials |
Lattice site descriptor61 | Yes | No | GAN | Generation of graphene/BN-mixed lattice structures |
Unit cell vectors and coordinates36,62 | Yes | No | GAN | Generation of inorganic crystals |
Non-structural descriptor-based ML models have been proposed.20,22 For example, Meredig et al.22 proposed a formation energy prediction model for ∼15000 materials existing in the ICSD16,17 using both data-driven heuristics utilizing the composition-weighted average of corresponding binary compound formation energies (MAE = 0.12 eV per atom) and ensembles of decision trees which take average atomic properties of constituent elements as input (MAE = 0.16 eV per atom). The proposed models were used to explore ∼1.6 million ternary compounds, and 4500 new stable materials were identified with the energy above convex hull ≤100 meV per atom. With the latter examples considering compositional information only, Seko et al.23 have shown that the inclusion of structural information such as radial distribution function could further improve the prediction accuracy significantly from RMSE = 0.249 to 0.045 eV per atom for a cohesive energy of 18000 inorganic compounds with kernel ridge regression.
ML models that encode the structural information of crystals for the prediction of energies and properties have also been proposed. Notably, Xie et al.24 proposed the symmetry invariant crystal graph convolutional neural network (CGCNN) to encode periodic crystal structures which showed very encouraging predictions for various properties including formation energies (MAE = 0.039 eV per atom) and band gaps (MAE = 0.388 eV). An improved version of CGCNN was also proposed by incorporating explicit 3-body correlations of neighboring atoms and applied to identify stable compounds out of 132600 structures obtained by tertiary elemental substitution of ThCr2Si2-structure prototype.25 Lately, the graph-based universal ML model that can treat both molecules and periodic crystals was proposed and demonstrated highly competitive accuracy across a wide range of 15–20 molecular and materials properties.24–27
While looking promising, one of the more practical challenges of ML-aided HTVS for crystals is that some property data is often limited in size to expect good predictive accuracy for model training across different chemistries.26,28 (A more general comparison between the inorganic crystal dataset and organic molecule database is discussed in more detail in a later section. See also Fig. 4) To address this small dataset size, algorithms such as transfer learning (i.e. using pre-trained parameters before training the model on the small-size of the database)29 and active learning (i.e. effectively sampling the training set from the whole database)30,31 could help. For example, one may build the ML model to predict computationally more difficult properties (e.g. band gap and bulk modulus) using model parameters trained on a relatively simple property (e.g. formation energy),26 and this would help prevent overfitting driven by using a smaller dataset for difficult properties.
Furthermore, it is important to note that most current ML models to predict energies for crystals can only evaluate energies on relaxed structures, but cannot (or have not been shown to) calculate forces. Thus, when elemental substitution (which requires geometry relaxation) is used to expand the search space, one cannot use aforementioned ML models and still must perform costly DFT structure relaxations for every substituted structure as shown in Fig. 2a. To address this, data-driven interatomic potential models32–34 that can compute forces and construct a continuous potential energy surface are particularly promising, although they have not been widely used for HTVS of crystals yet since potentials are often developed for particular systems and so not applicable for the screening of widely varying systems. Or, still using the energy-only ML model but quantifying uncertainty caused by using unrelaxed structures could be an alternative way to increase the practical efficiency of HTVS.35 In addition, since the substitution-based enumeration limits the structural diversity of the dataset, generative models which will be discussed in detail below can effectively expand the diversity by sampling the hidden portion of the chemical space36 as shown in Fig. 2b.
Amsler and Goedecker40 proposed the minima hopping method to discover new crystal structures by adapting the softening process which modifies initial molecular dynamic velocities to improve the search efficiency. The latter minima hopping approach was extended to design transition metal alloy-based magnetic materials (FeCr, FeMn, FeCo and FeNi) by combining with additional steps evaluating magnetic properties (i.e. magnetization and magnetic anisotropy energy).41 FeCr and FeMn were predicted as soft-magnetic materials while FeCo and FeNi were predicted as hard-magnetic materials.
Evolutionary algorithms use strategies inspired by biological evolution, such as reproduction, mutation, recombination, and selection, and they can be used to find new crystal structures with optimized properties. The properties to optimize can be stability only (called convex hull optimization) or both stability and desired chemical properties (called Pareto or multi-objective optimization, see ref. 37 for more technical details). Two popular approaches include the Oganov-Glass evolutionary algorithm42 and Wang's version of particle swarm optimization.43 While the detailed updating process of each algorithm is different,37 the two key steps are commonly shared: (1) generating a population consisting of randomly initialized atomic configurations and (2) updating the population after evaluating stability (or/and property) of each configuration existing in the population, using DFT calculations or ML-based methods for an accelerated search. One of the major advantages of EA-based models is their capability to generate completely new materials beyond existing databases and chemical intuitions.
For convex hull optimization, Kruglov et al.44 proposed new stable uranium polyhydrides (UxHy) as potential high-temperature superconductors. Zhu et al.45 systematically investigated the (V,Nb)-(Fe,Ru,Os)-(As,Sb,Bi) family of half-Heusler compounds where 6 compounds were identified as stable and entirely new structures, and 5 of them were experimentally verified as stable with a half-Heusler crystal structure. Multi-objective optimization led to the inverse discovery of new crystal structures with various properties in addition to stability. Zhang et al.46 proposed 24 promising electrides with an optimal degree of interstitial electron localization where 18 candidates were experimentally synthesized that have not been proposed as electrides previously. Xiang et al.47 discovered a cubic Si20 phase, a potential candidate for thin-film solar cells, with a quasi-direct band gap of 1.55 eV. Bedghiou et al.48 discovered new structures of rutile-TiO2 with the lowest direct band gap of 0.26 eV under (ultra)high pressure conditions (i.e. up to 300 GPa) by simultaneously optimizing the stability and band gap during the evolutionary algorithm.
As in HTVS, the large computational cost for property evaluations is a major bottleneck (99% of the entire cost49) in EA (or GO in general) and ML models can greatly help. Of course, the same property prediction ML models or interatomic potentials described in HTVS can also be used in EAs, as shown in Fig. 3a. In specific examples, Jennings et al.50 proposed an ML-based genetic algorithm framework by adapting on-the-fly a Gaussian process regression model to rapidly predict target properties (energy in this case). Here, for PtxAu147−x alloy nanoparticles, the genetic algorithm was shown to reduce the number of configurational visits (or DFT energy calculations) from 1044 (brute force combinatorial possibilities) to 16000, and with the Gaussian process model described above, the required DFT calculations were further reduced to 300, representing 50-fold reduction in cost due to ML. Avery et al.51 constructed a bulk modulus prediction ML model, trained with the database existing in the Automatic FLOW (AFLOW)52 library, and used it to predict new 43 superhard carbon-phases in their EA-based materials design. Podryabinkin et al.49 used a moment tensor potential-based ML interatomic potential to replace expensive DFT structure relaxations in their crystal structure prediction of carbon and boron allotropes using EAs. The authors were able to find all the main allotropes as well as to find a hitherto unknown 54-atom structure of boron with substantially low cost.
Since most EA-based methods need a fixed chemical composition as input, one often needs to try many different compositions or requires experts' guess for the initial composition. To address this computational difficulty of searching through a large composition space, the recently proposed ML-based53 and tensor decomposition-based54 chemical composition recommendation models are noteworthy since those models could provide promising unknown chemical compositions from prior knowledge of experimentally reported chemical composition. Furthermore, Halder et al.55 combined the classification ML model with EAs, in which the classification model selected potentially promising compositions that would go into the EA-based crystal structure prediction as shown in Fig. 3b. The authors applied the method to find new magnetic double perovskites (DPs). They first used the random forest to select elemental compositions in A2BB′O6 (A = Ca/Sr/Ba, B/B′ = transition metals) as potentially stable DPs (finding 33 compounds out of 412 unexplored compositions), and using EA and DFT calculations they subsequently identified new 21 DPs with various magnetic and electronic properties.
We first note that for organic molecules there are several string-based molecular representations that are symmetry-invariant and invertible as in SMILES63 and SELFIES,64 for which many language-based ML models such as RNN,65 Seq2Seq,66 and attention-based Transformer model67 can be applied.21,98 Furthermore, graph representation is another popular approach for organic molecules since chemical bonds between atoms in molecules can be explicitly defined and this can allow an inverse mapping from graph to molecular structure. Various implicit and explicit GMs68–70 have been proposed by adopting a graph convolutional network for organic molecules.71 However, in the case of crystal structures, currently there is no explicit rule to convert crystal structures into string-based representations, or vice versa. Although graph representation has been proposed with great success on property predictions, there is currently no explicit formulation on decoding the crystal graph back to the 3D crystal structure.
A low structural diversity (or data) per element for the inorganic crystal structure database is another critical bottleneck in establishing GMs (or in fact any ML models) for inorganic solids compared to organic molecules (see Fig. 4). This is because, for organic molecules, only a small number of main group elements can produce an enormous degree of chemical and structural diversity, but for inorganic crystals, the degree of structural diversity per chemical element is relatively low and not well balanced compared to molecules (for example, there are 2506 materials having ICSD-ids in MP that contain iron, but only 760 materials that contain scandium). This low structural diversity could bias the model during the training, and it may not able to generate so meaningful and very different new structures from existing materials. This makes a universal GM for inorganic crystals that covers the entire periodic table quite challenging.
Fig. 4 Distribution of elements existing in the crystal/molecular database. (a) Experimentally reported inorganic materials (# of data = 48567) taken from MP.10 They cover most elements in the periodic table (high elemental diversity), but the number of data per element is sparsely populated (low structural diversity). (b) Organic molecules taken from the subset of the ZINC database (# of data = 2077407).72 They cover very limited elements (low elemental diversity), but are densely populated (high structural diversity). |
Despite these challenges, there are some promising initial results for inorganic crystal generative models that addressed some of the aforementioned difficulties. Two concepts of GMs have been implemented for solid state materials recently (see Fig. 5a and b): variational autoencoder (VAE)73 and generative adversarial network (GAN).74 Here, we note that other generative frameworks (e.g. conditional VAE75/GAN,76 AAE,77 VAE-GAN,78etc.) derived from the latter two models can be applied depending on target objectives. VAE explicitly regularizes the latent space using known prior distributions such as Gaussian and Bernoulli distribution. Compared to VAE, the GAN implicitly learns the data distribution by iteratively checking the reality of the generated data from the known prior latent space distribution.
Fig. 5 (a) Variational autoencoder (VAE) learns materials chemical space under the density reconstruction scheme by explicitly constructing the latent space. Each point in the latent space represents a single material, and thus one can directly generate new materials with optimal functionality. (b) Generative adversarial network (GAN), however, learns materials chemical space under the implicit density prediction scheme which iteratively discriminates the reality of the data generated from the latent space. (c) A VAE-based crystal generative framework proposed by Noh et al.59 using an invertible 3D image representation for the unit cell and basis (adapted with permission from ref. 59 Copyright 2019 Elsevier Inc. Matter). |
Noh et al.59 proposed the first GM for inorganic solid-state materials structures using a 3D atomic image representation (Fig. 5c). Here, the stability-embedded latent space was constructed under the VAE scheme,73 and used to generate stable vanadium oxide crystal structures. In particular, due to a low structural diversity of the current inorganic dataset described above, the authors used the virtual V–O binary compound space as a restricted materials space to explore (instead of learning the crystal chemistry across the periodic table). This image-based GM then discovered several new compositions and meta-stable polymorphs of vanadium oxides that have been completely unknown. Hoffmann et al.79 proposed a general purpose encoding-decoding framework for 3D atomic density under the VAE formalism. The model was trained with atomic configurations taken from crystal structures reported in the ICSD16,17 (which does not impose a constraint in chemical composition), and an additional segmentation network80 was used to classify the elements information from the generated 3D representation. However, we note that the proposed model is focused on generating valid atomic configurations only, and thus an additional network which generates a unit cell associated with the generated atomic configuration would be required to generate new ‘materials’.
Despite those promising results, 3D image-based representations have a few limitations, a lack of invariance under symmetry operations and heuristic post-processing to clean up chemical bonds, for example. The former drawback can be approximately addressed by data augmentation,60 and for example, Kajita et al.81 showed that 3D representations with data augmentation yielded a reasonable prediction of the various electronic properties of 680 oxide materials. For the latter problem, a representation which does not require heuristic post-processing would be desirable.
Rather than using computationally burdensome 3D representations, Nouira et al.62 proposed to use unit cell vectors and fractional coordinates as input to generate new ternary hydride structures by learning the structures of binary hydrides inspired by a cross-domain learning strategy. Kim et al.36 proposed a GAN-based generative framework which uses a similar coordinate-based representation with symmetry invariance addressed with data augmentation and permutation invariance with symmetry operation as described in PointNet,82 and used it to generate new ternary Mg–Mn–O compounds suitable for photoanode applications. There are also examples in which generative frameworks are used to sample new chemical compositions for inorganic solid materials.83,84 For these studies, adding concrete structural information would be a desirable further development, similar to the work of Halder et al.,55 which also highlights the importance of invertible representations in GMs to predict crystal structures.
We note that, while GMs themselves offer essential architectures needed to inversely design materials with target properties by navigating the functional latent space, many of the present examples shown above currently deal with generating new stable structures, and one still needs to incorporate properties into the model for a practical inverse design beyond stability embedding. One can use conditional GMs in which the target function is used as a condition,68,85 or perform the optimization task on a continuous latent space as described in Gómez-Bombarelli et al.21 for organic molecules. For example, Dong et al.61 used a generative model to design graphene/boron nitride mixed lattice structures with the appropriate bandgap by adding a regression network within the GAN in combination with the simple lattice site representation. A similar crystal site-based representation20 which satisfies both invertibility and invariance (Table 1) can be used to generate new elemental combinations for the fixed structure template. Also Kim et al.60 used an image-based GAN model to inversely design zeolites with user-defined gas-adsorption properties by adding a penalty function that guides the target properties. Furthermore, Bhowmik et al.86 provided a perspective on using a generative model for inverse design of complex battery interphases, and suggested that utilizing data taken from multiple domains (i.e. simulations and experiments) would be critical for the development of rationale generative models to enable accelerated discovery of durable ultra-high performance batteries.
Another important missing ingredient is the lack of a model for synthesizability prediction for crystals. The screening and/or generation of hypothetical crystals produces a large number of promising candidates, but a significant number of them are not observed via experiments. Currently, hull energies (i.e. relative energy deviation from the ground state) are mostly used to evaluate the thermodynamic stability of crystals not because they are sufficient to predict synthesizability but mainly because they are simple quantities easily computable, but they are certainly insufficient to describe the complex phenomena of synthesizability of hypothetical materials.93 Developing a reliable model or a descriptor for synthesizability prediction is thus an urgent and essential area for accelerated inverse design of inorganic solid-state materials.
In the case of GMs, as mentioned in the ‘Generative models’ section, developing an invertible and invariant model is still of great challenge since there is currently no explicit approach that simultaneously satisfies the latter two conditions. There are several promising data-driven approaches along this direction. Thomas et al.94 proposed deep tensor field networks which have equivariance (i.e. generalized concept of invariance)95 under rotational and translational transformation for 3D point clouds. A recently proposed deep learning model, AlphaFold,96 predicting 3D protein structures from Euclidean distance geometry is also noteworthy since the distance between two atoms is an invariant quantity. Developing such invariant models and/or incorporating invariant features into 3D structures would thus be invaluable to develop more robust GMs for crystals.
This journal is © The Royal Society of Chemistry 2020 |