Advances of machine learning in materials science: Ideas and techniques

Sue Sin Chong , Yi Sheng Ng , Hui-Qiong Wang , Jin-Cheng Zheng

Front. Phys. ›› 2024, Vol. 19 ›› Issue (1) : 13501

PDF (11699KB)
Front. Phys. ›› 2024, Vol. 19 ›› Issue (1) : 13501 DOI: 10.1007/s11467-023-1325-z
REVIEW ARTICLE

Advances of machine learning in materials science: Ideas and techniques

Author information +
History +
PDF (11699KB)

Abstract

In this big data era, the use of large dataset in conjunction with machine learning (ML) has been increasingly popular in both industry and academia. In recent times, the field of materials science is also undergoing a big data revolution, with large database and repositories appearing everywhere. Traditionally, materials science is a trial-and-error field, in both the computational and experimental departments. With the advent of machine learning-based techniques, there has been a paradigm shift: materials can now be screened quickly using ML models and even generated based on materials with similar properties; ML has also quietly infiltrated many sub-disciplinary under materials science. However, ML remains relatively new to the field and is expanding its wing quickly. There are a plethora of readily-available big data architectures and abundance of ML models and software; The call to integrate all these elements in a comprehensive research procedure is becoming an important direction of material science research. In this review, we attempt to provide an introduction and reference of ML to materials scientists, covering as much as possible the commonly used methods and applications, and discussing the future possibilities.

Graphical abstract

Keywords

machine learning / materials science

Cite this article

Download citation ▾
Sue Sin Chong, Yi Sheng Ng, Hui-Qiong Wang, Jin-Cheng Zheng. Advances of machine learning in materials science: Ideas and techniques. Front. Phys., 2024, 19(1): 13501 DOI:10.1007/s11467-023-1325-z

登录浏览全文

4963

注册一个新账户 忘记密码

1 Introduction

Commonly recognized as the fourth paradigm of science [1-4], machine learning (ML) has played a crucial role in the development of the data-driven scientific process, shaping the changes in experimental methodology, from measurement to data analysis, assisting the proving of mathematical theorem, and finding new discoveries in areas that were once deemed impossible. In the field of computational material science, the methods have been enriched by the efficient high-throughput ML-aided simulation/data generation and data-driven discovery [5, 6]. In experiments and materials synthesis, the advances in ML have helped researchers to efficiently analyze the data and identify hidden features within the large dataset [7-9].

Discovery of new internal logics, patterns, or rules [10-13], and the study of complex systems, including nanostructures [14-29], alloys [30-35], superlattices [22, 36-38], surfaces [39-41], and interfaces [40, 42-46], as well as from materials to devices [47-48], are typical research topics in materials science. These areas could be addressed according to user specifications [49-50] by leveraging ML and big data statistical methods [51], which have advanced to a stage where users can utilize them to achieve large and complicated objectives with complex models. By breaking down broad objectives into smaller tasks, corresponding ML algorithms and objectives that are suitable can be identified and applied.

There are numerous comprehensive surveys on ML in material science [52-60]. In this review, we focus on the application of ML in material science, discussing the recent advances in ML, illustrating the basic principles of applying ML in materials science, and summarizing the current applications and briefly introducing the ML algorithms involved.

2 Basics on machine learning

ML has a long history [61-67]. However, it has only returned to the spotlight recently due to the compounding ability it has gained from the surge in big data and improving data infrastructure and computing power. Stemmed from statistical learning, ML has gained huge successes and popularity in many other tasks and has far-reaching influences in many fields, including physics, chemistry and material science. In this section, the basic ideas and concepts in ML and essential milestones in its illustrious history are covered.

ML can be broadly defined as computational methods using experience (available past information) to improve future performance or to make accurate predictions [68]. Typical ML methods involve three parts: the inputs (previously obtained data), outputs (predictions), and algorithms. The sample size (sample complexity) and the time & space complexity of algorithms are crucial for ML [68]. Therefore, the ML techniques are different from conventional methods such as experimental measurements or computer simulations, but are related to data analysis and statistics. In more general terms, ML techniques are data-driven methods, which combines the fundamental concepts of computer science with ideas from statistics, probability, and optimization. ML can be integrated with other disciplines, resulting in multi-discipline techiques such as quantum ML or physics-informed neural networks, materials science-based learning.

In ML, the “standard” or conventional learning tasks have been extensively studied, which include classification, regression, ranking, clustering, and dimensionality reduction or manifold learning [68]. The problems related to the above tasks are listed in Fig.1. The definitions and terminology commonly used in ML for different learning stages are listed in Fig.2. The typical stages of a learning process are also shown in Fig.3, which can be briefly described as follows: with a given collection of labeled examples, one can firstly divide the data/samples into three groups, namely, training samples, validation data and test samples, then the relevant features associated to the desired properties are chosen, which are next used to train the pre-determined learning algorithm. This is done by adjusting the hyperparameters Θ in order to ensure that the hypothesis Θ0 has the best performance on the validation sample. Typical learning scenarios include supervised learning, unsupervised learning, semi-supervised learning, transductive inference, on-line learning, reinforcement learning, active learning, and more other complex learning scenarios. Different from traditional data analysis, ML is fundamentally about generalization [68]. Spectacularly, the neural network-based ML is able to approximate functions in a very high dimension with unprecedented efficiency and accuracy [2], and therefore it can be used for complex tasks in a wide-range of applications.

3 Recent progress in machine learning

Recently, the ML community has seen breakthroughs in many traditional AI tasks and classical challenging scientific tasks. This leap of improvement is being powered by both the new grounds in the underlying theory, the overall implementation and architecture, and the massive surge in data and data infrastructures. This section covers the advancement of ideas from various areas of applications of Artificial Intelligence − Natural Language Processing (NLP), Computer Vision (CV), Reinforcement Learning (RL), Explainability Artificial Intellignce (XAI), etc.

3.1 Classical machine learning application areas

In the field of natural language processing and understanding, ML models have made huge progress with Attention Transformer networks [69] and pre-training techniques. SuperGLUE [70] is a natural language understanding benchmark consisting of many tasks, which requires in-depth understanding of short proses and sentences. With superhuman performances at SuperGLUE benchmarks, it has been demonstrated that ML is able to model both understanding of natural language and generation of relevant natural language in context. The technique that has led to this leap in performance is pre-training [71], which refers to “training a model with one task to help it form parameters that can be used in other tasks”. Prompt learning is a form of ML that works with large models, to learn knowledge from a language model simply by prompting the learnt model with various types of prompts. BERT-like models have also been extended to process data from realms outside natural language, like Programming Languages, e.g., CodeBERT [72], and Images [73], and had been very successful in these realms too. Tab.1 lists works relevant to several main ideas in machine learning for Natural Language Processing (NLP).

Unsupervised learning has made strides in computer vision tasks, with models being able to identify subject in video, or identify poses of objects from point cloud in a video without labels. In the area of unsupervised learning for time-series data, ML models are able to effectively identify features from time-series data for both classification and prediction. Recently, there has been many work that extends the use of transformers to the characterization and prediction of extra-long timeseries sequences, Informer [99], Longformer [100]. Tab.2 lists works relevant to several main ideas in ML for Computer Vision (CV).

Reinforcement learning is the training of agents to make a sequence of reward-optimal decisions in an environment, often modelled as maximizing reward expectation in a partial Markov Decision Process (MDP). In reinforcement learning thrust, there is a huge improvement in the ability of state-of-the-art models to effectively navigate extra large search space to search for sequential actions to maximize task goals. Most notably, models like AlphaGo [116] and AlphaHoldem [117] have been able to navigate extra large search spaces with Monte Carlo search tree methods on latent representation of state space and action. Classical methods like State–Action–Reward–State–Action (SARSA) [118], Q-learning [119], TD-learning explores action space with a reward function, and learns a matching of state space to action, a policy or the q-value of the actions. In 2012, apprenticeship learning [120] initiated by Abbeel, proposes to define the architecture such that agent is able to learn directly from observations of a task being completed, instead of specifying the steps of a task. There is also a trend to integrate reinforcement learning with meta-learning in order to train multi-tasks agents to perform a variety of tasks [121, 122]. Tab.3 lists works relevant to several main ideas in ML for Reinforcement Learning (RL).

Many human teaching or learning techniques have been the source of inspiration for advancement of this thrust. With the emergence of effective sampling techniques amongst others, the Efficient Zero [140] models have been able to progress by accumulating experiences through randomly playing against itself repeatedly, which improves the ability of game-playing models. Reinforcement learning has not only made breakthroughs in all-information open games; recently there has also been breakthrough in multi-player partial information games like Texas Hold’em and Alpha Hold’em [117]. For multi-agent reinforcement learning, the common benchmark task StarCraft Multi-Agent Challenge (SMAC) [141] can now be effectively completed by reinforcement learning models, which effectively decomposes the cooperation task into role-learning by large neural networks [142] amongst many other techniques [143,144]. This is a breakthrough for multi-agent reinforcement learning.

3.2 On quantum machine learning

Quantum ML is one of the big next steps of ML [145]. While error-correction still limits our ability to build a fully quantum computer, it is possible to innovate with hybrid algorithms that uses quantum sub-algorithms or components to speed-up, robustify ML or simply to expand the theoretical boundaries of ML with 2 norm probabilities. In quantum computing, we can compute the similarity between feature vectors with state overlaps (denoted by bra and ket) instead of kernels via inner product. Consider a simple quantum ML scenario below:

K (x,x)= ϕ(x ),ϕ (x )QKE (x,x)= ϕ(x) ϕ( x).

The feature space in quantum ML can be obtained by state preparation. For instance,

(3.2)ϕ :[x 1;x2]( x1|0 +x 2|1) (x1|0+ x2|1).

The corresponding circuit is denoted by

(3.3) Sx AB( |0 A|0B)=SxA|0A SxB|0B.

We can have quantum kernel estimation [see Eq. (1)] [146], or quantum feature spaces [147] in hybrid ML algorithms or intermediate scale hybrid machines [148]. This offers a new insight to the types of kernels and linear algebra that we can use to improve ML in the classical sense. Quantum physics or chemistry can be more effectively simulated in the primitive sense using quantum ML algorithm. The hybrid ML coupled with quantum material science is potentially an important stepping stone for material scientists and computer scientists alike to innovate and research more efficiently.

3.3 Theory, explainable AI and verification

In classical computer science, the very hard case of Travelling Salesman Problem (TSP), a classical NP problem, has been solved with very satisfactory result based on neural networks, which either blend with pre-training of a solver of a mini-TSP or a reinforcement learning-based [149] strategy selector combined with heuristic. Other prominent NP problems like Maximum Independent Set (MIS) or Satisfiability Modulo Test (SMT) have also been solved satisfactorily with ML-guided heuristic search [150]. This demonstrates that ML models have been able to push through boundaries that have been set forth by traditional theoretical computer science. This breakthrough has been made possible by effective latent representation learning of essential features of the problem itself and the solver.

Explainability XAI techniques like Integrated Gradients (IG) [151], Local Interpretable Model-agnostic Explanations (LIME)[152], Shapley Additive Explanations (SHAP)[153], SimplEx [154] and various others have gained much attention. LIME attempts to identify hot areas in the image responsible for features that result in the prediction. SimplEx [154] is an explainability technique that attempts to explain a prediction with linear combinations of samples drawn from the corpus of training data; the technique returns a combination of training samples that has contributed to the predictions. There are also efforts to incorporate explainability by adding a layer at the end of neural networks for capturing explainability information. Explainable Graph Neural Network NN) techniques that apply specifically to Graph Neural Networks are broadly classified into several classes: Gradients/features based Guided Back-propagation (BP) [155], Perturbation Based GNN Explainer [156], SubgraphX [157], Decomposition Based, Surrogates GraphLIME [158] and Generation [159]. These GNN XAI techniques are well-suited for explaining feature importance for predictions at either the node level, edge level or graph level.

Verification is important for protecting neural network models against adversarial behaviours; adversary behaviours can be characterized by ill-intent shifts of planes of separation in the model so that it is more likely to err on otherwise correctly classified samples or corrupting input samples with noise or otherwise. Neural network robustness verification techniques like Rectified Linear Unit-Plex (ReLUPlex) [160] and alpha-beta CROWN [161] have also made huge progress. It is a numerical bounds back-propagation technique where the score boundaries for each class are back-propagated throughout the network to determine the overlap between class scores. Specifically, in the non-linear portions of the neural network, the ReLU activation functions were bounded with linear functions. Safety-critical applications have also been secured with neural network verification techniques, and the Airborne Collision Avoidance System for Unmanned Aircraft (ACAS Xu) [162] is an ensemble of 45 neural networks whose purpose is to give anti-collision advice to flying planes, and utilize ReLUplex methods to make their advice robust.

3.4 Stack optimizations for deep learning

Graphical Processing Units (GPU) are processors capable of parallel processing instructions. Standard GPU deep learning speedup techniques include convolutional layer reuse, featuremap reuse and filter reuse, and memory access is a common bottleneck [163]. The basic idea is that functions that are computed many times should be optimized on all levels, from high to low, including the instruction set level. The entire software stack, compiler technologies, and code generation have been optimized for deep learning computations on GPU. Deep learning GPU is known for its high energy usage; reducing energy usage is an essential objective for GPU optimization research [164]. The requirement for the scale of hardware architecture for ML is also loosening up, as engineers are packing engineering insights from large systems into smaller and energy-conserving systems, TensorFlow Lite Micro [165].

ML theory and practice have made massive progress in recent years. It is now transforming the scientific methods and has become deeply integrated with many scientific and humanities [166] fields. Application-wise, ML models have been trusted to make more and more crucial decisions for the well-functioning of society. For instance, in the criminal justice setting [167], ML models have been used to set bail for defendants; in the finance sector, models can help make decisions [168]; in the energy sector, they predict power generation efficiency for wind power stations. While neural network might still be a black box and can be hard to verify at times, its effectiveness as a predictor and sometimes generator has already been relied upon by many societal sectors for greater efficiency and effectiveness.

4 Development trend of machine learning for materials science

ML has helped material scientist achieve their study aims in a wide variety of tasks, most prominently as a screening tool in the design of a large variety of materials, which include energy materials, semiconductors, polymer design, catalyst, high entropy alloy, etc. The trend of going from processing a single dataset to achieving a specific aim to learning a latent representation of the underlying structure, which can later be finetuned to perform specific tasks, such as predicting the energetically stable structure across datasets, is rather prominent.

4.1 From numerical analysis to feature engineering

Traditionally, ML has been used as an advanced numerical regression tool to analyse experimental data in material science and many other fields [169, 170]. The remarkable ability of ML to interpolate data has allowed scientists to explain phenomena and verify hypotheses effectively.

Traditional material science ML practitioners often concern themselves with explicit feature engineering of specific materials [171]. Bhadeshia [171] has outlined four categories of models in material science; traditionally ML models are “models used to express data, reveal patterns, or for implementation in control algorithm”. The classical works that involve material property prediction mostly fall into the fourth category. Fig.4 illustrates the feature engineering process for material science, which encompasses four stages: feature extraction; feature analysis; correlation and importance analysis; and feature selection [172].

In material space, there are many degrees of freedoms, such as the atomic coordinates, coordination numbers, interatomic distances, and the position of the various species. Often, they are impractical to be used as the direct inputs to the algorithms, as they are not invariant under translation and rotation. In feature extraction [see Fig.4(a)], we seek to convert them into descriptors, which extract the underlying symmetry and distinguish systems that are truly different and not just a product of translations and/or rotations.

After the features are extracted, they undergo a series of analysis to fine tune and reduce the dimensionality of the descriptors space. The four commonly used methods, shown in Fig.4(b), are the filter method, embedded method, wrapper method, and deep learning method. With the analysis process completed, a mapping, as illustrated in Fig.4(c), which relates the importance and correlations among the selected features, can be used to visualize their dependence. In turn, this aids the process of feature selection, in which many suitable subsets of features [see Fig.4(d)] are chosen to proceed to the next stage − fed into the ML algorithm and compared to obtain the best performing minimal subset.

4.2 From feature engineering to representation learning

While explicit feature engineering is a practical and valuable task, it often restricts the type of task that ML can perform and does not fully use its ability to learn a generalized representation or sound separation of features and ability to interpolate or extrapolate along those dimensions. Moreover, the task of sifting through a vast dataset is laborious and hard to manage for individuals. Furthermore, with the ever-expanding computing power, the dimensionality of the features that is computationally feasible also rapidly scales up, allowing the consideration of more factors, which ultimately improves the accuracy of the prediction while also widening the coverage of material types screened. Thus, there is a push towards representation learning, an automation of feature engineering of a large material dataset [173], which better captures the internal latent features [174]. This trend encouraged a deeper integration in both development trends in ML and material science, coupled with a concise selection of ML tools, which require an intuitive understanding of mathematical and theoretical computer science ideas behind these tools.

In representation learning, the features are automatically discovered and extracted from the raw data, and thus complicated patterns that are hidden from the human user but are highly relevant could boost the accuracy and effieciency of the ML model, which is highly dependent on the quality of the selected features. Therefore, representation learning excels in applications where the data dimensionality is high and features extraction is difficult, such as speech recognition and signal processing, object recognition, and natural language processing [175].

Neural networks can be packed into layers or attention blocks that can be integrated into a single neural network. Effective embedding of information that is a dimensional reduction tool reduces the complexity of the model, when upended upon the training pipeline, brings us to end-to-end learning. Fig.5 shows a simplified pipeline for material science end-to-end model, where datasets are turned into vectors by the encoder to use as the input for the surrogate model, which attempts to identify the latent representation that can be decoded to generate predictions.

Representation learning has been applied in materials science. By using the raw experimental X-ray absorption near edge structure (XANES) spectra, Routh et al. [176] managed to obtain latent features after performing unsupervised ML methods. The raw experimental data are fed into an autoencoder that includes the encoder and decoder, which uses the input data as the output data, while information is passed through a bottleneck layer, as illustrated in Fig.6.

4.3 From representation learning to inverse design

After learning the representations that are critical in influencing the functionality of the materials, we ought to think: could we use them inversely, to generate novel and maybe better materials? This question had been sought in 1999 by Franceschetti and Zunger [177], where they successfully searched for the alloy of fixed elements with targeted electronic structure, using Monte Carlo method only. This limited yet profound results show us the vast usefulness of solving the inverse problem. Now, armed with the computation power and advancement in ML, we are in a better position to answer this question. Generative models like Variational Autoencoders (VAE) and Generative Adversarial Networks (GAN) have been applied in the inverse design of molecules and solid-state crystal.

By combining the power of representation learning and generative models into a single extensive model, that is the joining of neural networks from several parts of the workflow into a single network, many benefits can be reaped. First of all, the ability of an extensive network to counter noise levels in the training dataset, resulting in better predictions or better-generated solutions. Secondly, the latent representation learnt from each part of the pipeline is more consistent with the final goal of experimentation or design. Thirdly, the absence of human error-prone non-ML intervention helps experimenters focus on the overall goal and architecture.

By using discriminative models, generative models, and rapid simulation, whether standalone or in combination, one can construct sophisticated models that tackle problems ranging from predicting density functional theory (DFT) properties to inverse device design with confidence. One can also explore material design at different scale and granularity with ML model as an aide. An example is shown in Fig.7, where both discriminative and generative models are used jointly to design photonics. When the dimensionality of the photonic structures involved is very low, at the order of 1, analytical methods are well-suited. However, as the dimensionality increases, the analytical methods are no longer feasible, and the ML methods are required. On its own, discriminative models are suitable at slightly larger parameters space, but when the degree of freedom scales up considerably, generative model can be employed to reduce the dimensionality.

5 Databases in material science

Data is prevalent in material science; data which originates from every aspect and process of material science research endeavour have varying types, accuracy and genre. Tab.4 lists the typical data types and database that are used in ML models. A material science task often includes processing a combination of data types listed.

The broad spectrum of data types and multi-modules of input data dictates that material science models need to learn to integrate multi-modal data to produce meaningful research results. This trend also means that the material science community needs to embrace the software and statistical revolution that will propel the field forward.

In order to use computer systems to process material information, material-related nomenclatures have to adapt to computer processing norms, like string. Both atomic and structure of molecules should be evident by parsing strings, e.g. Simplified molecular-input line-entry (SMILES), BigSMILES [197], Self-referencing embedded strings (SELFIES) [198], Physical Information File (PIF) [199]. Material Science datasets are often implemented in neural network data loaders like Deep Graph Library [200]. The ML community’s Datasets have codebases that organize information that eases software engineers to call and process with a library. Most quantum chemistry software is softwareengineering based Application Programming Interface (API) to share and process Quantum Chemistrydata; it is written to store quantum calculation data with well-tested and scalable database norms (like schema) and eases or speedup data batch processing. The basis for quantum chemistry libraries is standardized; typical ones include Gaussian Orbital Basis (GTO), Plane Wave Basis (PW), and Numerically Tabulated Atom-centered Orbitals (NAO). Tab.5 lists softwares which might be useful. The first portion lists general deep learning libraries (APIs), second portion lists useful libraries for machine learning tasks, third portion lists tools that might be useful to material science.

6 Machine learning descriptors for material science

The material science datasets are often comprised of atomistic information with the coordinates of atoms, the charges on the atoms, and their compositions. To capture the spatially invariant information, the local environment on atomic scale is oftenly extracted, such as the list of neighbouring atoms and their relative spatial positions. They are then compactified and propagated as descriptors in the form of a vector, in a neural network which maps this information to their properties that are of interests: total energy, mass density, bulk moduli, etc. In general, a good descriptor needs to have the following qualities:

i) Invariant under spatial transformation (arbitrary translations, rotations, and reflections)

ii) Invariant under permutation/exchange of atoms of identical species, i.e., only a unique representation for each arrangement of atoms.

iii) Computationally cheap and easy to implement.

iv) Minor deviation under small perturbations in the atomic structure.

Clearly, the Cartesian coordinates of the atoms do not satisfy the points i) and ii), even though it is the easiest imaginable method. There are many different descriptors that have been tried and tested in material science, which we will attempt to briefly summarize in this section, but it is by no means exhaustive. For further information and use examples on descriptors, the reader is recommended to the articles of Li et al. [215] and Schmidt et al. [216].

6.1 Pair-wise descriptor

Pair-wise descriptor is a type of descriptor that considers each and every possible pair of atoms in the system. Examples include Z-matrices, Weyl matrices, and more recently, the Coulomb matrices [216]. A figure briefly describing the Weyl matrices and Coulomb matrices are shown in Fig.8(a). In the work of Rupp et al. [217], Coulomb matrices were constructed for a set of organic molecules that are numerically extracted and sorted descendingly, then the Euclidean difference between the vectors of eigenvalues are computed and defined as the distance between two molecules (with different dimensions accounted for by adding trailing zeroes to vectors). Using this as the sole descriptor, they developed a ML approach for fast and accurate prediction of molecular atomization energy. The same eigenvalue-based method has also been used in a number of recent studies [218, 219]. The downsides of this method are the inabilility to differentiate enantiometer [220] and the loss of information, as the dimensions are reduced from N2 to N, which can sometimes be an advantage [219].

As described, the Coulomb matrices methods are only viable for finite system. To extend the pairwise descriptor to infinite periodic system, Faber et al. [220] proposed three different methods: Ewald sum matrices, Sine matrices, and Extended coulomb-like matrix, and their results show that Sine matrix is the most efficient and outputs the smallest error.

Another alternative, the partial radial distribution function (PRDF) was proposed by Schütt et al. [221] and used in their work to perform fast prediction of density of states at Fermi level for different types of solids. The pairwise distances dαβ between two atoms type are considered, in the following equation for PRDF:

(6.1)gαβ(r) = 1NαVri=1Nα j=1Nβθ (dαiβjr)θ(r +drdαiβj),

where θ (x) is the step function, Vr is the volume of the primitive cell, while Nα and Nβ are the number of atoms of types α and β. Only the atoms in the primitive cell are considered as the shell center, i.e., the atoms αi, see Fig.8(b). This function is invariant under translation, rotation, and different choice of the unit cell.

6.2 Local descriptor

The most intuitive methods to describe a system of atoms that also take into the geometrical aspect into account is the neighbour-based or local descriptor, as the electron density is only weakly affected by distant atoms. By considering the neighbouring atoms of a selected atom within a pre-determined cutoff radius, we can store the information about their bonds, such as the bond distance and angle.

Behler and Parinello [222] proposed the use of two symmetry functions, the radial symmetry function Gi1 and angular symmetry function Gi2:

(6.2) Gi1= j ia lle η(RijRs)2fc(R ij),

Gi2=21 ζ j,kiall (1+λ c osθij k)ζ× eη (Rij2+Rik2+ R jk 2)fc(R ij) fc(Rik )fc(Rjk ),

where Rij is the distance between atom i and j, θij k is the angle between the three atoms i,j,k. There are four free parameters in total, λ( =+1, 1), η, ζ, and the implicit Rc in fc, defined as

(6.4)fc(R ij)={ 0.5 ×[cos(πRijR c)+1]forRijRc;0forRij>Rc.

The symmetry functions capture the local environment of an atoms, are invariant to permutation, translation, rotation, and changes in coordination number. They have been used in reproducing potential energy surface (PES) at DFT accuracy. This formalism was extended and studied in extensive details in Behler [223], where the set of symmetry functions are coined the “Atom-centered Symmetry Functions (ACSFs)”. A further generalization was done by Seko et al. [224], which included basis functions other than the Gaussian in Eq. (2), such as Neumann functions and Bessel functions. They also introduced the use of the Least Absolute Shrinkage and Selection Operator (LASSO) technique to optimize the basis set and find the sparsest representation to speed up computation. This was successfully used to reproduce almost DFT-accuracy phonon dispersion and specific heat for hcp Mg. A more recent work to reduce the undesirable scaling in ASCF has also been discussed [225].

Another approach using bispectrum, a three-point correlation function, was introduced by Bartók et al. [226]. In this approach, they first construct local atomic density function for each atom i, as

(6.5)ρi(r) =δ(r)+ jδ(rrij) fc(| rij |),

where the δ(r)’s are the Dirac Delta function. This atomic density is then projected onto the surface of a 4D sphere, by expanding the atomic density using 4D spherical harmonics, Ummj (index i omitted):

(6.6)cj mm= Umm j | ρ,

and the bispectrum is then built from these coefficients, defined as

Bj1, j2, j3= m1,m1=j1 j1 m2,m2= j2 j2 m,m= jj1(cmmj) Cj1m1j 2 m2 jmCj1 m1j2m2 jm c m1m1 j1c m2m2 j2,

where the Cj1m1j2m2 jm’s are the ordinary Clebsch−Gordan coefficients of the SO(4) group.

The Smooth Overlap of Atomic Positions (SOAP) descriptor [226] uses the atomic density defined in Eq. (5), but with the Dirac Delta function replaced by the Gaussians, expanded in terms of spherical harmonics:

exp(α | r ri| 2)=4π e xp[α ( r2+ri2 )] lmhl(2αr ri)Ylm (r^ )Ylm( ri^),

where hl’s are the modified spherical Bessel functions of the first kind and Ylm is the spherical harmonics. A similarity kernel k (ρ, ρ)dR^ |ρ (r)ρ( R^r) dr |n was introduced to compare two different atomic environments, where n=2 is used in their study. The normalized kernel or SOAP kernel

(6.9)K(ρ,ρ)=(k (ρ, ρ)k (ρ,ρ)k (ρ, ρ ))ξ,

where ξ is any positive integer, chosen to control the sensitivity, goes into the PES of the form

(6.10)ε (q)=k=1NαkK(q,q(k)),

where the q(k) is the training set configurations. The SOAP descriptor is now widely adopted, especially in the machine-learning of potentials [227-230].

Based on the SOAP approach, Artrith et al. [216] introduced another descriptor for machine-learnt potentials, which does not scale with the number of chemical species, a feature that SOAP lacks. This is carried out by taking the union of a set of invariant coordinates which maps the atomic structure and another one that maps the chemical composition, which are both described by the radial and angular distribution functions:

(6.11)RDFi(r) = αca(2)ϕα (r), 0rRc,

(6.12)ADFi(r) = αca(3)ϕα (θ), 0θπ,

where Rc is the cutoff radius and the ϕα is a complete basis set, which in their work is the Chebyshev polynomials of the first kind.

6.3 Graph-based descriptor

By converting the atoms and bonds in a molecule into vertices and edges, we can turn the molecule into a graph as depicted in Fig.9(a). The information about the edges and the edge distance between vertices can then be encoded into the adjacency and distance matrices [231], shown in Fig.9(b). This graph-theoretic approach is known as structure graph, which has been devised long ago in 1863. Despite the simplicity and apparent loss of 3D information, structure graphs have seen widespread uses in comparing the structure of molecules.

The generalization of structure graph to periodic systems is the Universal Fragment Descriptor (UFD) [232], which uses the Voronoi tessellation [see Fig.9(c)] to determine the connectivity of atoms, in the following two steps:

i) The crystal is partitioned into atom-centered Vornoi−Dirichlet polyhedral;

ii) Atoms that share a perpendicular-bisecting Voronoi face with interatomic distance smaller than the Cordero covalent radii (with 0.25 Å tolerance) is determined to be connected. Periodic atoms are considered.

which defines the graph. Subgraphs are also constructed corresponding to the individual fragments, which include linear paths connecting at most 4 atoms and circular fragments, representing the coordination polyhedral of an atom. Then, an adjacency matrix A is constructed based on the determined connectivity, along with a reciprocal distance matrix ( Dij =1/rij2), which when multiplied together gives the Galvez matrix MA D. The information about the atomic/elemental reference property q (could be Mendeleev group and period number, number of valence electron, electronic affinity, covalent radii, etc.) is then incorporated in the pair of descriptors for a particular property q,

(6.13) TE= i= 1n 1 j=i+1n |qiqj|Mij,

(6.14) TbondE={i ,j}bo nd s| qi qj|Mij,

where the former runs over all pairs of atoms while the latter only considers bonded pairs of atoms.

Xie et al. [233] proposed a framework, the generalized crystal graph convolutional neural networks (CGCNN) which introduced another graph-based descriptor that is inspired by the UFD. Their construction of the crystal graph is illustrated in Fig.10, where the connectivity determination is the same as in UFD, but they used the one-hot encoding to encode the atom and bond properties in two separate feature vectors: node feature vectors and edge feature vectors. They are the descriptors, which are then sent through convolutional layer, which further extracts critical features while reducing the dimensions. Convolutional neural network is discussed in Section 7.

6.4 Topological descriptor

Topology famously does not differentiate between a donut and a coffee mug, as they both have a hole. This is because in topology, most of the geometrical features are stripped away, leaving only quantities that are invariant under continuous deformation. This seemingly bizarre concept has had deep impacts in physics and was the main theme of the works that won the 2014 Nobel Prize. A branch of topology, persistence homology, measures the topological features which persist across different scales or granularity, and encodes them into diagrams. This idea has already been used to classify and describe proteins [234, 235], and used as ML descriptor for crystalline [236] and armophous solids [237]. However, it has not been widely used due to its mathematically complicated nature, and the lack of physical and chemical intuition also further hinders the ability to interpret the results [216]. Here, we will attempt to provide a simplified and non-rigorous overview of persistent homology and the crystal topological descriptor of Jiang et al. [236]. For a rigorous and detailed introduction to persistent homology, Ref. [238] is recommended along with other works cited in this paragraph, especially Ref. [235].

The basic building blocks of persistent homology are the simplices (see Fig.11): a 0-simplex is a point, a 1-simplex is two connected points, a 2-simplex is a filled triangle, and a 3-simplex is a filled tetrahedron. A face can be a point, a line, or a 2D surface, depending on the number of points. A simplicial complexK is a collection of simplices which satisfies two conditions:

i) Faces of a simplex in K are also in K.

ii) Any intersection of two simplices in K is a face of both the simplices.

In a simplicial complex, holes are considered as voids that are bounded by simplices of different dimensions. In dimension 0, a connected component is counted as a hole; in dimension 1, a hole is a loop bounded by 1-simplices or edges; in dimension 2, hole is bounded by 2-simplices or triangles. The number of i-dimensional holes or voids in a simplicial complex is basically described by the i-th Betti numbers, βi(K), e.g., β0 is the number of connected components, β1 is the number of loops, and β2 is the number of cavities. An example is shown in Fig.11(b), where there are five 0-simplices or vertices, forming six 1-simplices, and one 2-simplices. Since all the points are connected, β0= 1; there is a square-shaped hole enclosed by the 1-simplices A, B, C, and F, giving β1=1. Note that the face of the triangle is filled, thus it is not counted as a hole.

To generate simplicial complex from the crystal data, we use the Vietoris−Rips (VR) filtration process, giving VR complex. This is carried out by increasing a filtration parameter, commonly the Euclidean distance cutoff between points, where points that are within cutoff distance of each other are connected. The filtration parameter used in the crystal descriptor is the radius measured from each atom, d, which is increased from 0 to 8 Å. As d increases, the simplicial complex also undergoes changes, where the Betti numbers of “holes” change. This can be quantitatively plotted using persistence barcodes for each of the Betti numbers, as can be seen in Fig.12, where each barcode represents each of the “hole” for each Betti number. As d reaches 4 Å, all the Betti 0 barcodes except one suddenly terminate, indicating that the points are now all connected as the Na atoms are separated by 4 Å. There is no Betti 1 barcode because the distances between any two Na atoms are the same, reflecting the structure symmetry.

To embed different elemental compositions, atom-wise chemical information is used, where a chosen atom is surrounded by atoms of a chosen type, such as in Fig.12(a) and (b), where the selected Na atom is surrounded by only Na atoms and Cl atoms, respectively. The birth, death, and the persistent length of the Betti barcodes are then encoded in a vector known as ASPH feature vector.

6.5 Reciprocal space-based descriptor

The reciprocal space is linked to the real space by the means of Fourier transform and can be mapped using X-ray diffraction (XRD) into 2D diffraction pattern [see Fig.13(a)], either experimentally or computationally. 2D XRD data has first been applied as descriptor by Ziletti et al. [239] to automatically classify crystal structures. In their work, they rotated the crystal 45o clockwise and counterclockwise about a chosen axis and superimposed the obtained XRD patterns. This is then carried out for the other two axes, with different colours of the RGB palette chosen for the patterns obtained from the rotation of different axis [e.g., red for x-axis, green for y-axis, and blue for z-axis, see Fig.13(b)]. The final obtained pattern is then used as the descriptor, fed into a convolutional network, similar to image-based object recognition. The benefits of the XRD descriptor are that the dimension is independent of the system size and very robust to defects [compare Fig.13(b) and Fig.13(c)].

The more conventional XRD is the 1D XRD, shown in Fig.13(d) for different crystals, which is obtained based on Bragg’s law, mapping the 3D structures into 1D fingerprints. 1D XRD based descriptor has been used to classify crystal structure [240] and predict their properties [241]. In the latter, the group used modified XRD, where only the anions sublattice is considerd with the cations removed, and the pymatgen package is used to generate the XRD computationally. They successfully distinguished solid-state lithium-ion conductors with this descriptor using unsupervised learning.

6.6 Reduction of descriptor dimension

In materials science, there are many possible combinations of various properties that can be used as descriptors. It is often difficult to select and fine-tune the descriptor space manually. This is a common problem in the field of ML, and several methods have been developed to tackle this issue: principal component analysis (PCA) [242, 243], least absolute shrinkage and selection operator (LASSO) [244], and sure independence screening and sparsifying operator (SISSO) [245]. However, they mainly work for models that are linear, hence not directly applicable for neural network-based models [239].

7 Machine learning algorithms for material science

In this section, we collate the recently developed ML-based tools and frameworks in materials science, grouping them together by the ML algorithms used. We then briefly describe the commonly used algorithms and also introduces some of the emerging algorithms, which could further unlock the potential of materials science ML applications.

7.1 Currently utilized algorithms

Tab.6 enumerates the ML algorithms used in relatively recent developed tools or framework in materials science. It can be seen that the convolutional and graph neural networks are the popular algorithms, with transfer learning also picking up pace. We will briefly introduce the algorithms, drawing examples from the materials science implementation.

7.1.1 Kernel-based linear algorithms

Support vector machine (SVM) and kernel ridge regression (KRR) are kernel-based ML algorithms, which utilize kernel functions K(x,x ) that allow high-dimensional feature to be used implicitly, without actually computing the feature coordinates explicitly, hence speeding up computation. Furthermore, it allows non-linear problem to be solved using linear algorithms by mapping the problem into a higher-dimensional space. Examples of commonly used kernel functions include

Linear kernel:

(7.1)K (x ,x)=( xi xj+θ);

Polynomial kernel:

(7.2)K (x ,x)=(xixj+θ )d;

Gaussian kernel/radial basis function (RBF):

(7.3)K (x ,x)=exp(|| xixj||2σ2),

where x is the input data, while θ and σ are adjustable parameters. SVM is used for both classification and regression problem, denoted as SVC and SVR, respectively. On the other hand, KKR is used only for regression problems and it is very similar to SVR, except for the different loss functions.

Applications of both types of SVM are demonstrated in the work of Lu et al. [246]. Using atomic parameters such as electronegativity, atomic radius, atomic mass, valence, and functions of these parameters, they constructed classifier for the formability of perovskite structure, and regression models to predict the band gaps of binary compounds, using the polynomial kernel [Eq. (7.2)] with d=2 and θ= 1.

Wu et al. [247] used KRR to assist in non-adiabatic molecular dynamics (NA-MD) simulations, particularly in the prediction of excitation energy and interpolate nonadiabatic coupling. KRR was chosen over neural networks because of the fewer hyperparameters KRR possessed, and KRR requires only the use of simple matrix operation to find the global minimum. By only providing a small fraction (4%) of sampled points, KKR gives a reliable estimate while saving MD computational effort of over an order of magnitude.

7.1.2 Neural network

Artificial Neural Networks (ANNs, shortened to NNs) is a type of ML architecture that aims to mimic the neural structure of the brain. In NNs, there are 3 types of layers consists of interconnected nodes: input layer, hidden layer(s), and output layer, as shown in Fig.14(a). The input layer receives the input raw data, which is then propagated to the hidden layer(s), where the nodes are functions of the backward-connected nodes and each connection is weighted. The function of a hidden layer node m with x being the node values of previous layer, takes the form:

(7.4)hm= σ(b+ inωixi),

where σ(z) is known as the activation function, where a common choice is the sigmoid function σ(z)=1/(1+e z) and b is bias term, and the ReLU (Rectified Linear Unit) function, simply defined as σ(z)=max(0, z). After the hidden layer(s), the information is then passed toward to the output layer, which is another function of the nodes of the final hidden layer. The outputs are measured against true value using a pre-defined cost function, with the simplest example for regression problem being the sum of squared error

J(ω )=1n i=1n12( y iy^i)2,

where yi and y^i are the true and predicted values respectively, and the sum is taken over the whole training set with n being the size of the training set. The weights are then optimized iteratively using the backpropagation method, which is a function of the gradient of the cost function. For a detailed discussion, the book [248] is recommended.

Deep NNs are NNs with more than one hidden layer [see Fig.14(b)]. By having more hidden layers, the model is better positioned to capture the nonlinearities in the data. However, having too many hidden layers can cause the convergence or learning to be slow and difficult, because the gradients used in backpropagation will tend to become vanishingly small. To overcome this issue, residual block has been devised [249], which introduces shortcut between layers.

SpookyNet [250] is a DNN-based model built to construct force fields that explicitly include nonlocal effects. In their DNN architecture, the generalized sigmoid linear unit (SiLU) activation function is used, which is given by s il u(x) =αx1+e βx, where both α and β are learneable parameters. They noted that a smooth activation function is crucial for the prediction of potential energies, as discontinuities in the atomic forces would be introduced otherwise. They introduced a loss function that has 3 components: energy, forces, and dipole moments, which is minimized by optimizing the weights using mini-batch gradient descent. They also incorporated residual block which allowed them to use a large number of hidden layers.

Convolutional NNs (CNNs) is primarily used in image pattern recognition, and is different from deep NNs by having a few extra layers, which are the convolutional and pooling layers. The extra layers filter and convolute the data to capture crucial features in the data and also reduce the input dimension, which scales quickly with resolution in image recognition problems. The work of Ziletti et al. [239] used CNN architecture, as depicted in Fig.15. The convolution layers capture elements that are discriminative and discard unimportant details.

Graph NNs is specifically designed for input data that are structured as graph, which contains nodes and edges, and can handle inputs of different sizes. There are several differrent types of graphs NNs, such as graph convolutional network (GCNNs), graph attention network, and Message Passing Neural Network.

7.1.3 Decision tree and ensembles

Decision tree is a supervised method for solving both classification and regression problems, which resembles a tree. A typical decision tree is shown in Fig.16(a), where each internal node represents a feature or attribute, each branch contains a decision rule, and each leaf node is a class label or a numerical value, depending on the type of problem solved. The number of node layers a decision tree contains is known as depth, which needs to be tuned. An important metric used in measuring the performance of a decision tree in classification is the information gain, which is defined as the difference of the information entropy between the parent and child node; while for regression problem, the variance reduction is the performance evaluation metric for a decision tree. Decision tree is advantageous when it comes to interpretability, but it suffers from overfitting, especially when the tree is too deep and complex. It can also be overly-sensitive to data changes.

Random forest is an algorithm that combines multiple decision tree, with each of them trained on randomized subsets of samples, where both training data and features are chosen random with replacement in a process known as bootstrapping. The final decision is then made by aggregating the results from each decision tree and taking the majority vote for classification or the average for regression. The steps taken above are collectively known as bagging, which help ensure that the random forest algorithm is less sensitive to changes in the dataset and more robust to overfitting.

Gradient Boosting Decision Tree (GDBT) is another method that uses ensembles of decision trees but in sequence rather than in parallel. GBDT works by adding decision trees iteratively, with each one attempts to improve upon the errors of the previous tree. The final output from the trees ensembles is then taken by using weighted average of the decision trees outputs.

Random forest models are used in the work of Zheng et al. [251], which predicts the atomic environment labels from the X-ray absorption near-edge structure (XANES). Using the random forest classifier of scikit-learn package, they found that 50 trees ensemble gave the best performance, even better than other classifiers, such as CNN and SVC. On the other hand, GBDT has been used for regression in the topology-based formation energy predictor [236]. Also using the scikit-learn package, they added a tree to their model one at a time and used bootstrapping to reduce overfitting. This topology-based model is able to achieve a high accuracy in cross-validation, with mean absolute error of only 61 meV/atom, outperforming previous works that uses Voronoi tessellations and Coulomb matrix method.

7.1.4 Unsupervised clustering

K-Means clustering is a popular unsupervised classification algorithm which aims to group similar data points together in K different clusters. K numbers of points that are known as cluster centroids are initialized randomly, and each data point is assigned to a cluster centroid that is closest in Euclidean distance to the data point. The centroids are then moved to a new location that is the arithmetic mean of the assigned data points. This repeats until convergence, i.e., there is no more movement among the centroids. The number K determines the number of classes in the data, which can be known before hand if the dataset has clear distinction, e.g., metal vs. non-metal, or can be optimized using the elbow method, which has an associated cost function J that is optimized by the best choice of K. Despite its popularity, K-Means clustering has some limitations, such as sensitivity to outliers, dependence on the centroids position initialization, ineffective for dataset with uneven distribution, and predetermined number of clusters.

Several alternatives have been: proposed which improves upon the limitations of K-Means clustering. Agglomerative hierarchical clustering (AHC), used in the work of Zhang et al. [241], is initialized by using each data point as a single cluster, then iteratively merged the clusters of the closest points until one big cluster is left. Then, a dendogram or a bottom-up hierarchical tree diagram, as show in Fig.16(b), which can be cut at a desired precision, as indicated in the figure via a dashed line, where 7 groups are obtained. To verify the results, they performed spectral clustering, which splits the samples into chosen K groups, based on the eigenvalues of the similarity matrix constructed from the data. This process is recursively applied bisectionally, and they obtained similar clusters as the AHC. There is also the mean-shift algorithm, utilized in Ref. [252].

7.1.5 Generative models (GAN and VAE)

Generative models attempt to learn the underlying distribution of a training dataset, and use that to generate new samples that resemble the original data. Two popular types of generative models are Generative Adversarial Networks (GAN) and Variational Autoencoders (VAE). As can be seen from Fig.17, there are two different neural networks in both of the models: GAN contains a discriminator and generator network, while VAE has a decoder and encoder network. In GAN, random noise is injected into the generator network and subsequently outputs a sample that is then fed to the discriminator network, which is then classified as real or generated sample. The networks are trained together until the generated samples are able to convince the discriminator that the samples are real and not generated. On the other hand, VAE tries to learn the latent representations from the training data and generate new samples based on them using probabilistic approach.

A variant of GAN, Wasserstein GAN, has been applied in the work of Kim et al. [253], which generate Mg−Mn−O ternary materials which can potentially be used as potential photoanode materials. The overview of their GAN architecture is shown in Fig.18(a), which after training, takes in random Gaussian noise vector Z and encoded composition vector, and spits out new unseen crystals. The new crystals are then passed to a critic and a classifier, where the former computes the Wasserstein distance that measures the dissimilarity between the generated and true data distributions, which are used to improve the realism of the generated materials, while the latter ensure that the generated materials meet the composition condition. Using this model, they found 23 previously unknown new crystals with suitable stability and band gap.

An example of inverse design using VAE was demonstrated in the work of Noh et al. [254], where their proposed a two-step VAE-based generator is shown in Fig.18(b). In the first step, the materials data is passed to a convolutional autoencoder, which contains 4 convolutional layers, outputting a compressed intermediate vector, which is then fed to a decoder that aims to maps the vector back to the input. The intermediate vector is fed into the VAE in 2nd step to learn about the latent materials space. To generate completely novel polymorphs, the materials space around known stable structure is sampled using random Gaussian distributed vectors and the resulting latent vectors are decoded in a series of steps to output new stable structures. The model is able to recover 25 out of 31 known structures that are not included in the training, and 47 new valid compositions are discovered that have eluded genetic algorithms.

7.1.6 Transfer learning

In materials science, high quality data for a specific type of materials is usually scarce, which severely impedes the applications of ML in generating high quality predictions [255]. Transfer learning (TL) is a method that can be applied to overcome this data scarcity issue. In transfer learning, the parameters of a model that has been pre-trained on a large data set but with different task/purpose, are used to initialize training on another data-scarce task, such as the parameters of the models used for predicting formation energy is later used to train another task of predicting band gaps.

Chang et al. [256] combined pairwise transfer learning and mixture of experts (MoEs) framework in their model. In pairwise transfer learning, a model is pre-trained on a source task (task designed for the large dataset) and a subset of the pre-trained model parameters is used to produce generalizable features of an atomic structure, defined as a feature extractor. This extractor extracts a feature vector from an atomic structure, which can be used to predict a scalar property after passing thorugh a neural network. On the other hand, MoE contains multiple neural network models that specialize in different regions of the input space, known as “experts”, and each of them are activated and controlled by a gating function. The outputs of the “experts” are then aggregated through an aggregation function. Using this architecture, the authors have performed many downstream, data-scarce tasks, such as predicting band gap, poisson ratio, 2D materials exfoliation energy, and experimental formation energies.

7.2 Emerging ML methods

7.2.1 Explainable AI (XAI) methods

The DNNs-based approaches discussed have proved to be of great help in assisting and speeding up materials research, but their black-box nature has made understanding and explaining the results difficult, which has also plagued the general ML community [302]. In systems that trust, fairness, and moral are highly critical, such as in healthcare, finance, and autonomous driving, the decisions made by AI cannot be blindly trusted without understanding the motivation and reasoning behind the choice. Furthermore, when the black box returns results that are erroneous and puzzling, it can be difficult to diagnose and correct without knowing what exactly went wrong. To overcome these issues, the XAI techniques were introduced, which try to explain the reasonings and connections behind a prediction or classification.

There are many post-hoc (i.e., applied after model fitting) XAI methods proposed for the general ML community [303], including gradient-based attribution (Gradients, Integrated gradients, and DeepLIFT), deconvolution-based methods (Guided backpropagation, Deconvolution, Class activation maps (CAM), Grad-CAM), model-agnostic techniques (Shapley additive explanations (SHAP), local interpretable model-agnostic explanations (LIME), Ancors).

Another type of XAI is the use of models that are inherently interpretable or explainable, which have one or more of these features [304]: sparsity, simulatability, and modularity. A model that has limited number of nonzero parameters is known as sparse, and this can be obtained by the LASSO method, whereas if a model can be easily comprehend and mentally simulate by the human user is simulatable, such as decision trees-based model. A modular model is a model that combines several modules which can be interpreted independently. In the field of materials science, the understanding of the physical and chemical intuition is paramount as it opens the door to understanding hidden connection and physics, and improve the efficiency of future studies by providing insights from previous work. The importance and implementation of XAI in materials ML tools (refer to Tab.6) have been discussed in the review of Oviedo et al. [58] and Zhong etal. [302]. Zhong et al. [302] presented an overview of DNNs-based XAI as shown in Fig.19(a), which highlight two fundamental motivations for XAI, which is the need for explaining how the results are obtained from the input (model processing), and what information is contained in the network (model representations). The design of an intrinsically explainable DNNs will prove important in answering the questions posed, but is itself a highly difficult task. In the following, we will illustrate some of the materials science XAI implementation, which is still in its infancy.

Kondo et al. [305] used heat maps to highlight the feature importance, particularly in identifying the positive and negative features that affect ionic conductivity in ceramics, using scanning electron microscope (SEM) images. Their CNN-based model used feature visualization method that is very similar to the deconvolution method used in CAM and Grad-CAM. By defining mask map, they obtained masked SEM images [see Fig.19(b)] that show features that determine low and high ionic conductivities.

A recent implementation of XAI for crystals is the CrysXPP [306] which is built upon an auto-encoder-based architure, CrysAE, that containing deep encoding module which is capable of capturing the important structural and composition information in crystal graph. The information learnt is then transferred to the GCNN contained within CrysXPP [shown in Fig.20(a)], which takes in feature selected from crystal graph. The feature selector contains trainable weights that selects weighted subset of important features, which is fine-tuned with LASSO to improve the sparsity of the features. An example of the explainable results obtained is shown in Fig.20(b), where features that affect the band gap of GaP crystal are weighted and compared.

Compositionally restricted attention-based network (CrabNet) [307, 308] is an example of explainable DNN in materials science that is based on the Transformer-based self-attention mechanism [69], a type of algorithm initially intended for NLP, but has exploded in popularity recently. Briefly, the transformer self-attention mechanism allows the model to focus on the different parts of the input and relate them with weights to encode a representation. In this way, the dependencies between the elements are better captured, even when some of the elements are present in very small amount, e.g., dopants, which even in small quantity can have tremendous effect on the properties of the materials.

7.2.2 Few-shot learning (FSL)

As mentioned above, high-quality data in materials science which are complete with proper labels are scarce. This issue is exarcebated when we look at experimental data, which unlike computationally-produced data, is plagued with issues stemming from the different experimental equipments and variable environments. Therefore, the few-shot learning (FSL), which specifically targets situation where data is limited, has enticed materials scientists, especially the experimentalists. There are several approaches to FSL, as discussed in Refs. [309, 310], such as metric-based, optimization-based, and model-based approach. FSL is still a relatively young and unrefined method, but it has already attracted a lot of attentions. FSL has been implemented in the prediction of molecular properties [311, 312], classification of space group from electron backscatter diffraction (EBSD) data [313], and segmenting electron microscopy data [314].

8 Machine learning tasks for material science

This section will discuss the coverage of materials science tasks that ML tools have been utilized to assists in tackling. The common ML tasks in material science often coincide with the traditional ML tasks, which have been extensively studied and optimized. The tasks of inference of material property given structural and compositional data, generative modelling from a latent representation of desired properties, and the generation of DFT functionals, are analogous to the tasks that ML has traditionally performed well, including object classification, image and text generation using text cues, and natural language processing (NLP).

8.1 Potentials, functionals, and parameters generation

Traditionally, the XC functionals used in DFT are generated through mathematical approaches, guided by empirical data, such as the Perdew−Zunger exchange, with the exact XC functional remains elusive. The search for an improved XC functional above the currently popular GGA on the Jacob’s ladder of Perdew [315] is desirable. The techniques of ML have been started to be utilized in the generation of new XC functionals [316-320], with the aim of improving the calculated accuracy while maintaining the efficiency. Transferability remains a huge challenge, which will need a huge and diverse dataset to achieve.

The potentials and force fields used in molecular dynamics (MD) are critical in determining the reliability and accuracy of the output [321]. MD that does not involve first-principles approach but rather fixed potentials are in general less accurate the ab initio MD (AIMD), but they can be applied on a large system and long time scale, where AIMD is too costly. As such, one would hope that the standard MD can bring about results similar to AIMD. Developed in 2017, DeePMD [322] accurately reproduced the water model obtained from DFT. The same team developed an open-source tool for the on-the-fly generation of MD potentials, known as DP-GEN [323], available on available on URL: github.com/deepmodeling/dpgen. In 2020, the team won the ACM Gordon Bell Prize for the DeePMD work, as it can be scaled efficiently on the best HPC. Similar works have also been carried out by other teams [324, 325].

Another material modeling technique is the Density Functional Tight Binding (DFTB) method, which is less computational expensive than DFT-based first principles calculations. Efforts have been carried out on applying ML to obtain the TB parameters [252, 326].

8.2 Screening of materials

There are many methods to compress design space. To name few, one could train a model that predicts material property given material information or performs ML guided simulation of new materials to predict material behaviour under certain circumstances. High-throughput screening eliminates most potential materials without actually performing actual experiments to verify their property and provide experimentalists with a minimal set of candidate materials to try out. Pivoting to a generative model perspective, one could also specify material properties and generate stable materials that are likely to fit the specified property [327-329].

There are many properties that are of interests, including band gaps, bulk and shear moduli, crystal structures, conductivity, and topological states, as discussed in details in Ref. [216]. These properties are usually computed via DFT, which could be computationally expensive depending on the system setup. Properly setup and trained ML models can produce DFT-level accuracy properties predictions, while at far lower computational time. Isayev et al. [232] managed to obtain prediction at 0.1 s for each structure, which amounts to 28 million structure in a day, as pictured in Fig.21. However, a well-trained and fully-transferable ML model requires the existence of high-quality large database and heavy computational power to optimize the model.

8.3 Novel material generation

The latent representations of common desired properties are of high interests among the community [331]. Based on the learnt latent representation, we can generate structure with similar desired properties at will. This is often carried out using the generative models, such as GAN and VAE, as demonstrated in the work of Dong et al. [332] and Pathak et al. [333].

8.4 Imaging data analysis

There are many imaging methods to capture the structure and fingerprints of a material experimentally at the atomic level, such as X-ray Diffraction (XRD), Fourier Transform Infrared Spectroscopy (FTIR), Atomic Force Microscopy (AFM), Transmission Electron Microscopy (TEM), and many others. Typically, they require laborious human interpretation to understand the meaning of the signals and whether they are due to noises and errors. This can be remedied with the help of ML, and thus far ML has been applied to:

• Identify symmetry and space group from XRD [334];

• Discover hidden atomic scale knowledge from XRD [297];

• Identify functional group in gaseous organic molecule [335];

• Analyze patterns and feature in AFM images, including domain wall and grain boundaries [336];

• Quantify nanoscale forces in dynamic AFM [337];

• Perform structural analysis and reconstruction in TEM [338];

• Identify chemistry and processing history from microstructure image [298];

• Characterize and analyze mechanical behaviour in microstructure image [339].

8.5 Natural language processing of material science literature

Nature language processing (NLP) refers to the ability of computer algorithm to understand spoken and written language. This technology has seen explosive development, with the recent GPT-3 [340] and now GPT-4 [341] models making strides not just academically, but used in just about everywhere. The trained AI models are able to hold realistic conversation with humans, take standardized exams [342], write codes in various programming language, and so on.

Most of the published results on materials science are not stored in a centralized database, which hinders the overall effort of ML applications. The NLP techniques can help this by scraping information from published literatures, such as the materials structure and properties. Tshitoyan et al. [343] demonstrated that an unsupervised learning modethat can capture complicated underlying knowledge from the literature and recommend materials for functional applications before their actual discovery. NLP also can help hypothesis formation and provide knowledge on the current trends in the field [344]. NLP methods can also serve as an efficient knowledge extractor from vast amount of material science literature, making the literature review process more efficient and thorough for researchers [345].

9 Perspectives on the integration of machine learning in materials science

In the following we will list perspectives on the integration of machine learning in materials science with materal science point of view and with machine learning point of view, respectively.

9.1 Perspectives from machine learning viewpoint

As ML techniques and ideals become ever more prevalent, we believe algorithmic templates and ML ideas will eventually become either the target modes of computation or the mode of guidelines which decides the permutation to which areas of material science garner attention and gain resources. Machine translation has evolved from a rule-based coupled with statistical model to a very data-driven approach, and researchers are discussing the translation task with less and less reference to a specific source and target languages, pivoting towards advancing mode of computation for the task as a whole.

9.1.1 More deep integrations

We might also observe the trend of attempting to learn descriptors for parts of complex systems with ML models to be either more computationally efficient or more human interpretable or editable. Instead of scientists attempting to describe a system with equations from first principles, ML models can help scientists discover a better set of descriptors for systems across all datasets. For example, descriptors could be descriptors for input data (atomistic information/reaction space) or labels (crystal structures). These discoveries can hugely impact physics and chemistry theory [346], experiments and research methods. Physics-informed neural networks [347] can both improve neural network performance and physics research efficiency.

Increasingly, it could be more and more about the mapping of descriptors. We can imagine that with the emergence of more sophisticated models, it is possible to advance a particular segment of study in material science, such as polymer design, by completing a well-defined sophisticated task with a model/model of computation, where the lack of relevant databases will limit its advancement. Tracing the development of computer science, sophisticated models which perform generic tasks in material science well will again be integrated into a giant multi-purpose model much like a generic processing chip, to which we can prompt for insights which was previously only gained by human experimentation at a much slower rate. ML models will bring material scientists closer to the many possibilities already inherent in big data itself, allowing us to explore and exploit the possibilities with greater efficiency. The task material scientist will be able to complete with the help of machine learning will become more integrated and sophisticated, from the screening of material to the design of material as a complete task. Then with the design of material as an atomic/primitivetransaction, we will be able to come out with new science on top of the material design as a whole.

9.1.2 Systematic generalization

In our stride towards autonomous general intelligence (AGI), researchers have drawn many parallels and inspirations from neuro-sciences [348] and how humans learn and teach each other to develop models which better generalizes to novel situations and objects well. We expect a body of material science knowledge and ideas to become generalized and accessible to other fields, conveyed by advanced models in the future, where we can generalize or verbalize properties of imagined materials or predict performance of material in novel situations with high accuracy with its formal deduction process generated by models. We can also observe the interaction, cooperation or contradiction between bodies of materials science knowledge for novel materials and circumstances, and perform research on the intersection of bodies of knowledge with more depth and rigor. ML models can also learn to identify potential directions for exploration, come up with a comprehensive experimentation plan and collaborate with human researchers as a navigation co-pilot. The novel direction identified will be novel and comprehensive because models can learn from passive observations of a large material science literature, its publication trend [349] and insight analysis of researchers.

9.1.3 Huge computational models

With the development of reaction environment models, one might also reasonably expect reinforcement learning game-play learning to learn an agent policy for a material, i.e., to first learn a material behaviour that is desirable for a particular purpose, followed by an automatic search/generation of material which suits the specification. In general, the compounding of learning methods to get a solution for an even more vaguely defined objective but more analyzable process for that solution results in a human-verifiable solution for large or vague problems. Moreover, the increasing synthesizability or explainability of the solution to vague problems will help material scientists navigate methods for solving huge overarching/generic problems with more finesse, evolution of subject through large models [350].

Model of computation might become the common language of material scientists and researchers from other fields. Task definition might become the lingua franca or the leading cause of concern for ML practitioners in material science. This broader definition of material science might then, in turn, propel the advancement of machine learning. In general, the barrier to entry to both advanced material science and advanced ML will be lowered, allowing more experts from other fields or individuals to contribute their efforts and ideas to the development of both fields.

Mechanisms in quantum ML will become readily-available to be integrated with quantum physics, chemistry and subsequently material science. As classical-quantum hybrid infrastructure and architectures [351] become more available, quantum learning for material science might incorporate mechanisms of both quantum computing and quantum analysis of materials as primitives. This trend is expected to speed up the inter-disciplinary mixing of these fields from both engineering and theoretical grounds.

The resulting phenomenon is the emergence of an ever more integrated huge ML model, a Super Deep-Learning Model, which will tackle most if not all of the most fundamental underlying problems in material science; it will integrate fundamental engineering ideas from computer science with domain-invariants of material science, which is designed to perform well for various tasks on both super-computing facilities, quantum or otherwise, and on limited resources devices, [352] scalable yet robust. Moreover, by integrating the best training and privacy practices from ML software and hardware development experience, future material scientists can quickly expect robust material science downstream models running smoothly and reliably as an application on widely available and portable devices like a cellphone.

9.2 Perspectives from material science viewpoint

Currently, one of the biggest challenges is the availability of high-quality data. The increasing number of research groups adapting the open data approach and the growing availability of internet of things (IoT) devices will solve this problem, albeit gradually. We have also discussed several possible methods to overcome the issues, which is the advancement of small training sample ML algorithms, such as transfer learning and few-shot learning algorithms will also be one of the possible solutions to this issue.

9.2.1 Theoretical and computational materials science

The various computational techniques in materials science, such as DFT, molecular dynamics in its various forms of molecular dynamics (MD), monte-carlo methods, and density functional tight binding method, has started to benefit from the application of ML and will continue to do so in a dramatic manner.

As of now, the Kohn−Sham DFT remains a reliable and popular method for determining various material properties. However, the accuracy of DFT calculations heavily relies on the quality of the approximations employed, such as the exchange-correlation (XC) functional. The search for improved approximations, including exact functionals, using ML has only recently commenced. Another area for improvement in DFT is reducing computational costs. Recently, ML-refined numerical techniques have emerged that offer faster speeds compared to their traditional counterparts [353-355]. It is hoped that these advancements can eventually be applied to accelerate DFT computations.

The integration of ML into MD, exemplified by methods like DeePMD, has demonstrated the potential to achieve DFT-level accuracy while maintaining the computational efficiency of classical MD. This breakthrough opens up new possibilities for conducting accurate calculations in ab initio molecular dynamics (AIMD) on extremely large systems (with over 100 million atoms) or over very long timescales (beyond 1 nanosecond) [333, 355]. By enabling adequate sampling of phase space, these advancements enable more comprehensive investigations across various applications, including (electro-)catalysis, sensors, fabrication, drug interactions, and more.

9.2.2 Experimental materials science

The availability of a vast number of predicted materials with desired properties is highly advantageous for experimentalists. With a large number of possible candidates, the experimentalist can focus on the materials that can be synthesized and tested on available facilities and equipments. Additionally, the automated learning of the fabrication parameters and conditions are on the rise recently [356-359]. The advancement of MD will also enable comprehensive simulations of fabrication process and finds out the best experimental conditions for successful synthesizes of new materials [360, 361]. Furthermore, the analysis of the data, a highly time-consuming and laborious task is being increasingly supported by ML algorithms. The implementation of on-the-fly accurate inference mechanism of experimental data will increase the producitivity and efficiency of fabrication process, enabling experimentalists to determine if the samples have been fabricated successfully and move on to the next attempt quickly.

9.2.3 Coupling of data-driven discovery with traditional techniques

The art of tailoring and creating materials with desired properties − materials engineering, includes techniques such as defect engineering [361-363], doping [44, 362, 364-369], fluorinating [370] or alloy engineering [30-34, 371] or salt engineering [35] by varying composition, strain engineering [11, 16, 26, 36, 170, 372-374] by applying mechanical load (such as hydro-static pressure [36, 375-379] or directional stress), and interfacial engineering [40, 42-46] by choosing different materials for forming interface with novel or exotic properties. These methods have been demonstrated to be very useful for tuning materials properties or creating new materials with advantageous properties. In quantum materials [14, 28, 37, 380], including strongly correlated functional materials and superconducting materials, the charge-spin-orbital engineering plays a crucial role in controlling the quantum behaviour.

The availability of advanced X-ray scattering and electron scattering techniques [14, 28, 191, 381, 382] such as synchrotron radiation and electron microscropy, and advancing nanotehnologies and simulation methods, has led to an increasingly growing amount of high quality experimental or simulated data, available for data-driven discovery and data mining. The integration of data-driven discovery with traditional techniques is expected to play an increasingly important role in materials science research at various length and time scales, ranging from microscropic scale to macroscropic scale.

10 Conclusion

In conclusion, this review briefly introduced basic concepts and history of machine learning, and provided detailed information of coupling between machine learning and materials science in fundamental and technical perspectives. The nuances of machine learning, from the descriptors to the various algorithms, have been discussed in the context of materials science. Besides, this review also covered the tasks or issues in materials scicence that has been tackled with the use of machine learning. We also discussed our vision for the future of materials science as the field matures with the integration of machine learning, which will be drastically different from what we know today.

References

[1]

E. Weinan. Machine learning and computational mathematics. Commun. Comput. Phys., 2020, 28: 1639

[2]

A. Agrawal, A. Choudhary. Perspective: Materials informatics and big data: Realization of the “fourth paradigm” of science in materials science. APL Mater., 2016, 4(5): 053208

[3]

Y. Xu, X. Liu, X. Cao, C. Huang, E. Liu. . Artificial intelligence: A powerful paradigm for scientific research. The Innovation, 2021, 2(4): 100179

[4]

G. Carleo, I. Cirac, K. Cranmer, L. Daudet, M. Schuld, N. Tishby, L. Vogt-Maranto, L. Zdeborová. Machine learning and the physical sciences. Rev. Mod. Phys., 2019, 91: 045002

[5]

G. R. Schleder, A. C. M. Padilha, C. M. Acosta, M. Costa, A. Fazzio. From DFT to machine learning: Recent approaches to materials science – A review. J. Phys.: Mater., 2019, 2(3): 032001

[6]

R. Potyrailo, K. Rajan, K. Stoewe, I. Takeuchi, B. Chisholm, H. Lam. Combinatorial and high-throughput screening of materials libraries: Review of state of the art. ACS Combin. Sci., 2011, 13(6): 579

[7]

K. Alberi, M. B. Nardelli, A. Zakutayev, L. Mitas, S. Curtarolo. . The 2019 materials by design roadmap. J. Phys. D, 2018, 52(1): 013001

[8]

S. Torquato. Optimal design of heterogeneous materials. Ann. Rev. Mater. Res., 2010, 40: 101

[9]

A. A. White. Big data are shaping the future of materials science. MRS Bull., 2013, 38(8): 594

[10]

Z. Fan, H. Q. Wang, J. C. Zheng. Searching for the best thermoelectrics through the optimization of transport distribution function. J. Appl. Phys., 2011, 109(7): 073713

[11]

J. C. Zheng, Y. Zhu. Searching for a higher superconducting transition temperature in strained MgB2. Phys. Rev. B, 2006, 73: 024509

[12]

J. C. Zheng. Asymmetrical transport distribution function: Skewness as a key to enhance thermoelectric performance. Research, 2022, 2022: 9867639

[13]

J. C. Zheng. Recent advances on thermoelectric materials. Front. Phys. China, 2008, 3(3): 269

[14]

J. C. Zheng, A. I. Frenkel, L. Wu, J. Hanson, W. Ku, E. S. Bozin, S. J. L. Billinge, Y. Zhu. Nanoscale disorder and local electronic properties of CaCu3Ti4O12: An integrated study of electron, neutron, and X-ray diffraction, X-ray absorption fine structure, and first-principles calculations. Phys. Rev. B, 2010, 81(14): 144203

[15]

N. Sa, S. S. Chong, H. Q. Wang, J. C. Zheng. Anisotropy engineering of ZnO nanoporous frameworks: A lattice dynamics simulation. Nanomaterials (Basel), 2022, 12(18): 3239

[16]

H.ChengJ. C. Zheng, Ab initio study of anisotropic mechanical and electronic properties of strained carbon-nitride nanosheet with interlayer bonding, Front. Phys. 16(4), 43505 (2021)

[17]

Y. Huang, C. Y. Haw, Z. Zheng, J. Kang, J. C. Zheng, H. Q. Wang. Biosynthesis of zinc oxide nanomaterials from plant extracts and future green prospects: A topical review. Adv. Sustain. Syst., 2021, 5(6): 2000266

[18]

Z.Q. WangH. ChengT.Y. LüH.Q. WangY.P. Feng J.C. Zheng, A super-stretchable boron nanoribbon network, Phys. Chem. Chem. Phys. 20(24), 16510 (2018)

[19]

Y. Li, H. Q. Wang, T. J. Chu, Y. C. Li, X. Li, X. Liao, X. Wang, H. Zhou, J. Kang, K. C. Chang, T. C. Chang, T. M. Tsai, J. C. Zheng. Tuning the nanostructures and optical properties of undoped and N-doped ZnO by supercritical fluid treatment. AIP Adv., 2018, 8(5): 055310

[20]

Y. L. Li, Z. Fan, J. C. Zheng. Enhanced thermoelectric performance in graphitic ZnO (0001) nanofilms. J. Appl. Phys., 2013, 113(8): 083705

[21]

J. He, I. D. Blum, H. Q. Wang, S. N. Girard, J. Doak, L. D. Zhao, J. C. Zheng, G. Casillas, C. Wolverton, M. Jose-Yacaman, D. N. Seidman, M. G. Kanatzidis, V. P. Dravid. Morphology control of nanostructures: Na-doped PbTe–PbS system. Nano Lett., 2012, 12(11): 5979

[22]

Z. Fan, J. Zheng, H. Q. Wang, J. C. Zheng. Enhanced thermoelectric performance in three-dimensional superlattice of topological insulator thin films. Nanoscale Res. Lett., 2012, 7(1): 570

[23]

N. Wei, H. Q. Wang, J. C. Zheng. Nanoparticle manipulation by thermal gradient. Nanoscale Res. Lett., 2012, 7(1): 154

[24]

N. Wei, Z. Fan, L. Q. Xu, Y. P. Zheng, H. Q. Wang, J. C. Zheng. Knitted graphene-nanoribbon sheet: A mechanically robust structure. Nanoscale, 2012, 4(3): 785

[25]

J. Q. He, J. R. Sootsman, L. Q. Xu, S. N. Girard, J. C. Zheng, M. G. Kanatzidis, V. P. Dravid. Anomalous electronic transport in dual-nanostructured lead telluride. J. Am. Chem. Soc., 2011, 133(23): 8786

[26]

N. Wei, L. Xu, H. Q. Wang, J. C. Zheng. Strain engineering of thermal conductivity in graphene sheets and nanoribbons: A demonstration of magic flexibility. Nanotechnology, 2011, 22(10): 105705

[27]

J. He, J. R. Sootsman, S. N. Girard, J. C. Zheng, J. Wen, Y. Zhu, M. G. Kanatzidis, V. P. Dravid. On the origin of increased phonon scattering in nanostructured PbTe-based thermoelectric materials. J. Am. Chem. Soc., 2010, 132(25): 8669

[28]

Y. Zhu, J. C. Zheng, L. Wu, A. I. Frenkel, J. Hanson, P. Northrup, W. Ku. Nanoscale disorder in CaCu3Ti4O12: A new route to the enhanced dielectric response. Phys. Rev. Lett., 2007, 99(3): 037602

[29]

J. C. Zheng, H. Q. Wang, A. T. S. Wee, C. H. A. Huan. Structural and electronic properties of Al nanowires: An ab initio pseudopotential study. Int. J. Nanosci., 2002, 01(02): 159

[30]

J. C. Zheng, H. Q. Wang, A. T. S. Wee, C. H. A. Huan. Possible complete miscibility of (BN)x(C2)1−x alloys. Phys. Rev. B, 2002, 66(9): 092104

[31]

J. C. Zheng, H. Q. Wang, C. H. A. Huan, A. T. S. Wee. The structural and electronic properties of (AlN)x(C2)1−x and (AlN)x(BN)1−x alloys. J. Phys.: Condens. Matter, 2001, 13(22): 5295

[32]

H. Q. Wang, J. C. Zheng, R. Z. Wang, Y. M. Zheng, S. H. Cai. Valence-band offsets of III−V alloy heterojunctions. Surf. Interface Anal., 1999, 28(1): 177

[33]

J. C. Zheng, R. Z. Wang, Y. M. Zheng, S. H. Cai. Valence offsets of three series of alloy heterojunctions. Chin. Phys. Lett., 1997, 14(10): 775

[34]

J. C. Zheng, Y. Zheng, R. Wang. Valence offsets of ternary alloy heterojunctions InxGa1-xAs/InxAl1-xAs. Chin. Sci. Bull., 1996, 41(24): 2050

[35]

L. Liu, T. Wang, L. Sun, T. Song, H. Yan, C. Li, D. Mu, J. Zheng, Y. Dai. Stable cycling of all‐solid‐state lithium metal batteries enabled by salt engineering of PEO‐based polymer electrolytes. Energy Environ. Mater., 2023, (Feb.): e12580

[36]

W. Zhang, F. Y. Du, Y. Dai, J. C. Zheng. Strain engineering of Li+ ion migration in olivine phosphate cathode materials LiMPO4 (M = Mn, Fe, Co) and (LiFePO4)n(LiMnPO4)m superlattices. Phys. Chem. Chem. Phys., 2023, 25(8): 6142

[37]

B. Zhang, L. Wu, J. Zheng, P. Yang, X. Yu, J. Ding, S. M. Heald, R. A. Rosenberg, T. V. Venkatesan, J. Chen, C. J. Sun, Y. Zhu, G. M. Chow. Control of magnetic anisotropy by orbital hybridization with charge transfer in (La0.67Sr0.33MnO3)n/(SrTiO3)n superlattice. NPG Asia Mater., 2018, 10(9): 931

[38]

L. Zhang, T. Y. Lü, H. Q. Wang, W. X. Zhang, S. W. Yang, J. C. Zheng. First principles studies on the thermoelectric properties of (SrO)m(SrTiO3)n superlattice. RSC Adv., 2016, 6(104): 102172

[39]

J. C. Zheng, C. H. A. Huan, A. T. S. Wee, M. A. V. Hove, C. S. Fadley, F. J. Shi, E. Rotenberg, S. R. Barman, J. J. Paggel, K. Horn, P. Ebert, K. Urban. Atomic scale structure of the 5-fold surface of a AlPdMn quasicrystal: A quantitative X-ray photoelectron diffraction analysis. Phys. Rev. B, 2004, 69(13): 134107

[40]

H. Q. Wang, J. Xu, X. Lin, Y. Li, J. Kang, J. C. Zheng. Determination of the embedded electronic states at nanoscale interface via surface-sensitive photoemission spectroscopy. Light Sci. Appl., 2021, 10(1): 153

[41]

M. A. Van Hove, K. Hermann, P. R. Watson. The NIST surface structure database – SSD version 4. Acta Crystallogr. B, 2002, 58(3): 338

[42]

H. Q. Wang, E. Altman, C. Broadbridge, Y. Zhu, V. Henrich. Determination of electronic structure of oxide-oxide interfaces by photoemission spectroscopy. Adv. Mater., 2010, 22: 2950

[43]

H. Zhou, L. Wu, H. Q. Wang, J. C. Zheng, L. Zhang, K. Kisslinger, Y. Li, Z. Wang, H. Cheng, S. Ke, Y. Li, J. Kang, Y. Zhu. Interfaces between hexagonal and cubic oxides and their structure alternatives. Nat. Commun., 2017, 8(1): 1474

[44]

J. D. Steiner, H. Cheng, J. Walsh, Y. Zhang, B. Zydlewski, L. Mu, Z. Xu, M. M. Rahman, H. Sun, F. M. Michel, C. J. Sun, D. Nordlund, W. Luo, J. C. Zheng, H. L. Xin, F. Lin. Targeted surface doping with reversible local environment improves oxygen stability at the electrochemical interfaces of nickel-rich cathode materials. ACS Appl. Mater. Interfaces, 2019, 11(41): 37885

[45]

J. C. Zheng, H. Q. Wang, A. T. S. Wee, C. H. A. Huan. Trends in bonding configuration at SiC/III–V semiconductor interfaces. Appl. Phys. Lett., 2001, 79(11): 1643

[46]

H.Q. WangJ. C. ZhengA.T. S. WeeC.H. A. Huan, Study of electronic properties and bonding configuration at the BN/SiC interface, J. Electron Spectrosc. Relat. Phenom. 114–116, 483 (2001)

[47]

S. Lin, B. Zhang, T. Y. Lü, J. C. Zheng, H. Pan, H. Chen, C. Lin, X. Li, J. Zhou. Inorganic lead-free B-γ-CsSnI 3 perovskite solar cells using diverse electron-transporting materials: A simulation study. ACS Omega, 2021, 6(40): 26689

[48]

F. Y. Du, W. Zhang, H. Q. Wang, J. C. Zheng. Enhancement of thermal rectification by asymmetry engineering of thermal conductivity and geometric structure for the multi-segment thermal rectifier. Chin. Phys. B, 2023, 32(6): 064402

[49]

M. Kulichenko, J. S. Smith, B. Nebgen, Y. W. Li, N. Fedik, A. I. Boldyrev, N. Lubbers, K. Barros, S. Tretiak. The rise of neural networks for materials and chemical dynamics. J. Phys. Chem. Lett., 2021, 12(26): 6227

[50]

W. Sha, Y. Guo, Q. Yuan, S. Tang, X. Zhang, S. Lu, X. Guo, Y. C. Cao, S. Cheng. Artificial intelligence to power the future of materials science and engineering. Adv. Intell. Syst., 2020, 2(4): 1900143

[51]

S.Leonelli, Scientific research and big data, in: The Stanford Encyclopedia of Philosophy, Summer 2020 Ed., edited by E. N. Zalta, Metaphysics Research Lab, Stanford University, 2020

[52]

J. Westermayr, M. Gastegger, K. T. Schütt, R. J. Maurer. Perspective on integrating machine learning into computational chemistry and materials science. J. Chem. Phys., 2021, 154(23): 230903

[53]

D. Morgan, R. Jacobs. Opportunities and challenges for machine learning in materials science. Annu. Rev. Mater. Res., 2020, 50(1): 71

[54]

C.ChenY. ZuoW.YeX.LiZ.Deng S.P. Ong, A critical review of machine learning of energy materials, Adv. Energy Mater. 10(8), 1903242 (2020)

[55]

J. Wei, X. Chu, X. Y. Sun, K. Xu, H. X. Deng, J. Chen, Z. Wei, M. Lei. Machine learning in materials science. InfoMat, 2019, 1(3): 338

[56]

G. Pilania. Machine learning in materials science: From explainable predictions to autonomous design. Comput. Mater. Sci., 2021, 193: 110360

[57]

K. T. Butler, D. W. Davies, H. Cartwright, O. Isayev, A. Walsh. Machine learning for molecular and materials science. Nature, 2018, 559(7715): 547

[58]

F. Oviedo, J. L. Ferres, T. Buonassisi, K. T. Butler. Interpretable and explainable machine learning for materials science and chemistry. Acc. Mater. Res., 2022, 3(6): 597

[59]

J. F. Rodrigues Jr, M. C. F. Florea, D. de Oliveira, D. Diamond, O. N. Oliveira Jr. Big data and machine learning for materials science. Discover Materials, 2021, 1(1): 12

[60]

K. Choudhary, B. DeCost, C. Chen, A. Jain, F. Tavazza, R. Cohn, C. W. Park, A. Choudhary, A. Agrawal, S. J. L. Billinge, E. Holm, S. P. Ong, C. Wolverton. Recent advances and applications of deep learning methods in materials science. npj Comput. Mater., 2022, 8: 59

[61]

L. Samuel. Some studies in machine learning using the game of checkers. IBM J. Res. Develop., 1959, 3(3): 210

[62]

L.BreimanJ. H. FriedmanR.A. OlshenC.J. Stone, Classification and Regression Trees, 1983

[63]

L.G. Valiant, A theory of the learnable, in: STOC ’84 Proceedings of the Sixteenth Annual ACM Symposium on Theory of Computing, pp 436–445, 1984

[64]

T. Mitchell, Machine Learning, New York, USA: McGrawHill, 1997

[65]

S.RoweisZ. Ghahramani, A unifying review of linear gaussian models, Neural Comput. 11(2), 305 (1999)

[66]

J. C. Zheng, J. Y. Chen, J. W. Shuai, S. H. Cai, R. Z. Wang. Storage capacity of the Hopfield neural network. Physica A, 1997, 246(3): 313

[67]

J. W. Shuai, J. C. Zheng, Z. X. Chen, R. T. Liu, B. X. Wu. The three-dimensional rotation neural network. Physica A, 1997, 238): 23

[68]

M.MohriA. RostamizadehA.Talwalkar, Foundations of Machine Learning, 2nd Ed. , Adaptive Computation and Machine Learning. Cambridge, MA: MIT Press, 2018

[69]

A.VaswaniN. ShazeerN.ParmarJ.UszkoreitL.Jones A.N. GomezL. KaiserI.Polosukhin, Attention is all you need, arXiv: 1706.03762 (2017)

[70]

A.WangY. PruksachatkunN.NangiaA.SinghJ.Michael F.HillO. LevyS.R. Bowman, SuperGLUE: A stickier benchmark for general-purpose language understanding systems, arXiv: 1905.00537 (2019)

[71]

D. Erhan, Y. Bengio, A. Courville, P. A. Manzagol, P. Vincent, S. Bengio. Why does unsupervised pre-training help deep learning. J. Mach. Learn. Res., 2010, 11: 625

[72]

Z.FengD. GuoD.TangN.DuanX.Feng M.GongL. ShouB.QinT.LiuD.Jiang M.Zhou, CodeBERT: A pre-trained model for programming and natural languages, arXiv: 2002.08155 (2020)

[73]

H.BaoL. DongF.Wei, BEIT: BERT pre-training of image transformers, arXiv: 2106.08254 (2021)

[74]

K.HakhamaneshiM.NassarM.Phielipp P.AbbeelV. Stojanović, Pretraining graph neural networks for few-shot analog circuit modeling and design, arXiv: 2203.15913 (2022)

[75]

J.LiD.Li C.XiongS. Hoi, BLIP: Bootstrapping language-image pre-training for unified vision-language understanding and generation, arXiv: 2201.12086 (2022)

[76]

K.LuA.Grover P.AbbeelI. Mordatch, Pretrained transformers as universal computation engines, arXiv: 2103.05247 (2021)

[77]

M.ReidY. YamadaS.S. Gu, Can Wikipedia help offline reinforcement learning? arXiv: 2201.12122 (2022)

[78]

C.SunX. QiuY.XuX.Huang, How to fine-tune BERT for text classification? arXiv: 1905.05583 (2019)

[79]

H. Liu, D. Tam, M. Muqeeth, J. Mohta, T. Huang, M. Bansal, C. Raffel. Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning. Advances in Neural Information Processing Systems, 2022, 35: 1950

[80]

J.DevlinM. W. ChangK.LeeK.Toutanova, BERT: Pre-training of deep bidirectional transformers for language understanding, arXiv: 1810.04805 (2018)

[81]

Z.LanM. ChenS.GoodmanK.GimpelP.Sharma R.Soricut, ALBERT: A lite BERT for self-supervised learning of language representations, arXiv: 1909.11942 (2019)

[82]

Y.LiuM. OttN.GoyalJ.DuM.Joshi D.ChenO. LevyM.LewisL.ZettlemoyerV.Stoyanov, ROBERTA: A robustly optimized BERT pretraining approach, arXiv: 1907.11692 (2019)

[83]

J.VigY. Belinkov, Analyzing the structure of attention in a transformer language model, arXiv: 1906.04284 (2019)

[84]

S.ZhangL. Xie, Improving attention mechanism in graph neural networks via cardinality preservation, in: Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, International Joint Conferences on Artificial Intelligence Organization, 2020, page 1395

[85]

Y. Tay, V. Q. Tran, M. Dehghani, J. Ni, D. Bahri, H. Mehta, Z. Qin, K. Hui, Z. Zhao, J. Gupta, T. Schuster, W. W. Cohen, D. Metzler. Transformer memory as a differentiable search index. Advances in Neural Information Processing Systems, 2022, 35: 21831

[86]

C. Raffel, N. Shazeer, A. Roberts, K. Lee, S. Narang, M. Matena, Y. Zhou, W. Li, P. J. Liu. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res., 2020, 21(1): 5485

[87]

T.ShinY. RazeghiR.L. L. IVE.WallaceS.Singh, AutoPrompt: Eliciting knowledge from language models with automatically generated prompts, arXiv: 2010.15980 (2020)

[88]

N.DingS. HuW.ZhaoY.ChenZ.Liu H.-T. ZhengM. Sun, Openprompt: An open-source framework for prompt-learning, arXiv: 2111.01998 (2021)

[89]

S.ZhangS. RollerN.GoyalM.ArtetxeM.Chen S.ChenC. DewanM.DiabX.LiX.V. Lin T.MihaylovM. OttS.ShleiferK.ShusterD.Simig P.S. KouraA. SridharT.WangL.Zettlemoyer, OPT: Open pre-trained transformer language models, arXiv: 2205.01068 (2022)

[90]

O.LieberO. SharirB.LenzY.Shoham, Jurassic-1: Technical Details and Evaluation, AI21 Labs, Tech. Rep., 2021

[91]

T.BrownB. MannN.RyderM.SubbiahJ.D. Kaplan, ., Language models are few-shot learners, in: Advances in Neural Information Processing Systems, edited by H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan, and H. Lin, 33 Curran Associates, Inc., 2020, pp 1877–1901, arXiv: 2005.14165

[92]

A.BapnaI. CaswellJ.KreutzerO.FiratD.van Esch, ., Building machine translation systems for the next thousand languages, arXiv: 2205.03983 (2022)

[93]

T.MikolovK. ChenG.CorradoJ.Dean, Efficient estimation of word representations in vector space, arXiv: 1301.3781 (2013)

[94]

J.PenningtonR.SocherC.Manning, GloVe: Global vectors for word representation, in: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Doha, Qatar: Association for Computational Linguistics, Oct. 2014, pp 1532–1543

[95]

O.MelamudJ. GoldbergerI.Dagan, Context2vec: Learning generic context embedding with bidirectional LSTM, in: Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning. Berlin, Germany: Association for Computational Linguistics, Aug. 2016, pp 51–61

[96]

H.DaiB. DaiL.Song, Discriminative embeddings of latent variable models for structured data, arXiv: 1603.05629 (2016)

[97]

J.YangR. ZhaoM.ZhuD.HallacJ.Sodnik J.Leskovec, Driver2vec: Driver identification from automotive data, arXiv: 2102.05234 (2021)

[98]

S.SchneiderA. BaevskiR.CollobertM.Auli, Wav2vec: Unsupervised pre-training for speech recognition, arXiv: 1904.05862 (2019)

[99]

H.ZhouS. ZhangJ.PengS.ZhangJ.Li H.XiongW. Zhang, Informer: Beyond efficient transformer for long sequence time-series forecasting, in: Proceedings of the AAAI Conference on Artificial Intelligence 35(12), 11106 (2021), arXiv: 2012.07436

[100]

I.BeltagyM. E. PetersA.Cohan, Longformer: The long-document transformer, arXiv: 2004.05150 (2020)

[101]

K.HanY. WangH.ChenX.ChenJ.Guo Z.LiuY. TangA.XiaoC.XuY.Xu Z.YangY. ZhangD.Tao, A survey on vision transformer, IEEE Trans. Pattern Anal. Mach. Intell. 45(1), 87 (2023)

[102]

J.B. AlayracJ.DonahueP.Luc A.MiechI. Barr, ., Flamingo: A visual language model for few-shot learning, Advances in Neural Information Processing Systems 35, 23716 (2022), arXiv: 2204.14198

[103]

J.YuZ.Wang V.VasudevanL. YeungM.SeyedhosseiniY.Wu, COCA: Contrastive captioners are image-text foundation models, arXiv: 2205.01917 (2022)

[104]

X.LiuC. GongL.WuS.ZhangH.Su Q.Liu, Fusedream: Training-free text-to-image generation with improved CLIP+GAN space optimization, arXiv: 2112.01573 (2021)

[105]

A.RadfordJ. W. KimC.HallacyA.RameshG.Goh S.AgarwalG. SastryA.AskellP.MishkinJ.Clark G.KruegerI. Sutskever, Learning transferable visual models from natural language supervision, arXiv: 2103.00020 (2021)

[106]

L.HeQ.Zhou X.LiL.Niu G.ChengX. LiW.LiuY.TongL.Ma L.Zhang, End-to-end video object detection with spatial-temporal transformers, in: Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp 1507–1516, arXiv: 2105.10920

[107]

X.ZhaiX. WangB.MustafaA.SteinerD.Keysers A.KolesnikovL.Beyer, LIT: Zero-shot transfer with locked-image text tuning, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp 18123–18133, arXiv: 2111.07991

[108]

A.TrockmanJ. Z. Kolter, Patches are all you need? arXiv: 2201.09792 (2022)

[109]

A.RameshM. PavlovG.GohS.GrayC.Voss A.RadfordM. ChenI.Sutskever, Zeroshot text-to-image generation, in: International Conference on Machine Learning, 2021, pp 8821–8831, arXiv: 2102.12092

[110]

A.TewariJ. ThiesB.MildenhallP.SrinivasanE.Tretschk Y.WangC. LassnerV.SitzmannR.Martin-BruallaS. LombardiT.SimonC.TheobaltM.Niessner J.T. BarronG. WetzsteinM.ZollhoeferV.Golyanik, Advances in neural rendering, Computer Graphics Forum 41(2), 703 (2022), arXiv: 2111.05849

[111]

B.MildenhallP.P. SrinivasanM.Tancik J.T. BarronR. RamamoorthiR.Ng, NERF: Representing scenes as neural radiance fields for view synthesis, Communications of the ACM 65(1), 99 (2021), arXiv: 2003.08934

[112]

S.ZhengJ. PanC.LuG.Gupta, Pointnorm: Normalization is all you need for point cloud analysis, arXiv: 2207.06324 (2022)

[113]

H.RanJ. LiuC.Wang, Surface representation for point clouds, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp 18942–18952, arXiv: 2205.05740

[114]

X.MaC.Qin H.YouH. RanY.Fu, Rethinking network design and local geometry in point cloud: A simple residual MLP framework, arXiv: 2202.07123 (2022)

[115]

Y.WangY. SunZ.LiuS.E. SarmaM.M. BronsteinJ.M. Solomon, Dynamic graph CNN for learning on point clouds, arXiv: 1801.07829 (2018)

[116]

D. Silver, J. Schrittwieser, K. Simonyan, I. Antonoglou, A. Huang, A. Guez, T. Hubert, L. Baker, M. Lai, A. Bolton, Y. Chen, T. Lillicrap, F. Hui, L. Sifre, G. van den Driessche, T. Graepel, D. Hassabis. Mastering the game of Go without human knowledge. Nature, 2017, 550(7676): 354

[117]

E.ZhaoR. YanJ.LiK.LiJ.Xing, Alphaholdem: High-performance artificial intelligence for heads-up no-limit poker via end-to-end reinforcement learning, in: Proceedings of the AAAI Conference on Artificial Intelligence 36(4), 4689 (2022)

[118]

S.ZouT. XuY.Liang, Finite-sample analysis for SARSA with linear function approximation, arXiv: 1902.02234 (2019)

[119]

C.J. C. H. WatkinsP.Dayan, Q-learning, Machine Learning 8(3), 279 (1992)

[120]

P.AbbeelA. Y. Ng, Apprenticeship learning via inverse reinforcement learning, in Proceedings of the Twenty-First International Conference on Machine Learning, Ser. ICML ’04. New York, NY, USA: Association for Computing Machinery, 2004

[121]

C.FinnP. AbbeelS.Levine, Model-agnostic meta-learning for fast adaptation of deep networks, In International conference on machine learning, 2017, pp 1126–1135, arXiv: 1703.03400

[122]

C.FiftyE. AmidZ.ZhaoT.YuR.Anil C.Finn, Efficiently identifying task groupings for multi-task learning, Advances in Neural Information Processing Systems 34, 27503 (2021), arXiv: 2109.04617

[123]

N.AnandD. Precup, Preferential temporal difference learning, arXiv: 2106.06508 (2021)

[124]

K.ChenR. CaoS.JamesY.LiY.H. Liu P.AbbeelQ. Dou, Sim-to-real 6d object pose estimation via iterative self-training for robotic bin-picking, in: Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XXXIX (pp 533−550). Cham: Springer Nature Switzerland, arXiv: 2204.07049

[125]

V.MnihK. KavukcuogluD.SilverA.GravesI.AntonoglouD.WierstraM.Riedmiller, Playing atari with deep reinforcement learning, arXiv: 1312.5602 (2013)

[126]

T.P. LillicrapJ.J. HuntA.Pritzel N.HeessT. ErezY.TassaD.SilverD.Wierstra, Continuous control with deep reinforcement learning, arXiv: 1509.02971 (2015)

[127]

D.YaratsD. BrandfonbrenerH.Liu M.LaskinP. AbbeelA.LazaricL.Pinto, Don’t change the algorithm, change the data: Exploratory data for offline reinforcement learning, arXiv: 2201.13425 (2022)

[128]

M.AhnA. BrohanN.BrownY.ChebotarO.Cortes, ., Do as I can, not as I say: Grounding language in robotic affordances, in: Conference on Robot Learning, 2023, pp 287–318, arXiv: 2204.01691

[129]

S.JamesP. Abbeel, Coarse-to-fine Q-attention with learned path ranking, arXiv: 2204.01571 (2022)

[130]

C.QiP.Abbeel A.Grover, Imitating, fast and slow: Robust learning from demonstrations via decision-time planning, arXiv: 2204.03597 (2022)

[131]

L.WangX. ZhangK.YangL.YuC.Li L.HongS. ZhangZ.LiY.ZhongJ.Zhu, Memory replay with data compression for continual learning, arXiv: 2202.06592 (2022)

[132]

L.ChenK. LuA.RajeswaranK.LeeA.Grover M.LaskinP. AbbeelA.SrinivasI.Mordatch, Decision transformer: Reinforcement learning via sequence modeling, Advances in Neural Information Processing Systems 34, 15084 (2021), arXiv: 2106.01345

[133]

J.Parker-HolderM.JiangM.Dennis M.SamvelyanJ. FoersterE.GrefenstetteT.Rocktäschel, Evolving curricula with regret-based environment design, in: International Conference on Machine Learning, 2022, pp 17473–17498, arXiv: 2203.01302

[134]

R. Wang, J. Lehman, J. Clune, and K. O. Stanley, Paired open-ended trailblazer (POET): Endlessly generating increasingly complex and diverse learning environments and their solutions, arXiv: 1901.01753 (2019)

[135]

Z.LiL.Li Z.MaP.Zhang J.ChenJ. Zhu, Read: Large-scale neural scene rendering for autonomous driving, arXiv: 2205.05509 (2022)

[136]

W.TangC. J. HoY.Liu, Bandit learning with delayed impact of actions, in: Advances in Neural Information Processing Systems, edited by A. Beygelzimer, Y. Dauphin, P. Liang, and J. W. Vaughan, 2021, arXiv: 1904.01763

[137]

Z.GaoY. HanZ.RenZ.Zhou, Batched multi-armed bandits problem, in: Advances in Neural Information Processing Systems, edited by H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, Curran Associates, Inc., 2019, arXiv: 1904.01763

[138]

Y. Yue, J. Broder, R. Kleinberg, T. Joachims. The k-armed dueling bandits problem. J. Comput. Syst. Sci., 2012, 78(5): 1538

[139]

A.CarpentierA.LazaricM.GhavamzadehR.MunosP.AuerA.Antos, Upperconfidence-bound algorithms for active learning in multi-armed bandits, in: Algorithmic Learning Theory: 22nd International Conference, ALT 2011, Espoo, Finland, October 5−7, 2011. Proceedings 22 (pp 189–203), Springer Berlin Heidelberg, arXiv: 1507.04523

[140]

W.YeS.Liu T.KurutachP. AbbeelY.Gao, Mastering Atari games with limited data, Advances in Neural Information Processing Systems 34, 25476 (2021), arXiv: 2111.00210

[141]

M.SamvelyanT. RashidC.Schroeder de WittG.FarquharN.Nardelli T.G. J. RudnerC.M. HungP.H. S. TorrJ.FoersterS.Whiteson, The StarCraft multi-agent challenge, in: Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems, 2019, arXiv: 1902.04043

[142]

T.WangT. GuptaA.MahajanB.PengS.Whiteson C.Zhang, Rode: Learning roles to decompose multi-agent tasks, arXiv: 2010.01523 (2020)

[143]

O. Vinyals, I. Babuschkin, W. M. Czarnecki, M. Mathieu, A. Dudzik. . Grandmaster level in StarCraft II using multi-agent reinforcement learning. Nature, 2019, 575(7782): 350

[144]

W.DuS.Ding, A survey on multi-agent deep reinforcement learning: From the perspective of challenges and applications, Artif. Intell. Rev. 54(5), 3215 (2021)

[145]

J. Biamonte, P. Wittek, N. Pancotti, P. Rebentrost, N. Wiebe, S. Lloyd. Quantum machine learning. Nature, 2017, 549: 195

[146]

Y.LiuS. ArunachalamK.Temme, A rigorous and robust quantum speed-up in supervised machine learning, Nat. Phys. 17(9), 1013 (2021)

[147]

V. Havlíček, A. D. Córcoles, K. Temme, A. W. Harrow, A. Kandala, J. M. Chow, J. M. Gambetta. Supervised learning with quantum-enhanced feature spaces. Nature, 2019, 567(7747): 209

[148]

S. Moradi, C. Brandner, C. Spielvogel, D. Krajnc, S. Hillmich, R. Wille, W. Drexler, L. Papp. Clinical data classification with noisy intermediate scale quantum computers. Sci. Rep., 2022, 12(1): 1851

[149]

J.ZhengK. HeJ.ZhouY.JinC.M. Li, Combining reinforcement learning with lin-kernighan-helsgaun algorithm for the traveling salesman problem, in: Proceedings of the AAAI Conference on Artificial Intelligence 35(14), 12445 (2021), arXiv: 2012.04461

[150]

Z.LiQ.Chen V.Koltun, Combinatorial optimization with graph convolutional networks and guided tree search, Advances in Neural Information Processing Systems 31, 2018, arXiv: 1810.10659

[151]

M.SundararajanA.TalyQ.Yan, Axiomatic attribution for deep networks, in: International Conference on Machine Learning, 2017, pp 3319–3328, arXiv: 1703.01365

[152]

M.T. RibeiroS.SinghC.Guestrin, Why Should I Trust You? Explaining the predictions of any classifier, in: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2016, pp 1135−1144, arXiv: 1602.04938

[153]

S.LundbergS. I. Lee, A unified approach to interpreting model predictions, arXiv: 1705.07874 (2017)

[154]

J.CrabbeZ. QianF.ImrieM.van der Schaar, Explaining latent representations with a corpus of examples, in: Advances in Neural Information Processing Systems, edited by M. Ranzato, A. Beygelzimer, Y. Dauphin, P. Liang, and J. W. Vaughan, Curran Associates, Inc., 2021, pp 12154–12166, arXiv: 2110.15355

[155]

J.T. SpringenbergA.DosovitskiyT.Brox M.Riedmiller, Striving for simplicity: The all convolutional net, arXiv: 1412.6806 (2014)

[156]

R.YingD. BourgeoisJ.YouM.ZitnikJ.Leskovec, Gnnexplainer: Generating explanations for graph neural networks, arXiv: 1903.03894 (2019)

[157]

H.YuanH. YuJ.WangK.LiS.Ji, On explainability of graph neural networks via subgraph explorations, in: International Conference on Machine Learning, 2021, pp 12241–12252, arXiv: 2102.05152

[158]

Q.HuangM. YamadaY.TianD.SinghD.Yin Y.Chang, GraphLIME: Local interpretable model explanations for graph neural networks, IEEE Transactions on Knowledge and Data Engineering, 35(7), 6968 (2023), arXiv: 2001.06216

[159]

H.YuanH. YuS.GuiS.Ji, Explainability in graph neural networks: A taxonomic survey, IEEE Transactions on Pattern Analysis and Machine Intelligence 45(5), 5782 (2023), arXiv: 2012.15445

[160]

G.KatzC. BarrettD.DillK.JulianM.Kochenderfer, ReLUPlex: An efficient smt solver for verifying deep neural networks, in: Computer Aided Verification: 29th International Conference, CAV 2017, Heidelberg, Germany, July 24−28, 2017, Proceedings, Part I 30, pp 97−117. Springer International Publishing, arXiv: 1702.01135

[161]

S.WangH. ZhangK.XuX.LinS.Jana C.J. HsiehJ. Z. Kolter, Beta-CROWN: Efficient bound propagation with per-neuron split constraints for complete and incomplete neural network verification, Advances in Neural Information Processing Systems 34, 2021, arXiv: 2103.06624

[162]

M.P. OwenA. PankenR.MossL.AlvarezC.Leeper, ACAS Xu: Integrated collision avoidance and detect and avoid capability for UAS, in: IEEE/AIAA 38th Digital Avionics Systems Conference (DASC), 2019

[163]

S.MittalS. Vaishay, A survey of techniques for optimizing deep learning on GPUS, J. Systems Archit. 99, 101635 (2019)

[164]

F. Wang, W. Zhang, S. Lai, M. Hao, Z. Wang. Dynamic GPU energy optimization for machine learning training workloads. IEEE Transactions on Parallel and Distributed Systems, 2022, 33(11): 2943

[165]

R.DavidJ. DukeA.JainV.Janapa ReddiN.Jeffries J.LiN.Kreeger I.NappierM. NatrajT.WangP.WardenR.Rhodes, Tensorflow lite micro: Embedded machine learning for tinyML systems, in: Proceedings of Machine Learning and Systems, edited by A. Smola, A. Dimakis, and I. Stoica, 2021, pp 800–811, arXiv: 2010.08678

[166]

C.TanasescuV. KesarwaniD.Inkpen, Metaphor detection by deep learning and the place of poetic metaphor in digital humanities, in: The Thirty-First International Flairs Conference, 2018

[167]

H.Surden, Machine learning and law, Wash. L. Rev. 89, 87 (2014)

[168]

J.De SpiegeleerD.B. MadanS.Reyners W.Schoutens, Machine learning for quantitative finance: Fast derivative pricing, hedging and fitting, Quantitative Finance 18(10), 1635–1643, 2018

[169]

W. Solano-Alvarez, M. Peet, E. Pickering, J. Jaiswal, A. Bevan, H. Bhadeshia. Synchrotron and neural network analysis of the influence of composition and heat treatment on the rolling contact fatigue of hypereutectoid pearlitic steels. Materials Science and Engineering A, 2017, 707: 259

[170]

J. J. Li, Y. Dai, J. C. Zheng. Strain engineering of ion migration in LiCoO2. Front. Phys., 2022, 17(1): 13503

[171]

H. K. D. H. Bhadeshia. Neural networks and information in materials science. Statistical Analysis and Data Mining, 2009, 1: 296

[172]

Y. Liu, O. C. Esan, Z. Pan, L. An. Machine learning for advanced energy materials. Energy and AI, 2021, 3: 100049

[173]

S. R. Kalidindi. Feature engineering of material structure for AI-based materials knowledge systems. J. Appl. Phys., 2020, 128(4): 041103

[174]

Z.XiangM. FanG.Vázquez Tovar W.TrehengekrnB.J. YoonX.Qian R.ArroyaveX. Qian, Physics-constrained automatic feature engineering for predictive modeling in materials science, in: Proceedings of the AAAI Conference on Artificial Intelligence 35(12), pp 10414–10421 (2021)

[175]

Y. Bengio, A. Courville, P. Vincent. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell., 2013, 35(8): 1798

[176]

P. K. Routh, Y. Liu, N. Marcella, B. Kozinsky, A. I. Frenkel. Latent representation learning for structural characterization of catalyst. J. Phys. Chem. Lett., 2021, 12(8): 2086

[177]

A. Franceschetti, A. Zunger. The inverse band-structure problem of finding an atomic configuration with given electronic properties. Nature, 1999, 402(6757): 6757

[178]

Z. Liu, D. Zhu, L. Raju, W. Cai. Tackling photonic inverse design with machine learning. Adv. Sci., 2021, 8: 2002923

[179]

J. E. Saal, S. Kirklin, M. Aykol, B. Meredig, C. Wolverton. Materials design and discovery with high-throughput density functional theory: The open quantum materials database (OQMD). JOM, 2013, 65(11): 1501

[180]

S. Kirklin, J. E. Saal, B. Meredig, A. Thompson, J. W. Doak, M. Aykol, S. Rühl, C. Wolverton. The open quantum materials database (OQMD): Assessing the accuracy of DFT formation energies. npj Comput. Mater., 2015, 1(1): 15010

[181]

A.JainS. P. OngG.HautierW.ChenW.D. RichardsS.DacekS.CholiaD.Gunter D.SkinnerG. CederK.Persson, The materials project: A materials genome approach to accelerating materials innovation, APL Mater. 1(1), 011002 (2013)

[182]

K. Choudhary, K. F. Garrity, A. C. E. Reid, B. DeCost, A. J. Biacchi. . The joint automated repository for various integrated simulations (JARVIS) for data-driven materials design. npj Comput. Mater., 2020, 6(1): 173

[183]

[184]

[185]

[186]

[187]

[188]

J.ZhouL. ShenM.D. CostaK.A. PerssonS.P. Ong P.HuckY. LuX.MaY.ChenH.Tang Y.P. Feng, 2dmatpedia, an open computational database of two-dimensional materials from top-down and bottom-up approaches, Scientific Data 6, 86, June 2019

[189]

M. Hellenbrandt. The inorganic crystal structure database (ICSD) — Present and future. Crystallography Rev., 2004, 10(1): 17

[190]

S. Gražulis, A. Daškevič, A. Merkys, D. Chateigner, L. Lutterotti, M. Quirós, N. R. Serebryanaya, P. Moeck, R. T. Downs, A. Le Bail. Crystallography Open Database (COD): An open-access collection of crystal structures and platform for world-wide collaboration. Nucleic Acids Research, 2011, 40(D1): D420

[191]

J. C. Zheng, L. Wu, Y. Zhu. Aspherical electron scattering factors and their parameterizations for elements from H to Xe. Journal of Applied Crystallography, 2009, 42: 1043

[192]

J.JumperR. EvansA.PritzelT.GreenM.Figurnov, ., Highly accurate protein structure prediction with alphafold, Nature 596(7873), 583 (2021)

[193]

A.DunnQ. WangA.GanoseD.DoppA.Jain, Benchmarking materials property prediction methods: The matbench test set and automatminer reference algorithm, npj Comput. Mater. 6(1), 138 (2020)

[194]

R. Lin, R. Zhang, C. Wang, X. Q. Yang, H. L. Xin. Temimagenet training library and atomsegnet deep-learning models for high-precision atom segmentation, localization, denoising, and deblurring of atomic-resolution images. Sci. Rep., 2021, 11(1): 5386

[195]

L. Han, H. Cheng, W. Liu, H. Li, P. Ou, R. Lin, H.-T. Wang, C.-W. Pao, A. R. Head, C.-H. Wang, X. Tong, C.-J. Sun, W.-F. Pong, J. Luo, J.-C. Zheng, H. L. Xin. A single-atom library for guided monometallic and concentration-complex multimetallic designs. Nat. Mater., 2022, 21: 681

[196]

D. Mrdjenovich, M. K. Horton, J. H. Montoya, C. M. Legaspi, S. Dwaraknath, V. Tshitoyan, A. Jain, K. A. Persson. Propnet: A knowledge graph for materials science. Matter, 2020, 2(2): 464

[197]

T. S. Lin, C. W. Coley, H. Mochigase, H. K. Beech, W. Wang, Z. Wang, E. Woods, S. L. Craig, J. A. Johnson, J. A. Kalow, K. F. Jensen, B. D. Olsen. Bigsmiles: A structurally-based line notation for describing macromolecules. ACS Cent. Sci., 2019, 5(9): 1523

[198]

M. Krenn, Q. Ai, S. Barthel, N. Carson, A. Frei. . Selfies and the future of molecular string representations. Patterns, 2022, 3(10): 100588

[199]

K. Michel, B. Meredig. Beyond bulk single crystals: A data format for all materials structure–property–processing relationships. MRS Bull., 2016, 41(8): 617

[200]

M.WangD. ZhengZ.YeQ.GanM.Li X.SongJ. ZhouC.MaL.YuY.Gai T.XiaoT. HeG.KarypisJ.LiZ.Zhang, Deep graph library: A graph-centric, highly-performant package for graph neural networks, arXiv: 1909.01315 (2019)

[201]

I.BabuschkinK.BaumliA.Bell S.BhupatirajuJ.Bruce, ., The DeepMind JAX Ecosystem, 2020

[202]

F.Chollet, ., Keras, URL: github.com/fchollet/keras (2015)

[203]

A.PaszkeS. GrossS.ChintalaG.ChananE.Yang Z.DeVitoZ. LinA.DesmaisonL.AntigaA.Lerer, Automatic differentiation in PYTORCH, 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 2017

[204]

M.AbadiA. AgarwalP.BarhamE.BrevdoZ.Chen, ., TensorFlow: Large-scale machine learning on heterogeneous systems, 2015, URL: www.tensorflow.org

[205]

T.WolfL. DebutV.SanhJ.ChaumondC.Delangue A.MoiP. CistacT.RaultR.LoufM.FuntowiczJ.DavisonS.Shleifer)vonPlaten C.MaY.Jernite J.PluC. XuT.L. ScaoS.GuggerM.Drame Q.LhoestA. M. Rush, Huggingface’s transformers: State-of-the-art natural language processing, arXiv: 1910.03771 (2019)

[206]

Openrefine: A free, open source, powerful tool for working with messy data, 2022

[207]

2022

[208]

[209]

2022

[210]

2020

[211]

L. Himanen, M. O. Jäger, E. V. Morooka, F. F. Canova, Y. S. Ranawat, D. Z. Gao, P. Rinke, A. S. Foster. Dscribe: Library of descriptors for machine learning in materials science. Comput. Phys. Commun., 2020, 247: 106949

[212]

W.HuM.Fey M.ZitnikY. DongH.RenB.LiuM.Catasta J.Leskovec, Open graph benchmark: Datasets for machine learning on graphs, Advances in neural information processing systems 33, 22118 (2020), arXiv: 2005.00687

[213]

O. Source, Rdkit: Open-source cheminformatics software, URL: www.rdkit.org, 2022

[214]

D. Grattarola, Spektral, URL: graphneural.network, 2022

[215]

S.LiY.Liu D.ChenY. JiangZ.NieF.Pan, Encoding the atomic structure for machine learning in materials science, Wiley Interdiscip. Rev. Comput. Mol. Sci. 12(1) (2022)

[216]

J. Schmidt, M. R. G. Marques, S. Botti, M. A. L. Marques. Recent advances and applications of machine learning in solid-state materials science. npj Comput. Mater., 2019, 5(1): 83

[217]

M. Rupp, A. Tkatchenko, K. R. Müller, O. A. von Lilienfeld. Fast and accurate modeling of molecular atomization energies with machine learning. Phys. Rev. Lett., 2012, 108(5): 058301

[218]

J. Schrier. Can one hear the shape of a molecule (from its Coulomb matrix eigenvalues). J. Chem. Inf. Model., 2020, 60(8): 3804

[219]

M. McCarthy, K. L. K. Lee. Molecule identification with rotational spectroscopy and probabilistic deep learning. J. Phys. Chem. A, 2020, 124(15): 3002

[220]

F. Faber, A. Lindmaa, O. A. von Lilienfeld, R. Armiento. Crystal structure representations for machine learning models of formation energies. Int. J. Quantum Chem., 2015, 115(16): 1094

[221]

K. T. Schütt, H. Glawe, F. Brockherde, A. Sanna, K. R. Müller, E. K. U. Gross. How to represent crystal structures for machine learning: Towards fast prediction of electronic properties. Phys. Rev. B, 2014, 89(20): 205118

[222]

J. Behler, M. Parrinello. Generalized neural-network representation of high-dimensional potential-energy surfaces. Phys. Rev. Lett., 2007, 98(14): 146401

[223]

J. Behler. Atom-centered symmetry functions for constructing high-dimensional neural network potentials. J. Chem. Phys., 2011, 134(7): 074106

[224]

A. Seko, A. Takahashi, I. Tanaka. Sparse representation for a potential energy surface. Phys. Rev. B, 2014, 90(2): 024101

[225]

M.GasteggerL. SchwiedrzikM.BittermannF.BerzsenyiP.Marquetand, WACSF — Weighted atom-centered symmetry functions as descriptors in machine learning potentials, J. Chem. Phys. 148(24), 241709 (2018)

[226]

A. P. Bartók, R. Kondor, G. Csányi. On representing chemical environments. Phys. Rev. B, 2013, 87(18): 184115

[227]

C. W. Rosenbrock, E. R. Homer, G. Csányi, G. L. W. Hart. Discovering the building blocks of atomic systems using machine learning: Application to grain boundaries. npj Comput. Mater., 2017, 3: 29

[228]

F. M. Paruzzo, A. Hofstetter, F. Musil, S. De, M. Ceriotti, L. Emsley. Chemical shifts in molecular solids by machine learning. Nat. Commun., 2018, 9(1): 4501

[229]

A. S. Rosen, S. M. Iyer, D. Ray, Z. Yao, A. Aspuru-Guzik, L. Gagliardi, J. M. Notestein, R. Q. Snurr. Machine learning the quantum-chemical properties of metal–organic frameworks for accelerated materials discovery. Matter, 2021, 4(5): 1578

[230]

Z. Fan, Z. Zeng, C. Zhang, Y. Wang, K. Song, H. Dong, Y. Chen, T. Ala-Nissila. Neuroevolution machine learning potentials: Combining high accuracy and low cost in atomistic simulations and application to heat transport. Phys. Rev. B, 2021, 104(10): 104309

[231]

Z.MihalićN.Trinajstić, A graph-theoretical approach to structure-property relationships, J. Chem. Educ. 69(9), 701 (1992)

[232]

O. Isayev, C. Oses, C. Toher, E. Gossett, S. Curtarolo, A. Tropsha. Universal fragment descriptors for predicting properties of inorganic crystals. Nat. Commun., 2017, 8(1): 15679

[233]

T. Xie, J. C. Grossman. Crystal graph convolutional neural networks for an accurate and interpretable prediction of material properties. Phys. Rev. Lett., 2018, 120(14): 145301

[234]

K. Xia, G. W. Wei. Persistent homology analysis of protein structure, flexibility and folding. Int. J. Numer. Methods Biomed. Eng., 2014, 30(8): 814

[235]

Z.CangL. MuK.WuK.OpronK.Xia G.W. Wei, A topological approach for protein classification, Comput. Math. Biophys. 3(1) (2015)

[236]

Y. Jiang, D. Chen, X. Chen, T. Li, G.-W. Wei, F. Pan. Topological representations of crystalline compounds for the machine-learning prediction of materials properties. npj Comput. Mater., 2021, 7: 28

[237]

E. Minamitani, T. Shiga, M. Kashiwagi, I. Obayashi. Topological descriptor of thermal conductivity in amorphous Si. J. Chem. Phys., 2022, 156(24): 244502

[238]

M. E. Aktas, E. Akbas, A. E. Fatmaoui. Persistence homology of networks: Methods and applications. Appl. Netw. Sci., 2019, 4(1): 1

[239]

A. Ziletti, D. Kumar, M. Scheffler, L. M. Ghiringhelli. Insightful classification of crystal structures using deep learning. Nat. Commun., 2018, 9(1): 2775

[240]

W. B. Park, J. Chung, J. Jung, K. Sohn, S. P. Singh, M. Pyo, N. Shin, K. S. Sohn. Classification of crystal structure using a convolutional neural network. IUCrJ, 2017, 4(4): 486

[241]

Y. Zhang, X. He, Z. Chen, Q. Bai, A. M. Nolan, C. A. Roberts, D. Banerjee, T. Matsunaga, Y. Mo, C. Ling. Unsupervised discovery of solid-state lithium ion conductors. Nat. Commun., 2019, 10(1): 5260

[242]

S. C. Sieg, C. Suh, T. Schmidt, M. Stukowski, K. Rajan, W. F. Maier. Principal component analysis of catalytic functions in the composition space of heterogeneous catalysts. QSAR Comb. Sci., 2007, 26(4): 528

[243]

R. Tranås, O. M. Løvvik, O. Tomic, K. Berland. Lattice thermal conductivity of half-Heuslers with density functional theory and machine learning: Enhancing predictivity by active sampling with principal component analysis. Comput. Mater. Sci., 2022, 202: 110938

[244]

L. M. Ghiringhelli, J. Vybiral, E. Ahmetcik, R. Ouyang, S. V. Levchenko, C. Draxl, M. Scheffler. Learning physical descriptors for materials science by compressed sensing. New J. Phys., 2017, 19(2): 023017

[245]

R.OuyangS. CurtaroloE.AhmetcikM.SchefflerL.M. Ghiringhelli, SISSO: A compressed-sensing method for identifying the best low-dimensional descriptor in an immensity of offered candidates, Phys. Rev. Mater. 2, 083802 (2018)

[246]

W. C. Lu, X. B. Ji, M. J. Li, L. Liu, B. H. Yue, L. M. Zhang. Using support vector machine for materials design. Adv. Manuf., 2013, 1(2): 151

[247]

Y. Wu, N. Prezhdo, W. Chu. Increasing efficiency of nonadiabatic molecular dynamics by Hamiltonian interpolation with kernel ridge regression. J. Phys. Chem. A, 2021, 125(41): 9191

[248]

T.HastieR. TibshiraniJ.H. Friedman, The elements of statistical learning: Data mining, inference, and prediction, 2nd Ed., in: Springer series in statistics, NY: Springer, 2009

[249]

K.HeX.Zhang S.RenJ. Sun, Deep residual learning for image recognition, in: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp 770–778

[250]

O. T. Unke, S. Chmiela, M. Gastegger, K. T. Schütt, H. E. Sauceda, K.-R. Müller. Spookynet: Learning force fields with electronic degrees of freedom and nonlocal effects. Nat. Commun., 2021, 12: 7273

[251]

C. Zheng, C. Chen, Y. Chen, S. P. Ong. Random forest models for accurate identification of coordination environments from X-ray absorption near-edge structure. Patterns, 2020, 1(2): 100013

[252]

J. J. Kranz, M. Kubillus, R. Ramakrishnan, O. A. von Lilienfeld, M. Elstner. Generalized density-functional tight-binding repulsive potentials from unsupervised machine learning. J. Chem. Theory Comput., 2018, 14(5): 2341

[253]

S. Kim, J. Noh, G. H. Gu, A. Aspuru-Guzik, Y. Jung. Generative adversarial networks for crystal structure prediction. ACS Cent. Sci., 2020, 6(8): 1412

[254]

J. Noh, J. Kim, H. S. Stein, B. Sanchez-Lengeling, J. M. Gregoire, A. Aspuru-Guzik, Y. Jung. Inverse design of solid-state materials via a continuous representation. Matter, 2019, 1(5): 1370

[255]

M.L. HutchinsonE.AntonoB.M. GibbonsS.ParadisoJ.LingB.Meredig, Overcoming data scarcity with transfer learning, arXiv: 1711.05099 (2017)

[256]

R. Chang, Y.-X. Wang, E. Ertekin. Towards overcoming data scarcity in materials science: Unifying models and datasets with a mixture of experts framework. npj Comput. Mater., 2022, 8: 242

[257]

M.A. Nielsen, Neural Networks and Deep Learning, Determination Press, 2015

[258]

A. Akbari, L. Ng, B. Solnik. Drivers of economic and financial integration: A machine learning approach. J. Empir. Finance, 2021, 61: 82

[259]

L. Weng, Flow-based deep generative models, URL: lilianweng.github.io, 2018

[260]

P. Raccuglia, K. C. Elbert, P. D. F. Adler, C. Falk, M. B. Wenny, A. Mollo, M. Zeller, S. A. Friedler, J. Schrier, A. J. Norquist. Machine-learning-assisted materials discovery using failed experiments. Nature, 2016, 533(7601): 73

[261]

A. O. Oliynyk, L. A. Adutwum, J. J. Harynuk, A. Mar. Classifying crystal structures of binary compounds AB through cluster resolution feature selection and support vector machine analysis. Chem. Mater., 2016, 28(18): 6672

[262]

J.TangQ. CaiY.Liu, Prediction of material mechanical properties with support vector machine, in: 2010 International Conference on Machine Vision and Human-machine Interface, Aprl 2010, pp 592–595

[263]

D. C. Elton, Z. Boukouvalas, M. S. Butrico, M. D. Fuge, P. W. Chung. Applying machine learning techniques to predict the properties of energetic materials. Sci. Rep., 2018, 8(1): 9059

[264]

D. Hu, Y. Xie, X. Li, L. Li, Z. Lan. Inclusion of machine learning kernel ridge regression potential energy surfaces in on-the-fly nonadiabatic molecular dynamics simulation. J. Phys. Chem. Lett., 2018, 9(11): 2725

[265]

K. T. Schütt, F. Arbabzadah, S. Chmiela, K. R. Müller, A. Tkatchenko. Quantum-chemical insights from deep tensor neural networks. Nat. Commun., 2017, 8(1): 13890

[266]

D. Jha, L. Ward, A. Paul, W.-K. Liao, A. Choudhary, C. Wolverton, A. Agrawal. Elemnet: Deep learning the chemistry of materials from only elemental composition. Sci. Rep., 2018, 8: 17593

[267]

D.JhaL. WardZ.YangC.WolvertonI.Foster W.K. LiaoA. ChoudharyA.Agrawal, IRNet: A general purpose deep residual regression framework for materials discovery, in: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp 2385–2393, 2019

[268]

O. T. Unke, M. Meuwly. Physnet: A neural network for predicting energies, forces, dipole moments, and partial charges. J. Chem. Theo. Comput., 2019, 15(6): 3678

[269]

Z. Liu, L. Lin, Q. Jia, Z. Cheng, Y. Jiang, Y. Guo, J. Ma. Transferable multilevel attention neural network for accurate prediction of quantum chemistry properties via multitask learning. J. Chem. Inform. Model., 2021, 61(3): 1066

[270]

A. M. Krajewski, J. W. Siegel, J. Xu, Z. K. Liu. Extensible structure-informed prediction of formation energy with improved accuracy and usability employing neural networks. Comput. Mater. Sci., 2022, 208: 111254

[271]

K.T. SchüttP.J. KindermansH.E. SaucedaS.ChmielaA.TkatchenkoK.R. Müller, SchNet: A continuous-filter convolutional neural network for modeling quantum interactions, in Proceedings of the 31st International Conference on Neural Information Processing Systems, in NIPS’17. Red Hook, NY, USA: Curran Associates Inc., Dec. 2017, pp 992–1002

[272]

J. Jung. . Super-resolving material microstructure image via deep learning for microstructure characterization and mechanical behavior analysis. npj Comput. Mater., 2021, 7: 96

[273]

A. A. K. Farizhandi, O. Betancourt, M. Mamivand. Deep learning approach for chemistry and processing history prediction from materials microstructure. Sci. Rep., 2022, 12(1): 4552

[274]

T. Xie, J. C. Grossman. Crystal graph convolutional neural networks for an accurate and interpretable prediction of material properties. Phys. Rev. Lett., 2018, 120(14): 145301

[275]

C. Chen, W. Ye, Y. Zuo, C. Zheng, S. P. Ong. Graph networks as a universal machine learning framework for molecules and crystals. Chem. Mater., 2019, 31(9): 3564

[276]

S. Y. Louis, Y. Zhao, A. Nasiri, X. Wang, Y. Song, F. Liu, J. Hu. Graph convolutional neural networks with global attention for improved materials property prediction. Phys. Chem. Chem. Phys., 2020, 22(32): 18141

[277]

Z. Qiao, M. Welborn, A. Anandkumar, F. R. Manby, T. F. Miller. OrbNet: Deep learning for quantum chemistry using symmetry-adapted atomic-orbital features. J. Chem. Phys., 2020, 153(12): 124111

[278]

J.GasteigerJ. GroßS.Günnemann, Directional message passing for molecular graphs, arXiv: 2003.03123 (2020)

[279]

K. Choudhary, B. DeCost. Atomistic line graph neural network for improved materials property predictions. npj Comput. Mater., 2021, 7(1): 185

[280]

S.ZhangY. LiuL.Xie, Molecular mechanics-driven graph neural network with multiplex graph for molecular structures, arXiv: 2011.07457 (2020)

[281]

M. Ghorbani, S. Prasad, J. B. Klauda, B. R. Brooks. GraphVAMPNet, using graph neural networks and variational approach to Markov processes for dynamical modeling of biomolecules. J. Chem. Phys., 2022, 156(18): 184103

[282]

T. Xie, A. France-Lanord, Y. Wang, Y. Shao-Horn, J. C. Grossman. Graph dynamical networks for unsupervised learning of atomic scale dynamics in materials. Nat. Commun., 2019, 10(1): 2667

[283]

S.BatznerA. MusaelianL.SunM.GeigerJ.P. MailoaM.KornbluthN.MolinariT.E. Smidt B.Kozinsky, E(3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials, Nat. Commun. 13, 2453 (2022)

[284]

K.T. SchüttO.T. UnkeM.Gastegger, Equivariant message passing for the prediction of tensorial properties and molecular spectra, in: International Conference on Machine Learning, pp 9377–9388, 2021

[285]

Y. Jiang, Z. Yang, J. Guo, H. Li, Y. Liu, Y. Guo, M. Li, X. Pu. Coupling complementary strategy to flexible graph neural network for quick discovery of coformer in diverse co-crystal materials. Nat. Commun., 2021, 12(1): 5950

[286]

C. W. Park, C. Wolverton. Developing an improved crystal graph convolutional neural network framework for accelerated materials discovery. Phys. Rev. Mater., 2020, 4(6): 063801

[287]

G. P. Ren, Y. J. Yin, K. J. Wu, Y. He. Force field-inspired molecular representation learning for property prediction. J. Cheminform., 2023, 15(1): 17

[288]

C. Chen, S. P. Ong. AtomSets as a hierarchical transfer learning framework for small and large materials datasets. npj Comput. Mater., 2021, 7: 173

[289]

H. Yamada, C. Liu, S. Wu, Y. Koyama, S. Ju, J. Shiomi, J. Morikawa, R. Yoshida. Predicting materials properties with little data using shotgun transfer learning. ACS Cent. Sci., 2019, 5(10): 1717

[290]

S.FengH. FuH.ZhouY.WuZ.Lu H.Dong, A general and transferable deep learning framework for predicting phase formation in materials, npj Comput. Mater. 7(1), 10 (2021)

[291]

V. Gupta, K. Choudhary, F. Tavazza, C. Campbell, W. K. Liao, A. Choudhary, A. Agrawal. Cross-property deep transfer learning framework for enhanced predictive analytics on small materials data. Nat. Commun., 2021, 12: 6595

[292]

V. Stanev, C. Oses, A. G. Kusne, E. Rodriguez, J. Paglione, S. Curtarolo, I. Takeuchi. Machine learning modeling of superconducting critical temperature. npj Comput. Mater., 2018, 4(1): 29

[293]

D. S. Palmer, N. M. O’Boyle, R. C. Glen, J. B. O. Mitchell. Random forest models to predict aqueous solubility. J. Chem. Inform. Model., 2007, 47(1): 150

[294]

P. Banerjee, R. Preissner. Bittersweetforest: A random forest based binary classifier to predict bitterness and sweetness of chemical compounds. Front. Chem., 2018, 6: 93

[295]

P. Raccuglia, K. C. Elbert, P. D. F. Adler, C. Falk, M. B. Wenny, A. Mollo, M. Zeller, S. A. Friedler, J. Schrier, A. J. Norquist. Machine-learning-assisted materials discovery using failed experiments. Nature, 2016, 533(7601): 7601

[296]

L. Chen, B. Xu, J. Chen, K. Bi, C. Li, S. Lu, G. Hu, Y. Lin. Ensemble-machine-learning-based correlation analysis of internal and band characteristics of thermoelectric materials. J. Mater. Chem. C, 2020, 8(37): 13079

[297]

J. Venderley, K. Mallayya, M. Matty, M. Krogstad, J. Ruff, G. Pleiss, V. Kishore, D. Mandrus, D. Phelan, L. Poudel, A. G. Wilson, K. Weinberger, P. Upreti, M. Norman, S. Rosenkranz, R. Osborn, E. A. Kim. Harnessing interpretable and unsupervised machine learning to address big data from modern X-ray diffraction. Proc. Natl. Acad. Sci. USA, 2022, 119(24): e2109665119

[298]

R. Cohn, E. Holm. Unsupervised machine learning via transfer learning and k-means clustering to classify materials image data. Integr. Mater. Manuf. Innov., 2021, 10(2): 231

[299]

R. E. A. Goodall, A. A. Lee. Predicting materials properties without crystal structure: Deep representation learning from stoichiometry. Nat. Commun., 2020, 11(1): 6280

[300]

K. Muraoka, Y. Sada, D. Miyazaki, W. Chaikittisilp, T. Okubo. Linking synthesis and structure descriptors from a large collection of synthetic records of zeolite materials. Nat. Commun., 2019, 10(1): 4459

[301]

D. Jha, K. Choudhary, F. Tavazza, W. Liao, A. Choudhary, C. Campbell, A. Agrawal. Enhancing materials property prediction by leveraging computational and experimental data using deep transfer learning. Nat. Commun., 2019, 10(1): 5316

[302]

X. Zhong, B. Gallagher, S. Liu, B. Kailkhura, A. Hiszpanski, T. Y.-J. Han. Explainable machine learning in materials science. npj Comput. Mater., 2022, 8: 204

[303]

P. Linardatos, V. Papastefanopoulos, S. Kotsiantis. Explainable AI: A review of machine learning interpretability methods. Entropy (Basel), 2020, 23(1): 18

[304]

W. J. Murdoch, C. Singh, K. Kumbier, R. Abbasi-Asl, B. Yu. Definitions, methods, and applications in interpretable machine learning. Proc. Natl. Acad. Sci. USA, 2019, 116(44): 22071

[305]

R. Kondo, S. Yamakawa, Y. Masuoka, S. Tajima, R. Asahi. Microstructure recognition using convolutional neural networks for prediction of ionic conductivity in ceramics. Acta Mater., 2017, 141: 29

[306]

K. Das, B. Samanta, P. Goyal, S.-C. Lee, S. Bhattacharjee, N. Ganguly. CrysXPP: An explainable property predictor for crystalline materials. npj Comput. Mater., 2022, 8: 43

[307]

A. Y. T. Wang, S. K. Kauwe, R. J. Murdock, T. D. Sparks. Compositionally restricted attention-based network for materials property predictions. npj Comput. Mater., 2021, 7(1): 77

[308]

A. Y. T. Wang, M. S. Mahmoud, M. Czasny, A. Gurlo. CrabNet for explainable deep learning in materials science: bridging the gap between academia and industry. Integr. Mater. Manuf. Innov., 2022, 11(1): 41

[309]

A.ParnamiM. Lee, Learning from few examples: A summary of approaches to few-shot learning, arXiv: 2203.04291 (2023)

[310]

Y. Wang, Q. Yao, J. T. Kwok, L. M. Ni. Generalizing from a few examples: A survey on few-shot learning. ACM Comput. Surv., 2020, 53(3): 63

[311]

Y.WangA. AbuduweiliQ.YaoD.Dou, Property-aware relation networks for few-shot molecular property prediction, arXiv: 2107.07994 (2021)

[312]

Z.Guo, ., Few-shot graph learning for molecular property prediction, in: Proceedings of the Web Conference 2021, in: WWW ’21. New York, USA: Association for Computing Machinery, June 2021, pp 2559–2567

[313]

K. Kaufmann, H. Lane, X. Liu, K. S. Vecchio. Efficient few-shot machine learning for classification of EBSD patterns. Sci. Rep., 2021, 11(1): 8172

[314]

S. Akers. . Rapid and flexible segmentation of electron microscopy data using few-shot machine learning. npj Comput. Mater., 2021, 7: 187

[315]

J. P. Perdew, K. Schmidt. Jacob’s ladder of density functional approximations for the exchange-correlation energy. AIP Conf. Proc., 2001, 577: 1

[316]

S. Dick, M. Fernandez-Serra. Machine learning accurate exchange and correlation functionals of the electronic density. Nat. Commun., 2020, 11(1): 3509

[317]

R. Nagai, R. Akashi, O. Sugino. Completing density functional theory by machine learning hidden messages from molecules. npj Comput. Mater., 2020, 6(1): 43

[318]

J. Kirkpatrick, B. McMorrow, D. H. P. Turban, A. L. Gaunt, J. S. Spencer, A. G. D. G. Matthews, A. Obika, L. Thiry, M. Fortunato, D. Pfau, L. R. Castellanos, S. Petersen, A. W. R. Nelson, P. Kohli, P. Mori-Sánchez, D. Hassabis, A. J. Cohen. Pushing the frontiers of density functionals by solving the fractional electron problem. Science, 2021, 374(6573): 1385

[319]

J. C. Snyder, M. Rupp, K. Hansen, K. R. Müller, K. Burke. Finding density functionals with machine learning. Phys. Rev. Lett., 2012, 108(25): 253002

[320]

X. Lei, A. J. Medford. Design and analysis of machine learning exchange-correlation functionals via rotationally invariant convolutional descriptors. Phys. Rev. Mater., 2019, 3(6): 063801

[321]

Z.FanY. WangP.YingK.SongJ.Wang Y.WangZ. ZengK.XuE.LindgrenJ.M. Rahm A.J. GabourieJ.LiuH.Dong J.WuY.Chen Z.ZhongJ. SunP.ErhartY.SuT.Ala-Nissila, GPUMD: A package for constructing accurate machine-learned potentials and performing highly efficient atomistic simulations, J. Chem. Phys. 157(11), 114801 (2022)

[322]

H. Wang, L. Zhang, J. Han, W. E. DeePMD-kit: A deep learning package for many-body potential energy representation and molecular dynamics. Comput. Phys. Commun., 2018, 228: 178

[323]

Y.ZhangH. WangW.ChenJ.ZengL.Zhang H.WangW. E, DP-GEN: A concurrent learning platform for the generation of reliable deep learning based potential energy models, Comput. Phys. Commun. 253, 107206 (2020)

[324]

P. Pattnaik, S. Raghunathan, T. Kalluri, P. Bhimalapuram, C. V. Jawahar, U. D. Priyakumar. Machine learning for accurate force calculations in molecular dynamics simulations. J. Phys. Chem. A, 2020, 124(34): 6954

[325]

J. Westermayr, P. Marquetand. Machine learning and excited-state molecular dynamics. Mach. Learn.: Sci. Technol., 2020, 1(4): 043001

[326]

G. Fan, A. McSloy, B. Aradi, C. Y. Yam, T. Frauenheim. Obtaining electronic properties of molecules through combining density functional tight binding with machine learning. J. Phys. Chem. Lett., 2022, 13(43): 10132

[327]

Z. Ahmad, T. Xie, C. Maheshwari, J. C. Grossman, V. Viswanathan. Machine learning enabled computational screening of inorganic solid electrolytes for suppression of dendrite formation in lithium metal anodes. ACS Cent. Sci., 2018, 4(8): 996

[328]

S. Gong, S. Wang, T. Zhu, X. Chen, Z. Yang, M. J. Buehler, Y. Shao-Horn, J. C. Grossman. Screening and understanding Li adsorption on two-dimensional metallic materials by learning physics and physics-simplified learning. JACS Au, 2021, 1(11): 1904

[329]

T. Xie, A. France-Lanord, Y. Wang, J. Lopez, M. A. Stolberg, M. Hill, G. M. Leverick, R. Gomez-Bombarelli, J. A. Johnson, Y. Shao-Horn, J. C. Grossman. Accelerating amorphous polymer electrolyte screening by learning to reduce errors in molecular dynamics simulated properties. Nat. Commun., 2022, 13(1): 3415

[330]

K. Gubaev, E. V. Podryabinkin, G. L. Hart, A. V. Shapeev. Accelerating high-throughput searches for new alloys with active learning of interatomic potentials. Comput. Mater. Sci., 2019, 156: 148

[331]

T.XieX. FuO.E. GaneaR.BarzilayT.Jaakkola, Crystal diffusion variational autoencoder for periodic material generation, arXiv: 2110.06197 (2021)

[332]

Y. Dong, D. Li, C. Zhang, C. Wu, H. Wang, M. Xin, J. Cheng, J. Lin. Inverse design of two-dimensional graphene/h-BN hybrids by a regressional and conditional GAN. Carbon, 2020, 169: 9

[333]

Y. Pathak, K. S. Juneja, G. Varma, M. Ehara, U. D. Priyakumar. Deep learning enabled inorganic material generator. Phys. Chem. Chem. Phys., 2020, 22(46): 26935

[334]

Y. Suzuki, H. Hino, T. Hawai, K. Saito, M. Kotsugi, K. Ono. Symmetry prediction and knowledge discovery from X-ray diffraction patterns using an interpretable machine learning approach. Sci. Rep., 2020, 10(1): 21790

[335]

A. A. Enders, N. M. North, C. M. Fensore, J. Velez-Alvarez, H. C. Allen. Functional group identification for FTIR spectra using image-based machine learning models. Anal. Chem., 2021, 93(28): 9711

[336]

B. Huang, Z. Li, J. Li. An artificial intelligence atomic force microscope enabled by machine learning. Nanoscale, 2018, 10(45): 21320

[337]

A. Chandrashekar, P. Belardinelli, M. A. Bessa, U. Staufer, F. Alijani. Quantifying nanoscale forces using machine learning in dynamic atomic force microscopy. Nanoscale Adv., 2022, 4(9): 2134

[338]

S. V. Kalinin, C. Ophus, P. M. Voyles, R. Erni, D. Kepaptsoglou, V. Grillo, A. R. Lupini, M. P. Oxley, E. Schwenker, M. K. Y. Chan, J. Etheridge, X. Li, G. G. D. Han, M. Ziatdinov, N. Shibata, S. J. Pennycook. Machine learning in scanning transmission electron microscopy. Nat. Rev. Methods Primers, 2022, 2(1): 11

[339]

J. Jung. . Super-resolving material microstructure image via deep learning for microstructure characterization and mechanical behavior analysis. npj Comput. Mater., 2021, 7: 96

[340]

L.FloridiM. Chiriatti, GPT-3: Its nature, scope, limits, and consequences, Minds Mach. 30(4), 681 (2020)

[341]

OpenAI, GPT-4 Technical Report, arXiv: 2303.08774 (2023)

[342]

D.M. KatzM. J. BommaritoS.GaoP.Arredondo, GPT-4 passes the bar exam, Rochester, NY, Mar. 15, 2023

[343]

V. Tshitoyan, J. Dagdelen, L. Weston, A. Dunn, Z. Rong, O. Kononova, K. A. Persson, G. Ceder, A. Jain. Unsupervised word embeddings capture latent knowledge from materials science literature. Nature, 2019, 571(7763): 95

[344]

E. A. Olivetti, J. M. Cole, E. Kim, O. Kononova, G. Ceder, T. Y.-J. Han, A. M. Hiszpanski. Data-driven materials research enabled by natural language processing and information extraction. Appl. Phys. Rev., 2020, 7(4): 041317

[345]

P. Shetty, R. Ramprasad. Automated knowledge extraction from polymer literature using natural language processing. iScience, 2021, 24(1): 101922

[346]

A. Davies, P. Veličković, L. Buesing, S. Blackwell, D. Zheng, N. Tomašev, R. Tanburn, P. Battaglia, C. Blundell, A. Juhász, M. Lackenby, G. Williamson, D. Hassabis, P. Kohli. Advancing mathematics by guiding human intuition with AI. Nature, 2021, 600(7887): 70

[347]

G. E. Karniadakis, I. G. Kevrekidis, L. Lu, P. Perdikaris, S. Wang, L. Yang. Physics informed machine learning. Nat. Rev. Phys., 2021, 3(6): 422

[348]

A.GoyalY. Bengio, Inductive biases for deep learning of higher-level cognition, Proc. R. Soc. A 478(2266), 20210068 (2022)

[349]

B.BakerI. AkkayaP.ZhokhovJ.HuizingaJ.Tang A.EcoffetB.HoughtonR.Sampedro J.Clune, Video pretraining (VPT): Learning to act by watching unlabeled online videos, Advances in Neural Information Processing Systems 35, 24639 (2022)

[350]

J.LehmanJ. GordonS.JainK.NdousseC.Yeh K.O. Stanley, Evolution through large models, arXiv: 2206.08896 (2022)

[351]

M.S. Anis, ., Qiskit: An open-source framework for quantum computing, 2021

[352]

C. Wu, F. Wu, L. Lyu, Y. Huang, X. Xie. Communication-efficient federated learning via knowledge distillation. Nat. Commun., 2022, 13: 2032

[353]

H. G. Yu. Neural network iterative diagonalization method to solve eigenvalue problems in quantum mechanics. Phys. Chem. Chem. Phys., 2015, 17(21): 14071

[354]

S. K. Ghosh, D. Ghosh. Machine learning matrix product state Ansatz for strongly correlated systems. J. Chem. Phys., 2023, 158(6): 064108

[355]

P. C. H. Nguyen, J. B. Choi, H. S. Udaykumar, S. Baek. Challenges and opportunities for machine learning in multiscale computational modeling. J. Comput. Inf. Sci. Eng., 2023, 23(6): 060808

[356]

H. Wahab, V. Jain, A. S. Tyrrell, M. A. Seas, L. Kotthoff, P. A. Johnson. Machine-learning-assisted fabrication: Bayesian optimization of laser-induced graphene patterning using in-situ Raman analysis. Carbon, 2020, 167: 609

[357]

A. Tayyebi, A. S. Alshami, X. Yu, E. Kolodka. Can machine learning methods guide gas separation membranes fabrication. J. Membrane Sci. Lett., 2022, 2(2): 100033

[358]

Y. T. Chen, M. Duquesnoy, D. H. S. Tan, J. M. Doux, H. Yang, G. Deysher, P. Ridley, A. A. Franco, Y. S. Meng, Z. Chen. Fabrication of high-quality thin solid-state electrolyte films assisted by machine learning. ACS Energy Lett., 2021, 6(4): 1639

[359]

W. Li, L. Liang, S. Zhao, S. Zhang, J. Xue. Fabrication of nanopores in a graphene sheet with heavy ions: A molecular dynamics study. J. Appl. Phys., 2013, 114(23): 234304

[360]

L. L. Safina, J. A. Baimova. Molecular dynamics simulation of fabrication of Ni-graphene composite: Temperature effect. Micro & Nano Lett., 2020, 15(3): 176

[361]

B. Zhao, C. Shen, H. Yan, J. Xie, X. Liu, Y. Dai, J. Zhang, J. Zheng, L. Wu, Y. Zhu, Y. Jiang. Constructing uniform oxygen defect engineering on primary particle level for high-stability lithium-rich cathode materials. Chem. Eng. J., 2023, 465: 142928

[362]

X. X. Liao, H. Q. Wang, J. C. Zheng. Tuning the structural, electronic, and magnetic properties of strontium titanate through atomic design: A comparison between oxygen vacancies and nitrogen doping. J. Am. Ceram. Soc., 2013, 96(2): 538

[363]

H. Xing, H. Q. Wang, T. Song, C. Li, Y. Dai, G. Fu, J. Kang, J. C. Zheng. Electronic and thermal properties of Ag-doped single crystal zinc oxide via laser-induced technique. Chin. Phys. B, 2023, 32(6): 066107

[364]

L. Wu, J. C. Zheng, J. Zhou, Q. Li, J. Yang, Y. Zhu. Nanostructures and defects in thermoelectric AgPb18SbTe20 single crystal. J. Appl. Phys., 2009, 105(9): 094317

[365]

H. Zeng, M. Wu, H. Q. Wang, J. C. Zheng, J. Kang. Tuning the magnetic and electronic properties of strontium titanate by carbon doping. Front. Phys., 2021, 16(4): 43501

[366]

D. Li, H. Q. Wang, H. Zhou, Y. P. Li, Z. Huang, J. C. Zheng, J. O. Wang, H. Qian, K. Ibrahim, X. Chen, H. Zhan, Y. Zhou, J. Kang. Influence of nitrogen and magnesium doping on the properties of ZnO films. Chin. Phys. B, 2016, 25(7): 076105

[367]

R. Wang, J. C. Zheng. Promising transition metal decorated borophene catalyst for water splitting. RSC Advances, 2023, 13(14): 9678

[368]

J. He, L. D. Zhao, J. C. Zheng, J. W. Doak, H. Wu, H. Q. Wang, Y. Lee, C. Wolverton, M. G. Kanatzidis, V. P. Dravid. Role of sodium doping in lead chalcogenide thermoelectrics. J. Am. Chem. Soc., 2013, 135(12): 4624

[369]

L. D. Cooley, A. J. Zambano, A. R. Moodenbaugh, R. F. Klie, J. C. Zheng, Y. Zhu, Inversion of two-band superconductivity at the critical electron doping of (Mg. Al)B2. Phys. Rev. Lett., 2005, 95(26): 267002

[370]

H. Yan, T. Wang, L. Liu, T. Song, C. Li, L. Sun, L. Wu, J. C. Zheng, Y. Dai. High voltage stable cycling of all-solid-state lithium metal batteries enabled by top-down direct fluorinated poly (ethylene oxide)-based electrolytes. J. Power Sources, 2023, 557: 232559

[371]

J. C. Zheng, C. H. A. Huan, A. T. S. Wee, R. Z. Wang, Y. M. Zheng. Ground-state properties of cubic C-BN solid solutions. J. Phys.: Condens. Matter, 1999, 11(3): 927

[372]

Z. Huang, T. Y. Lü, H. Q. Wang, S. W. Yang, J. C. Zheng, Electronic properties of the group-III nitrides (BN. AlN and GaN) atomic sheets under biaxial strains. Comput. Mater. Sci., 2017, 130: 232

[373]

T. Y. Lü, X. X. Liao, H. Q. Wang, J. C. Zheng, Tuning the indirect–direct band gap transition of SiC. GeC and SnC monolayer in a graphene-like honeycomb structure by strain engineering: A quasiparticle GW study. J. Mater. Chem., 2012, 22(19): 10062

[374]

J. C. Zheng, J. W. Davenport. Ferromagnetism and stability of half-metallic MnSb and MnBi in the strained zinc-blende structure: Predictions from full potential and pseudopotential calculations. Phys. Rev. B, 2004, 69(14): 144415

[375]

L. Xu, H. Q. Wang, J. C. Zheng, Thermoelectric properties of PbTe. SnTe, and GeTe at high pressure: An ab initio study. J. Electron. Mater., 2011, 40(5): 641

[376]

L. Xu, Y. Zheng, J. C. Zheng. Thermoelectric transport properties of PbTe under pressure. Phys. Rev. B, 2010, 82(19): 195102

[377]

J. C. Zheng, Superhard hexagonal transition metal, its carbide. OsC, and OsN. Phys. Rev. B, 2005, 72(5): 052105

[378]

T. Sun, K. Umemoto, Z. Wu, J. C. Zheng, R. M. Wentzcovitch. Lattice dynamics and thermal equation of state of platinum. Phys. Rev. B, 2008, 78(2): 024304

[379]

Z. Wu, R. M. Wentzcovitch, K. Umemoto, B. Li, K. Hirose, J. C. Zheng. Pressure-volume-temperature relations in MgO: An ultrahigh pressure-temperature scale for planetary sciences applications. J. Geophys. Res., 2008, 113(B6): B06204

[380]

S. Deng, L. Wu, H. Cheng, J. C. Zheng, S. Cheng, J. Li, W. Wang, J. Shen, J. Tao, J. Zhu, Y. Zhu. Charge-lattice coupling in hole-doped LuFe2O4+δ: The origin of second-order modulation. Phys. Rev. Lett., 2019, 122(12): 126401

[381]

J. C. Zheng, L. Wu, Y. Zhu, J. W. Davenport. On the sensitivity of electron and X-ray scattering factors to valence charge distribution. J. Appl. Crystall., 2005, 38: 648

[382]

J. C. Zheng, H. Q. Wang. Principles and applications of a comprehensive characterization method combining synchrotron radiation technology, transmission electron microscopy, and density functional theory. Sci. Sin. - Phys. Mech. & Astron., 2021, 51(3): 030007

RIGHTS & PERMISSIONS

The Authors

AI Summary AI Mindmap
PDF (11699KB)

3111

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/