
Citation: | Tongqi Wen, Linfeng Zhang, Han Wang, Weinan E, David J Srolovitz. Deep potentials for materials science[J]. Materials Futures, 2022, 1(2): 022601. DOI: 10.1088/2752-5724/ac681d |
Atomistic simulations are playing an increasingly important role in materials science and changing how research in this heavily-experimental field is conducted [1]. All atomistic simulations in materials modelling require the input of some form of potential energy surface (PES) to describe how atoms interact; from which atomic forces are determined. The most accurate way to obtain the PES, within the Born-Oppenheimer approximation [2], is by solving the Schrdinger equation based on a quantum mechanical treatment of the electronic structure for fixed atomic nuclei coordinates [3]. However, in the most-widely applied electronic structure approach, density function theory (DFT) [4], the scaling is normally
Many approaches have been proposed to strike a balance between accuracy and efficiency in atomistic simulations. One approach to overcoming the low efficiency of widely-used DFT methods was the development of the ONESTEP program [11] in which plane-wave DFT calculations are performed on parallel computers, leading to computational costs that are linear in the number of atoms. DFT calculations have also been implemented on GPU machines [12-15] leading to accelerations by a factor of more than 20 compared to CPU machines [15]. The accuracy of empirical interatomic potentials can be improved, to some extent, by developing analytical functional forms that include many additional parameters to account for additional physical insights. A typical example is the modified embedded atom method (MEAM) potential [10] which extends the embedded atom method (EAM) [9] by considering the angular nature of electron density distributions. However, this tradeoff leads to a decrease in computational speed of the MEAM potentials compared to the simpler EAM potentials; such angular potentials are also more difficult to accelerate than simpler potentials. Nonetheless, the progress from these two approaches are encouraging and have led to many applications. Nevertheless, as materials systems of interest become larger, more complex and demands on predictability more severe, obtaining a better balance between accuracy and efficiency is both an urgent and challenging problem. This challenge requires fundamentally new approaches, rather than tweaking existing techniques. New insights may be garnered from recent progress in other disciplines.
Machine learning (ML) is well known for its surprising successes in, for example, pattern recognition and, as a result many different ML methods were developed in recent years [16-18]. ML potentials may be viewed as versatile descriptions of the PES parameterised using a flexible ML-based analytical form. The flexibility of ML potential increases the representability (accuracy) compared to empirical interatomic potentials and the analytical form significantly improves the efficiency relative to DFT calculations. In this perspective, ML potentials, fit to DFT results, have the potential for achieving DFT accuracy and empirical interatomic potential efficiency. Developing such ML potentials is a challenge.
Since the pioneering work of Blank et al [19] where ML neural network methods for describing the PES were first introduced, a variety of ML potentials have been proposed. Behler and Parrinello (BP) [20] introduced ML neural network potential (NNP) approach in which radial and angular symmetry functions are used as atomic environment descriptors; NNP has found applications in bulk silicon [20], carbon [21], TiO2 [22] and many other materials [23]. BP-NNPs were categorised into four generations [23] for which there are several recent reviews [23-25]. Schtt et al [26, 27] developed SchNet and the SchNetPack package based upon a neural network framework to model the chemical properties and PES of molecular materials. Many other NNPs exist and are popular for different material systems [28-31]. Apart from NNP and SchNet, several other types of ML potentials were introduced; Gaussian approximation potentials [32-35], moment tensor potentials (MTP) [36-38], spectral neighbour analysis potentials [39-41], gradient-domain ML [42, 43], etc [44]. A comprehensive comparison of the major ML potentials in terms of accuracy and efficiency can be found in a recent paper [45].
Deep potentials (DP) [46-49] are of the NNP type that were first published in 2017. This approach has been used extensively for different material systems. The underlying theory has also developed continuously pushing these potentials to increasingly favourable combinations of accuracy and efficiency. Recently, DP has been applied to MD simulations of more than 100 million atoms with ab initio accuracy on a state-of-the-art supercomputer [50]. This is a good example of the power of integrating physical modelling, ML and high-performance computing. In this review, we focus on the application of DP in materials science and discuss a vision for future DPs. The paper is organised as follows. In section 2, we review the basic theory underlying the DP method (section 2.1), demonstrate the steps for developing DPs and their application for atomistic simulation (section 2.2), introduce the extant software for DP development (section 2.3) and the DP Library (section 2.4), and discuss how to make DP more practical for atomistic simulations (section 2.5). Section 3 lists many examples of DP applications in materials science, covering elemental bulk systems (section 3.1), multi-element bulk systems (section 3.2), aqueous systems (section 3.3), and other applications (section 3.4). Next, we discuss the efficiency and accuracy of DPs in practice (section 4.1) and a comparison of the computational speed of DP versus other approaches (section 4.2). We conclude section 5 with an assessment of where DP is going in the near future.
Consider a system of |
(1) |
Training of the ML potential belongs to the category of classical supervised ML. We first obtain the total energy, atomic forces and virial tensors of a number of different system configurations described by the atomic coordinates and use these data as the training labels. Then the ML potential is trained on these labels. We denote a ML potential as |
(3) |
In the ML potential, the extensibility of the total energy is preserved upon decomposition into atomic energies as follows: Ew=∑Ni=1Ewi=∑Ni=1Ewi(Ri), where |
(7) |
The local (neighbourhood) dependence of the atomic energy is an assumption. There are, however, non-local (long-range) interactions, arising mainly from Coulombic interactions within the electron density distribution. For metallic systems, the local dependence assumption is reasonable as a result of shielding effects. For homogeneous materials, the long-range interactions attenuates rapidly with increasing atomic separation such that a sufficiently large rc can always satisfy the local dependence assumption. For materials where long-range interactions dominate, these must be explicitly considered in the model construction. Although there are research reports which introduce long-range interaction in the model construction [28, 30, 51-59], there is still no widely accepted method which can handle this interaction appropriately. In the DP method, we focus on the most common situation, in which
The total energy of the material system is invariant under a set of symmetry operations that may include translations, rotations, and permutations: E(R)=E(UR), where |
(9) |
The atomic energy can thus be written as: Ewi(Ri)=F(D(Ri)), where |
(12) |
Different types of ML potentials employ different descriptors. Some examples are the symmetry functions from BP [20], the Smooth Overlap of Atomic Positions from Bartok et al [34], the SchNet descriptors’ from Schtt et al [26, 27], and the moment tensor from Shapeev [36]. In the following, we introduce the DNN and then focus on the construction and physical meanings of the descriptors employed in DPs.
Deep neural networks (DNN) and kernel methods are the two dominant methods employed in constructing ML potentials. Kernel-based methods usually construct descriptors explicitly, while DNN methods can be used to learn’ descriptors. These are both broad areas, so we refer interested readers to a standard reference [18]. Due to the robust fitting ability of DNN for high-dimensional and nonlinear properties, the DP uses DNNs both as a fitting method for
The DNN is used as the fitting net |
(13) |
Quantification of the representability of a DNN is an active area of research. Barron [62, 63] proved that a neural network with only 1 hidden layer (P = 1) and an arbitrary number of neurons can approximate a class of functions with arbitrary precision. Many researchers have subsequently investigated the approximation ability of DNNs with P>1 [64-67] and explained why the DNN is more successful (and widely used) than wide neural networks (many neurons in one hidden layer). E et al [68, 69] explained and demonstrated why DNNs are suitable for high dimensional problems.
There are two classes of descriptors for DPs, namely non-smooth [ |
(16) |
A schematic of the descriptor for the non-smooth DP is shown for a water molecule in figure 1. The red and blue spheres denote oxygen and hydrogen atoms, (ex, ey, ez) is the local frame of atom i and its neighbour atom j, and
The advantage of non-smooth descriptor is that all neighbour information is preserved. However, due to uncertainty in the choice of neighbour atoms a(i) and b(i), the descriptor is non-smooth. In practice, a(i) is picked as the nearest neighbour and b(i) as the second nearest neighbour. Continuous change in atom positions can thus result in a discontinuous change of the atom number, the local frame, and the local coordinates. In addition, the sorting operation for the other neighbours introduces additional discontinuities in the descriptor and its derivatives (see equation (18)).
The workflow for the DP-Smooth Edition (DP-SE) model [ |
(19) |
Next, the two-body embedding matrix |
(23) |
Finally, from equations (23) and (24), smooth descriptors can be constructed from the two-body and three-body embedding matrix and tensor: D(2,r)i=1Ni∑j(G(2)i)j,D(2,a)i=1N2i(G(2),M<i)T˜Ri(˜Ri)TG(2)i,D(3)i=1N2i(˜Ri(˜Ri)T):G(3)i, where |
(26) |
After illustrating the descriptors and DNN fitting in the DP method, we now outline the main steps we apply to develop a DP: preparing training datasets, the training process, and model validation. We also explain how to apply DPs in atomistic simulations.
The preparation of the training data for a DP has two parts: (1) providing atomic configurations, (i.e. the coordinates of atoms and the cell shape tensor) and (2) labelling (i.e. calculating the energy, atomic forces and virial tensor for the configuration). Labelling is done through a DFT calculation; e.g. well-established DFT packages such as the Vienna Ab initio Simulation Package (VASP) [71, 72], Quantum ESPRESSO (QE) [73], and Atomic-orbital Based Ab-initio Computation at UStc (ABACUS) [74]. Since DPs are trained using a DFT training set, the accuracy of a DP will never exceed that of its training data. The errors in DFT calculations have two main sources. The first is the error introduced by the approximate form of the exchange-correlation functional. This type of error may be reduced by moving up the Jacob’s ladder [75] of increasingly accurate exchange correlation functionals (usually at higher computational cost). It is also possible to use the high order post-Hartree Fock methods like the Mller-Plesset perturbation [76], coupled cluster [77], and configuration interaction [78] for labelling. The other source of error in DFT calculations is numerical, i.e. error introduced by the numerical discretisation of wave functions in real and k spaces and convergence. This type of error can be systematically controlled with the use of more complete basis sets (increasing the energy cutoff in plane-wave approximations), reducing k-space mesh spacing and using stricter stop criteria for self-consistent field iterations. Generally speaking, increasing label quality implies larger computational demands. One usually seeks for a balance between quality and cost.
The training dataset is another critical issue for generating a DP; here, the two main issues are (1) completeness and (2) compactness. By completeness, we mean the training datasets need to sample the relevant configuration space as completely as possible. Increasing the diversity in the members of the training dataset helps increase the transferability of the DP. By compactness, we mean the training data should be the minimal subset of the sampled configurations from which a model with uniform accuracy on the sampled configurations is trained. This is important for minimising DFT computational time. Different approaches have been used to sample the configuration space including MD simulations, genetic algorithms [79, 80], enhanced sampling methods [81], active learning [82] and concurrent learning schemes [83, 84]. Among these, the concurrent learning scheme we refer to as the deep potential generator (DP-GEN) is found to be particularly effective at generating training datasets that satisfy both the completeness and compactness conditions (see section 2.3.3).
We employ the DeePMD-kit package [49] to train a DP. A python package, dpdata (web link 1 in section S2 of the supplementary material (SM) available online at stacks.iop.org/MF/1/022601/mmedia) is used together with the DeePMD-kit to transform the DFT labels from DFT software to the data format accepted by the DeePMD-kit package. DeePMD-kit and dpdata package details are discussed in section 2.3.
There are several important issues in the training process. First, the DP should be trained such as to avoid under-fitting and over-fitting. Under-fitting implies that the DP performs poorly for both the training and validating datasets. Increasing the number of fitting parameters (wider or deeper DNN) or adjusting the ML algorithms can help avoid under-fitting. Under-fitting is easily detected in the training process. On the other hand, over-fitting suggests that a DP is very good at reproducing the training data, but poor at prediction. This may be solved by increasing the training dataset size, decreasing the number of fitting parameters, or using the force and virial as labels (see below). Over-fitting is more difficult to identify; hence, comprehensive testing is necessary.
Second, while training may involve the energy, force and virial labels (normally from DFT calculations), not all of these are necessary. However, it is strongly recommended to use force labels because (a) force has more information: 3N vs. 1 compared to the energy label and (b) training with gradient information helps avoid over-fitting. The schematic plot in figure 3 demonstrates that the gradients of an over-fit model deviate from the ground truth, thus training with forces (gradients of energy with respect to coordinates) and virials (gradients of energy with respect to the cell tensor) as labels helps avoid over-fitting.
Third, there are many tunable hyper-parameters in the training process in the DeePMD-kit, including neural network size, learning rate, prefactors in the loss functions (equation (3)). In practice, we observe that the quality of the DP is not very sensitive to these hyper-parameters and the default settings in the DeePMD-kit routinely provide reasonable accuracy. In some cases, high accuracy for energy is required and fine tuning’ with larger energy prefactors helps. A more detailed discussion of the hyper-parameters in DeePMD-kit is in section 2.3.2.
After the training process is complete, it is advisable to validate the obtained DP to decide whether additional training datasets are required prior to atomistic simulation applications. There are two main approaches to test the DP performance. (1) We can construct a small set of datasets which are unrelated to the training datasets. The DP is used to predict the total energy, atomic forces, and virial tensors of the constructed datasets and the results are compared with DFT results for the same atomic configuration. Experience shows that the root mean square error (RMSE) for energy and forces for a good DP should be smaller than 10 meVatom-1 and 100 meV Å-1. It is not uncommon to achieve 1 meVatom-1 and
Model inference (a commonly used term in the ML community), is the process of providing live data to the ML model to obtain an output. In the context of ML models for the potential energy surface, the inference means taking the configuration (atom coordinates and cell tensor) as input to calculate the energy, force and virial tensor. With the interfaces provided by the DeePMD-kit package, one can easily make inferences about the DP in a Python or C++ programming environment. This makes it possible to use DPs in various molecular simulation tasks such as MD, Monte Carlo, geometric optimisation, by interfacing DeePMD-kit with molecular simulation packages, such as the Large-scale Atomic/Molecular Massively Parallel Simulator (LAMMPS) [85], the atomic simulation environment (ASE) [86], i-PI [87], and GROMACS [88].
There are three main packages for DP development: (1) dpdata for converting the output of DFT software to the data format accepted by DeePMD-kit; (2) DeePMD-kit for DP training and inference; (3) DP-GEN for efficient sampling and labelling of the training data. The DP-GEN package integrates with DeePMD-kit and DFT calculation packages to automatically generate and test DPs.
dpdata is a python package which converts and manipulates training data in different DFT package formats to the compressed format used by DeePMD-kit. A typical dpdata workflow is as follows. (1) Load data from data files. Data files can be written in one of the following package formats: VASP [71, 72], LAMMPS [85], Gaussian [89], SIESTA [90], CP2K [91], QE [73], FHI-aims [92], QUIP (web link 2 in section S2 of the SM), PWmat [14, 15], AMBER [93], GROMACS [88], and ABACUS [74]. (2) Data may be manipulated through operations including replication of atom configurations in a supercell, perturbation of the cell vector and atom positions, and replacement of a number of one type of atoms with others. (3) Output data in a format of one of the aforementioned software. Here we give a simple example of transforming the OUTCAR file from VASP into a training dataset that DeePMD-kit can read:
After execution of these commands, a directory named
The DeePMD-kit package [49] (first made publicly in November 2018) is available on GitHub (web link 3 in section S2 of the SM) under the GNU Lesser General Public License (LGPL). DeePMD-kit has continued to evolve since first becoming available. The DeePMD-kit package interfaces with TensorFlow [94] to make the training and inference codes more efficient and automated. The training is integrated and made available via a command line interface. Model inference is provided through C++ and Python interfaces, which accept atom positions and the cell tensor and return the energy, force and virial for this configuration. These interfaces may be used by MD and molecular simulation packages written in C, C++ or Python, as discussed in section 2.2.4. The DeePMD-kit supports GPU accelerated training and inference. When interfaced with the LAMMPS MD package, parallel and distributed computations accelerated by GPUs are available. (On Summit, one of the most powerful supercomputers in the world, DeePMD-kit has pushed the limit of MD with ab initio accuracy to 100 million atoms and achieved a peak performance of 91 PetaFLOPS in double precision (45.5% of the theoretical peak) [50]). The DeePMD-kit package is now well-developed, been installed over 30000 times (GitHub) and received over 650 GitHub stars.
DeePMD-kit can now be easily installed via off-line packages and package managers such as Conda and Docker. A detailed introduction can be found at web link 4 in section S2 of the SM. The brief release history and key milestones of DeePMD-kit can be found in the SM.
After installation of DeePMD-kit package, we follow the path in section 2.2 to develop a DP. First, dpdata is used to generate the training datasets to be read by DeePMD-kit package (see section 2.3.1). Then, DP training is started with the following command:
After the training process, the architecture and the parameters of the DP can be abstracted from the checkpoint file and saved in the DP file in the Google’s protobuf format
The adjustment of hyper-parameters in ML is never a trivial issue. Fortunately, the efficiency of training a DP is not very sensitive to hyper-parameters and the default settings in DeePMD-kit usually yield a DP of reasonable accuracy. Detailed documentation of all available training parameters can be accessed at web link 6 in section S2 of the SM. On occasion, the hyper-parameters settings may not be satisfactory; e.g. if one wants a DP with higher energy and virial accuracy without a large penalty in force accuracy, fine-tuning of the hyper-parameters may be necessary.
To fine-tune a DP, one may adjust the prefactors of the energy ( |
(29) |
Model fine-tuning restarts training. In the fine-tuning, the model parameter are initialised from the DP trained with the default setting, the start learning rate is decreased to 10-4 while the stop learning rate is kept, and the loss prefactors set to
Another hyper-parameter critical to the accuracy of the DP is the number of training steps. Different settings may be employed for different DP purposes. In the concurrent learning procedure, (for example DP-GEN see section 2.3.3), 0.4-2 million training steps should suffice. If the DP is intended for high-accuracy atomistic simulations, over 8 million training steps are often employed for from-scratch training or fine-tuning.
Finally, the
After a DP is trained and tested by DeePMD-kit, application requires linking with other software. DeePMD-kit provides Python and C++ interfaces for model inference, which is helpful for calculating energy, atomic forces, and virial tensors with input atomic coordinates and the cell tensor. DeePMD-kit also interfaces with ASE [86], LAMMPS [85], i-PI [87], and GROMACS [88] for DP-based atomistic simulations. The detailed use of DP for these can be found at web link 7 in section S2 of the SM.
Here, we explicitly discuss using DPs in LAMMPS. The installation package for DeePMD-kit already incorporates the latest stable version of LAMMPS. Running MD or other molecular simulations with DP in LAMMPS is very simple. The user only needs to add two lines of commands specifying the interatomic interaction with the LAMMPS command
Deep Potential GENerator (DP-GEN) is a software package that implements the concurrent learning framework for generating high quality DPs.
As discussed in section 2.2.1, generating a complete and compact dataset is critical for training high quality DPs. A straightforward way to generate data is to run finite-temperature ab initio MD (AIMD) simulations and use the configurations along the trajectories with labels (energy, forces, and virial tensors) as training datasets. However, this method of data generation is not particularly efficient. First, AIMD simulations are computationally expensive since labels are calculated at each time step by DFT. Second, computationally affordable AIMD simulations are typically very short and do not effectively explore phase space. Many important phenomena, such as phase transformations are difficult to be observed on AIMD time scales. Also configurations generated at successive MD time steps are extremely similar to one the other.
The DP-GEN concurrent learning scheme is designed to overcome these difficulties by generating a much more complete and compact training dataset for DP training. The DP-GEN is an iterative scheme. In each iteration it performs exploration, labelling and training. The training dataset is gradually enriched, and the quality of the DP improves on each iteration. The DP-GEN scheme is deemed converged when all relevant configurations are explored and the DP is uniformly accurate on the explored configurations. The three steps in each DP-GEN iteration are as follows.
(a) Exploration. We start with an ensemble of DPs. One of these DPs is used in the exploration step (figure 4(b)) to efficiently sample the relevant configuration space. The sampler typically performs several DP MD simulations (different initial configurations) for a set of thermodynamics conditions. In principle, sampling can also be performed using Monte Carlo simulations, enhanced sampling MD simulations [98], and any molecular simulation method that explores configuration space.
The ML potential models cannot be extrapolated’ to the configurations that they are not trained on [34], thus the sampling should generate training datasets that explore the relevant configuration space as completely as possible. In the context of the DP-GEN scheme, the configurations in the training dataset are proposed in the exploration step, thus it should explore the relevant configurations as complete as possible by exploiting the high efficiency of the DP itself. The design of the exploration strategy depends on the applications for which the DP is intended. For example, a DP for liquid water properties need not fully explore ice configurations. A DP for chemical reactions may need enhanced sampling techniques to explore reaction pathways usually not available by standard MD simulations [99].
For each explored configuration, the difference between the DP prediction and the ground truth (i.e. error) is |
(31) |
How the model deviation indicates the error is shown in the schematic figure 5. Four deep learning (DL) models (potentials), DL 1 to 4, are trained from the same training data (blue crosses) with random initialisation of the model parameters. At times, DL1 to DL4 all show small training error in the region covered’ by the training data, but the predicting error is large in the region far away from the training data. The standard deviation of the DL model predictions follows exactly the same trend. In the region covered’ by the training data, all models/potentials are trained against the same target, thus their predictions agree with each other, while in the region far from the dataset the different DL potentials (training from different initialisations lead to different potentials) yield different predictions. Therefore, the model deviation is a good indicator of the DL prediction error.
Upper and lower trust levels,
(b) Labelling. In the labelling step (figure 4(c)), a DFT software package is called to calculate the energy, force and virial tensor on the selected configurations. The labelled data are added to the existing training dataset.
(c) Training. In the training step (figure 4(d)), an ensemble of new potentials is trained based on the same training dataset but with different and random initialisation of the potential parameters. Relatively short trainings are performed at this step, because the configurations are selected for labelling by the force criteria, and reasonably accurate DP forces can be obtained with relatively short trainings. Experience shows that 400000 to 2000000 training steps are appropriate. One may also initialise the potential parameters in the ensemble of potentials from the previous iteration, and start the training in the next iteration with a relatively small initial learning rate (e.g. 10-4), and with a bias to the new training data.
(d) Initial datasets for starting the DP-GEN workflow. The DP-GEN workflow introduced above is automatic, but human intervention is required at the beginning when no DP exists. Generally, we generate some initial datasets and train an ensemble of initial DPs to kick off the workflow. In metals and alloys, the generation of initial bulk datasets usually consists of the following steps. (a) Start from a supercell (
Here we provide a practical guide for generating a general-purpose DP for a bulk metal or alloy using the DP-GEN scheme.
(a) Generation of initial datasets. Initial datasets are needed to generate the first ensemble of DPs. Because initial datasets are only used to start DP-GEN, datasets with only several hundred different configurations should suffice. The accuracy of the DPs is not sensitive to the choice of the initial datasets and would be gradually improved as iterations increase assuming that the DP-GEN run parameters are set appropriately.
(b) Exploration. The exploration strategy is not unique and is determined by the user according to the applications of interest. Here, we focus on the general exploration strategy using, again, bulk metals or alloys as our example. DP-based MD simulations at different temperatures and pressures are employed to explore the configurational space. For efficient sampling, we increase the temperature during the exploration. If the highest melting point of the element in an alloy is Tm, we usually divide the temperature range from 50 K to 2Tm into 4 regions. (a) [50 K, (0.1 , 0.2 , 0.3 , 0.4)Tm] (b) [0.5, 0.6, 0.7, 0.8, 0.9]Tm; (c) [1.0, 1.1, 1.2, 1.3, 1.4]Tm; (d) [1.5, 1.6, 1.7, 1.8, 1.9]Tm. For each temperature region, the pressure range is varied over [0.001, 0.01, 0.1, 1, 5, 10, 20, 50] kBar. In this case, there are
After the MD runs, candidate systems are selected based upon the trust levels
System | [ |
Mg [83] | [0.03,0.13] |
Al [83] & Al-Mg [83] | [0.05,0.15] |
Cu [84] | [0.05,0.20] |
Mg-Al-Cu [100] | [0.05,0.20] |
Ti [101] | [0.10,0.25] at |
exploration and [0.15,0.30] elsewhere | |
W [102] | [0.20,0.35] |
Ag-Au [103] | [0.05,0.20] |
water [95] | [0.15,0.25] in first 24 iterations |
[0.18,0.32] in iterations 25 to 32 | |
[0.20,0.35] in iterations 33 to 36 | |
SiC [104] | [0.15,0.30] |
Li10(Ge,Si, or Sn)P2S12 [105] | [0.12,0.25] |
Tm is 1941 K, which is the experimental melting point for Ti. |
(c) Labelling. After candidate structures are selected from the exploration step, users can choose the maximal number of structures sent to DFT labelling in each iteration. On the one hand, if there is a large number of structures for DFT labelling in one iteration, the computational cost for labelling would be high and this dataset of configurations may be redundant, leading to wasted computational resources. On the other hand, if the number of structures for DFT labelling is too small (e.g. one) in one iteration, the datasets after DP-GEN run would be very compact, requiring many iterations, which again would be computationally expensive for training in each iteration. In this case, the maximal number of structures for DFT labelling in one iteration should be set appropriately, we typically set this number to 100.
(d) Training. The prefactor of the virial tensor in the loss function (equation (3)) is usually set to zero in the DP-GEN training loop for two reasons. First, only the model deviation of the forces is used to select candidates in the exploration step; this is irrelevant for the virial tensors. Second, longer training is needed to obtain a DP with good accuracy on both forces and virial tensors. The training step in each iteration of DP-GEN should be small, so we do not initially train on the virial tensors.
(e) DP-GEN convergence. In principle, DP-GEN is considered converged when the model deviation () on each structure is smaller than the trust level
After one or a series of DPs is trained, we employ the Autotest package, to calculate a simple set of properties and/or perform tests for comparison with DFT and/or empirical interatomic potentials (EAM, MEAM, and etc). Because the DP was obtained by fitting DFT calculation results, the degree of agreement between DP and DFT for a series of properties is usually excellent; if not, this comparison provides a guide for further training and parameter settings. Autotest, as part of the DP-GEN package, has standardised the calculation of some critical physical properties and provides a series of reliable benchmark testing to better evaluate the performance of the DPs. As of this writing, Autotest includes the calculation of the following set of properties (additional properties are added continuously): (a) equilibrium structural parameters (relaxation), (b) equation of state (eos), (c) elastic constants (elastic), (d) vacancy formation energy (vacancy), (e) interstitial formation energy (interstitial), and (f) surface formation energy (surface).
Autotest can also use LAMMPS or VASP to reproduce and refine previous calculation results. The current package is mainly targeted for simple crystal structures (metals and alloys), but is extensible for users to implement new features. The detailed manual and framework of Autotest package is available at web link 8 in section S2 of the SM.
The DP-GEN package provides a relatively automatic routine for generating DPs, and the DP Library project is a place for sharing and publishing the DPs and the training datasets. First, DP Library is a place for model sharing, which is similar to other interatomic potential repositories like the NIST (web link 9 in section S2 of the SM) and OpenKIM (web link 10 in section S2 of the SM). Second, DP Library provides an opportunity for data sharing. One can contribute and download DFT datasets used to train a published DP. If a DP needs to be refined or improved, one can first add new data to the downloaded training datasets from DP Library and then perform training using DeePMD-kit. For example, one may develop a DP for A-B alloys based on the training data for element A and element B, instead of generating the datasets by him/herself. The settings (usually settings to use a DFT software) used to generate the dataset are asked to provide, so the new data can be generated in the same way as the downloaded dataset. The shared training data makes the published DPs reproducible and improvable. DP Library can be accessed at web link 11 in section S2 of the SM and an image of the website is shown in figure 6. In the periodic table, the available DP for elements is in black and details about the DP and DFT data can be found by just several clicks. For more details, please refer to the manual of DP Library at web link 12 in section S2 of the SM.
To make DP more practical for applications in different materials systems, efficiency and accuracy must be balanced.
Although DP is very much faster than DFT, it is still much slower than empirical interatomic potentials like EAM and MEAM. This is not surprising given the vast number of parameters in the neural net of the DP (compared to simple empirical potentials such as EAM). Hence, optimisation of the efficiency in use of the DP can greatly enhance the practical usability. Recently, a highly optimised GPU version of DeePMD-kit pushed the limit of MD with ab initio accuracy to 100 million atoms [50, 106] (recognised by the 2020 Gordon Bell Prize). Further optimisations are still possible based on advances in neural networks.
DPs generated by the example exploration protocol (section 2.3.3) may not be accurate for all applications. For example, the Ti DP yields screw
We first introduce DP model compression which can easily accelerate DP by a factor of 4 to 18, based on experience. Next, we discuss DP specialisation and consider the example of Ti, mentioned above.
The most computationally intensive part of using the DP is the evaluation of the embedding net (equation (23) or (24)) and the assembly of the descriptor by equations (26)-(28) [107]. The goal of DP compression is to reduce the computational and memory overhead associated with the embedding net with little loss of accuracy. The evaluation of the DP involved the mapping from a scalar to a vector of dimension M, each dimension of which can be approximated by a piece-wise fifth order interpolating polynomials. The range of the embedding net is first discretised by nodes
The DP compression was benchmarked for Cu, H2O and Al-Cu-Mg ternary alloy DPs [107]. The model inference was accelerated by 9.7, 4.3 and 18.0 times on a CPU and by 9.7, 3.7 and 16.2 times on an Nvidia V100 GPU for Cu, H2O and Al-Cu-Mg, respectively. The maximum number of atoms handled by one GPU also increased from 12, 49 and 5 103 to 129, 246 and 61 103, respectively.
Figures 7(a) and (b) display the speed comparison of the compressed Ti DP [101] with an EAM [108], and an MEAM potential [109] on a CPU and GPU machine. Note that the DP has a larger radius cutoff distance than EAM and MEAM in this case. On CPUs, the compressed DP is 200-300 times slower than EAM potentials and 30-40 times slower than the MEAM potential. On GPUs, the compressed DP is 20-30 times slower than the EAM potential (MEAM is currently not ported to GPU in LAMMPS). All potentials show a linear scaling with the number of atoms. Because of this linearity and speed, the compressed DP can be used to perform large scale MD simulations to investigate a wide range of properties with ab initio accuracy; e.g. defect properties, phase transformations, and transport coefficients.
DP compression is supported in the DeePMD-kit package (releases beyond 2.0.0) and the compressed model can be easily generated using the following command:
Using the optional
The DP generated from the DP-GEN scheme described above may not be sufficiently accurate for some complex phenomena and specialisation of this general-purpose DP may be required. This is not surprising since the exploration strategy may not provide a sufficient sampling of the relevant local structures that are inherent to the complex phenomena of interest. Therefore, some special structures should be added to the training process to better represent the requisite subtle properties. The workflow for specialising DP is shown in figure 8. The initialisation and DP-GEN loop steps were discussed in section 2.3.3.
In the specialisation step, Special’ structures are first created based on the special properties/structures of interest. For example, if the DP elastic constants are not sufficiently accurate, Special’ structures may be created corresponding to a range of different crystal deformations. If the specialised DP is to describe dislocation properties in complex crystal, then Special’ structures that include sheared structures akin to those used in determining the generalised stacking fault energy
Here, we show an example of DP specialisation for the mechanical response of Ti [101]. Figures 9(a) and (b) present a comparison of different empirical interatomic potentials with DFT on the
In the past three years, DPs have been applied in a number of systems in materials science including (1) elemental bulk systems, (2) multi-element bulk systems, (3) aqueous systems, (4) molecular systems and clusters, and (5) surfaces and low-dimensional systems. Table 2 shows a list of the material systems to which DPs have been applied (as of the writing of this paper). We choose several examples from each category to briefly discuss the corresponding DP application and how DP aids materials science research.
System | Reference |
Elemental bulk systems | |
Al | [83, 114-118] |
Mg | [83] |
Cu | [84] |
Ti, W | [101, 102] |
Ag, Au | [103, 119, 120] |
Li | [121] |
Be | [122] |
Ga | [123] |
Sb | [124] |
C | [125] |
Si | [126, 127] |
P | [98] |
Multi-element bulk systems | |
Al-Mg, Al-Cu-Mg | [83, 100, 128, 129] |
Al-Cu, Al-Zn-Mg | [130, 131] |
Al-Cu-Ni | [132] |
Ag-Au | [103, 119] |
Pd-Si, Nb5Si3, Zr77Rh23, Bi2Te3 | [133-136] |
Al90X10 (X = Tb, Cr, or Ce) | [137-140] |
(Pd, Pt) | [141] |
P2Sn5 | [142] |
Silica, silicate | [143-146] |
SiC | [104, 147] |
B4C | [148] |
Molten salt LiF, FLiBe, and chloride | [149-157] |
Li or Na-based battery materials | [105, 158-162] |
TiO2 | [163] |
[164] | |
Ferroelectrics HfO2 | [165] |
Ag2S | [166] |
MoS2 | [167] |
SnSe | [168] |
Zr | [169] |
(Hf0.2Zr0.2Ta0.2Nb0.2Ti0.2)X (X = C or B2) | [170, 171] |
Aqueous systems | |
Water | [95, 172-183] |
Zinc ion in water | [184] |
Water-vapour interface | [185, 186] |
Water-TiO2 interface | [187] |
Ice | [188, 189] |
Molecular systems and clusters | |
Organic molecules | [99, 190-195] |
Metal and alloy clusters | [119, 196] |
Surfaces and low-dimensional systems | |
Metal and alloy surfaces | [103, 119, 129] |
Graphane | [125, 197] |
Monolayer In2Se3 | [198] |
2D Co-Fe-B | [199] |
To date, DPs have been applied to a wide-range of pure systems, including Al, Mg, Cu, Ti, W, Ga, C, Si, as shown in table 2. Al was the first metal system to which DP was applied and a general-purpose DP developed [83]. It accurately reproduces the lattice parameter, elastic constants, vacancy and interstitial formation energies, surface energies, twin and stacking fault energies, melting point, enthalpy of fusion, diffusion coefficient. The general-purpose DP for other metal elements is also accurate for the same properties. For other properties, not included in the training datasets, DP is in better agreement with DFT than MEAM [10] for phonon dispersion relations, equations of state, and the liquid state radial distribution function [83]. Based on this, Wang et al [114] smoothly interpolated the Ziegler-Biersack-Littmark (ZBL) screened nuclear repulsion potential with a DP to obtained a DP-ZBL model for irradiation damage simulations, surpassing the widely adopted ZBL MEAM [200] or EAM [201] potentials. Later, a DP was developed for warm dense Al to simulate ion dynamics near the hydrodynamic limit [115], structural and dynamic properties [116], and electronic and ionic thermal conductivities [117]. A DP was also developed for high temperature and high pressure liquid Al to calculate shear velocity [118].
Dislocation properties play important roles in the plastic response of most structural materials, including Ti and W. The specialised Ti DP accurately depicts the
The DP approach has also been applied to Ag and Au (widely used in catalytic applications). Andolina et al [119] and Wang et al [103] developed DP for Ag and Au that is accurate at lattice parameters, elastic constants, surface formation energies, interstitial and vacancy formation energies, etc. Furthermore, Andolina et al [119] got accurate adsorption energy and diffusion barriers for adatoms on {100}, {110}, and {111} compared to DFT results. Wang et al [103] presented a comprehensive study of the Au {111} surface reconstruction using a DP that yields excellent agreement with DFT results. From another perspective, Chen et al [120] used DP MD to illustrate the dynamics compression process of Au. The developed DP could accurately reproduce the experimentally determined phase boundaries and the short-to-medium range orders are proposed to reduce the Gibbs free energies of the shocked structures. The examples above for Ag and Au validate the applications of DP in both catalytic and shock compression areas.
In addition, DP has also been applied in many other elemental bulk systems and here we only list some of the examples in table 2. For Li, which is an important element for battery, Jiao et al [121] developed DP to reveal self-healing mechanisms in a large Li-metal system. For Ga, Niu et al [123] used DP to construct the phase diagram of liquid Ga,
The applications in elemental bulk systems are among the first applications of DPs in materials science. Their successes for different crystal structures and various classes of properties (mechanical, catalytic, irradiation properties, phase transformation, thermal conductivity, etc) encouraged their extension to multi-element bulk systems and increasingly complex phenomena.
Al-Mg was the first alloy system for which an accurate DP was developed [83]. This DP was used to describe the 28 crystalline Al-Mg alloys structures in the Materials Project (MP) database [202]; include accurate prediction of formation energies, equilibrium volumes, elastic constants, vacancy and interstitial formation energies, and unrelaxed surface energies. Wang [128] applied this Al-Mg DP and validated its reliability for crystal structure prediction by using DP+CALYPSO. This work interfaces the DeePMD-kit package and crystal structure prediction software (e.g. CALYPSO [203], USPEX [80], and Pychemia (web link 13 in section S2 of the SM)). Andolina et al [129] developed a Al-Mg DP based on the original DP to investigate anisotropic surface segregation. Based on the Al-Mg DP, a Al-Cu-Mg ternary DP was developed for the entire compositional space [100]. 2.73 billion alloy configurations were explored in the DP-GEN process. The resulting DP yields more accurate results for energetic, mechanical, and defect properties of 58 crystalline structures as compared with MEAM potentials [10]. The multi-component DP approach can readily be applied to high-entropy alloys for which adequate empirical potentials are difficult to obtain.
Bourgeois et al [130] built an Al-Cu DP to simulate the aggregation of vacancies around embedded
Tang et al [137-140] performed DP MD simulations of a series Al-based alloys; we focus now on Al-Cr quasicrystals [139] as an example and a demonstration of how DPs can be used together with experimental studies. Dendritic growth of metastable quasicrystals were observed in the Al13Cr2 approximant phase (formed from Al90Cr10 thin film) by pulsed laser deposition [139] which is structurally similar to quasicrystal of the Al13Cr2 matrix. The Al-Cr DP was used to simulate the quenching of the Al90Cr10 alloy from 2200 to 700 K at 1011 K s-1. There are three types of 13-atom icosahedra in the approximant Al13Cr2 phase and one icosahedral Al-Cr quasicrystal motif. These 4 icosahedral motifs are similar despite slightly different Cr-Al bond lengths. All 4 types of 13-atom icosahedral motifs were Cr-centred. The icosahedral motif appears in both the quasicrystal and approximant structures, which results in the survival of the 13-atom icosahedron after laser irradiation. This was observed in both simulation and experiment; the success of the DP was attributed to the excellent reproduction of the liquid structure in the DP-based simulations.
A DP was developed to simulate liquid and glassy silica which proved to have satisfactory accuracy based upon a relatively small training dataset [143]. Other DPs were developed to calculate transport properties of silicate in the mantle [144-146]. DPs were also employed in large-scale calculations of thermodynamic, transport, and structural properties in different molten salts [149-157].
Another exciting class of DP applications was for Li and/or Na-based battery materials [105, 158-162]; here we focus on the example of Li10GeP2S12-type superionic conductors [105]. The DP-GEN scheme was used to generate DPs for three solid-state electrolyte materials (Li10GeP2S12, Li10SiP2S12, and Li10SnP2S12) and applied to diffusion over a wide temperature range with 1000 atoms. The predicted diffusion coefficients slightly overestimated the experimental values but were within the experimental uncertainty. These DP-based simulations provided a starting point for large size scale and long time scale MD investigations of solid-state electrolyte materials.
Additional DPs were developed for a wide-range of other multi-element bulk systems, including metal oxide [163-165], metal sulfide [166, 167], thermoelectric SnSe materials [168], metal borides [169, 171], and metal carbide [170] systems.
Since the original DP paper [48], water has been widely studied with DPs. Ko et al [172] applied DPs to perform extensive sampling of thermal and nuclear quantum fluctuations on an accurate PES. In particular, a DP was used to investigate isotopic effects on structural properties of liquid water (H2O and D2O). To understand the experimentally observed isotope effect in the x-ray absorption spectra between liquid H2O and D2O, DP-based, path-integral MD simulations were performed [174]. A more comprehensive study by the same group [177] examined isotope effects on radial distribution functions, O-O-O triplet angular distributions, density and found that the DP-based simulations were consistent with the experimental observations. Recently, Calio et al [183] performed DP-based simulations to interpret experimental observations on the hydrated excess proton in water.
Sommers et al [173] trained a DP to predict the polarisability of liquid water with ab initio accuracy in order to calculate the Raman spectra in long time scale. Gartner et al [175] trained a DP for water to examine the questions of the existence of a liquid-liquid transition in water. Andreani et al [176] combined neutron scattering experiments and DP MD to investigate hydrogen dynamics in supercritical water. The vibrational densities of states observed in DP MD showed coupling between intramolecular vibrational and intermolecular librational and rotational motions. Piaggi et al [178] used a DP to investigate ice nucleation in water, hexagonal ice, and cubic ice and obtained quantitative agreement between DP and experiment (better than the state-of-the-art semi-empirical potentials). A more complete description of the phase equilibrium between different phases of H2O was achieved through constructing the DP phase diagram from low temperature and pressure to 2400 K and 50 GPa [95]. Although both the DP and empirical TIP4P/2005 potential [204] work well at low and intermediate pressures, DP outperforms TIP4P/2005 at higher pressure compared to experimental results, especially at the phase boundaries between ice VIII, VII, and VI (figure 11). Unlike in experiments, the TIP4P/2005 predicts a first-order transition from ice VII to a plastic phase (no such phase transformation occurs using DP). This is a milestone for DP considering the importance of water, the vast range of temperatures and pressures, and the high accuracy required for free energy representation.
Tisi et al [179] calculated the thermal conductivity of water using both DFT (SCAN) and DP. Interestingly, both approaches yield the same conductivity which was 50% higher than the experimental value. Zhang et al [180] applied DP to help improve the exchange functional in DFT from SCAN to SCAN0 for water. Similarly, Torres et al [181] evaluated the errors of DFT-based simulations on structural and dynamical properties due to time- and size-scale limitations by using DP MD. In these two examples, the high efficiency and accuracy of DP provided rapid screening of different properties to feedback into DFT exchange-correlation functional optimisation. Shi et al [182] extended DP to produce accurate molecular multipole moments in the bulk and near interfaces consistent with AIMD simulations. These moments were used to compute the electrostatic potential at the centre of a molecular-sized hydrophobic cavity in water.
The DP approach has also been used in a wide range of aqueous systems. Xu et al [184] developed a DP to perform MD study of zinc ions in liquid water. The experimentally observed zinc-water radial distribution function, as well as the x-ray absorption near edge structure spectrum, was well-reproduced by the DP MD simulation. Recently, the Limmer group [185, 186] applied DP to study liquid-vapour interfaces. They found that the DP yielded accurate interfacial properties by incorporating explicit models of the slowly varying long-ranged interactions and training neural networks only on the short-ranged components [185]. In addition, they trained a DP for solvated N2O5 and bulk ambient water and applied DP MD and importance sampling to study the uptake of N2O5 into an aqueous aerosol [186]. In contrast to the previous understanding that the uptake process occurs within the bulk of an aerosol, interfacial processes dominate the uptake process due to facile hydrolysis at the liquid-vapour interface and competitive re-evaporation. This work not only brings new insights to a long-standing questions, but also extends the application of DPs to the liquid-vapour interface. Other examples of DP applications in aqueous systems include TiO2-water interfaces [187], the ice Ih/XI transition [188], and dynamical states of high-pressure ice VII [189].
Here we briefly list several applications of DPs to other classes of systems including molecular systems, clusters, surfaces, and low-dimensional systems. Jiang et al [190] developed DPs for sulfuric acid-sulfuric acid, dimethylamine-dimethylamine, and sulfuric aciddimethylamine organic molecular systems to investigate the atmospheric aerosol nucleation process. Zeng et al [191] trained a DP based on a dataset for the pyrolysis of n-dodecane and performed a reactive DP MD simulation to reveal the detailed pyrolysis mechanism, in good agreement with experiment. Chen et al [192] used a DP to accurately represent the ground- and excited-state PES of CN2NH. This DP accurately reproduced excited-state topological structures, photo-isomerisation paths, and conical band structure intersections. Yang et al [99] first generated training datasets through active learning with enhanced sampling and then developed a DP to study the urea decomposition process in water. Wang et al [194] presented a data-driven coarse-grained simulation of polymers in solution and validated the accuracy of this method with DPs to construct a coarse-grained potential. Pan et al [195] extended the DP-approach to incorporate external electrostatic potentials in a molecular system; the resultant DP was accurate for energies and forces of representative configurations along the Menshutkin and chorismate mutase reactions pathways.
A study of metal and alloy clusters and surfaces demonstrated a conflict between Al bulk and cluster energies [196]. This indicates that the compromise between properties are, on occasion, necessary and suggest that DPs should be developed for target properties (this is the specialisation discussed above). Andolina et al applied a DP to study the nucleation and growth of seeded core-shell Ag and Au clusters [119] and predict the anisotropic surface segregation for Al-Mg alloys [129]. Wang et al [103] successfully applied DPs in Ag-Au, Au {111} surface reconstruction and segregation of Au on the Ag-Au nanoalloy surfaces.
Achar et al [197] proposed a DP for graphane and showed that it outperforms empirical interatomic potentials for phonon density of states, thermodynamic properties, velocity autocorrelation function, and stress-strain curve up to the yield point. Wu et al [198] developed a DP for In2Se3 monolayers and used it to predict a range of thermodynamic properties of In2Se3 polymorphs and lattice dynamics of ferroelectric In2Se3 with ab initio accuracy. Chen et al [199] applied a DP to simulate the synthesis of amorphous CoFeB during a rapid cooling process. The applications of DPs in low-dimensional systems are in their early stage, but existing evidence suggests that the DP method is promising for the simulation of low-dimensional materials.
The summary of recent applications of the DP method in different systems suggests the wide-applicability and high accuracy of DPs. Nonetheless, it is appropriate to return to the competition between DP accuracy vs the efficiency of DP-based simulations. On the accuracy issue, we reflect on what we have learned from the applications performed to-date and our experience, to address: (1) How good is DP? (2) What have we learned? (3) When can we rely on DPs and when can we not? On the efficiency issue, we summarise how fast DPs are by comparison with other approaches from pair potentials to DFT.
Broadly speaking, DPs are more accurate than other types of empirical interatomic potentials; this was the main reason behind the development of the DP method and its application to a wide range of systems. The improvement in accuracy often leads to qualitatively’ new results. We return to two examples where DPs are qualitatively’ better. The first example is metal Ti [101]. Previous experiments and DFT calculations [113] confirmed that the screw
Another example of qualitative’ improvements made possible by a DP is the calculation of the phase diagram of water [95]. TIP4P/2005 is one of the most accurate empirical water models available today for phase diagram prediction [95]. At high temperature and pressure, TIP4P/2005 predicts a first-order transition from ice VII to a plastic phase, in which the BCC oxygen sublattice coexists with freely rotating molecules. This prediction has not been experimentally confirmed. The DP for water predicts a super-ionic ice VII in that region, in agreement with recent experimental observations [206]. DP also better reproduces the phase boundaries between ice VI, VII, and VIII, at high pressures better than TIP4P/2005. These are two of many examples of where the improved quantitative predictions of DP enable qualitatively’ correct phenomena not accessible through other empirical interatomic potentials.
The accuracy of DP is routinely compared with DFT results especially when benchmarks from widely accepted empirical interatomic potentials do not exist. In most examples (and our own experience), the agreement between energy and force (RMSE) values obtained using a DP and DFT is commonly smaller than 10 meVatom-1 and 100 meV Å-1, respectively. In some systems, such as the water phase diagram, the error is typically 1 meVmolecule-1. This accuracy should meet the requirements for most applications. In particular, almost half of the examples listed in table 2 focus on non-crystalline systems (liquids, amorphous systems; including liquid metallic alloys, and metallic glasses). DP-based predictions are especially accurate for liquid properties (DFT references) on liquid structures, diffusion coefficients, thermal conductivities.
While the flexibility of DPs is desirable for describing complex potential energy surfaces, the ultimate accuracy of DPs are often limited by the accuracy of the DFT training set. Since DFT is itself an approximation, perfect agreement with nature is not expected. Improvements are possible with improved exchange and correlation DFT functionals; climbing the Jacob’s ladder from LDA to GGA to meta-GGA to hybrid-GGA to fully non-local approaches [75]. In practice, the ultimate limit of DP accuracy can often be associated with the DP training sets, the accuracies of which are often associated with the choice of the DFT functionals. This choice is often dictated by the associated computational cost which rises rapidly on climbing the rungs of the Jacob’s ladder.
The two examples, above, demonstrate that DPs can lead to very high accuracy results; comparable to the underlying DFT approaches and higher than the vast majority of empirical interatomic potentials. The examples, above, also show us that DPs can be employed to add new understanding and insight in situations which were previously inaccessible to other computational approaches and experimental observation. The DP approach is also useful in the development of new ab initio methods and pseudopotentials [180, 181]. Finally, the examples showed how DPs can be readily specialised to describe phenomena for which general purpose DPs do not suffice.
At early stages of DP development and for simple applications (e.g. simulation of liquid state structure) where the demands on accuracy are not too high, sampling efficiency is not so critical. For such simple applications, an initial DP based on a small set of AIMD trajectory training sets, often leads to efficient and accurate results. However, sampling efficiency is critical when a general-purpose DP is required or target properties are subtle (e.g. where energy differences between phases are very small or some defect properties). AIMD may not adequately/efficiently sample atomic configurations that represent those associated with properties of interest. This may be addressed by starting with the general purposed DP developed through DP-GEN and tweaking it through identification of property-appropriate and incorporation in the specialisation step, as described above. The specialisation process varies with systems and properties of interest and may be viewed as the art of tweaking DPs based upon physical understanding.
After performing our normal suite of property testing on a DP (e.g. see [101]), experience shows that such DPs yield reliable results in atomistic simulationsespecially compared with DFT calculations that are of insufficient spatial scales or empirical interatomic potentials that are of insufficient accuracy. Critical issues for all ML potentials are representability and transferability. Representability implies the ability of the functional form to accurately reproduce the target properties. Transferability is the ability of a potential to describe the properties which were not included in the training process.
The DP approach usually performs well from a representability perspective; DPs are usually able to provide fits that adequately represent all of the training datasets. In some cases DP failed to distinguish similar configurations (e.g. the configurations along the transition path of the screw dislocation in BCC W); the representability of the DP can be improved by using more expressive descriptors, such as a three-body embedding descriptor [102]. On the other hand, transferability can be non-trivial and subtle for DPs. Transferability can be classified as in-distribution and out-of-distribution transferability. In-distribution or out-of-distribution transferability is the ability of model to interpolate within or extrapolate out of the sampled configuration distribution, respectively. The in-distribution transferability of DPs trained with the DP-GEN scheme is generally quite good, providing reliable and accurate predictions of configurations similar to those in the sampled distribution. However, ML potentials will fail in the out-of-distribution transferability where the explored configuration is far’ from those sampled configurations used in training [34]. A simple example of an out-of-distribution issue is a DP trained using only liquid datasets; such a DP normally shows poor transferability with respect to crystal datasets because there is little overlap between the liquid and solid configuration distributions. For defect properties, we do not know a priori whether defect configurations represent an in- or out-of the distribution with respect to those sampled by DP-GEN. Model deviation (equation (31)) serves as a good indicator of the transferability of a DP (without the need for additional DFT calculations). In the case of transferability failure, the DP can be specialised by adding configurations to the training set that more closely represent the configuration of interest; this is converting out-of-distribution transferability failure to in-distribution transferability agreement.
In the cases when a user is not confident whether a DP is transferable, the DP can be used in conjunction with model deviation after validating against relevant DFT or experimental benchmarks. DPs are, like other empirical potentials subject to the adages that interatomic potentials will work only before they fail’ or will work until they do not’. From this perspective, the replacement of DFT by ML potentials including DP are not completely reliable; DFT will remain the method of choice where very high accuracy property prediction is necessary. The combination of DFT and DP provides a practical strategy for the needs of the materials science community, providing the implicit trade-offs between accuracy and computational efficiency.
Using the previous DP application examples, we see that (1) on CPUs, compressed DPs are faster than DFT by a factor of over 106 and slower than empirical interatomic potentials such as EAM (MEAM) by 100 (10) times; (2) on GPUs, DP compression model is slower than potentials like EAM by, of order 10 times. Of course, the actual efficiency is application dependent (especially for comparison with DFT). Compressed DPs can be faster than the original DPs by a factor of over 10 and consume an order of magnitude less memory. Additional optimisation of the neural networks at the heart of DPs is possible by optimisation of different operators on the computational graph and through hardware changes [107]. Both DPs and empirical potentials show linear scaling with the number of atoms on both CPU and GPU machines. This linear scaling is the enabler of large-scale atomistic simulations. The lower speed of DPs, compared with empirical potentials, is reasonable considering the vast number of parameters in DPs (often at the order of 105). Because DPs are and will continue to be slower than empirical potentials, simpler, empirical potentials will continue to play an important role in materials science. The competition’ between MLs and simpler, empirical potentials drives the continuous improvement of each. The ML potential community continues to focus on improving the potential efficiency (computational speed), while empirical potential development continues to develop new formalism increasing accuracy. The concept of ML potentials guided by the physics in the empirical potentials’ is also an exciting area; e.g. see the recent work of Mishin et al [31, 207].
With increasing need for atomistic simulations with higher accuracy, larger length scales, longer time scales and computational efficiency, ML-based interatomic potentials are rapidly gaining acceptance in the broad materials science community. This is especially true in areas where the phenomena of interest are subtle and those in which the material system is complex. In this review, we examined the DP approach (for ML potentials); summarising the basic theory, how to develop DPs and apply DP-development software and database, how to make DPs more efficient in applications, how to specialise DPs for subtle application, reviewed several DP applications, and discussed DP accuracy and efficiency. After several years of evolution, the DP method is now relatively mature, yet continuing to improve in both accuracy and efficiency within an open-source community framework. We envision the DP method to continue developing in the coming years and the continued expansion of the data base of useable DPs.
Continued development will likely proceed along several avenues. The first is the development of new and more intelligent descriptors for better predictability. We see from the W example [102] that the Peierls barrier (the barrier for dislocation glide) can only be accurately reproduced by expanding the DP descriptor to include three-body embedding. We suspect that such examples will continue to arise as DPs are expanded to include a broader set of applications in different materials. Another issue is related to magnetism; how can magnetic moment degrees of freedom be incorporated into DPs? Empirical EAM and MEAM potentials deal with this issue through a set of assumptions and approximations. More intelligent descriptors also improve the ease of DP training. For example, questions arise for the current hybrid descriptors, with two-body and three-body embeddings about how much to weight three-body embeddings in the hybrid descriptors. Current strategies in this area tend to be based largely on empirical experience; hence, there are opportunities to transform this into a machine-driven process.
An important second area of development will be improvement of the automation of DP training and specialisation. In DeePMD-kit and DP-GEN software, different settings do, on occasion, influence the performance of trained DPs. Although we presented our experience on choosing these settings in section 2.3, ideally this experience should be replaced in future generations of the DeePMD-kit and DP-GEN software. The automatic selection of the trust levels in DP-GEN is already in the testing stage. The specialisation step should also be more automated to reduce user intervention in determining: (1) what types of specialisation datasets are needed? (2) how many specialisation datasets are needed to combine with DP-GEN datasets? (3) when to include specialisation datasets and how to modify DeePMD-kit and DP-GEN settings? More automated training and specialisation schemes would accelerate the development of new DPs for more systems and applications.
A third area of DP future development is further optimisation of the computational speed of DP. Currently, many empirical potentials are faster than DP (at least a factor of 10) which leads potential users to prefer empirical potentials when accuracy demands are not high. In principle, DPs cannot be faster than empirical potentials considering the vast number of parameters involved, but decreasing the computation efficiency loss in using a DP can change the speedaccuracy tradeoff and enable more effective material simulations. As seen above, the greatly improved accuracy of DPs over most empirical potentials opens the door to applications where empirical potentials are simply qualitatively’ incorrect.
A fourth area is the combination of DP and transformer [208]. Transformers were introduced in 2017 and are now commonly used in the fields of natural language processing (NLP) and computer vision. Large pre-trained models based on the transformer architecture (e.g. BERT [209] and GPT-3 [210]) are changing the research paradigm of NLP. Traditionally, models for a specific task are trained from scratch. The Transformer approach is based upon fine-tuning pre-trained models such that new records are being established for specific tasks [211]. For DP, pre-trained models are trained using a large database based upon different elements in the periodic table; learning the embedding mapping from the element type to the abstract space. For example, when we have training datasets for binary systems and want to train a DP model for a ternary system, the traditional step would be using DPGEN to generate datasets for ternary system from scratch. However, with the introduction of pre-trained binary system DP models, the number of additional datasets required to train a ternary system DP model is greatly decreased. From this perspective, pre-trained models can accelerate the training process and decrease the computational time required for generating the requisite training datasets.
Finally, the open-source DP Library (a database for DP including training datasets, training schemes, DP, and testing results) should be greatly expanded to include most of the periodic table, figure 6 and alloys. This will be an on-going effort requiring contributions from the entire user group. Another task is to make the DP Library easier to use. Both of these will be enabled by improvements in the openness of scientific computing community (e.g. appropriate acknowledgment of contributions).
The development of the DP approach (and related ML potential approaches) represents an important milestone for the field of atomistic simulations of materials that rests on advances in ML technology and descriptors of atomic environments. DPs routinely provide high (near DFT) accuracy with reasonable computational efficiency, as compared with empirical potentials. The accuracy and efficiency of DPs open the door to qualitatively’ new applications of atomistic simulation.
T W and D J S gratefully acknowledge the support of the Research Grants Council, Hong Kong SAR, through the Collaborative Research Fund Project No. 8730054. The work of H W is supported by the National Science Foundation of China under Grant Nos. 11871110 and 12122103. The work of W E is supported in part by a gift from iFlytek to Princeton University.
Authors to whom any correspondence should be addressed.
[1] |
Hafner J 2000 Atomic-scale computational materials science Acta Mater. 48 71 DOI: 10.1016/S1359-6454(99)00288-8
|
[2] |
Born M, Oppenheimer R 1927 Zur quantentheorie der molekeln Ann. Phys., Lpz. 389 457 DOI: 10.1002/andp.19273892002
|
[3] |
Dirac P A M 1929 Quantum mechanics of many-electron systems Proc. R. Soc. A 123 714 DOI: 10.1098/rspa.1929.0094
|
[4] |
Kohn W, Sham L J 1965 Self-consistent equations including exchange and correlation effects Phys. Rev. 140 A1133 DOI: 10.1103/PhysRev.140.A1133
|
[5] |
Verlet L 1967 Computer experiments on classical fluids. I. Thermodynamical properties of Lennard-Jones molecules Phys. Rev. 159 98 DOI: 10.1103/PhysRev.159.98
|
[6] |
Zwanzig R W 1954 High-temperature equation of state by a perturbation method. I. Nonpolar gases J. Chem. Phys. 22 1420 DOI: 10.1063/1.1740409
|
[7] |
Tersoff J 1989 Modeling solid-state chemistry: interatomic potentials for multicomponent systems Phys. Rev. B 39 5566 DOI: 10.1103/PhysRevB.39.5566
|
[8] |
Vink R, Barkema G, van der Weg W, Mousseau N 2001 Fitting the Stillinger-Weber potential to amorphous silicon J. Non-Cryst. Solids 282 248 DOI: 10.1016/S0022-3093(01)00342-8
|
[9] |
Daw M S, Baskes M I 1984 Embedded-atom method: derivation and application to impurities, surfaces and other defects in metals Phys. Rev. B 29 6443 DOI: 10.1103/PhysRevB.29.6443
|
[10] |
Baskes M I 1992 Modified embedded-atom potentials for cubic materials and impurities Phys. Rev. B 46 2727 DOI: 10.1103/PhysRevB.46.2727
|
[11] |
Prentice J C A, et al 2020 The ONETEP linear-scaling density functional theory program J. Chem. Phys. 152 174111 DOI: 10.1063/5.0004445
|
[12] |
Hacene M, Anciaux-Sedrakian A, Rozanska X, Klahr D, Guignon T, Fleurat-Lessard P 2012 Accelerating VASP electronic structure calculations using graphic processing units J. Comput. Chem. 33 2581 DOI: 10.1002/jcc.23096
|
[13] |
Hutchinson M, Widom M 2012 VASP on a GPU: application to exact-exchange calculations of the stability of elemental boron Comput. Phys. Commun. 183 1422 DOI: 10.1016/j.cpc.2012.02.017
|
[14] |
Jia W, Cao Z, Wang L, Fu J, Chi X, Gao W, Wang L-W 2013 The analysis of a plane wave pseudopotential density functional theory code on a GPU machine Comput. Phys. Commun. 184 9 DOI: 10.1016/j.cpc.2012.08.002
|
[15] |
Jia W, Fu J, Cao Z, Wang L, Chi X, Gao W, Wang L-W 2013 Fast plane wave density functional theory molecular dynamics calculations on multi-GPU machines J. Comput. Phys. 251 102 DOI: 10.1016/j.jcp.2013.05.005
|
[16] |
Bishop C M 2006 Pattern Recognition and Machine LearningNew YorkSpringer
|
[17] |
Jordan M I, Mitchell T M 2015 Machine learning: trends, perspectives and prospects Science 349 255 DOI: 10.1126/science.aaa8415
|
[18] |
Mahesh B 2020 Machine learning algorithms-a review Int. J. Sci. Res. 9 381 DOI: 10.21275/ART20203995
|
[19] |
Blank T B, Brown S D, Calhoun A W, Doren D J 1995 Neural network models of potential energy surfaces J. Chem. Phys. 103 4129 DOI: 10.1063/1.469597
|
[20] |
Behler J, Parrinello M 2007 Generalized neural-network representation of high-dimensional potential-energy surfaces Phys. Rev. Lett. 98 146401 DOI: 10.1103/PhysRevLett.98.146401
|
[21] |
Khaliullin R Z, Eshet H, Khne T D, Behler J, Parrinello M 2011 Nucleation mechanism for the direct graphite-to-diamond phase transition Nat. Mater. 10 693 DOI: 10.1038/nmat3078
|
[22] |
Artrith N, Urban A 2016 An implementation of artificial neural-network potentials for atomistic materials simulations: performance for TiO2 Comput. Mater. Sci. 114 135 DOI: 10.1016/j.commatsci.2015.11.047
|
[23] |
Behler J 2021 Four generations of high-dimensional neural network potentials Chem. Rev. 121 10037 DOI: 10.1021/acs.chemrev.0c00868
|
[24] |
Behler J 2016 Perspective: machine learning potentials for atomistic simulations J. Chem. Phys. 145 170901 DOI: 10.1063/1.4966192
|
[25] |
Behler J 2017 First principles neural network potentials for reactive simulations of large molecular and condensed systems Angew. Chem., Int. Ed. 56 12828 DOI: 10.1002/anie.201703114
|
[26] |
Schtt K T, Sauceda H E, Kindermans P-J, Tkatchenko A, Mller K-R 2018 SchNeta deep learning architecture for molecules and materials J. Chem. Phys. 148 241722 DOI: 10.1063/1.5019779
|
[27] |
Schtt K T, Kessel P, Gastegger M, Nicoli K A, Tkatchenko A, Mller K-R 2019 SchNetPack: a deep learning toolbox for atomistic systems J. Chem. Theory Comput. 15 448 DOI: 10.1021/acs.jctc.8b00908
|
[28] |
Ghasemi S A, Hofstetter A, Saha S, Goedecker S 2015 Interatomic potentials for ionic systems with density functional accuracy based on charge densities obtained by a neural network Phys. Rev. B 92 045131 DOI: 10.1103/PhysRevB.92.045131
|
[29] |
Hy T S, Trivedi S, Pan H, Anderson B M, Kondor R 2018 Predicting molecular properties with covariant compositional networks J. Chem. Phys. 148 241745 DOI: 10.1063/1.5024797
|
[30] |
Unke O T, Meuwly M 2019 Physnet: a neural network for predicting energies, forces, dipole moments and partial charges J. Chem. Theory Comput. 15 3678 DOI: 10.1021/acs.jctc.9b00181
|
[31] |
Purja Pun G P, Batra R, Ramprasad R, Mishin Y 2019 Physically informed artificial neural networks for atomistic modeling of materials Nat. Commun. 10 2339 DOI: 10.1038/s41467-019-10343-5
|
[32] |
Bartk A P, Payne M C, Kondor R, Csnyi G 2010 Gaussian approximation potentials: the accuracy of quantum mechanics, without the electrons Phys. Rev. Lett. 104 136403 DOI: 10.1103/PhysRevLett.104.136403
|
[33] |
Dragoni D, Daff T D, Csnyi G, Marzari N 2018 Achieving DFT accuracy with a machine-learning interatomic potential: thermomechanics and defects in bcc ferromagnetic iron Phys. Rev. Mater. 2 013808 DOI: 10.1103/PhysRevMaterials.2.013808
|
[34] |
Bartk A P, Kermode J, Bernstein N, Csnyi G 2018 Machine learning a general-purpose interatomic potential for silicon Phys. Rev. X 8 041048 DOI: 10.1103/PhysRevX.8.041048
|
[35] |
Deringer V L, Bartk A P, Bernstein N, Wilkins D M, Ceriotti M, Csnyi G 2021 Gaussian process regression for materials and molecules Chem. Rev. 121 10073 DOI: 10.1021/acs.chemrev.1c00022
|
[36] |
Shapeev A V 2016 Moment tensor potentials: a class of systematically improvable interatomic potentials Multiscale Model. Simul. 14 1153 DOI: 10.1137/15M1054183
|
[37] |
Podryabinkin E V, Shapeev A V 2017 Active learning of linearly parametrized interatomic potentials Comput. Mater. Sci. 140 171 DOI: 10.1016/j.commatsci.2017.08.031
|
[38] |
Podryabinkin E V, Tikhonov E V, Shapeev A V, Oganov A R 2019 Accelerating crystal structure prediction by machine-learning interatomic potentials with active learning Phys. Rev. B 99 064114 DOI: 10.1103/PhysRevB.99.064114
|
[39] |
Chen C, Deng Z, Tran R, Tang H, Chu I-H, Ong S P 2017 Accurate force field for molybdenum by machine learning large materials data Phys. Rev. Mater. 1 043603 DOI: 10.1103/PhysRevMaterials.1.043603
|
[40] |
Li X-G, Hu C, Chen C, Deng Z, Luo J, Ong S P 2018 Quantum-accurate spectral neighbor analysis potential models for Ni-Mo binary alloys and fcc metals Phys. Rev. B 98 094104 DOI: 10.1103/PhysRevB.98.094104
|
[41] |
Deng Z, Chen C, Li X-G, Ong S P 2019 An electrostatic spectral neighbor analysis potential for lithium nitride npj Comput. Mater. 5 75 DOI: 10.1038/s41524-019-0212-1
|
[42] |
Sauceda H E, Chmiela S, Poltavsky I, Mller K-R, Tkatchenko A 2019 Molecular force fields with gradient-domain machine learning: construction and application to dynamics of small molecules with coupled cluster forces J. Chem. Phys. 150 114102 DOI: 10.1063/1.5078687
|
[43] |
Chmiela S, Sauceda H E, Poltavsky I, Mller K-R, Tkatchenko A 2019 sGDML: constructing accurate and data efficient molecular force fields using machine learning Comput. Phys. Commun. 240 38 DOI: 10.1016/j.cpc.2019.02.007
|
[44] |
Unke O T, Chmiela S, Sauceda H E, Gastegger M, Poltavsky I, Schtt K T, Tkatchenko A, Mller K-R 2021 Machine learning force fields Chem. Rev. 121 10142 DOI: 10.1021/acs.chemrev.0c01111
|
[45] |
Zuo Y, et al 2020 Performance and cost assessment of machine learning interatomic potentials J. Phys. Chem. A 124 731 DOI: 10.1021/acs.jpca.9b08723
|
[46] |
Han J, Zhang L, Car R, E W 2017 Deep potential: a general representation of a many-body potential energy surface arXiv:1707.01478 [physics.comp-ph])
|
[47] |
Han J, Zhang L, Car R, E W 2018 Deep potential: a general representation of a many-body potential energy surface Commun. Comput. Phys. 23 629 DOI: 10.4208/cicp.OA-2017-0213
|
[48] |
Zhang L, Han J, Wang H, Car R, E W 2018 Deep potential molecular dynamics: a scalable model with the accuracy of quantum mechanics Phys. Rev. Lett. 120 143001 DOI: 10.1103/PhysRevLett.120.143001
|
[49] |
Wang H, Zhang L, Han J, E W 2018 Deepmd-kit: a deep learning package for many-body potential energy representation and molecular dynamics Comput. Phys. Commun. 228 178 DOI: 10.1016/j.cpc.2018.03.016
|
[50] |
Jia W, Wang H, Chen M, Lu D, Lin L, Car R, E W, Zhang L 2020 Pushing the limit of molecular dynamics with ab initio accuracy to 100 million atoms with machine learning SC20: Int. Conf. for High Performance Computing, Networking, Storage and Analysis pp 1-14
|
[51] |
Artrith N, Morawietz T, Behler J 2011 High-dimensional neural-network potentials for multicomponent systems: applications to zinc oxide Phys. Rev. B 83 153101 DOI: 10.1103/PhysRevB.83.153101
|
[52] |
Bereau T, Andrienko D, von Lilienfeld O A 2015 Transferable atomic multipole machine learning models for small organic molecules J. Chem. Theory Comput. 11 3225 DOI: 10.1021/acs.jctc.5b00301
|
[53] |
Bereau T, DiStasio R A, Tkatchenko A, von Lilienfeld O A 2018 Non-covalent interactions across organic and biological subsets of chemical space: Physics-based potentials parametrized from machine learning J. Chem. Phys. 148 241706 DOI: 10.1063/1.5009502
|
[54] |
Nebgen B, Lubbers N, Smith J S, Sifain A, Lokhov A, Isayev O, Roitberg A, Barros K, Tretiak S 2018 Transferable molecular charge assignment using deep neural networks arXiv:1803.04395 [physics.chem-ph])
|
[55] |
Sifain A E, Lubbers N, Nebgen B T, Smith J S, Lokhov A Y, Isayev O, Roitberg A E, Barros K, Tretiak S 2018 Discovering a transferable charge assignment model using machine learning J. Phys. Chem. Lett. 9 4495 DOI: 10.1021/acs.jpclett.8b01939
|
[56] |
Ko T W, Finkler J A, Goedecker S, Behler J 2021 A fourth-generation high-dimensional neural network potential with accurate electrostatics including non-local charge transfer Nat. Commun. 12 398 DOI: 10.1038/s41467-020-20427-2
|
[57] |
Ko T W, Finkler J A, Goedecker S, Behler J 2021 General-purpose machine learning potentials capturing nonlocal charge transfer Acc. Chem. Res. 54 808 DOI: 10.1021/acs.accounts.0c00689
|
[58] |
Grisafi A, Ceriotti M 2019 Incorporating long-range physics in atomic-scale machine learning J. Chem. Phys. 151 204105 DOI: 10.1063/1.5128375
|
[59] |
Grisafi A, Nigam J, Ceriotti M 2021 Multi-scale approach for the prediction of atomic scale properties Chem. Sci. 12 2078 DOI: 10.1039/D0SC04934D
|
[60] |
Frenkel D, Smit B 2002 Understanding Molecular Simulation From Algorithms to ApplicationsNew YorkAcademic
|
[61] |
He K, Zhang X, Ren S, Sun J 2016 Deep residual learning for image recognition Proc. Conf. on Computer Vision and Pattern Recognition (CVPR)
|
[62] |
Barron A 1993 Universal approximation bounds for superpositions of a sigmoidal function IEEE Trans. Inf. Theory 39 930 DOI: 10.1109/18.256500
|
[63] |
Barron A R 1994 Approximation and estimation bounds for artificial neural networks Mach. Learn. 14 115 DOI: 10.1007/BF00993164
|
[64] |
Liang S, Srikant R 2017 Why deep neural networks for function approximation? arXiv:1610.04161 [cs.LG])
|
[65] |
Telgarsky M 2016 benefits of depth in neural networks 29th Conf. on Learning Theory (Proc. Machine Learning Research) (PMLR)vol 49ed(V Feldman, A Rakhlin and O Shamir)(New York: Columbia University) 1517-39
|
[66] |
Yarotsky D 2017 Error bounds for approximations with deep relu networks Neural Netw. 94 103 DOI: 10.1016/j.neunet.2017.07.002
|
[67] |
Lu J, Shen Z, Yang H, Zhang S 2021 Deep network approximation for smooth functions SIAM J. Math. Anal. 53 5465 DOI: 10.1137/20M134695X
|
[68] |
E W, Ma C, Wu L 2019 A priori estimates of the population risk for two-layer neural networks Commun. Math. Sci. 17 1407-25 DOI: 10.4310/CMS.2019.v17.n5.a11
|
[69] |
E W, Ma C, Wu L 2022 The barron space and the flow-induced function spaces for neural network models Constructive Approx. 55 369-406 DOI: 10.1007/s00365-021-09549-y
|
[70] |
Zhang L, Han J, Wang H, Saidi W, Car R, E W 2018 End-to-end symmetry preserving inter-atomic potential energy model for finite and extended systems Advances in Neural Information Processing Systemsvol 31Bengio S, Wallach H, Larochelle H, Grauman K, Cesa-Bianchi N, Garnett R 2018 (Curran Associates, Inc.) pp 4436-46
|
[71] |
Kresse G, Furthmller J 1996 Efficiency of ab-initio total energy calculations for metals and semiconductors using a plane-wave basis set Comput. Mater. Sci. 6 15 DOI: 10.1016/0927-0256(96)00008-0
|
[72] |
Kresse G, Furthmller J 1996 Efficient iterative schemes for ab initio total-energy calculations using a plane-wave basis set Phys. Rev. B 54 11169 DOI: 10.1103/PhysRevB.54.11169
|
[73] |
Giannozzi P, et al 2009 QUANTUM ESPRESSO: a modular and open-source software project for quantum simulations of materials J. Phys.: Condens. Matter. 21 395502 DOI: 10.1088/0953-8984/21/39/395502
|
[74] |
Chen M, Guo G-C, He L 2010 Systematically improvable optimized atomic basis sets for ab initio calculations J. Phys.: Condens. Matter. 22 445501 DOI: 10.1088/0953-8984/22/44/445501
|
[75] |
Perdew J P, Schmidt K 2001 Jacob’s ladder of density functional approximations for the exchange-correlation energy Conf. Proc.vol 577 p 1
|
[76] |
Mller C, Plesset M S 1934 Note on an approximation treatment for many-electron systems Phys. Rev. 46 618 DOI: 10.1103/PhysRev.46.618
|
[77] |
ek J 1996 On the correlation problem in atomic and molecular systems. Calculation of wavefunction components in ursell-type expansion using quantum-field theoretical methods J. Chem. Phys. 45 4256 DOI: 10.1063/1.1727484
|
[78] |
Fano U 1961 Effects of configuration interaction on intensities and phase shifts Phys. Rev. 124 1866 DOI: 10.1103/PhysRev.124.1866
|
[79] |
Deaven D M, Ho K M 1995 Molecular geometry optimization with a genetic algorithm Phys. Rev. Lett. 75 288 DOI: 10.1103/PhysRevLett.75.288
|
[80] |
Glass C W, Oganov A R, Hansen N 2006 USPEX-evolutionary crystal structure prediction Comput. Phys. Commun. 175 713 DOI: 10.1016/j.cpc.2006.07.020
|
[81] |
Laio A, Parrinello M 2002 Escaping free-energy minima Proc. Natl Acad. Sci. 99 12562 DOI: 10.1073/pnas.202427399
|
[82] |
Cohn D, Atlas L, Ladner R 1994 Improving generalization with active learning Mach. Learn. 15 201 DOI: 10.1007/BF00993277
|
[83] |
Zhang L, Lin D-Y, Wang H, Car R, E W 2019 Active learning of uniformly accurate interatomic potentials for materials simulation Phys. Rev. Mater. 3 023804 DOI: 10.1103/PhysRevMaterials.3.023804
|
[84] |
Zhang Y, Wang H, Chen W, Zeng J, Zhang L, Wang H, E W 2020 DP-GEN: a concurrent learning platform for the generation of reliable deep learning based potential energy models Comput. Phys. Commun. 253 107206 DOI: 10.1016/j.cpc.2020.107206
|
[85] |
Plimpton S 1995 Fast parallel algorithms for short-range molecular dynamics J. Comput. Phys. 117 1 DOI: 10.1006/jcph.1995.1039
|
[86] |
Larsen A H, et al 2017 The atomic simulation environmenta python library for working with atoms J. Phys.: Condens. Matter. 29 273002 DOI: 10.1088/1361-648x/aa680e
|
[87] |
Ceriotti M, More J, Manolopoulos D E 2014 i-PI: a python interface for ab initio path integral molecular dynamics simulations Comput. Phys. Commun. 185 1019 DOI: 10.1016/j.cpc.2013.10.027
|
[88] |
Van Der Spoel D, Lindahl E, Hess B, Groenhof G, Mark A E, Berendsen H J C 2005 Gromacs: fast, flexible and free J. Comput. Chem. 26 1701 DOI: 10.1002/jcc.20291
|
[89] |
Gaussian 16 Revision C.01 2016 Gaussian Inc. Wallingford CT
|
[90] |
Soler J M, Artacho E, Gale J D, Garca A, Junquera J, Ordejn P, Snchez-Portal D 2002 The SIESTA method for ab initio order-N materials simulation J. Phys.: Condens. Matter. 14 2745 DOI: 10.1088/0953-8984/14/11/302
|
[91] |
Khne T D, et al 2020 Cp2k: an electronic structure and molecular dynamics software packagequickstep: efficient and accurate electronic structure calculations J. Chem. Phys. 152 194103 DOI: 10.1063/5.0007045
|
[92] |
Blum V, Gehrke R, Hanke F, Havu P, Havu V, Ren X, Reuter K, Scheffler M 2009 Ab initio molecular simulations with numeric atom-centered orbitals Comput. Phys. Commun. 180 2175 DOI: 10.1016/j.cpc.2009.06.022
|
[93] |
Case D A, Cheatham III T E, Darden T, Gohlke H, Luo R, Merz Jr. K M, Onufriev A, Simmerling C, Wang B, Woods R J 2005 The amber biomolecular simulation programs J. Comput. Chem. 26 1668 DOI: 10.1002/jcc.20290
|
[94] |
Abadi M, et al 2015 TensorFlow: large-scale machine learning on heterogeneous systems software available fromwww.tensorflow.org/
|
[95] |
Zhang L, Wang H, Car R, E W 2021 Phase diagram of a deep potential water model Phys. Rev. Lett. 126 236001 DOI: 10.1103/PhysRevLett.126.236001
|
[96] |
Schimka L, Gaudoin R, Klime J, Marsman M, Kresse G 2013 Lattice constants and cohesive energies of alkali, alkaline-earth and transition metals: random phase approximation and density functional theory results Phys. Rev. B 87 214102 DOI: 10.1103/PhysRevB.87.214102
|
[97] |
Kittel C 2005 Introduction to Solid State Physics8th ednNew YorkWiley
|
[98] |
Yang M, Karmakar T, Parrinello M 2021 Liquid-liquid critical point in phosphorus Phys. Rev. Lett. 127 080603 DOI: 10.1103/PhysRevLett.127.080603
|
[99] |
Yang M, Bonati L, Polino D, Parrinello M 2022 Using metadynamics to build neural network potentials for reactive events: the case of urea decomposition in water Catal. Today 387 143 DOI: 10.1016/j.cattod.2021.03.018
|
[100] |
Jiang W, Zhang Y, Zhang L, Wang H 2021 Accurate Deep Potential model for the Al-Cu-Mg alloy in the full concentration space Chin. Phys. B 30 050706 DOI: 10.1088/1674-1056/abf134
|
[101] |
Wen T, Wang R, Zhu L, Zhang L, Wang H, Srolovitz D J, Wu Z 2021 Specialising neural network potentials for accurate properties and application to the mechanical response of titanium npj Comput. Mater. 7 206 DOI: 10.1038/s41524-021-00661-y
|
[102] |
Wang X, Wang Y, Zhang L, Dai F, Wang H 2021 A tungsten deep potential with high accuracy and generalization ability based on a newly designed three-body embedding formalism arXiv:2111.04281 [cond-mat.mtrl-sci])
|
[103] |
Wang Y, Zhang L, Xu B, Wang X, Wang H 2022 A generalizable machine learning potential of Ag-Au nanoalloys and its application to surface reconstruction, segregation and diffusion Modelling Simul. Mater. Sci. Eng. 30 025003 DOI: 10.1088/1361-651X/ac4002
|
[104] |
Fu B, Sun Y, Zhang L, Wang H, Xu B 2021 Deep learning inter-atomic potential for thermal and phonon behaviour of silicon carbide with quantum accuracy arXiv:2110.10843 [cond-mat.mtrl-sci])
|
[105] |
Huang J, Zhang L, Wang H, Zhao J, Cheng J, E W 2021 Deep potential generation scheme and simulation protocol for the Li10GeP2S12-type superionic conductors J. Chem. Phys. 154 094703 DOI: 10.1063/5.0041849
|
[106] |
Lu D, Wang H, Chen M, Lin L, Car R, E W, Jia W, Zhang L 2021 86 PFLOPS deep potential molecular dynamics simulation of 100 million atoms with ab initio accuracy Comput. Phys. Commun. 259 107624 DOI: 10.1016/j.cpc.2020.107624
|
[107] |
Lu D, Jiang W, Chen Y, Zhang L, Jia W, Wang H, Chen M 2021 DP train, then DP compress: model compression in deep potential molecular dynamics arXiv:2107.02103 [physics.comp-ph])
|
[108] |
Mendelev M, Underwood T, Ackland G 2016 Development of an interatomic potential for the simulation of defects, plasticity and phase transformations in titanium J. Chem. Phys. 145 154102 DOI: 10.1063/1.4964654
|
[109] |
Hennig R, Lenosky T, Trinkle D, Rudin S, Wilkins J W 2008 Classical potential describes martensitic phase transformations between the
|
[110] |
Vtek V 1968 Intrinsic stacking faults in body-centred cubic crystals Phil. Mag. 18 773 DOI: 10.1080/14786436808227500
|
[111] |
Ko W-S, Grabowski B, Neugebauer J 2015 Development and application of a Ni-Ti interatomic potential with high predictive accuracy of the martensitic phase transition Phys. Rev. B 92 134107 DOI: 10.1103/PhysRevB.92.134107
|
[112] |
Dickel D, Barrett C, Carino R, Baskes M, Horstemeyer M 2018 Mechanical instabilities in the modeling of phase transitions of titanium Modelling Simul. Mater. Sci. Eng. 26 065002 DOI: 10.1088/1361-651X/aac95d
|
[113] |
Clouet E, Caillard D, Chaari N, Onimus F, Rodney D 2015 Dislocation locking versus easy glide in titanium and zirconium Nat. Mater. 14 931 DOI: 10.1038/nmat4340
|
[114] |
Wang H, Guo X, Zhang L, Wang H, Xue J 2019 Deep learning inter-atomic potential model for accurate irradiation damage simulations Appl. Phys. Lett. 114 244101 DOI: 10.1063/1.5098061
|
[115] |
Zeng Q, Yu X, Yao Y, Gao T, Chen B, Zhang S, Kang D, Wang H, Dai J 2021 Ab initio validation on the connection between atomistic and hydrodynamic description to unravel the ion dynamics of warm dense matter Phys. Rev. Res. 3 033116 DOI: 10.1103/PhysRevResearch.3.033116
|
[116] |
Liu Q, Lu D, Chen M 2020 Structure and dynamics of warm dense aluminum: a molecular dynamics study with density functional theory and deep potential J. Phys.: Condens. Matter. 32 144002 DOI: 10.1088/1361-648x/ab5890
|
[117] |
Liu Q, Li J, Chen M 2021 Thermal transport by electrons and ions in warm dense aluminum: a combined density functional theory and deep potential study Matter Radiat. Extremes 6 026902 DOI: 10.1063/5.0030123
|
[118] |
Cheng Y, et al 2021 Deep-learning potential method to simulate shear viscosity of liquid aluminum at high temperature and high pressure by molecular dynamics AIP Adv. 11 015043 DOI: 10.1063/5.0036298
|
[119] |
Andolina C M, Bon M, Passerone D, Saidi W A 2021 Robust, multi-length-scale, machine learning potential for Ag-Au bimetallic alloys from clusters to bulk materials J. Phys. Chem. C 125 17438 DOI: 10.1021/acs.jpcc.1c04403
|
[120] |
Chen B, Zeng Q, Wang H, Zhang S, Kang D, Lu D, Dai J 2021 Atomistic mechanism of phase transition in shock compressed gold revealed by deep potential arXiv:2006.13136 [cond-mat.mtrl-sci])
|
[121] |
Jiao J 2021 Self-healing mechanism of lithium in lithium metal batteries arXiv:2106.10979 [cond-mat.mtrl-sci])
|
[122] |
Zhang Y, Gao C, Liu Q, Zhang L, Wang H, Chen M 2020 Warm dense matter simulation via electron temperature dependent deep potential molecular dynamics Phys. Plasmas 27 122704 DOI: 10.1063/5.0023265
|
[123] |
Niu H, Bonati L, Piaggi P M, Parrinello M 2020 Ab initio phase diagram and nucleation of gallium Nat. Commun. 11 1 DOI: 10.1038/s41467-020-16372-9
|
[124] |
Shi M, Li J, Tao M, Zhang X, Liu J 2021 Artificial intelligence model for efficient simulation of monatomic phase change material antimony Mater. Sci. Semicond. Process. 136 106146 DOI: 10.1016/j.mssp.2021.106146
|
[125] |
Wang J, Shen H, Yang R, Xie K, Zhang C, Chen L, Ho K-M, Wang C-Z, Wang S 2022 A deep learning interatomic potential developed for atomistic simulation of carbon materials Carbon 186 1 DOI: 10.1016/j.carbon.2021.09.062
|
[126] |
Bonati L, Parrinello M 2018 Silicon liquid structure and crystal nucleation from ab Initio deep metadynamics Phys. Rev. Lett. 121 265701 DOI: 10.1103/PhysRevLett.121.265701
|
[127] |
Li R, Lee E, Luo T 2020 A unified deep neural network potential capable of predicting thermal conductivity of silicon in different phases Mater. Today Phys. 12 100181 DOI: 10.1016/j.mtphys.2020.100181
|
[128] |
Wang H, Zhang Y, Zhang L, Wang H 2020 Crystal structure prediction of binary alloys via deep potential Front. Chem. 8 895 DOI: 10.3389/fchem.2020.589795
|
[129] |
Andolina C M, Wright J G, Das N, Saidi W A 2021 Improved Al-Mg alloy surface segregation predictions with a machine learning atomistic potential Phys. Rev. Mater. 5 083804 DOI: 10.1103/PhysRevMaterials.5.083804
|
[130] |
Bourgeois L, Zhang Y, Zhang Z, Chen Y, Medhekar N V 2020 Transforming solid-state precipitates via excess vacancies Nat. Commun. 11 1 DOI: 10.1038/s41467-020-15087-1
|
[131] |
Cheng B, Zhao X, Zhang Y, Chen H, Polmear I, Nie J-F 2020 Co-segregation of Mg and Zn atoms at the planar
|
[132] |
Ryltsev R E, Chtchelkatchev N M 2021 Deep machine learning potentials for multicomponent metallic melts: development, predictability and compositional transferability arXiv:2110.14006 [cond-mat.mtrl-sci])
|
[133] |
Wen T, et al 2019 Development of a deep machine learning interatomic potential for metalloid-containing Pd-Si compounds Phys. Rev. B 100 174101 DOI: 10.1103/PhysRevB.100.174101
|
[134] |
Wang Q, Zhai B, Wang H P, Wei B 2021 Atomic structure of liquid refractory Nb5Si3 intermetallic compound alloy based upon deep neural network potential J. Appl. Phys. 130 185103 DOI: 10.1063/5.0067157
|
[135] |
Guo Y, et al 2019 Bergman-type medium range order in amorphous Zr77Rh23 alloy studied by ab initio molecular dynamics simulations J. Alloys Compd. 790 675 DOI: 10.1016/j.jallcom.2019.03.197
|
[136] |
Guo Y, et al 2019 Temperature dependence of structural, dynamical and electronic properties of amorphous Bi2Te3: an ab initio study New J. Phys. 21 093062 DOI: 10.1088/1367-2630/ab4535
|
[137] |
Tang L, Yang Z J, Wen T Q, Ho K M, Kramer M J, Wang C Z 2020 Development of interatomic potential for Al-Tb alloys using a deep neural network learning method Phys. Chem. Chem. Phys. 22 18467 DOI: 10.1039/D0CP01689F
|
[138] |
Tang L, Yang Z, Wen T, Ho K M, Kramer M J, Wang C Z 2021 Short- and medium-range orders in Al90Tb10 glass and their relation to the structures of competing crystalline phases Acta Mater. 204 116513 DOI: 10.1016/j.actamat.2020.116513
|
[139] |
Han I, McKeown J T, Tang L, Wang C-Z, Parsamehr H, Xi Z, Lu Y-R, Kramer M J, Shahani A J 2020 Dynamic observation of dendritic quasicrystal growth upon laser-induced solid-state transformation Phys. Rev. Lett. 125 195503 DOI: 10.1103/PhysRevLett.125.195503
|
[140] |
Tang L, Ho K M, Wang C Z 2021 Molecular dynamics simulation of metallic Al-Ce liquids using a neural network machine learning interatomic potential J. Chem. Phys. 155 194503 DOI: 10.1063/5.0066061
|
[141] |
Daniels C L, Liu D-J, Adamson M A S, Knobeloch M, Vela J 2021 Azo(xy) vs aniline selectivity in catalytic nitroarene reduction by intermetallics: experiments and simulations J. Phys. Chem. C 125 24440 DOI: 10.1021/acs.jpcc.1c08569
|
[142] |
Zhang C, Sun Y, Wang H-D, Zhang F, Wen T-Q, Ho K-M, Wang C-Z 2021 Crystallization of the P3Sn4 phase upon cooling P2Sn5 liquid by molecular dynamics simulation using a machine learning interatomic potential J. Phys. Chem. C 125 3127 DOI: 10.1021/acs.jpcc.0c08873
|
[143] |
Balyakin I A, Rempel S V, Ryltsev R E, Rempel A A 2020 Deep machine learning interatomic potential for liquid silica Phys. Rev. E 102 052125 DOI: 10.1103/PhysRevE.102.052125
|
[144] |
Deng J, Stixrude L 2021 Thermal conductivity of silicate liquid determined by machine learning potentials Geophys. Res. Lett. 48 e2021GL093806 DOI: 10.1029/2021GL093806
|
[145] |
Luo H, Karki B B, Ghosh D B, Bao H 2021 Anomalous behavior of viscosity and electrical conductivity of MgSiO3 melt at mantle conditions Geophys. Res. Lett. 48 e2021GL093573 DOI: 10.1029/2021GL093573
|
[146] |
Luo H, Karki B B, Ghosh D B, Bao H 2021 Deep neural network potentials for diffusional lithium isotope fractionation in silicate melts Geochim. Cosmochim. Acta 303 38 DOI: 10.1016/j.gca.2021.03.031
|
[147] |
Chen W, Li L-S 2021 The study of the optical phonon frequency of 3C-SiC by molecular dynamics simulations with deep neural network potential J. Appl. Phys. 129 244104 DOI: 10.1063/5.0049464
|
[148] |
An Q 2021 Mitigating amorphization in superhard boron carbide by microalloying-induced stacking fault formation Phys. Rev. Mater. 5 103602 DOI: 10.1103/PhysRevMaterials.5.103602
|
[149] |
Rodriguez A, Lam S, Hu M 2021 Thermodynamic and transport properties of LiF and FLiBe molten salts with deep learning potentials ACS Appl. Mater. Interfaces 13 55367-79 DOI: 10.1021/acsami.1c17942
|
[150] |
Liang W, Lu G, Yu J 2021 Theoretical prediction on the local structure and transport properties of molten alkali chlorides by deep potentials J. Mater. Sci. Technol. 75 78 DOI: 10.1016/j.jmst.2020.09.040
|
[151] |
Liang W, Lu G, Yu J 2020 Molecular dynamics simulations of molten magnesium chloride using machine-learning-based deep potential Adv. Theory Simul. 3 2000180 DOI: 10.1002/adts.202000180
|
[152] |
Pan G, Chen P, Yan H, Lu Y 2020 A DFT accurate machine learning description of molten ZnCl2 and its mixtures: 1. Potential development and properties prediction of molten ZnCl2 Comput. Mater. Sci. 185 109955 DOI: 10.1016/j.commatsci.2020.109955
|
[153] |
Pan G, Ding J, Du Y, Lee D-J, Lu Y 2021 A DFT accurate machine learning description of molten ZnCl2 and its mixtures: 2. Potential development and properties prediction of ZnCl2-NaCl-KCl ternary salt for CSP Comput. Mater. Sci. 187 110055 DOI: 10.1016/j.commatsci.2020.110055
|
[154] |
Liang W, Lu G, Yu J 2021 Machine-learning-driven simulations on microstructure and thermophysical properties of MgCl2-KCl eutectic ACS Appl. Mater. Interfaces 13 4034 DOI: 10.1021/acsami.0c20665
|
[155] |
Bu M, Liang W, Lu G, Yu J 2021 Local structure elucidation and properties prediction on KCl-CaCl2 molten salt: a deep potential molecular dynamics study Sol. Energy Mater. Sol. Cells 232 111346 DOI: 10.1016/j.solmat.2021.111346
|
[156] |
Zhao J, Liang W, Lu G 2021 Theoretical prediction on the redox potentials of rare-earth ions by deep potentials Ionics 27 2079 DOI: 10.1007/s11581-021-03988-0
|
[157] |
Zhang J, Fuller J, An Q 2021 Coordination and thermophysical properties of transition metal chlorocomplexes in LiCl-KCl eutectic J. Phys. Chem. B 125 8876 DOI: 10.1021/acs.jpcb.1c03748
|
[158] |
Xu N, Shi Y, He Y, Shao Q 2020 A deep-learning potential for crystalline and amorphous Li-Si alloys J. Phys. Chem. C 124 16278 DOI: 10.1021/acs.jpcc.0c03333
|
[159] |
Marcolongo A, Binninger T, Zipoli F, Laino T 2019 Simulating diffusion properties of solid-state electrolytes via a neural network potential: performance and training scheme arXiv:1910.10090 [physics.comp-ph])
|
[160] |
Gupta M K, Ding J, Osti N C, Abernathy D L, Arnold W, Wang H, Hood Z, Delaire O 2021 Fast Na diffusion and anharmonic phonon dynamics in superionic Na3PS4 Energy Environ. Sci. 14 6554 DOI: 10.1039/D1EE01509E
|
[161] |
Li H-X, Zhou X-Y, Wang Y-C, Jiang H 2021 Theoretical study of Na+ transport in the solid-state electrolyte Na3OBr based on deep potential molecular dynamics Inorg. Chem. Front. 8 425 DOI: 10.1039/D0QI00921K
|
[162] |
Lin M, Liu X, Xiang Y, Wang F, Liu Y, Fu R, Cheng J, Yang Y 2021 Unravelling the fast alkali-ion dynamics in paramagnetic battery materials combined with NMR and deep-potential molecular dynamics simulation Angew. Chem., Int. Ed. 60 12547 DOI: 10.1002/anie.202102740
|
[163] |
Calegari Andrade M F, Selloni A 2020 Structure of disordered TiO2 phases from ab initio based deep neural network simulations Phys. Rev. Mater. 4 113803 DOI: 10.1103/PhysRevMaterials.4.113803
|
[164] |
Li R, Liu Z, Rohskopf A, Gordiz K, Henry A, Lee E, Luo T 2020 A deep neural network interatomic potential for studying thermal conductivity of
|
[165] |
Wu J, Zhang Y, Zhang L, Liu S 2021 Deep learning of accurate force field of ferroelectric HfO2 Phys. Rev. B 103 024108 DOI: 10.1103/PhysRevB.103.024108
|
[166] |
Balyakin I, Sadovnikov S 2022 Deep learning potential for superionic phase of Ag2S Comput. Mater. Sci. 202 110963 DOI: 10.1016/j.commatsci.2021.110963
|
[167] |
Wang H, Guo X, Xue J 2020 Deep-learning interatomic potential for irradiation damage simulations in MoS2 with ab initial accuracy arXiv:2010.09547 [cond-mat.mtrl-sci])
|
[168] |
Guo D, Li C, Li K, Shao B, Chen D, Ma Y, Sun J, Cao X, Zeng W, Chang X 2021 The thermoelectric performance of new structure SnSe studied by quotient graph and deep learning potential Mater. Today Energy 20 100665 DOI: 10.1016/j.mtener.2021.100665
|
[169] |
Dai F-Z, Wen B, Xiang H, Zhou Y 2020 Grain boundary strengthening in ZrB2 by segregation of W: atomistic simulations with deep learning potential J. Eur. Ceram. Soc. 40 5029 DOI: 10.1016/j.jeurceramsoc.2020.06.007
|
[170] |
Dai F-Z, Wen B, Sun Y, Xiang H, Zhou Y 2020 Theoretical prediction on thermal and mechanical properties of high entropy (Zr0.2Hf0.2Ti0.2Nb0.2Ta0.2)C by deep learning potential J. Mater. Sci. Technol. 43 168 DOI: 10.1016/j.jmst.2020.01.005
|
[171] |
Dai F-Z, Sun Y, Wen B, Xiang H, Zhou Y 2021 Temperature dependent thermal and elastic properties of high entropy (Ti0.2Zr0.2Hf0.2Nb0.2Ta0.2)B2: molecular dynamics simulation by deep learning potential J. Mater. Sci. Technol. 72 8 DOI: 10.1016/j.jmst.2020.07.014
|
[172] |
Ko H-Y, Zhang L, Santra B, Wang H, E W, DiStasio Jr R A, Car R 2019 Isotope effects in liquid water via deep potential molecular dynamics Mol. Phys. 117 3269 DOI: 10.1080/00268976.2019.1652366
|
[173] |
Sommers G M, Calegari Andrade M F, Zhang L, Wang H, Car R 2020 Raman spectrum and polarizability of liquid water from deep neural networks Phys. Chem. Chem. Phys. 22 10592 DOI: 10.1039/D0CP01893G
|
[174] |
Zhang C, Zhang L, Xu J, Tang F, Santra B, Wu X 2020 Isotope effects in x-ray absorption spectra of liquid water Phys. Rev. B 102 115155 DOI: 10.1103/PhysRevB.102.115155
|
[175] |
Gartner T E, Zhang L, Piaggi P M, Car R, Panagiotopoulos A Z, Debenedetti P G 2020 Signatures of a liquid-liquid transition in an ab initio deep neural network model for water Proc. Natl Acad. Sci. 117 26040 DOI: 10.1073/pnas.2015440117
|
[176] |
Andreani C, Romanelli G, Parmentier A, Senesi R, Kolesnikov A I, Ko H-Y, Calegari Andrade M F, Car R 2020 Hydrogen dynamics in supercritical water probed by neutron scattering and computer simulations J. Phys. Chem. Lett. 11 9461 DOI: 10.1021/acs.jpclett.0c02547
|
[177] |
Xu J, Zhang C, Zhang L, Chen M, Santra B, Wu X 2020 Isotope effects in molecular structures and electronic properties of liquid water via deep potential molecular dynamics based on the SCAN functional Phys. Rev. B 102 214113 DOI: 10.1103/PhysRevB.102.214113
|
[178] |
Piaggi P M, Panagiotopoulos A Z, Debenedetti P G, Car R 2021 Phase equilibrium of water with hexagonal and cubic ice using the SCAN functional J. Chem. Theory Comput. 17 3065 DOI: 10.1021/acs.jctc.1c00041
|
[179] |
Tisi D, Zhang L, Bertossa R, Wang H, Car R, Baroni S 2021 Heat transport in liquid water from first-principles and deep-neural-network simulations arXiv:2108.10850 [cond-mat.mtrl-sci])
|
[180] |
Zhang C, Tang F, Chen M, Xu J, Zhang L, Qiu D Y, Perdew J P, Klein M L, Wu X 2021 Modeling liquid water by climbing up Jacob’s ladder in density functional theory facilitated by using deep neural network potentials J. Phys. Chem. B 125 11444 DOI: 10.1021/acs.jpcb.1c03884
|
[181] |
Torres A, Pedroza L S, Fernandez-Serra M, Rocha A R 2021 Using neural network force fields to ascertain the quality of ab initio simulations of liquid water J. Phys. Chem. B 125 10772 DOI: 10.1021/acs.jpcb.1c04372
|
[182] |
Shi Y, Doyle C C, Beck T L 2021 Condensed phase water molecular multipole moments from deep neural network models trained on ab initio simulation data J. Phys. Chem. Lett. 12 10310 DOI: 10.1021/acs.jpclett.1c02328
|
[183] |
Calio P B, Li C, Voth G A 2021 Resolving the structural debate for the hydrated excess proton in water J. Am. Chem. Soc. 143 18672 DOI: 10.1021/jacs.1c08552
|
[184] |
Xu M, Zhu T, Zhang J Z H 2019 Molecular dynamics simulation of zinc ion in water with an ab initio based neural network potential J. Phys. Chem. A 123 6587 DOI: 10.1021/acs.jpca.9b04087
|
[185] |
Niblett S P, Galib M, Limmer D T 2021 Learning intermolecular forces at liquid-vapor interfaces J. Chem. Phys. 155 164101 DOI: 10.1063/5.0067565
|
[186] |
Galib M, Limmer D T 2021 Reactive uptake of N2O5 by atmospheric aerosol is dominated by interfacial processes Science 371 921 DOI: 10.1126/science.abd7716
|
[187] |
Andrade M F C, Ko H-Y, Zhang L, Car R, Selloni A 2020 Free energy of proton transfer at the water-TiO2 interface from ab initio deep potential molecular dynamics Chem. Sci. 11 2335 DOI: 10.1039/C9SC05116C
|
[188] |
Piaggi P M, Car R 2021 Enhancing the formation of ionic defects to study the ice Ih/XI transition with molecular dynamics simulations Mol. Phys. 119 e1916634 DOI: 10.1080/00268976.2021.1916634
|
[189] |
Ye Q-J, Zhuang L, Li X-Z 2021 Dynamic nature of high-pressure ice VII Phys. Rev. Lett. 126 185501 DOI: 10.1103/PhysRevLett.126.185501
|
[190] |
Jiang S, Liu Y-R, Huang T, Feng Y-J, Wang C-Y, Wang Z-Q, Huang W 2021 Towards fully ab initio simulation of atmospheric aerosol nucleation arXiv:2107.04802 [physics.atm-clus])
|
[191] |
Zeng J, Zhang L, Wang H, Zhu T 2021 Exploring the chemical space of linear alkane pyrolysis via deep potential generator Energy Fuels 35 762 DOI: 10.1021/acs.energyfuels.0c03211
|
[192] |
Chen W-K, Liu X-Y, Fang W-H, Dral P O, Cui G 2018 Deep learning for nonadiabatic excited-state dynamics J. Phys. Chem. Lett. 9 6702 DOI: 10.1021/acs.jpclett.8b03026
|
[193] |
Zhang L, Wang H, E W 2018 Reinforced dynamics for enhanced sampling in large atomic and molecular systems J. Chem. Phys. 148 124113 DOI: 10.1063/1.5019675
|
[194] |
Wang S, Ma Z, Pan W 2020 Data-driven coarse-grained modeling of polymers in solution with structural and dynamic properties conserved Soft Matter 16 8330 DOI: 10.1039/D0SM01019G
|
[195] |
Pan X, Yang J, Van R, Epifanovsky E, Ho J, Huang J, Pu J, Mei Y, Nam K, Shao Y 2021 Machine-learning-assisted free energy simulation of solution-phase and enzyme reactions J. Chem. Theory Comput. 17 5745 DOI: 10.1021/acs.jctc.1c00565
|
[196] |
Tuo P, Ye X B, Pan B C 2020 A machine learning based deep potential for seeking the low-lying candidates of Al clusters J. Chem. Phys. 152 114105 DOI: 10.1063/5.0001491
|
[197] |
Achar S K, Zhang L, Johnson J K 2021 Efficiently trained deep learning potential for graphane J. Phys. Chem. C 125 14874 DOI: 10.1021/acs.jpcc.1c01411
|
[198] |
Wu J, Bai L, Huang J, Ma L, Liu J, Liu S 2021 Accurate force field of two-dimensional ferroelectrics from deep learning Phys. Rev. B 104 174107 DOI: 10.1103/PhysRevB.104.174107
|
[199] |
Chen H, Chen J, Ning P, Chen X, Liang J, Yao X, Chen D, Qin L, Huang Y, Wen Z 2021 2D heterostructure of amorphous CoFeB coating black phosphorus nanosheets with optimal oxygen intermediate absorption for improved electrocatalytic water oxidation ACS Nano 15 12418 DOI: 10.1021/acsnano.1c04715
|
[200] |
Pascuet M, Fernndez J 2015 Atomic interaction of the MEAM type for the study of intermetallics in the Al-U alloy J. Nucl. Mater. 467 229 DOI: 10.1016/j.jnucmat.2015.09.030
|
[201] |
Jacobsen K W, Norskov J K, Puska M J 1987 Interatomic interactions in the effective-medium theory Phys. Rev. B 35 7423 DOI: 10.1103/PhysRevB.35.7423
|
[202] |
Jain A, et al 2013 Commentary: The materials project: a materials genome approach to accelerating materials innovation APL Mater. 1 011002 DOI: 10.1063/1.4812323
|
[203] |
Wang Y, Lv J, Zhu L, Ma Y 2012 Calypso: a method for crystal structure prediction Comput. Phys. Commun. 183 2063 DOI: 10.1016/j.cpc.2012.05.008
|
[204] |
Aragones J L, Conde M M, Noya E G, Vega C 2009 The phase diagram of water at high pressures as obtained by computer simulations of the tip4p/2005 model: the appearance of a plastic crystal phase Phys. Chem. Chem. Phys. 11 543 DOI: 10.1039/B812834K
|
[205] |
Poschmann M, Asta M, Chrzan D C 2018 Convergence of calculated dislocation core structures in hexagonal close packed titanium Modelling Simul. Mater. Sci. Eng. 26 014003 DOI: 10.1088/1361-651X/aa9ba9
|
[206] |
Queyroux J-A, et al 2020 Melting curve and isostructural solid transition in superionic ice Phys. Rev. Lett. 125 195501 DOI: 10.1103/PhysRevLett.125.195501
|
[207] |
Mishin Y 2021 Machine-learning interatomic potentials for materials science Acta Mater. 214 116980 DOI: 10.1016/j.actamat.2021.116980
|
[208] |
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez A N, Kaiser L, Polosukhin I 2017 Attention is all you need available at: https://doi.org/10.48550/ARXIV.1706.03762
|
[209] |
Devlin J, Chang M-W, Lee K, Toutanova K 2019 Bert: pre-training of deep bidirectional transformers for language understanding Proc. 2019th Conf. North American Chapter of the Association for Computational Linguistics: Human Language Technologies, (Long and Short Papers Vol 1)Minneapolis, MN(Association for Computational Linguistics) pp 4171-86
|
[210] |
Brown T, et al 2020 Language models are few-shot learners Adv. Neural Inf. Process. Syst. 33 1877-901
|
[211] |
Min B, Ross H, Sulem E, Veyseh A P B, Nguyen T H, Sainz O, Agirre E, Heinz I, Roth D 2021 Recent advances in natural language processing via large pre-trained language models: a survey available at: https://doi.org/10.48550/ARXIV.2111.01243
|
[1] | Shuai Wang, ChiYung Yam, LiHong Hu, Faan-Fung Hung, Shuguang Chen, Chi-Ming Che, GuanHua Chen. A general protocol for phosphorescent platinum(II) complexes: generation, high throughput virtual screening and highly accurate predictions[J]. Materials Futures, 2025, 4(2): 025601. DOI: 10.1088/2752-5724/adb320 |
[2] | Xiang Xu, Yi Wang, Siyu Zhu, Qian Xu, Zulan Liu, Guotao Cheng, Dingpei Long, Lan Cheng, Fangyin Dai. Functionalized bio-spinning silk fiber scaffolds containing Mg2+ with osteoimmunomodulatory and osteogenesis abilities for critical-sized bone defect regeneration[J]. Materials Futures, 2025, 4(1): 015401. DOI: 10.1088/2752-5724/ad9e09 |
[3] | Wenqi Qian, Haiyi Liu, Guangyi Tao, Fangxun Liu, Sihan Lin, Tengteng Gao, Xueying Wang, Qihong Hu, Dalin Zhang, Dong Xiang, Lie Lin, Pengfei Qi, Zheyu Fang, Weiwei Liu. Exciton funneling in 2D artificial potential landscapes decorated by reassembled micro-bubbles[J]. Materials Futures. DOI: 10.1088/2752-5724/adc8c1 |
[4] | Zhongshu Ren, Jiayun Shao, Haolin Liu, Samuel J Clark, Lin Gao, Lilly Balderson, Kyle Mumm, Kamel Fezzaa, Anthony D Rollett, Levent Burak Kara, Tao Sun. Sub-millisecond keyhole pore detection in laser powder bed fusion using sound and light sensors and machine learning[J]. Materials Futures, 2024, 3(4): 045001. DOI: 10.1088/2752-5724/ad89e2 |
[5] | Fei-Teng Wang, Xiandong Liu, Jun Cheng. Water structures and anisotropic dynamics at Pt(211)/water interface revealed by machine learning molecular dynamics[J]. Materials Futures, 2024, 3(4): 041001. DOI: 10.1088/2752-5724/ad7619 |
[6] | Daniel Wines, Kamal Choudhary. Data-driven design of high pressure hydride superconductors using DFT and deep learning[J]. Materials Futures, 2024, 3(2): 025602. DOI: 10.1088/2752-5724/ad4a94 |
[7] | Guanghui Cai, Zhendong Cao, Fankai Xie, Huaxian Jia, Wei Liu, Yaxian Wang, Feng Liu, Xinguo Ren, Sheng Meng, Miao Liu. Predicting structure-dependent Hubbard U parameters via machine learning[J]. Materials Futures, 2024, 3(2): 025601. DOI: 10.1088/2752-5724/ad19e2 |
[8] | Hui-Ru Zhang, Liang Gao, Yu-Hao Ye, Jia-Xin Zhang, Tao Zhang, Qing-Zhou Bu, Qun Yang, Zeng-Wei Zhu, Shuai Wei, Hai-Bin Yu. Fragility crossover mediated by covalent-like electronic interactions in metallic liquids[J]. Materials Futures, 2024, 3(2): 025002. DOI: 10.1088/2752-5724/ad4404 |
[9] | Renrui Fang, Woyu Zhang, Kuan Ren, Peiwen Zhang, Xiaoxin Xu, Zhongrui Wang, Dashan Shang. In-materio reservoir computing based on nanowire networks: fundamental, progress, and perspective[J]. Materials Futures, 2023, 2(2): 022701. DOI: 10.1088/2752-5724/accd87 |
[10] | Juefan Wang, Abhishek A Panchal, Pieremanuele Canepa. Strategies for fitting accurate machine-learned inter-atomic potentials for solid electrolytes[J]. Materials Futures, 2023, 2(1): 015101. DOI: 10.1088/2752-5724/acb506 |
1. | Zhang, J., Pagotto, J., Gould, T. et al. Scalable and accurate simulation of electrolyte solutions with quantum chemical accuracy. Machine Learning: Science and Technology, 2025, 6(1): 015053. DOI:10.1088/2632-2153/adaf76 | |
2. | Zhang, J.-Y., Huynh, G., Albaret, T. et al. Transformation-induced plasticity in CeO2-ZrO2 ceramics: Atomic-scale insights using a deep neural network potential. Acta Materialia, 2025. DOI:10.1016/j.actamat.2024.120661 | |
3. | Wu, Y., Chu, W., Wang, B. et al. Atomistic Origin of Microsecond Carrier Lifetimes at Perovskite Grain Boundaries: Machine Learning-Assisted Nonadiabatic Molecular Dynamics. Journal of the American Chemical Society, 2025, 147(6): 5449-5458. DOI:10.1021/jacs.4c18223 | |
4. | Stolte, N., Daru, J., Forbert, H. et al. Random Sampling Versus Active Learning Algorithms for Machine Learning Potentials of Quantum Liquid Water. Journal of Chemical Theory and Computation, 2025, 21(2): 886-899. DOI:10.1021/acs.jctc.4c01382 | |
5. | Hu, H., Zhang, C., Yue, R. et al. Unraveling ductility enhancement mechanisms in W-Ta alloys using machine-learning potential. International Journal of Mechanical Sciences, 2025. DOI:10.1016/j.ijmecsci.2024.109911 | |
6. | Hu, J., Zhou, L., Jiang, J. Efficient Machine Learning Force Field for Large-Scale Molecular Simulations of Organic Systems. CCS Chemistry, 2025, 7(3): 716-730. DOI:10.31635/ccschem.024.202404785 | |
7. | Watanabe, N., Hori, Y., Sugisawa, H. et al. A machine learning potential construction based on radial distribution function sampling. Journal of Computational Chemistry, 2024, 45(32): 2949-2958. DOI:10.1002/jcc.27497 | |
8. | Wu, Y., Liu, D., Chu, W. et al. Point defects at grain boundaries can create structural instabilities and persistent deep traps in metal halide perovskites. Nanoscale, 2024, 17(4): 2224-2234. DOI:10.1039/d4nr03424d | |
9. | Wang, L., Zhou, X., Luo, Z. et al. Review of External Field Effects on Electrocatalysis: Machine Learning Guided Design. Advanced Functional Materials, 2024, 34(49): 2408870. DOI:10.1002/adfm.202408870 | |
10. | Zhang, D., Liu, X., Zhang, X. et al. DPA-2: a large atomic model as a multi-task learner. npj Computational Materials, 2024, 10(1): 293. DOI:10.1038/s41524-024-01493-2 | |
11. | Li, A.C., Li, B., González-Cataldo, F. et al. Diamond under extremes. Materials Science and Engineering R: Reports, 2024. DOI:10.1016/j.mser.2024.100857 | |
12. | Zhang, H., Cao, B., Huang, L. et al. Machine learning force field study of carboxylate ligands on the surface of zinc-blende CdSe quantum dots. Nano Research, 2024, 17(12): 10685-10693. DOI:10.1007/s12274-024-6983-9 | |
13. | Xue, H.-T., Chang, Z., Li, J. et al. Molecular dynamics simulations of the shear and tensile mechanical properties of rare-earth metal erbium based on deep-learning potential. Materials Today Communications, 2024. DOI:10.1016/j.mtcomm.2024.110485 | |
14. | Gong, X., Li, Z., Pattamatta, A.S.L.S. et al. An accurate and transferable machine learning interatomic potential for nickel. Communications Materials, 2024, 5(1): 157. DOI:10.1038/s43246-024-00603-3 | |
15. | Shakya, A., Ghosh, D.B., Jackson, C. et al. Insights into core-mantle differentiation from bulk Earth melt simulations. Scientific Reports, 2024, 14(1): 18739. DOI:10.1038/s41598-024-69873-8 | |
16. | Fu, F., Wang, X., Hu, T. et al. What drives the heterogeneous interdiffusion in the Li-Si interfacial region of Si anodes: the Li flux or the Si flux?. npj Computational Materials, 2024, 10(1): 131. DOI:10.1038/s41524-024-01327-1 | |
17. | Li, S., Zhang, C., Wang, X. et al. A machine learning methodology for investigating the liquid-liquid transition of hydrogen under high-pressure. International Journal of Modern Physics C, 2024, 35(12): 2450152. DOI:10.1142/S0129183124501523 | |
18. | Urata, S., Bertani, M., Pedone, A. Applications of machine-learning interatomic potentials for modeling ceramics, glass, and electrolytes: A review. Journal of the American Ceramic Society, 2024, 107(12): 7665-7691. DOI:10.1111/jace.19934 | |
19. | Zhang, D., Bi, H., Dai, F.-Z. et al. Pretraining of attention-based deep learning potential model for molecular simulation. npj Computational Materials, 2024, 10(1): 94. DOI:10.1038/s41524-024-01278-7 | |
20. | Liu, M., Wang, J., Hu, J. et al. Layer-by-layer phase transformation in Ti3O5 revealed by machine-learning molecular dynamics simulations. Nature Communications, 2024, 15(1): 3079. DOI:10.1038/s41467-024-47422-1 | |
21. | Sowa, J.K., Rossky, P.J. A Bond-Based Machine Learning Model for Molecular Polarizabilities and A Priori Raman Spectra. Journal of Chemical Theory and Computation, 2024, 20(22): 10071-10079. DOI:10.1021/acs.jctc.4c01086 | |
22. | Jiang, L., Bai, X., Zhi, X. et al. Advancing electrochemical N2 reduction: interfacial electrolyte effects and operando computational approaches. EES Catalysis, 2024, 3(1): 57-79. DOI:10.1039/d4ey00197d | |
23. | Khazieva, E.O., Chtchelkatchev, N.M., Ryltsev, R.E. Transfer learning for accurate description of atomic transport in Al-Cu melts. Journal of Chemical Physics, 2024, 161(17): 174101. DOI:10.1063/5.0222355 | |
24. | Wu, F., Wu, S., Wang, C.-Z. et al. Melting temperature of iron under the Earth's inner core condition from deep machine learning. Geoscience Frontiers, 2024, 15(6): 101925. DOI:10.1016/j.gsf.2024.101925 | |
25. | Yu, W., Cheng, S., Li, Z. et al. The application of multi-scale simulation in advanced electronic packaging. Fundamental Research, 2024, 4(6): 1442-1454. DOI:10.1016/j.fmre.2024.04.004 | |
26. | Luo, K., Zhou, R., Martin, S.W. et al. Flexible doorway controlled Na+ ion diffusion in NaPSO glassy electrolytes from machine-learning force field simulations. Journal of Materials Chemistry A, 2024, 12(48): 33518-33525. DOI:10.1039/d4ta05071a | |
27. | Malenfant-Thuot, O., Ryczko, K., Tamblyn, I. et al. Efficient determination of Born-effective charges, LO-TO splitting, and Raman tensors of solids with a real-space atom-centered deep learning approach. Journal of Physics Condensed Matter, 2024, 36(42): 425901. DOI:10.1088/1361-648X/ad64a2 | |
28. | Wen, T., Liu, H., Gong, X. et al. Deep Potentials for Materials Science | [深度势能方法在材料科学中的应用]. Jinshu Xuebao/Acta Metallurgica Sinica, 2024, 60(10): 1299-1311. DOI:10.11900/0412.1961.2024.00141 | |
29. | Chen, H.M., Wang, Q., Xiao, R.L. et al. Liquid thermophysical properties of Ag-Si alloy based on deep learning potential. Computational Materials Science, 2024. DOI:10.1016/j.commatsci.2024.113293 | |
30. | Huang, W., Dong, Z., Lin, L. Density Functional Theory and Machine Learning of Transition Metals in Mo2C for Gas Sensors. ACS Applied Nano Materials, 2024, 7(18): 22189-22199. DOI:10.1021/acsanm.4c04274 | |
31. | Wen, M., Shi, J., Chang, X. et al. Uncovering the decomposition mechanism of nitrate ester plasticized polyether (NEPE): a neural network potential simulation. Physical Chemistry Chemical Physics, 2024, 26(39): 25719-25730. DOI:10.1039/d4cp02223h | |
32. | Xie, Y., Deng, B., Zhang, Y. et al. Intelligent R&D of battery design automation in the era of artificial intelligence | [AI for Science时代下的电池平台化智能研发]. Energy Storage Science and Technology, 2024, 13(9): 3182-3197. DOI:10.19799/j.cnki.2095-4239.2024.0698 | |
33. | Zhang, X., Li, B., Zou, J. et al. Deep-learning enabled atomic insights into the phase transitions and nanodomain topology of lead-free (K, Na)NbO3 ferroelectrics | [深度学习分子动力学模拟助力无铅(K, Na)NbO3铁电 材料相变及纳米畴结构研究]. Science China Materials, 2024, 67(9): 3029-3038. DOI:10.1007/s40843-024-2999-8 | |
34. | Zhang, H., Thong, H.-C., Bastogne, L. et al. Finite-temperature properties of the antiferroelectric perovskite PbZrO3 from a deep-learning interatomic potential. Physical Review B, 2024, 110(5): 054109. DOI:10.1103/PhysRevB.110.054109 | |
35. | Liu, R.-L., Wang, J., Shen, Z.-H. et al. AI for dielectric capacitors. Energy Storage Materials, 2024. DOI:10.1016/j.ensm.2024.103612 | |
36. | Zhang, P., Feng, M., Xu, X. Double-Layer Distribution of Hydronium and Hydroxide Ions in the Air-Water Interface. ACS Physical Chemistry Au, 2024, 4(4): 336-346. DOI:10.1021/acsphyschemau.3c00076 | |
37. | Liu, J., Zhang, X., Chen, T. et al. Machine-Learning-Based Interatomic Potentials for Group IIB to VIA Semiconductors: Toward a Universal Model. Journal of Chemical Theory and Computation, 2024, 20(13): 5717-5731. DOI:10.1021/acs.jctc.3c01320 | |
38. | Bhullar, M., Bai, Z., Akinpelu, A. et al. Phase Transition in Silicon from Machine Learning Informed Metadynamics. ChemPhysChem, 2024, 25(13): e202400090. DOI:10.1002/cphc.202400090 | |
39. | Al-Maeeni, A., Lazarev, M., Kazeev, N. et al. Review on automated 2D material design. 2D Materials, 2024, 11(3): 032002. DOI:10.1088/2053-1583/ad4661 | |
40. | Wu, F., Sun, Y., Wan, T. et al. Deep-Learning-Based Prediction of the Tetragonal → Cubic Transition in Davemaoite. Geophysical Research Letters, 2024, 51(12): e2023GL108012. DOI:10.1029/2023GL108012 | |
41. | Fan, M., Wen, T., Chen, S. et al. Perspectives Toward Damage-Tolerant Nanostructure Ceramics. Advanced Science, 2024, 11(24): 2309834. DOI:10.1002/advs.202309834 | |
42. | Zhang, J.-Y., Huynh, G., Dai, F.-Z. et al. A deep-neural network potential to study transformation-induced plasticity in zirconia. Journal of the European Ceramic Society, 2024, 44(6): 4243-4254. DOI:10.1016/j.jeurceramsoc.2024.01.007 | |
43. | Ding, C.-J., Lei, Y.-W., Wang, X.-Y. et al. A deep learning interatomic potential suitable for simulating radiation damage in bulk tungsten. Tungsten, 2024, 6(2): 304-322. DOI:10.1007/s42864-023-00230-4 | |
44. | Sowa, J.K., Cadena, D.M., Mehmood, A. et al. IR Spectroscopy of Carboxylate-Passivated Semiconducting Nanocrystals: Simulation and Experiment. Journal of Physical Chemistry C, 2024, 128(21): 8724-8731. DOI:10.1021/acs.jpcc.4c01988 | |
45. | Klatt, M.A., Kim, J., Gartner, T.E. et al. Local number fluctuations in ordered and disordered phases of water across temperatures: Higher-order moments and degrees of tetrahedrality. Journal of Chemical Physics, 2024, 160(20): 204502. DOI:10.1063/5.0204696 | |
46. | Selvaraj, S.C., Koverga, V., Ngo, A.T. Exploring Li-Ion Transport Properties of Li3TiCl6: A Machine Learning Molecular Dynamics Study. Journal of the Electrochemical Society, 2024, 171(5): 050544. DOI:10.1149/1945-7111/ad4ac9 | |
47. | Shi, Y., He, R., Zhang, B. et al. Revisiting the phase diagram and piezoelectricity of lead zirconate titanate from first principles. Physical Review B, 2024, 109(17): 174104. DOI:10.1103/PhysRevB.109.174104 | |
48. | Nagare, B.J., Chacko, S., Kanhere, D.G. Machine-Learned Potential Energy Surfaces for Free Sodium Clusters with Density Functional Accuracy: Applications to Melting. Physica Scripta, 2024, 99(5): 056005. DOI:10.1088/1402-4896/ad37ab | |
49. | Shuai, C., Xu, W., He, H. et al. Layered co-continuous structure in bone scaffold fabricated by laser additive manufacturing for enhancing electro-responsive shape memory properties. Journal of Materials Research and Technology, 2024. DOI:10.1016/j.jmrt.2024.03.050 | |
50. | Agrawal, S., Wang, B., Wu, Y. et al. Photocatalytic activity of dual defect modified graphitic carbon nitride is robust to tautomerism: machine learning assisted ab initio quantum dynamics. Nanoscale, 2024, 16(18): 8986-8995. DOI:10.1039/d4nr00606b | |
51. | Ran, N., Yin, L., Qiu, W. et al. Recent advances in machine learning interatomic potentials for cross-scale computational simulation of materials | [机器学习原子间势在材料跨尺度计算模拟中的最新进展]. Science China Materials, 2024, 67(4): 1082-1100. DOI:10.1007/s40843-023-2836-0 | |
52. | Zhou, L., Tian, P., Zhang, B. et al. Data-driven rational design of single-atom materials for hydrogen evolution and sensing. Nano Research, 2024, 17(4): 3352-3358. DOI:10.1007/s12274-023-6137-5 | |
53. | Ren, F., Wu, Y., Zuo, W. et al. Visualizing the SEI formation between lithium metal and solid-state electrolyte. Energy and Environmental Science, 2024, 17(8): 2743-2752. DOI:10.1039/d3ee03536k | |
54. | Wan, T., Luo, C., Sun, Y. et al. Thermoelastic properties of bridgmanite using deep-potential molecular dynamics. Physical Review B, 2024, 109(9): 094101. DOI:10.1103/PhysRevB.109.094101 | |
55. | Zhang, S., Meng, F., Fu, R. et al. Highly efficient and transferable interatomic potentials for α-iron and α-iron/hydrogen binary systems using deep neural networks. Computational Materials Science, 2024. DOI:10.1016/j.commatsci.2024.112843 | |
56. | Bin Faheem, A., Lee, K.-K. Development of a Neural Network Potential for Modeling Molten LiCl/KCl Salts: Bridging Efficiency and Accuracy. Journal of Physical Chemistry C, 2024, 128(5): 2163-2178. DOI:10.1021/acs.jpcc.3c07010 | |
57. | Gong, Z., Liu, J.Z., Ding, X. et al. Strain-induced two-dimensional relaxor ferroelectrics in Se-doped PbTe. Physical Review B, 2024, 109(5): 054117. DOI:10.1103/PhysRevB.109.054117 | |
58. | Ding, Y., Huang, J. Implementation and Validation of an OpenMM Plugin for the Deep Potential Representation of Potential Energy. International Journal of Molecular Sciences, 2024, 25(3): 1448. DOI:10.3390/ijms25031448 | |
59. | Kempt, R., Kuc, A., Brumme, T. et al. Edge Conductivity in PtSe2 Nanostructures. Small Structures, 2024, 5(2): 2300222. DOI:10.1002/sstr.202300222 | |
60. | Wen, M., Chang, X., Xu, Y. et al. Determining the mechanical and decomposition properties of high energetic materials (α-RDX, β-HMX, and ϵ-CL-20) using a neural network potential. Physical Chemistry Chemical Physics, 2024. DOI:10.1039/d4cp00017j | |
61. | Chen, T., Liu, Q., Liu, Y. et al. Combining stochastic density functional theory with deep potential molecular dynamics to study warm dense matter. Matter and Radiation at Extremes, 2024, 9(1): 015604. DOI:10.1063/5.0163303 | |
62. | Bodenschatz, C.J., Saidi, W.A., Stokes, J.L. et al. Theoretical Prediction of Thermal Expansion Anisotropy for Y2Si2O7 Environmental Barrier Coatings Using a Deep Neural Network Potential and Comparison to Experiment. Materials, 2024, 17(2): 286. DOI:10.3390/ma17020286 | |
63. | Kamaeva, L.V., Tsiok, E.N., Chtchelkatchev, N.M. Deep machine learning, molecular dynamics and experimental studies of liquid Al-Cu-Co alloys. Journal of Molecular Liquids, 2024. DOI:10.1016/j.molliq.2023.123659 | |
64. | Jin, X., Li, H. SONets: Sub-operator learning enhanced neural networks for solving parametric partial differential equations. Journal of Computational Physics, 2023. DOI:10.1016/j.jcp.2023.112536 | |
65. | Liang, F., Ding, J., Wei, X. et al. Interfacial heat and mass transfer at silica/binary molten salt interface from deep potential molecular dynamics. International Journal of Heat and Mass Transfer, 2023. DOI:10.1016/j.ijheatmasstransfer.2023.124705 | |
66. | Liu, A., Wen, T., Han, J. et al. Finite-temperature screw dislocation core structures and dynamics in α-titanium. npj Computational Materials, 2023, 9(1): 228. DOI:10.1038/s41524-023-01181-7 | |
67. | Lan, J., Palizhati, A., Shuaibi, M. et al. AdsorbML: a leap in efficiency for adsorption energy calculations using generalizable machine learning potentials. npj Computational Materials, 2023, 9(1): 172. DOI:10.1038/s41524-023-01121-5 | |
68. | Khazieva, E.O., Shchelkatchev, N.M., Tipeev, A.O. et al. Accuracy, Performance, and Transferability of Interparticle Potentials for Al–Cu Alloys: Comparison of Embedded Atom and Deep Machine Learning Models. Journal of Experimental and Theoretical Physics, 2023, 137(6): 864-877. DOI:10.1134/S1063776123120208 | |
69. | Sun, Y., Wan, K., Shen, W. et al. Modeling Exchange Reactions in Covalent Adaptable Networks with Machine Learning Force Fields. Macromolecules, 2023, 56(21): 9003-9013. DOI:10.1021/acs.macromol.3c01377 | |
70. | Shi, G.-Y., Sun, H.-J., Wang, S.-Y. et al. Simulation studies of the stability and growth kinetics of Pt-Sn phases using a machine learning interatomic potential. Computational Materials Science, 2023. DOI:10.1016/j.commatsci.2023.112388 | |
71. | Gong, Z., Liu, J.Z., Ding, X. et al. Strain-aided room-temperature second-order ferroelectric phase transition in monolayer PbTe: Deep potential molecular dynamics simulations. Physical Review B, 2023, 108(13): 134112. DOI:10.1103/PhysRevB.108.134112 | |
72. | Tipeev, A.O., Ryltsev, R.E., Chtchelkatchev, N.M. et al. Machine learning-assisted MD simulation of melting in superheated AlCu validates the Classical Nucleation Theory. Journal of Molecular Liquids, 2023. DOI:10.1016/j.molliq.2023.122606 | |
73. | Wu, C., Liu, T., Ran, X. et al. Advancing Accurate and Efficient Surface Behavior Modeling of Al Clusters with Machine Learning Potential. Journal of Physical Chemistry C, 2023, 127(38): 19115-19126. DOI:10.1021/acs.jpcc.3c03229 | |
74. | Yu, W., Zhang, Z., Wan, X. et al. High-Accuracy Machine-Learned Interatomic Potentials for the Phase Change Material Ge3Sb6Te5. Chemistry of Materials, 2023, 35(17): 6651-6658. DOI:10.1021/acs.chemmater.3c00524 | |
75. | Guo, Y.-X., Zhuang, Y.-B., Shi, J. et al. ChecMatE: A workflow package to automatically generate machine learning potentials and phase diagrams for semiconductor alloys. Journal of Chemical Physics, 2023, 159(9): 094801. DOI:10.1063/5.0166858 | |
76. | Feng, Y., Feng, Y., Peng, H.-L. Ab initio study of chemical effect on structural properties of Ti-Al melts. Chinese Physics B, 2023, 32(10): 106101. DOI:10.1088/1674-1056/acd2c0 | |
77. | Massaro, A.. Intelligent Materials and Nanomaterials Improving Physical Properties and Control Oriented on Electronic Implementations. Electronics (Switzerland), 2023, 12(18): 3772. DOI:10.3390/electronics12183772 | |
78. | Ayadim, A., Levrel, L., Amokrane, S. Structure and thermodynamics of liquid ruthenium and ruthenium-based alloys from ab initio and classical molecular dynamics with embedded atom model potentials. Modelling and Simulation in Materials Science and Engineering, 2023, 31(6): 065001. DOI:10.1088/1361-651X/acda50 | |
79. | Chtchelkatchev, N.M., Ryltsev, R.E., Magnitskaya, M.V. et al. Local structure, thermodynamics, and melting of boron phosphide at high pressures by deep learning-driven ab initio simulations. Journal of Chemical Physics, 2023, 159(6): 064507. DOI:10.1063/5.0165948 | |
80. | Coletti, M., Sedova, A., Chahal, R. et al. Multiobjective Hyperparameter Optimization for Deep Learning Interatomic Potential Training Using NSGA-II. 2023. DOI:10.1145/3605731.3608931 | |
81. | Piaggi, P.M., Gartner, T.E., Car, R. et al. Melting curves of ice polymorphs in the vicinity of the liquid-liquid critical point. Journal of Chemical Physics, 2023, 159(5): 054502. DOI:10.1063/5.0159288 | |
82. | Zeng, J., Zhang, D., Lu, D. et al. DeePMD-kit v2: A software package for deep potential models. Journal of Chemical Physics, 2023, 159(5): 054801. DOI:10.1063/5.0155600 | |
83. | Andolina, C.M., Saidi, W.A. Highly transferable atomistic machine-learning potentials from curated and compact datasets across the periodic table. Digital Discovery, 2023, 2(4): 1070-1077. DOI:10.1039/d3dd00046j | |
84. | Wang, Z., Sun, Z., Yin, H. et al. The role of machine learning in carbon neutrality: Catalyst property prediction, design, and synthesis for carbon dioxide reduction. eScience, 2023, 3(4): 100136. DOI:10.1016/j.esci.2023.100136 | |
85. | Zhang, C., Feng, Y., Liu, H. et al. Local structures and undercooling ability of Zr–Ti melts. Journal of Non-Crystalline Solids, 2023. DOI:10.1016/j.jnoncrysol.2023.122348 | |
86. | Xie, J.-Z., Jiang, H. Revealing Carbon Vacancy Distribution on α-MoC1-x Surfaces by Machine-Learning Force-Field-Aided Cluster Expansion Approach. Journal of Physical Chemistry C, 2023, 127(27): 13228-13237. DOI:10.1021/acs.jpcc.3c01941 | |
87. | Wan, X., Yu, W., Wang, A. et al. High-Throughput Screening of Gas Sensor Materials for Decomposition Products of Eco-Friendly Insulation Medium by Machine Learning. ACS Sensors, 2023, 8(6): 2319-2330. DOI:10.1021/acssensors.3c00376 | |
88. | Kondratyuk, N., Ryltsev, R., Ankudinov, V. et al. First-principles calculations of the viscosity in multicomponent metallic melts: Al-Cu-Ni as a test case. Journal of Molecular Liquids, 2023. DOI:10.1016/j.molliq.2023.121751 | |
89. | Tao, J.-L., Yang, Z.-Y., Wang, Y.-J. Structural rationale for boson peak in metallic glass informed by an interpretable neural network model. Physical Review Research, 2023, 5(2): 023113. DOI:10.1103/PhysRevResearch.5.023113 | |
90. | Liu, C., Zhang, Z., Ding, J. et al. On the reliability of using reverse Monte Carlo simulations to construct the atomic structure model of metallic glasses. Scripta Materialia, 2023. DOI:10.1016/j.scriptamat.2022.115159 | |
91. | Zeng, J., Tao, Y., Giese, T.J. et al. QDπ: A Quantum Deep Potential Interaction Model for Drug Discovery. Journal of Chemical Theory and Computation, 2023, 19(4): 1261-1275. DOI:10.1021/acs.jctc.2c01172 | |
92. | Gomes-Filho, M.S., Torres, A., Reily Rocha, A. et al. Size and Quality of Quantum Mechanical Data Set for Training Neural Network Force Fields for Liquid Water. Journal of Physical Chemistry B, 2023, 127(6): 1422-1428. DOI:10.1021/acs.jpcb.2c09059 | |
93. | Fidalgo Cândido, V., Matusalem, F., De Koning, M. Melting conditions and entropies of superionic water ice: Free-energy calculations based on hybrid solid/liquid reference systems. Journal of Chemical Physics, 2023, 158(6): 064502. DOI:10.1063/5.0138987 | |
94. | Luo, X., Wang, Z., Gao, P. et al. Review on Machine Learning Accelerated Crystal Structure Prediction | [机器学习加速理论晶体结构预测研究进展]. Kuei Suan Jen Hsueh Pao/Journal of the Chinese Ceramic Society, 2023, 51(2): 552-560. DOI:10.14062/j.issn.0454-5648.20220835 | |
95. | Käser, S., Vazquez-Salazar, L.I., Meuwly, M. et al. Neural network potentials for chemistry: concepts, applications and prospects. Digital Discovery, 2023, 2(1): 28-58. DOI:10.1039/d2dd00102k | |
96. | Pitike, K.C., Setyawan, W. Accurate Fe–He machine learning potential for studying He effects in BCC-Fe. Journal of Nuclear Materials, 2023. DOI:10.1016/j.jnucmat.2022.154183 | |
97. | Wu, Y., Liu, D., Chu, W. et al. Fluctuations at Metal Halide Perovskite Grain Boundaries Create Transient Trap States: Machine Learning Assisted Ab Initio Analysis. ACS Applied Materials and Interfaces, 2022, 14(50): 55753-55761. DOI:10.1021/acsami.2c16203 | |
98. | Gartner, T.E., Piaggi, P.M., Car, R. et al. Liquid-Liquid Transition in Water from First Principles. Physical Review Letters, 2022, 129(25): 255702. DOI:10.1103/PhysRevLett.129.255702 | |
99. | Li, W., Ou, Q., Chen, Y. et al. DeePKS + ABACUS as a Bridge between Expensive Quantum Mechanical Models and Machine Learning Potentials. Journal of Physical Chemistry A, 2022, 126(49): 9154-9164. DOI:10.1021/acs.jpca.2c05000 | |
100. | Zhang, X., Zhang, J., Wang, H. et al. Defect-characterized phase transition kinetics. Applied Physics Reviews, 2022, 9(4): 041311. DOI:10.1063/5.0117234 | |
101. | Matusalem, F., Rego, J.S., de Koning, M. Plastic deformation of superionic water ices. Proceedings of the National Academy of Sciences of the United States of America, 2022, 119(45): e2203397119. DOI:10.1073/pnas.2203397119 | |
102. | He, R., Wu, H., Lu, Y. et al. Origin of negative thermal expansion and pressure-induced amorphization in zirconium tungstate from a machine-learning potential. Physical Review B, 2022, 106(17): 174101. DOI:10.1103/PhysRevB.106.174101 | |
103. | Liu, X., Peng, R., Sun, Z. et al. Moiré Phonons in Magic-Angle Twisted Bilayer Graphene. Nano Letters, 2022, 22(19): 7791-7797. DOI:10.1021/acs.nanolett.2c02010 | |
104. | Pedone, A., Bertani, M., Brugnoli, L. et al. Interatomic potentials for oxide glasses: Past, present, and future. Journal of Non-Crystalline Solids: X, 2022. DOI:10.1016/j.nocx.2022.100115 | |
105. | Zhang, J., Pagotto, J., Duignan, T.T. Towards predictive design of electrolyte solutions by accelerating ab initio simulation with neural networks. Journal of Materials Chemistry A, 2022, 10(37): 19560-19571. DOI:10.1039/d2ta02610d |
System | [ |
Mg [83] | [0.03,0.13] |
Al [83] & Al-Mg [83] | [0.05,0.15] |
Cu [84] | [0.05,0.20] |
Mg-Al-Cu [100] | [0.05,0.20] |
Ti [101] | [0.10,0.25] at |
exploration and [0.15,0.30] elsewhere | |
W [102] | [0.20,0.35] |
Ag-Au [103] | [0.05,0.20] |
water [95] | [0.15,0.25] in first 24 iterations |
[0.18,0.32] in iterations 25 to 32 | |
[0.20,0.35] in iterations 33 to 36 | |
SiC [104] | [0.15,0.30] |
Li10(Ge,Si, or Sn)P2S12 [105] | [0.12,0.25] |
Tm is 1941 K, which is the experimental melting point for Ti. |
System | Reference |
Elemental bulk systems | |
Al | [83, 114-118] |
Mg | [83] |
Cu | [84] |
Ti, W | [101, 102] |
Ag, Au | [103, 119, 120] |
Li | [121] |
Be | [122] |
Ga | [123] |
Sb | [124] |
C | [125] |
Si | [126, 127] |
P | [98] |
Multi-element bulk systems | |
Al-Mg, Al-Cu-Mg | [83, 100, 128, 129] |
Al-Cu, Al-Zn-Mg | [130, 131] |
Al-Cu-Ni | [132] |
Ag-Au | [103, 119] |
Pd-Si, Nb5Si3, Zr77Rh23, Bi2Te3 | [133-136] |
Al90X10 (X = Tb, Cr, or Ce) | [137-140] |
(Pd, Pt) | [141] |
P2Sn5 | [142] |
Silica, silicate | [143-146] |
SiC | [104, 147] |
B4C | [148] |
Molten salt LiF, FLiBe, and chloride | [149-157] |
Li or Na-based battery materials | [105, 158-162] |
TiO2 | [163] |
[164] | |
Ferroelectrics HfO2 | [165] |
Ag2S | [166] |
MoS2 | [167] |
SnSe | [168] |
Zr | [169] |
(Hf0.2Zr0.2Ta0.2Nb0.2Ti0.2)X (X = C or B2) | [170, 171] |
Aqueous systems | |
Water | [95, 172-183] |
Zinc ion in water | [184] |
Water-vapour interface | [185, 186] |
Water-TiO2 interface | [187] |
Ice | [188, 189] |
Molecular systems and clusters | |
Organic molecules | [99, 190-195] |
Metal and alloy clusters | [119, 196] |
Surfaces and low-dimensional systems | |
Metal and alloy surfaces | [103, 119, 129] |
Graphane | [125, 197] |
Monolayer In2Se3 | [198] |
2D Co-Fe-B | [199] |