A Comparative Study of Machine Learning Models Predicting Energetics of Interacting Defects
Abstract
Interacting defect systems are ubiquitous in materials under realistic scenarios, yet gaining an atomic-level understanding of these systems from a computational perspective is challenging - it often demands substantial resources due to the necessity of employing supercell calculations. While machine learning techniques have shown potential in accelerating materials simulations, their application to systems involving interacting defects remains relatively rare. In this work, we present a comparative study of three different methods to predict the free energy change of systems with interacting defects. We leveraging a limited dataset from Density Functional Theory(DFT) calculations to assess the performance models using materials descriptors, graph neural networks and cluster expansion. Our findings indicate that the cluster expansion model can achieve precise energetics predictions even with this limited dataset. Furthermore, with synthetic data generate from cluster expansion model at near-DFT levels, we obtained enlarged dataset to assess the data requirements for training accurate prediction models using graph neural networks for systems featuring interacting defects. A brief discussion of the computational cost for each method is provided at the end. This research provide a preliminary evaluation of applying machine learning techniques in imperfect surface systems.
1 Introduction
Defects in materials profoundly influence their properties. Understanding defects interactions is essential for optimizing material properties. Phenomena arising from defect interactions can occur from atomic-level to mesoscale and modeling across the scales is computationally challenging. Recently, machine learning potentials have shown success in accelerating materials simulations for molecules and crystals[Noé et al., 2020]. However, surface systems with a number of defects are rarely studied. Except defects interactions are complex, simulations of surface systems are costly since calculation for large supercells are needed [Sun and Ceder, 2013]. These prevent the progress of studies on complex phenomena in surface systems.
To explore the potential of applying machine learning techniques to study interacting defects on surfaces at atomic level, we provided as case study of single-type defects on pure element crystal. Driven from possible applications in real-life problem related to lithium battery[Liu and Lu, 2017], we focus on different numbers of vacancies on (100) surface of lithium (Figure 1). We defined the defect concentration on surface as
(1) |
The prediction task is, given the structure of defects distribute on surface to predict the free energy change when vacancies present on the surface. This free energy change is defined as
(2) |
where is the ground state energy of lithium surface with defects from Density Functional Theory calculation(DFT), is the number of vacancies in this system and is the chemical potential of lithium (atomic energy of lithium in its most stable form), and is the ground state energy of a pristine lithium supercell.
Our initial data set calculated by DFT has 88 structures with concentration from 6.25% to 50%. We split the training set and test set by concentration to make the training set and test set dissimilar both in configurations and concentrations. The training set contains structures with concentration less than 40%, and there are 73 structures. The test set contains the rest 15 structures with concentration higher than 40%. We first study the model performance on initial test data with three possible approaches. Cluster expansion(CE) is the the approach we showed it can predict the target at high accuracy. We further generate more surface systems with different defect configurations and their energy using this cluster expansion model trained with initial data, and use these configuration-energy pairs to test the potential of applying graph neural network(GNN) in this type of systems. Contributions of this work include (1) Qualitative reveal the limitation of machine learning potentials for systems with interacting defects and provide a model to alleviate; (2) Establish a computationally feasible way to test the potential of GNN for surface with intereacting defects.

2 Methodology
2.1 Atomic Representations
Within the machine learning framework of predicting property of a material system , where is the matrix of atomic positions and is atomic numbers, it is assumed that the property has contribution from each atom in the system[Bader, 1991]. It can be formulated into regression task in two ways, (1) transform to descriptors and use ML algorithms learn from these representations; (2) transform as a graph with nodes and edges represented in adjacency matrix and then apply graph neural network. Coulomb matrix (CM), Smooth Overlap of Atomic Positions(SOAP) and Many-body Tensor Representation (MBTR) are commonly used descriptors [Laakso et al., 2023, Himanen et al., 2020]. Descriptors are encoded and limited within domain-knowledge from experts, it is effective when training data is little. Graphs are naturally good representation of atomic structures, and graph neural networks are used for tasks where graph-structured data has inherent relationships. To obtain a good GNN, large amount of data related to this task is needed. We used the framework provided by the open catalyst project[Chanussot* et al., 2021] conduct experiments.
2.2 Configurational Cluster Expansion
When the property of a material system is related to the configurational degrees of freedom, configurational cluster expansion is a way to build a model that can extrapolate predictions to unseen configurations. In our case, The lithium surface systems we consider only differs at surface configurations and there are configurations, is the sites of perfect crystal on surface. This collection is denoted as The energy dependency with configurations in cluster expansion can be formulated with where refers to the point, pair, triplet, etc. clusters within the crystal, is the effective interaction strength of a cluster and is the crystal basis function. We can obtain these quantities when we construct a cluster expansion model provided few structures in the configurational space[Ångqvist et al., 2019].
3 Results



MAE (Train/Test in ) | |||
---|---|---|---|
Model Name/ Layers | 2 | 4 | 6 |
MBTR | 0.006/0.032 | 0.012/0.038 | 0.014/0.147 |
DimeNet | 0.020/0.117 | 0.028/0.138 | 0.117/0.219 |
Cluster Expansion | 0.007/0.021 | 0.007/0.024 | 0.007/0.024 |
The models were trained on systems with low defect concentrations and then tested on systems with high defect concentrations. We explored 3 kinds approaches: (1) descriptors + ML algorithms, (2) graph neural networks, (3) cluster expansion. For the first two approaches, we presented the prediction statistics of the most representative models, they are (1) MBTR[Huo and Rupp, 2022] with linear regression, (2) DimeNetGasteiger et al. [2022]. We tested the best number of atoms to include in representation and the results in mean absolute error(MAE) and coefficient of determination() are in Table 1 and Figure 2. All the models training with representations building from 2 layers of atoms, which are surface and subsurface atoms, have the lowest test MAE. Cluster expansion model can predict the energy with high concentration defects most accurately and achieving negligible error compare to DFT level accuracy. Linear regression model with MBTR also has low test MAE, however, the prediction error is higher when the concentration of defect is higher. DimeNet has the worst performance compare to the last two models, this is due to the training data is limited. Since the most important factor that impact the energy of this kind of extended system is surface configuration, when building a prediction model, information about the atoms far from surface makes less contribution in determining the free energy change of the system. This match well with the results in Figure 1 comparing column-wise. Notably, in graph neural network, if we represented the supercell structures as graphs while only surfaces differ, the model will treat every structure the same.
Since cluster expansion model is superior in prediction accuracy and uncertainty, we use this cluster expansion model to generate more target energies with different surface configurations at low concentrations and support the training of DimeNet. Using the energies predicted by cluster expansion as surrogates, it bypass the need of computational cost for thousands of supercell of DFT calculationsto test the potential of DimeNet applied at systems with interacting defects. We trained DimeNet with 500 and 5,000 CE generated energies respectively and the results are in Figure 3 and Table 2. As there are more training data points, DimeNet gains more expressive representation. When training with 5,000 configuration, DimeNet is able to predict as good as MBTR designed by domain expert. This finding motivate us to do further studies on how DimeNet captures relationship between defects and interactions by utilizing more low concentration defect configurations and whether the representation generalize to high concentration systems.


Number of Training Configurations | 73 | 500 | 5000 |
---|---|---|---|
Test MAE () | 0.117 | 0.052 | 0.033 |
4 Conclusion
In this work, we explored three approaches to predict the free energy change of surface systems with interacting defects with a small data set. While materials descriptor with classical machine learning algorithm performs well with limited data, cluster expansion is more accurate and less biased in prediction. Within the scope of interacting defects scattered on surface, cluster expansion model can serve as a means to provide training data at low cost to survey the potential of applying deep neural networks in the task of property prediction for surface with interacting defects. On the other hand, it is worth considering the necessity of using machine learning with descriptors or deep learning to study the defects interactions at surface only when cluster expansion model performs well with purchasable cost to obtain training data.
5 Disclaimer
This project was conducted independently during my early stage of graduate studies and may not have received formal supervision. If you come across this report, please feel free to contact me with any issues or suggestions you may have.
References
- Noé et al. [2020] Frank Noé, Alexandre Tkatchenko, Klaus-Robert Müller, and Cecilia Clementi. Machine learning for molecular simulation. Annual Review of Physical Chemistry, 71(1):361–390, 2020. doi: 10.1146/annurev-physchem-042018-052331. URL https://doi.org/10.1146/annurev-physchem-042018-052331. PMID: 32092281.
- Sun and Ceder [2013] Wenhao Sun and Gerbrand Ceder. Efficient creation and convergence of surface slabs. Surface Science, 617:53–59, 2013. ISSN 0039-6028. doi: https://doi.org/10.1016/j.susc.2013.05.016. URL https://www.sciencedirect.com/science/article/pii/S003960281300160X.
- Liu and Lu [2017] Guangyu Liu and Wei Lu. A model of concurrent lithium dendrite growth, sei growth, sei penetration and regrowth. Journal of The Electrochemical Society, 164(9):A1826, jun 2017. doi: 10.1149/2.0381709jes. URL https://dx.doi.org/10.1149/2.0381709jes.
- Bader [1991] Richard F. W. Bader. A quantum theory of molecular structure and its applications. Chemical Reviews, 91(5):893–928, 1991. doi: 10.1021/cr00005a013. URL https://doi.org/10.1021/cr00005a013.
- Laakso et al. [2023] Jarno Laakso, Lauri Himanen, Henrietta Homm, Eiaki V. Morooka, Marc O. J. Jäger, Milica Todorović, and Patrick Rinke. Updates to the DScribe library: New descriptors and derivatives. The Journal of Chemical Physics, 158(23):234802, 06 2023. ISSN 0021-9606. doi: 10.1063/5.0151031. URL https://doi.org/10.1063/5.0151031.
- Himanen et al. [2020] Lauri Himanen, Marc O.J. Jäger, Eiaki V. Morooka, Filippo Federici Canova, Yashasvi S. Ranawat, David Z. Gao, Patrick Rinke, and Adam S. Foster. Dscribe: Library of descriptors for machine learning in materials science. Computer Physics Communications, 247:106949, 2020. ISSN 0010-4655. doi: https://doi.org/10.1016/j.cpc.2019.106949. URL https://www.sciencedirect.com/science/article/pii/S0010465519303042.
- Chanussot* et al. [2021] Lowik Chanussot*, Abhishek Das*, Siddharth Goyal*, Thibaut Lavril*, Muhammed Shuaibi*, Morgane Riviere, Kevin Tran, Javier Heras-Domingo, Caleb Ho, Weihua Hu, Aini Palizhati, Anuroop Sriram, Brandon Wood, Junwoong Yoon, Devi Parikh, C. Lawrence Zitnick, and Zachary Ulissi. Open catalyst 2020 (oc20) dataset and community challenges. ACS Catalysis, 2021. doi: 10.1021/acscatal.0c04525.
- Ångqvist et al. [2019] Mattias Ångqvist, William A. Muñoz, J. Magnus Rahm, Erik Fransson, Céline Durniak, Piotr Rozyczko, Thomas H. Rod, and Paul Erhart. Icet – a python library for constructing and sampling alloy cluster expansions. Advanced Theory and Simulations, 2(7):1900015, 2019. doi: https://doi.org/10.1002/adts.201900015. URL https://onlinelibrary.wiley.com/doi/abs/10.1002/adts.201900015.
- Huo and Rupp [2022] Haoyan Huo and Matthias Rupp. Unified representation of molecules and crystals for machine learning. Machine Learning: Science and Technology, 3(4):045017, nov 2022. doi: 10.1088/2632-2153/aca005. URL https://dx.doi.org/10.1088/2632-2153/aca005.
- Gasteiger et al. [2022] Johannes Gasteiger, Janek Groß, and Stephan Günnemann. Directional message passing for molecular graphs, 2022.