toplogo
Đăng nhập

Cartesian Atomic Cluster Expansion for Machine Learning Interatomic Potentials


Khái niệm cốt lõi
Proposing a Cartesian-based atomic density expansion as an alternative to spherical harmonics for accurate and efficient interatomic potentials.
Tóm tắt

Machine learning interatomic potentials are transforming material science and chemistry by utilizing atomic cluster expansion or message passing frameworks. The proposed Cartesian Atomic Cluster Expansion (CACE) offers a complete set of independent features while integrating low-dimensional embeddings and inter-atomic message passing. CACE demonstrates accuracy, stability, and generalizability across diverse systems like bulk water, small molecules, and high-entropy alloys. The method avoids the complexities of spherical harmonics and Clebsch-Gordan contraction, providing a more efficient approach to symmetrization in Cartesian coordinates. By incorporating message passing mechanisms, CACE enhances the predictive capabilities of the potential model. The framework is stable, scalable, and capable of extrapolating to unseen elements or high temperatures with impressive accuracy.

edit_icon

Customize Summary

edit_icon

Rewrite with AI

edit_icon

Generate Citations

translate_icon

Translate Source

visual_icon

Generate MindMap

visit_icon

Visit Source

Thống kê
rcut = 5.5 ˚A, 6 Bessel radial functions c = 12, lmax = 3, νmax = 3, Nembedding = 3 Training time on GeForce GTX 1080 Ti GPU: about two days for water dataset CACE model parameters: 24,572 trainable parameters for T=0 model; 69,320 trainable parameters for T=1 model MD simulations on Nvidia A100 GPU: varying system sizes run at different speeds per hour
Trích dẫn
"The resulting potential named Cartesian Atomic Cluster Expansion (CACE) exhibits good accuracy, stability, and generalizability." "CACE descriptors are low-dimensional and can be computed independently and efficiently." "CACE potential shows high stability and extrapolatability in various datasets."

Thông tin chi tiết chính được chắt lọc từ

by Bingqing Che... lúc arxiv.org 03-19-2024

https://arxiv.org/pdf/2402.07472.pdf
Cartesian atomic cluster expansion for machine learning interatomic  potentials

Yêu cầu sâu hơn

How does the efficiency of the CACE potential compare to other MLIP methods in terms of training time

The efficiency of the Cartesian Atomic Cluster Expansion (CACE) potential in terms of training time is noteworthy compared to other Machine Learning Interatomic Potentials (MLIPs). The training time for CACE on datasets like liquid water, small molecules, and high-entropy alloys is relatively short. For instance, training the CACE potential for the water dataset took about two days on a GeForce GTX 1080 Ti GPU card. Similarly, small molecule datasets can be trained on a laptop within one or two days. This indicates that CACE has a competitive advantage in terms of training speed when compared to other MLIP methods.

What are the implications of using low-dimensional embeddings in CACE for scalability and generalization

The use of low-dimensional embeddings in CACE has significant implications for scalability and generalization. By incorporating low-dimensional embeddings for various chemical elements, CACE reduces the complexity of feature representations while maintaining essential information about atomic environments. This streamlined approach not only enhances computational efficiency but also improves scalability by reducing memory requirements during model training and inference. Additionally, these compact embeddings contribute to better generalization capabilities across diverse material systems as they capture essential chemical information without unnecessary complexity.

How might the alchemical learning capability of CACE impact its applicability across diverse material systems

The alchemical learning capability embedded within the CACE potential opens up new possibilities for its applicability across diverse material systems. Alchemical learning allows the model to generalize well beyond its initial training data by extrapolating to unseen elements or configurations with minimal additional tuning or retraining efforts. In practical terms, this means that CACE can adapt effectively to novel scenarios or materials where traditional models might struggle due to lack of prior exposure during training. This flexibility makes CACE a versatile tool for exploring complex material properties and behaviors across different domains with enhanced accuracy and stability.
0
star