- Introduction to HamGNN
- Requirements
- Installation
- Usage
- The support for ABACUS software
- Diagonalizing Hamiltonian matrices for large scale systems
- Explanation of the parameters in config.yaml
- Minimum irreps for node and edge features in config.yaml
- References
- Code contributors
- Project leaders
The HamGNN model is an E(3) equivariant graph neural network designed for the purpose of training and predicting tight-binding (TB) Hamiltonians of molecules and solids. Currently, HamGNN can be used in common ab initio DFT software that is based on numerical atomic orbitals, such as OpenMX, Siesta, and Abacus. HamGNN supports predictions of SU(2) equivariant Hamiltonians with spin-orbit coupling effects. HamGNN not only achieves a high fidelity approximation of DFT but also enables transferable predictions across material structures, making it suitable for high-throughput electronic structure calculations and accelerating computations on large-scale systems.
We recommend using the Python 3.9 interpreter. HamGNN needs the following python libraries:
- numpy == 1.21.2
- PyTorch == 1.11.0
- PyTorch Geometric == 2.0.4
- pytorch_lightning == 1.5.10
- e3nn == 0.5.0
- pymatgen == 2022.3.7
- tensorboard == 2.8.0
- tqdm
- scipy == 1.7.3
- yaml
The user can create a Python environment for HamGNN using conda env create -f environment.yaml
.
A another way to set up the Python environment for HamGNN is to use the HamGNN conda environment I have uploaded to this website. Users can simply extract this conda environment directly to their own conda/envs
directory.
HamGNN aims to fit the TB Hamiltonian generated by OpenMX
. The user need to know the basic OpenMX parameters and how to use them properly. OpenMX can be downloaded from this site.
openmx_postprocess is a modified OpenMX package used for computing overlap matrices and other Hamiltonian matrices that can be calculated analytically. The data computed by openmx_postprocess will be stored in a binary file overlap.scfout
. The installation and usage of openmx_postprocess is essentially the same as that of OpenMX. To install openmx_postprocess, you need to install the GSL library first.Then enter the openmx_postprocess directory and modify the following parameters in the makefile:
GSL_lib
: The lib path of GSLGSL_include
: The include path of GSLMKLROOT
: The intel MKL pathCMPLR_ROOT
: The path where the intel compiler is installed
After modifying the makefile, you can directly execute the make command to generate two executable programs, openmx_postprocess
and read_openmx
.
read_openmx is a binary executable that can be used to export the matrices from the binary file overlap.scfout to a file called HS.json
.
Run the following command to install HamGNN:
git clone https://github.com/QuantumLab-ZY/HamGNN.git
cd HamGNN
python setup.py install
Please note that if you have previously installed an older version of HamGNN, you need to uninstall the old version first by using pip uninstall HamGNN
. After uninstalling, there may still be a residual installation package in the site-packages
directory named 'HamGNN-x.x.x-py3.9.egg/HamGNN'
. In this case, you need to manually delete this directory before installing the new version of HamGNN; otherwise, the new version of HamGNN may call functions from the old one.
First, generate a set of structure files (POSCAR or CIF files) using molecular dynamics or random perturbation. After setting the appropriate path parameters in the poscar2openmx.yaml
file,
run poscar2openmx --config path/to/the/poscar2openmx.yaml
to convert these structures into OpenMX's .dat
file format. Run OpenMX to perform static calculations on these structure files and obtain the .scfout
binary files, which store the Hamiltonian and overlap matrix information for each structure. These files serve as the target Hamiltonians during training. Next, run openmx_postprocess
to process each structure and obtain the overlap.scfout
file, which contains the Hamiltonian matrix H0 that is independent of the self-consistent charge density. If the constructed dataset is only used for prediction purposes and not for training (i.e., no target Hamiltonian is needed), run openmx_postprocess
to obtain the overlap.scfout
file merely. openmx_postprocess
is executed similarly to OpenMX and supports MPI parallelism.
After setting the appropriate path information in a graph_data_gen.yaml
file, run graph_data_gen --config graph_data_gen.yaml
to package the structural information and Hamiltonian data from all .scfout
files into a single graph_data.npz
file, which serves as the input data for the HamGNN network.
Prepare the config.yaml
configuration file and set the network parameters, training parameters, and other details in this file. To run HamGNN, simply enter HamGNN --config config.yaml
. Running tensorboard --logdir train_dir
allows real-time monitoring of the training progress, where train_dir
is the folder where HamGNN saves the training data, corresponding to the train_dir
parameter in config.yaml
. To enhance the transferability and prediction accuracy of the network, the training is divided into two steps. The first step involves training with only the loss value of the Hamiltonian in the loss function until the Hamiltonian training converges or the error reaches around 10^-5 Hartree, at which point the training can be stopped. Then, the band energy error is added to the loss function, and the network parameters obtained from the previous step are loaded for further training. After obtaining the final network parameters, the network can be used for prediction. First, convert the structures to be predicted into the input data format (graph_data.npz
) for the network, following similar steps and procedures as preparing the training set. Then, in the config.yaml
file, set the checkpoint_path
to the path of the network parameter file and set the stage
parameter to test
. After configuring the parameters in config.yaml
, running HamGNN --config config.yaml
will perform the prediction.
Several pre-trained models and the config.yaml
file for the test examples are available on Zenodo (https://doi.org/10.5281/zenodo.8147631).
When the training of the Hamiltonian matrix is completed in the first step, it is necessary to use the trained network weights to initialize the HamGNN network and start training for the energy bands. The parameters related to energy band training are as follows:
checkpoint_path
parameter should be set to path of the weight file obtained after training on the Hamiltonian matrix in the first step.- Set
load_from_checkpoint
to True lr
should not be too large, it is recommended to use 0.0001.- In
losses_metrics
andmetrics
, remove the commented section forband_energy
. - Set
calculate_band_energy
to True and specify the parametersnum_k
,band_num
, andk_path
.
After setting the above parameters, start the training again.
Set the parameters in band_cal.yaml, mainly the path to the Hamiltonian data, then run band_cal --config band_cal.yaml
The utilities to support ABACUS software have been uploaded in the utils_abacus
directory. Users need to modify the parameters in the scripts within this directory. The code for abacus_postprocess
in utils_abacus/abacus_H0_export
is derived from modifying the abacus
program based on ABACUS-3.5.3. The function of this tool is similar to openmx_postprocess
and it is used to export the Hamiltonian part H0
, which is independent of the self-consistent field (SCF) charge density. Compilation of abacus-postprocess
is the same as that of the original ABACUS.
poscar2abacus.py
and graph_data_gen_abacus.py
scripts are respectively utilized for generating ABACUS structure files and packaging the Hamiltonian matrix into the graph_data.npz
file. Users can explore the usage of these tools independently. Later on, I'll briefly introduce the meanings of the parameters within these scripts.
For crystal structures containing thousands of atoms, diagonalizing the Hamiltonian matrix using the serial band_cal
script can be quite challenging. To address this, we've introduced a multi-core parallel band_cal_parallel
script within band_cal_parallel directory. Note: In certain MKL environments, using the band_cal_parallel
may trigger a bug that reports the error message 'Intel MKL FATAL ERROR: Cannot load symbol MKLMPI_Get_wrappers'. Users can try the solutions provided in Issues #18 and #12 to resolve this issue (thanks to the help from flamingoXu
and newplay
).
pip install mpitool-0.0.1-cp39-cp39-manylinux1_x86_64.whl
pip install band_cal_parallel-0.1.12-py3-none-any.whl
In the Python environment with band_cal_parallel
installed, execute the following command with multiple cpus to compute the band structure:
mpirun -np ncpus band_cal_parallel --config band_cal_parallel.yaml
The input parameters in config.yaml are divided into different modules, which mainly include 'setup'
, 'dataset_params'
, 'losses_metrics'
, 'optim_params'
and network-related parameters ('HamGNN_pre'
and 'HamGNN_out'
). Most of the parameters work well using the default values. The following introduces some commonly used parameters in each module.
-
setup
:stage
: Select the state of the network: training (fit
) or testing (test
).GNN_Net
: UseHamGNN_pre
for normal Hamiltonian fitting and use 'HamGNN_pre_charge' for fitting the Hamiltonian of charged defects.property
:Select the type of physical quantity to be output by the network, generally set tohamiltonian
num_gpus
: number of gpus to train on (int
) or which GPUs to train on (list
orstr
) applied per node.resume
: resume training (true
) or start from scratch (false
).checkpoint_path
: Path of the checkpoint from which training is resumed (stage
=fit
) or path to the checkpoint you wish to test (stage
=test
).load_from_checkpoint
: If set totrue
, the model will be initialized with weights from the checkpoint_path.
-
dataset_params
:graph_data_path
: The directory where the processed compressed graph data files (grah_data.npz
) are stored.batch_size
: The number of samples or data points that are processed together in a single forward and backward pass during the training of a neural network. defaut: 1.train_ratio
: The proportion of the training samples in the entire data set.val_ratio
: The proportion of the validation samples in the entire data set.test_ratio
:The proportion of the test samples in the entire data set.
-
losses_metrics
:losses
: define multiple loss functions and their respective weights in the total loss value. Currently, HamGNN supportsmse
,mae
, andrmse
.metrics
:A variety of metric functions can be defined to evaluate the accuracy of the model on the validation set and test set.
-
optim_params
:min_epochs
: Force training for at least these many epochs.max_epochs
: Stop training once this number of epochs is reached.lr
:learning rate, the default value is 0.001.
-
profiler_params
:train_dir
: The folder for saving training information and prediction results. This directory can be read by tensorboard to monitor the training process.
-
HamGNN_pre
: The representation network to generate the node and pair interaction featuresnum_types
:The maximum number of atomic types used to build the one-hot vectors for atomscutoff
: The cutoff radius adopted in the envelope function for interatomic distances.cutoff_func
: which envelope function is used for interatomic distances. Options:cos
refers to cosine envelope function,pol
refers to the polynomial envelope function.rbf_func
: The radial basis function type used to expand the interatomic distancesnum_radial
: The number of Bessel basis.num_interaction_layers
: The number of interaction layers or orbital convolution layers.add_edge_tp
: Whether to utilize the tensor product of i and j to construct pair interaction features. This option requires a significant amount of memory, but it can sometimes improve accuracy.irreps_edge_sh
: Spherical harmonic representation of the orientation of an edgeirreps_node_features
: O(3) irreducible representations of the initial atomic featuresirreps_edge_output
: O(3) irreducible representations of the edge features to outputirreps_node_output
: O(3) irreducible representations of the atomic features to outputfeature_irreps_hidden
: intermediate O(3) irreducible representations of the atomic features in convelutionirreps_triplet_output(deprecated)
: O(3) irreducible representations of the triplet features to outputinvariant_layers
: The layers of the MLP used to map the invariant edge embeddings to the weights of each tensor product pathinvariant_neurons
: The number of the neurons of the MLP used to map the invariant edge embeddings to the weights of each tensor product pathnum_charge_attr_feas
: The number of features used for doping charge whenGNN_Net
is set to 'HamGNN_pre_charge'
-
HamGNN_out
: The output layer to transform the representation of crystals into Hamiltonian matrixnao_max
: It is modified according to the maximum number of atomic orbitals in the data set, which can be14
,19
,26
.For short-period elements such as C, Si, O, etc., a nao_max of 14 is sufficient; the number of atomic bases for most common elements does not exceed 19. Setting nao_max to 26 would allow the description of all elements supported by OpenMX. For the Hamiltonian of ABACUS,nao_max
can be set to either27
(without Al, Hf, Ta, W) or40
(supporting all elements in ABACUS).add_H0
: Generally true, the complete Hamiltonian is predicted as the sum of H_scf plus H_nonscf (H0)symmetrize
:if set to true, the Hermitian symmetry constraint is imposed on the Hamiltoniancalculate_band_energy
: Whether to calculate the energy bands to train the modelnum_k
: When calculating the energy bands, the number of K points to useband_num_control
:dict
: controls how many orbitals are considered for each atom in energy bands;int
: [vbm-num, vbm+num];null
: all bandsk_path
:auto
: Automatically determine the k-point path;null
: random k-point path;list
: list of k-point paths provided by the usersoc_switch
: if true, Fit the SOC Hamiltoniannonlinearity_type
:norm
activation orgate
activation as the nonlinear activation function
from e3nn import o3
row=col=o3.Irreps("1x0e+1x0e+1x0e+1x1o+1x1o+1x2e+1x2e") # for 'sssppd'
ham_irreps_dim = []
ham_irreps = o3.Irreps()
for _, li in row:
for _, lj in col:
for L in range(abs(li.l-lj.l), li.l+lj.l+1):
ham_irreps += o3.Irrep(L, (-1)**(li.l+lj.l))
print(ham_irreps.sort()[0].simplify())
Output: 17x0e+20x1o+8x1e+8x2o+20x2e+8x3o+4x3e+4x4e
The papers related to HamGNN:
[1] Transferable equivariant graph neural networks for the Hamiltonians of molecules and solids
[2] Universal Machine Learning Kohn-Sham Hamiltonian for Materials
[4] Topological interfacial states in ferroelectric domain walls of two-dimensional bismuth
[5] Transferable Machine Learning Approach for Predicting Electronic Structures of Charged Defects
- Yang Zhong (Fudan University)
- Changwei Zhang (Fudan University)
- Zhenxing Dai (Fudan University)
- Shixu Liu (Fudan University)
- Hongyu Yu (Fudan University)
- Yuxing Ma (Fudan University)
- Hongjun Xiang (Fudan University)
- Xingao Gong (Fudan University)