A collection of Neural Network Models for chemistry
- Quantum Chemistry Method
- Force Field Method
- Semi-Empirical Method
- Coarse-Grained Method
- Enhanced Sampling Method
- QM/MM Method
- Charge Method
-
DeePKS, DeePHF
DeePKS-kit is a program to generate accurate energy functionals for quantum chemistry systems, for both perturbative scheme (DeePHF) and self-consistent scheme (DeePKS). -
NeuralXC
Implementation of a machine-learned density functional. -
MOB-ML
Machine Learning for Molecular Orbital Theory. -
DM21
Pushing the Frontiers of Density Functionals by Solving the Fractional Electron Problem. -
NN-GGA, NN-NRA, NN-meta-GGA, NN-LSDA
Completing density functional theory by machine-learning hidden messages from molecules. -
FemiNet
FermiNet is a neural network for learning highly accurate ground state wavefunctions of atoms and molecules using a variational Monte Carlo approach. -
DeePQMC
DeepQMC implements variational quantum Monte Carlo for electrons in molecules, using deep neural networks written in PyTorch as trial wave functions. -
PauliNet
PauliNet builds upon HF or CASSCF orbitals as a physically meaningful baseline and takes a neural network approach to the SJB wavefunction in order tocorrect this baseline towards a high-accuracy solution. -
DeePErwin
DeepErwin is python package that implements and optimizes wave function models for numerical solutions to the multi-electron Schrödinger equation. -
Jax-DFT
JAX-DFT implements one-dimensional density functional theory (DFT) in JAX. It uses powerful JAX primitives to enable JIT compilation, automatical differentiation, and high-performance computation on GPUs. -
sns-mp2
Improving the accuracy of Moller-Plesset perturbation theory with neural networks -
DeepH-pack
Deep neural networks for density functional theory Hamiltonian. -DeepH-E3
General framework for E(3)-equivariant neural network representation of density functional theory Hamiltonian -
kdft
The Kernel Density Functional (KDF) code allows generating ML-based DFT functionals. -
ML-DFT
ML-DFT: Machine learning for density functional approximations This repository contains the implementation for the kernel ridge regression based density functional approximation method described in the paper "Quantum chemical accuracy from density functional approximations via machine learning". -
D4FT
this work proposed a deep-learning approach to KS-DFT. First, in contrast to the conventional SCF loop, directly minimizing the total energy by reparameterizing the orthogonal constraint as a feed-forward computation. They prove that such an approach has the same expressivity as the SCF method yet reduces the computational complexity from O(N^4) to O(N^3) -
SchOrb
Unifying machine learning and quantum chemistry with a deep neural network for molecular wavefunctions
- CiderPress
Tools for training and evaluating CIDER functionals for use in Density Functional Theory calculations.
- ML-RPA
This work demonstrates how machine learning can extend the applicability of the RPA to larger system sizes, time scales, and chemical spaces. - ΔOF-MLFF
a Δ-machine learning model for obtaining Kohn–Sham accuracy from orbital-free density functional theory (DFT) calculations
-
PairNet
A molecular orbital based machine learning model for predicting accurate CCSD(T) correlation energies. The model, named as PairNet, shows excellent transferability on several public data sets using features inspired by pair natural orbitals(PNOs). -
SPAHM(a,b)
SPAHM(a,b): encoding the density information from guess Hamiltonian in quantum machine learning representations -
GradDFT
GradDFT is a JAX-based library enabling the differentiable design and experimentation of exchange-correlation functionals using machine learning techniques. -
lapnet
A JAX implementation of the algorithm and calculations described in Forward Laplacian: A New Computational Framework for Neural Network-based Variational Monte Carlo. -
M-OFDFT
M-OFDFT is a deep-learning implementation of orbital-free density functional theory that achieves DFT-level accuracy on molecular systems but with lower cost complexity, and can extrapolate to much larger molecules than those seen during training
- ANN for Schrodinger
Artificial neural networks (NN) are universal function approximators and have shown great ability in computing the ground state energy of the electronic Schrödinger equation, yet NN has not established itself as a practical and accurate approach to solving the vibrational Schrödinger equation for realistic polyatomic molecules to obtain vibrational energies and wave functions for the excited states - equivariant_electron_density
Generate and predict molecular electron densities with Euclidean Neural Networks - DeePDFT
This is the official Implementation of the DeepDFT model for charge density prediction. - DFA_recommeder
System-specific density functional recommender - EG-XC
The accuracy of density functional theory hinges on the approximation of nonlocal contributions to the exchange-correlation (XC) functional. To date, machine-learned and human-designed approximations suffer from insufficient accuracy, limited scalability, or dependence on costly reference data. To address these issues, we introduce Equivariant Graph Exchange Correlation (EG-XC), a novel non-local XC functional based on equivariant graph neural network - scdp
Machine learning methods are promising in significantly accelerating charge density prediction, yet existing approaches either lack accuracy or scalability. They propose a recipe that can achieve both. In particular, they identify three key ingredients: (1) representing the charge density with atomic and virtual orbitals (spherical fields centered at atom/virtual coordinates); (2) using expressive and learnable orbital basis sets (basis function for the spherical fields); and (3) using high-capacity equivariant neural network architecture - physics-informed-DFT
We have developed an approach for physics-informed training of flexible empirical density functionals. In this approach, the “physics knowledge” is transferred from PBE, or any other exact-constraints-based functional, using local exchange−correlation energy density regularization, i.e., by adding its local energies into the training set
- DeepGreen
The many-body Green's function provides access to electronic properties beyond density functional theory level in ab inito calculations. It present proof-of-concept benchmark results for both molecules and simple periodic systems, showing that our method is able to provide accurate estimate of physical observables such as energy and density of states based on the predicted Green's function.
- wigner_kernel
They propose a novel density-based method which involves computing “Wigner kernels”.
- DeePMD
A package designed to minimize the effort required to build deep learning based model of interatomic potential energy and force field and to perform molecular dynamics. - Torch-ANI
TorchANI is a pytorch implementation of ANI model. - mdgrad
Pytorch differentiable molecular dynamics - PESPIP
Mathematica programs for choosing the best basis of permutational invariant polynomials for fitting a potential energy surface - Schrodinger-ANI
A neural network potential energy function for use in drug discovery, with chemical element support extended from 41% to 94% of druglike molecules based on ChEMBL. - NerualForceFild
The Neural Force Field (NFF) code is an API based on SchNet, DimeNet, PaiNN and DANN. It provides an interface to train and evaluate neural networks for force fields. It can also be used as a property predictor that uses both 3D geometries and 2D graph information. - NNPOps
The goal of this project is to promote the use of neural network potentials (NNPs) by providing highly optimized, open-source implementations of bottleneck operations that appear in popular potentials. - RuNNer
A program package for constructing high-dimensional neural network potentials,4G-HDNNPs,3G-HDNNPs. - aenet
The Atomic Energy NETwork (ænet) package is a collection of tools for the construction and application of atomic interaction potentials based on artificial neural networks. - sGDML
Symmetric Gradient Domain Machine Learning - GAP
This package is part of QUantum mechanics and Interatomic Potentials - QUIP
The QUIP package is a collection of software tools to carry out molecular dynamics simulations. It implements a variety of interatomic potentials and tight binding quantum mechanics, and is also able to call external packages, and serve as plugins to other software such as LAMMPS, CP2K and also the python framework ASE. - NNP-MM
NNP/MM embeds a Neural Network Potential into a conventional molecular mechanical (MM) model. - GAMD
Data and code for Graph neural network Accelerated Molecular Dynamics. - PFP
Here we report a development of universal NNP called PreFerred Potential (PFP), which is able to handle any combination of 45 elements. Particular emphasis is placed on the datasets, which include a diverse set of virtual structures used to attain the universality. - TeaNet
universal neural network interatomic potential inspired by iterative electronic relaxations. - n2p2
This repository provides ready-to-use software for high-dimensional neural network potentials in computational physics and chemistry. - AIMNET
This repository contains reference AIMNet implementation along with some examples and menchmarks. - AIMNet2
A general-purpose neural netrork potential for organic and element-organic molecules. - aevmod
This package provides functionality for computing an atomic environment vector (AEV), as well as its Jacobian and Hessian. - charge3net
Official implementation of ChargeE3Net, introduced in Higher-Order Equivariant Neural Networks for Charge Density Prediction in Materials. - jax-nb
This is a JAX implementation of Polarizable Charge Equilibrium (PQEq) and DFT-D3 dispersion correction.
-
Nequip
NequIP is an open-source code for building E(3)-equivariant interatomic potentials. -
E3NN
Euclidean neural networks,The aim of this library is to help the development of E(3) equivariant neural networks. It contains fundamental mathematical operations such as tensor products and spherical harmonics. -
SchNet
SchNet is a deep learning architecture that allows for spatially and chemically resolved insights into quantum-mechanical observables of atomistic systems. -
SchNetPack
SchNetPack aims to provide accessible atomistic neural networks that can be trained and applied out-of-the-box, while still being extensible to custom atomistic architectures. containsschnet
,painn
,filedschnet
,so3net
-
XequiNet XequiNet is an equivariant graph neural network for predicting the properties of chemical molecules or periodical systems.
-
G-SchNet
Implementation of G-SchNet - a generative model for 3d molecular structures. -
PhysNet
PhysNet: A Neural Network for Predicting Energies, Forces, Dipole Moments and Partial Charges. -
DimeNet
Directional Message Passing Neural Network. -
GemNet
Universal Directional Graph Neural Networks for Molecules. -
DeePMoleNet
DeepMoleNet is a deep learning package for molecular properties prediction. -
AirNet
A new GNN-based deep molecular model by MindSpore. -
TorchMD-Net
TorchMD-NET provides graph neural networks and equivariant transformer neural networks potentials for learning molecular potentials. -
AQML
AQML is a mixed Python/Fortran/C++ package, intends to simulate quantum chemistry problems through the use of the fundamental building blocks of larger systems. -
TensorMol
A pakcages of NN model chemistry, contains Behler-Parrinello with electrostatics, Many Body Expansion Bonds in Molecules NN, Atomwise, Forces, Inductive Charges. -
charge_transfer_nnp
About Graph neural network potential with charge transfer with nequip model. -
AMP
Amp: A modular approach to machine learning in atomistic simulations(https://github.com/ulissigroup/amptorch) -
SCFNN
A self consistent field neural network (SCFNN) model. -
jax-md
JAX MD is a functional and data driven library. Data is stored in arrays or tuples of arrays and functions transform data from one state to another. -
EANN
Embedded Atomic Neural Network (EANN) is a physically-inspired neural network framework. The EANN package is implemented using the PyTorch framework used to train interatomic potentials, dipole moments, transition dipole moments and polarizabilities of various systems. -
REANN
Recursively embedded atom neural network (REANN) is a PyTorch-based end-to-end multi-functional Deep Neural Network Package for Molecular, Reactive and Periodic Systems. -
FIREANN
Field-induced Recursively embedded atom neural network (FIREANN) is a PyTorch-based end-to-end multi-functional Deep Neural Network Package for Molecular, Reactive and Periodic Systems under the presence of the external field with rigorous rotational equivariance. -
MDsim
Training and simulating MD with ML force fields -
ForceNet
We demonstrate that force-centric GNN models without any explicit physical constraints are able to predict atomic forces more accurately than state-of-the-art energy centric GNN models, while being faster both in training and inference. -
DIG
A library for graph deep learning research. -
scn
Spherical Channels for Modeling Atomic Interactions -
spinconv
Rotation Invariant Graph Neural Networks using Spin Convolutions. -
HIPPYNN
a modular library for atomistic machine learning with pytorch. -
VisNet
a scalable and accurate geometric deep learning potential for molecular dynamics simulation -
flare
FLARE is an open-source Python package for creating fast and accurate interatomic potentials.) -
alignn
The Atomistic Line Graph Neural Network (https://www.nature.com/articles/s41524-021-00650-1) introduces a new graph convolution layer that explicitly models both two and three body interactions in atomistic systems. -
So3krates
Repository for training, testing and developing machine learned force fields using the So3krates model. -
spice-model-five-net
Contains the five equivariant transformer models about the spice datasets(https://github.com/openmm/spice-dataset/releases/tag/1.1). -
sake
Spatial Attention Kinetic Networks with E(n)-Equivariance -
eqgat
Pytorch implementation for the manuscript Representation Learning on Biomolecular Structures using Equivariant Graph Attention -
phast
PyTorch implementation for PhAST: Physics-Aware, Scalable and Task-specific GNNs for Accelerated Catalyst Design -
GNN-LF
Graph Neural Network With Local Frame for Molecular Potential Energy Surface -
Cormorant
We propose Cormorant, a rotationally covariant neural network architecture for learning the behavior and properties of complex many-body physical systems. -
LieConv
Generalizing Convolutional Neural Networks for Equivariance to Lie Groups on Arbitrary Continuous Data -
torchmd-net/ET
Neural network potentials based on graph neural networks and equivariant transformers -
GemNet
GemNet: Universal Directional Graph Neural Networks for Molecules -
equiformer
Equiformer: Equivariant Graph Attention Transformer for 3D Atomistic Graphs -
VisNet-LSRM
Inspired by fragmentation-based methods, we propose the Long-Short-Range Message-Passing (LSR-MP) framework as a generalization of the existing equivariant graph neural networks (EGNNs) with the intent to incorporate long-range interactions efficiently and effectively. -
AP-net
AP-Net: An atomic-pairwise neural network for smooth and transferable interaction potentials -
MACE
MACE provides fast and accurate machine learning interatomic potentials with higher order equivariant message passing. -
MACE-OFF23
This repository contains the MACE-OFF23 pre-traained transferable organic force fields. -
Unimol+ Uni-Mol+ first generates a raw 3D molecule conformation from inexpensive methods such as RDKit. Then, the raw conformation is iteratively updated to its target DFT equilibrium conformation using neural networks, and the learned conformation will be used to predict the QC properties.
-
ColfNet
Inspired by differential geometry and physics, we introduce equivariant local complete frames to graph neural networks, such that tensor information at given orders can be projected onto the frames. -
AIRS
AIRS is a collection of open-source software tools, datasets, and benchmarks associated with our paper entitled “Artificial Intelligence for Science in Quantum, Atomistic, and Continuum Systems”. -
nnp-pre-training
Synthetic pre-training for neural-network interatomic potentials -
AlF_dimer
a global potential for AlF-AlF dimer -
q-AQUA,q-AQUA-pol
CCSD(T) potential for water, interfaced with TTM3-F -
LeftNet
A New Perspective on Building Efficient and Expressive 3D Equivariant Graph Neural Networks -
mlp-train
General machine learning potentials (MLP) training for molecular systems in gas phase and solution -
ARROW-NN
The simulation conda package contains the InterX ARBALEST molecular dynamics simulation software along with all the necessary database files to run ARROW-NN molecular simulations -
SO3krates with transformer
we propose a transformer architecture called SO3krates that combines sparse equivariant representations -
AMOEBA+NN
It present an integrated non-reactive hybrid model, AMOEBA+NN, which employs the AMOEBA potential for the short- and long-range non-bonded interactions and an NNP to capture the remaining local (covalent) contributions -
LEIGNN A lightweight equivariant interaction graph neural network (LEIGNN) that can enable accurate and efficient interatomic potential and force predictions in crystals. Rather than relying on higher-order representations, LEIGNN employs a scalar-vector dual representation to encode equivariant feature.
-
Arrow NN
A hybrid wide-coverage intermolecular interaction model consisting of an analytically polarizable force field combined with a short-range neural network correction for the total intermolecular interaction energy. -
PAMNet
PAMNet(Physics-aware Multiplex Graph Neural Network) is an improved version of MXMNet and outperforms state-of-the-art baselines regarding both accuracy and efficiency in diverse tasks including small molecule property prediction, RNA 3D structure prediction, and protein-ligand binding affinity prediction. -
Multi-fidelity GNNs
Multi-fidelity GNNs for drug discovery and quantum mechanics -
GPIP
GPIP: Geometry-enhanced Pre-training on Interatomic Potentials.they propose a geometric structure learning framework that leverages the unlabeled configurations to improve the performance of MLIPs. Their framework consists of two stages: firstly, using CMD simulations to generate unlabeled configurations of the target molecular system; and secondly, applying geometry-enhanced self-supervised learning techniques, including masking, denoising, and contrastive learning, to capture structural information -
ictp
Official repository for the paper "Higher Rank Irreducible Cartesian Tensors for Equivariant Message Passing". It is built upon the ALEBREW repository and implements irreducible Cartesian tensors and their products. -
CHGNet
A pretrained universal neural network potential for charge-informed atomistic modeling (see publication) -
GPTFF
GPTFF: A high-accuracy out-of-the-box universal AI force field for arbitrary inorganic materials -
rascaline
Rascaline is a library for the efficient computing of representations for atomistic machine learning also called "descriptors" or "fingerprints". These representations can be used for atomistic machine learning (ml) models including ml potentials, visualization or similarity analysis.
- PairNet-OPs/PairFE-Net
In PairFE-Net, an atomic structure is encoded using pairwise nuclear repulsion forces
- SpookyNet
Spookynet: Learning force fields with electronic degrees of freedom and nonlocal effects. - trip
Transformer Interatomic Potential (TrIP): a chemically sound potential based on the SE(3)-Transformer
- grappa
A machine-learned molecular mechanics force field using a deep graph attentional network - espaloma
Extensible Surrogate Potential of Ab initio Learned and Optimized by Message-passing Algorithm. - FeNNol
FeNNol is a library for building, training and running neural network potentials for molecular simulations. It is based on the JAX library and is designed to be fast and flexible.
- OrbNet; OrbNet Denali
OrbNet Denali: A machine learning potential for biological and organic chemistry with semi-empirical cost and DFT accuracy.
-
OrbNet-Equi
INFORMING GEOMETRIC DEEP LEARNING WITH ELECTRONIC INTERACTIONS TO ACCELERATE QUANTUM CHEMISTRY -
OrbNet-Spin
OrbNet-Spin incorporates a spin-polarized treatment into the underlying semiempirical quantum mechanics orbital featurization and adjusts the model architecture accordingly while maintaining the geometrical constraints. -
AIQM1
Artificial intelligence-enhanced quantum chemical method with broad applicability.
- BpopNN
Incorporating Electronic Information into Machine Learning Potential Energy Surfaces via Approaching the Ground-State Electronic Energy as a Function of Atom-Based Electronic Populations. - Delfta
The DelFTa application is an easy-to-use, open-source toolbox for predicting quantum-mechanical properties of drug-like molecules. Using either ∆-learning (with a GFN2-xTB baseline) or direct-learning (without a baseline), the application accurately approximates DFT reference values (ωB97X-D/def2-SVP). - PYSEQM
PYSEQM is a Semi-Empirical Quantum Mechanics package implemented in PyTorch. - DFTBML
DFTBML provides a systematic way to parameterize the Density Functional-based Tight Binding (DFTB) semiempirical quantum chemical method for different chemical systems by learning the underlying Hamiltonian parameters rather than fitting the potential energy surface directly. - mopac-ml
MOPAC-ML implements the PM6-ML method, a semiempirical quantum-mechanical computational method that augments PM6 with a machine learning (ML) correction. It acts as a wrapper calling a modified version of MOPAC, to which it provides the ML correction.
- cgnet
Coarse graining for molecular dynamics - SchNet-CG
We explore the application of SchNet models to obtain a CG potential for liquid benzene, investigating the effect of model architecture and hyperparameters on the thermodynamic, dynamical, and structural properties of the simulated CG systems, reporting and discussing challenges encountered and future directions envisioned. - CG-SchNET
By combining recent deep learning methods with a large and diverse training set of all-atom protein simulations, we here develop a bottom-up CG force field with chemical transferability, which can be used for extrapolative molecular dynamics on new sequences not used during model parametrization. - torchmd-protein-thermodynamics
This repository contains code, data and tutarial for reproducing the paper "Machine Learning Coarse-Grained Potentials of Protein Thermodynamics". https://arxiv.org/abs/2212.07492 - torchmd-exp This repository contains a method for training a neural network potential for coarse-grained proteins using unsupervised learning
- AICG
Learning coarse-grained force fields for fibrogenesis modeling(https://doi.org/10.1016/j.cpc.2023.108964)
-
VES-NN
Neural networks-based variationallyenhanced sampling -
Accelerated_sampling_with_autoencoder
Accelerated sampling framework with autoencoder-based method
- Enhanced Sampling with Machine Learning: A Review
we highlight successful strategies like dimensionality reduction, reinforcement learning, and fl ow-based methods. Finally, we discuss open problems at the exciting ML-enhanced MD interface - mlcolvar
mlcolvar is a Python library aimed to help design data-driven collective-variables (CVs) for enhanced sampling simulations.
- NNP-MM
NNP/MM embeds a Neural Network Potential into a conventional molecular mechanical (MM) model. We have implemented this using the Custom QM/MM features of NAMD 2.13, which interface NAMD with the TorchANI NNP python library developed by the Roitberg and Isayev groups. - DeeP-HP
Scalable hybrid deep neural networks/polarizable potentials biomolecular simulations including long-range effects
- PairF-Net
Here, we further develop the PairF-Net model to intrinsically incorporate energy conservation and couple the model to a molecular mechanical (MM) environment within the OpenMM package - embedding
This work presents a variant of an electrostatic embedding scheme that allows the embedding of arbitrary machine learned potentials trained on molecular systems in vacuo. - field_schnet
FieldSchNet provides a deep neural network for modeling the interaction of molecules and external environments as described.
- gimlet
Graph Inference on Molecular Topology. A package for modelling, learning, and inference on molecular topological space written in Python and TensorFlow.