MOFormer: Self-Supervised Transformer model for Metal-Organic Framework Property Prediction

10/25/2022
by   Zhonglin Cao, et al.
0

Metal-Organic Frameworks (MOFs) are materials with a high degree of porosity that can be used for applications in energy storage, water desalination, gas storage, and gas separation. However, the chemical space of MOFs is close to an infinite size due to the large variety of possible combinations of building blocks and topology. Discovering the optimal MOFs for specific applications requires an efficient and accurate search over an enormous number of potential candidates. Previous high-throughput screening methods using computational simulations like DFT can be time-consuming. Such methods also require optimizing 3D atomic structure of MOFs, which adds one extra step when evaluating hypothetical MOFs. In this work, we propose a structure-agnostic deep learning method based on the Transformer model, named as MOFormer, for property predictions of MOFs. The MOFormer takes a text string representation of MOF (MOFid) as input, thus circumventing the need of obtaining the 3D structure of hypothetical MOF and accelerating the screening process. Furthermore, we introduce a self-supervised learning framework that pretrains the MOFormer via maximizing the cross-correlation between its structure-agnostic representations and structure-based representations of crystal graph convolutional neural network (CGCNN) on >400k publicly available MOF data. Using self-supervised learning allows the MOFormer to intrinsically learn 3D structural information though it is not included in the input. Experiments show that pretraining improved the prediction accuracy of both models on various downstream prediction tasks. Furthermore, we revealed that MOFormer can be more data-efficient on quantum-chemical property prediction than structure-based CGCNN when training data is limited. Overall, MOFormer provides a novel perspective on efficient MOF design using deep learning.

READ FULL TEXT

page 7

page 12

page 15

page 16

research
04/27/2022

Masked Spectrogram Prediction For Self-Supervised Audio Pre-Training

Transformer-based models attain excellent results and generalize well wh...
research
08/17/2021

RRLFSOR: An Efficient Self-Supervised Learning Strategy of Graph Convolutional Networks

To further improve the performance and the self-learning ability of GCNs...
research
10/29/2020

Graph Neural Network for Metal Organic Framework Potential Energy Approximation

Metal-organic frameworks (MOFs) are nanoporous compounds composed of met...
research
08/16/2023

Is Self-Supervised Pretraining Good for Extrapolation in Molecular Property Prediction?

The prediction of material properties plays a crucial role in the develo...
research
02/19/2020

Molecule Attention Transformer

Designing a single neural network architecture that performs competitive...
research
08/05/2021

Self-supervised optimization of random material microstructures in the small-data regime

While the forward and backward modeling of the process-structure-propert...
research
02/21/2022

Ligandformer: A Graph Neural Network for Predicting Compound Property with Robust Interpretation

Robust and efficient interpretation of QSAR methods is quite useful to v...

Please sign up or login with your details

Forgot password? Click here to reset