SOTAVerified

GIT-Mol: A Multi-modal Large Language Model for Molecular Science with Graph, Image, and Text

2023-08-14Code Available1· sign in to hype

PengFei Liu, Yiming Ren, Jun Tao, Zhixiang Ren

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Large language models have made significant strides in natural language processing, enabling innovative applications in molecular science by processing textual representations of molecules. However, most existing language models cannot capture the rich information with complex molecular structures or images. In this paper, we introduce GIT-Mol, a multi-modal large language model that integrates the Graph, Image, and Text information. To facilitate the integration of multi-modal molecular data, we propose GIT-Former, a novel architecture that is capable of aligning all modalities into a unified latent space. We achieve a 5%-10% accuracy increase in properties prediction and a 20.2% boost in molecule generation validity compared to the baselines. With the any-to-language molecular translation strategy, our model has the potential to perform more downstream tasks, such as compound name recognition and chemical reaction prediction.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
BACEGIT-Mol(G+S)AUC0.81Unverified
BBBPGIT-Mol(G+S)AUC0.74Unverified
clintoxGIT-Mol(G+S)AUC0.88Unverified
SIDERGIT-Mol(G+S)AUC0.63Unverified
Tox21GIT-Mol(G+S)AUC0.76Unverified
ToxCastGIT-Mol(G+S)AUC0.67Unverified

Reproductions