MolReGPT: Empowering Molecule Discovery for Molecule-Caption Translation with Large Language Models: A ChatGPT Perspective
The official repo of MolReGPT
- "Empowering Molecule Discovery for Molecule-Caption Translation with Large Language Models: A ChatGPT Perspective". (Accepted by IEEE TKDE (To appear))
Name | Affiliation | |
---|---|---|
Jiatong Li | The Hong Kong Polytechnic University | [email protected] |
Yunqing Liu | The Hong Kong Polytechnic University | [email protected] |
Wenqi Fan | The Hong Kong Polytechnic University | [email protected] |
Xiao-Yong Wei | The Hong Kong Polytechnic University & Sichuan University | [email protected] |
Hui Liu | Michigan State University | [email protected] |
Jiliang Tang | Michigan State University | [email protected] |
Qing Li | The Hong Kong Polytechnic University | [email protected] |
If you encounter any problems, please feel free to contact us.
- The official weights of ICMA are released via the hugging face repo, please refer to ICMA model weights
- We release the remaining query code in another GitHub repo. Please refer to MolReGPT TKDE
- Our paper has been accepted by IEEE TKDE and will appear soon~ 🎉
- We release our second work In-Context Molecule Adaptation (ICMA), which aims to further improve the performance of MolReGPT by adapting the parameters of LLMs via in-context molecule tuning Paper Link.
- We release the results of MolReGPT (GPT-4-0314) for Mol2Cap and Cap2Mol tasks and update the evaluations folder for convenient testing.
- Codes, data, and demo will be available soon~ We will follow our roadmap to release the results. Please stay tuned!
Our paper is accpeted by IEEE TKDE and we have updated the accepted version~ 🎉
We will soon release all the codes, data, and demo for MolReGPT. Please stay tuned! Thanks for your patience!
We first released the results of zero_shot and ten shot results ten_shot_bm25, ten_shot_morgan.
We would like to announce our second work, In-Context Molecule Adaptation (ICMA), which aims to further improve the performance of MolReGPT by adapting the parameters of LLMs via in-context molecule tuning Paper Link.
MolReGPT aims to create a foundation method for molecule discovery by leveraging large language models (LLMs). Thus, we focus on two crucial aspects:
- molecule understanding
- text-conditioned molecule generation
To this end, we focus on a specific task, molecule-caption translation
, the two sub-tasks of which exactly corresponds to the two aspects.
molecule2caption (i.e., Mol2Cap)
aims to generate a caption for a given molecule to describe its structure, properties, and functions.caption2molecule (i.e., Cap2Mol)
aims to generate a molecule for a given caption, which could help researchers customize their molecules for specific purposes.
Molecule discovery plays a crucial role in various scientific fields, advancing the design of tailored materials and drugs. Traditional methods for molecule discovery follow a trial-and-error process, which are both time-consuming and costly, while computational approaches such as artificial intelligence (AI) have emerged as revolutionary tools to expedite various tasks, like molecule-caption translation. Despite the importance of molecule-caption translation for molecule discovery, most of the existing methods heavily rely on domain experts, require excessive computational cost, and suffer from poor performance. On the other hand, Large Language Models (LLMs), like ChatGPT, have shown remarkable performance in various cross-modal tasks due to their great powerful capabilities in natural language understanding, generalization, and reasoning, which provides unprecedented opportunities to advance molecule discovery. To address the above limitations, in this work, we propose a novel LLMs-based framework (MolReGPT) for molecule-caption translation, where a retrieval-based prompt paradigm is introduced to empower molecule discovery with LLMs like ChatGPT without fine-tuning. More specifically, MolReGPT leverages the principle of molecular similarity to retrieve similar molecules and their text descriptions from a local database to ground the generation of LLMs through in-context few-shot molecule learning. We evaluate the effectiveness of MolReGPT via molecule-caption translation, which includes molecule understanding and text-based molecule generation. Experimental results show that MolReGPT outperforms fine-tuned models like MolT5-base without any additional training. To the best of our knowledge, MolReGPT is the first work to leverage LLMs in molecule-caption translation for advancing molecule discovery.
We apply the same dataset used in MolT5, which is the dataset of ChEBI-20
Results have been released!👏
Method | BLEU-2 |
BLEU-4 |
ROUGEL-1 |
ROUGEL-2 |
ROUGEL-L |
METEOR |
Text2Mol |
---|---|---|---|---|---|---|---|
Transformer | 0.061 | 0.027 | 0.204 | 0.087 | 0.186 | 0.114 | 0.057 |
GPT-3.5-turbo (zero_shot) | 0.103 | 0.050 | 0.261 | 0.088 | 0.204 | 0.161 | 0.352 |
T5-base | 0.511 | 0.423 | 0.607 | 0.451 | 0.550 | 0.539 | 0.523 |
MolT5-base | 0.540 | 0.457 | 0.634 | 0.485 | 0.578 | 0.569 | 0.547 |
MolReGPT (GPT-3.5-turbo) | 0.565 | 0.482 | 0.623 | 0.450 | 0.543 | 0.585 | 0.560 |
T5-large | 0.558 | 0.467 | 0.630 | 0.478 | 0.569 | 0.586 | 0.563 |
MolT5-large | 0.594 | 0.508 | 0.654 | 0.510 | 0.594 | 0.614 | 0.582 |
MolReGPT (GPT-4-0314) | 0.607 | 0.525 | 0.634 | 0.476 | 0.562 | 0.610 | 0.585 |
Method | BLEU |
EM |
Levenshtein |
MACCS FTS |
RDK FTS |
Morgan FTS |
FCD |
Text2Mol |
Validity |
---|---|---|---|---|---|---|---|---|---|
Transformer | 0.499 | 0.000 | 57.66 | 0.480 | 0.320 | 0.217 | 11.32 | 0.277 | 0.906 |
GPT-3.5-turbo (zero_shot) | 0.489 | 0.019 | 52.13 | 0.705 | 0.462 | 0.367 | 2.05 | 0.479 | 0.802 |
T5-base | 0.762 | 0.069 | 24.950 | 0.731 | 0.605 | 0.545 | 2.48 | 0.499 | 0.660 |
MolT5-base | 0.769 | 0.081 | 24.458 | 0.721 | 0.588 | 0.529 | 2.18 | 0.496 | 0.772 |
MolReGPT (GPT-3.5-turbo) | 0.790 | 0.139 | 24.91 | 0.847 | 0.708 | 0.624 | 0.57 | 0.571 | 0.887 |
T5-large | 0.854 | 0.279 | 16.721 | 0.823 | 0.731 | 0.670 | 1.22 | 0.552 | 0.902 |
MolT5-large | 0.854 | 0.311 | 16.071 | 0.834 | 0.746 | 0.684 | 1.20 | 0.554 | 0.905 |
MolReGPT (GPT-4-0314) | 0.857 | 0.280 | 17.14 | 0.903 | 0.805 | 0.739 | 0.41 | 0.593 | 0.899 |
transformers == 4.30.0
torch == 1.13.1+cu117
rdkit == 2022.09.5
fcd == 1.1
rank_bm25 == 0.2.2
sentence_transformers == 2.2.2
openai == 0.27.2
Thanks for your attention! The codes will be released ASAP!
Currently, you can customize your own prompt via our released jupyter-notebook demo
We will also provide web-page for a more user-friendly demo.
@article{molregpt,
title={Empowering Molecule Discovery for Molecule-Caption Translation with Large Language Models: A ChatGPT Perspective},
author={Jiatong Li and Yunqing Liu and Wenqi Fan and Xiao-Yong Wei and Hui Liu and Jiliang Tang and Qing Li},
journal={arXiv preprint arXiv:2306.06615},
year={2023}
}