Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Vision Language Model is NOT All You Need: Augmentation Strategies for Molecule Language Models (2407.09043v3)

Published 12 Jul 2024 in cs.AI

Abstract: Recently, there has been a growing interest among researchers in understanding molecules and their textual descriptions through molecule LLMs (MoLM). However, despite some early promising developments, the advancement of MoLM still trails significantly behind that of vision LLMs (VLM). This is because unique challenges exist apart from VLM in the field of MoLM due to 1) a limited amount of molecule-text paired data and 2) missing expertise that occurred due to the specialized areas of focus among the experts. To this end, we propose AMOLE, which 1) augments molecule-text pairs with structural similarity preserving loss, and 2) transfers the expertise between the molecules. Specifically, AMOLE enriches molecule-text pairs by sharing descriptions among structurally similar molecules with a novel structural similarity preserving loss. Moreover, we propose an expertise reconstruction loss to transfer knowledge from molecules that have extensive expertise to those with less expertise. Extensive experiments on various downstream tasks demonstrate the superiority of AMOLE in comprehending molecules and their descriptions, highlighting its potential for application in real-world drug discovery. The source code for AMOLE is available at https://github.com/Namkyeong/AMOLE.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Namkyeong Lee (21 papers)
  2. Siddhartha Laghuvarapu (3 papers)
  3. Chanyoung Park (83 papers)
  4. Jimeng Sun (181 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets