Advancements in Computational Models for Graph Structures, Organic Synthesis, and Materials Science

The recent developments in the research area highlight a significant shift towards integrating advanced computational models with domain-specific knowledge to tackle complex problems. A notable trend is the application of Large Language Models (LLMs) to understand and process graph-structured data, aiming to overcome limitations in modeling high-order graph structures and capturing temporal patterns. Innovations in this space include translating graphs into a graph language corpus for LLMs and integrating temporal graph learning into LLM-based models for temporal knowledge graphs. Another advancement is seen in the field of organic synthesis, where a novel transformer-based approach incorporates 3D conformer data to enhance retrosynthesis predictions. Representation learning on heterogeneous text-rich networks has also seen progress with the introduction of a pure PLM-based framework that models text data and graph structures seamlessly. In materials science, a new learning paradigm constructs crossmodal embedding spaces between crystal structures and texts, facilitating intuitive material retrieval. Lastly, a novel representation of molecules through Algebraic Data Types offers a comprehensive approach to modeling complex molecular structures, addressing limitations of existing representations.

Noteworthy Papers

  • Graph-defined Language Learning with LLMs: Introduces GDL4LLM, enabling LLMs to transfer language understanding capabilities to graph-structured data, outperforming existing methods in node classification tasks.
  • Integrate Temporal Graph Learning into LLM-based Temporal Knowledge Graph Model: Proposes TGL-LLM, a framework that enhances LLM-based models for temporal knowledge graphs by capturing temporal and relational patterns effectively.
  • Enhancing Retrosynthesis with Conformer: Presents a template-free method incorporating 3D conformer data, setting a new benchmark for retrosynthesis predictions.
  • HierPromptLM: A pure PLM-based framework for representation learning on heterogeneous text-rich networks, achieving significant improvements in node classification and link prediction.
  • Contrastive Language-Structure Pre-training Driven by Materials Science Literature: Introduces CLaSP, a learning paradigm for constructing crossmodal embedding spaces between crystal structures and texts, facilitating intuitive material retrieval.
  • SMILES has to go: Proposes a novel representation of molecules through Algebraic Data Types, offering a comprehensive approach to modeling complex molecular structures.

Sources

Graph-defined Language Learning with LLMs

Integrate Temporal Graph Learning into LLM-based Temporal Knowledge Graph Model

Enhancing Retrosynthesis with Conformer: A Template-Free Method

HierPromptLM: A Pure PLM-based Framework for Representation Learning on Heterogeneous Text-rich Networks

Contrastive Language-Structure Pre-training Driven by Materials Science Literature

SMILES has to go : Representation of Molecules via Algebraic Data Types

Built with on top of