Integrating Chemistry Knowledge in Large Language Models via Prompt Engineering

This paper presents a study on the integration of domain-specific knowledge in prompt engineering to enhance the performance of large language models (LLMs) in scientific domains. A benchmark dataset is curated to encapsulate the intricate physical-chemical properties of small molecules, their drugability for pharmacology, alongside the functional attributes of enzymes and crystal materials, underscoring the relevance and applicability across biological and chemical domains.The proposed domain-knowledge embedded prompt engineering method outperforms traditional prompt engineering strategies on various metrics, including capability, accuracy, F1 score, and hallucination drop. The effectiveness of the method is demonstrated through case studies on complex materials including the MacMillan catalyst, paclitaxel, and lithium cobalt oxide. The results suggest that domain-knowledge prompts can guide LLMs to generate more accurate and relevant responses, highlighting the potential of LLMs as powerful tools for scientific discovery and innovation when equipped with domain-specific prompts. The study also discusses limitations and future directions for domain-specific prompt engineering development..

Medienart:

Preprint

Erscheinungsjahr:

2024

Erschienen:

2024

Enthalten in:

arXiv.org - (2024) vom: 22. Apr. Zur Gesamtaufnahme - year:2024

Sprache:

Englisch

Beteiligte Personen:

Liu, Hongxuan [VerfasserIn]
Yin, Haoyu [VerfasserIn]
Luo, Zhiyao [VerfasserIn]
Wang, Xiaonan [VerfasserIn]

Links:

Volltext [kostenfrei]

Themen:

000
Computer Science - Artificial Intelligence
Computer Science - Computation and Language

Förderinstitution / Projekttitel:

PPN (Katalog-ID):

XAR04336814X