TY - JOUR
T1 - 32 examples of LLM applications in materials science and chemistry: towards automation, assistants, agents, and accelerated scientific discovery
AU - Zimmermann, Yoel
AU - Bazgir, Adib
AU - Al-Feghali, Alexander
AU - Ansari, Mehrad
AU - Bocarsly, Joshua
AU - Brinson, L Catherine
AU - Chiang, Yuan
AU - Circi, Defne
AU - Chiu, Min-Hsueh
AU - Daelman, Nathan
AU - Evans, Matthew L
AU - Gangan, Abhijeet S
AU - George, Janine
AU - Harb, Hassan
AU - Khalighinejad, Ghazal
AU - Takrim Khan, Sartaaj
AU - Klawohn, Sascha
AU - Lederbauer, Magdalena
AU - Mahjoubi, Soroush
AU - Mohr, Bernadette
AU - Mohamad Moosavi, Seyed
AU - Naik, Aakash
AU - Beste Ozhan, Aleyna
AU - Plessers, Dieter
AU - Roy, Aritra
AU - Schöppach, Fabian
AU - Schwaller, Philippe
AU - Terboven, Carla
AU - Ueltzen, Katharina
AU - Wu, Yue
AU - Zhu, Shang
AU - Janssen, Jan
AU - Li, Calvin
AU - Foster, Ian
AU - Blaiszik, Ben
PY - 2025/9/30
Y1 - 2025/9/30
N2 - Large language models (LLMs) are reshaping many aspects of materials science and chemistry research, enabling advances in molecular property prediction, materials design, scientific automation, knowledge extraction, and more. Recent developments demonstrate that the latest class of models are able to integrate structured and unstructured data, assist in hypothesis generation, and streamline research workflows. To explore the frontier of LLM capabilities across the research lifecycle, we review applications of LLMs through 32 total projects developed during the second annual LLM hackathon for applications in materials science and chemistry, a global hybrid event. These projects spanned seven key research areas: (1) molecular and material property prediction, (2) molecular and material design, (3) automation and novel interfaces, (4) scientific communication and education, (5) research data management and automation, (6) hypothesis generation and evaluation, and (7) knowledge extraction and reasoning from the scientific literature. Collectively, these applications illustrate how LLMs serve as versatile predictive models, platforms for rapid prototyping of domain-specific tools, and much more. In particular, improvements in both open source and proprietary LLM performance through the addition of reasoning, additional training data, and new techniques have expanded effectiveness, particularly in low-data environments and interdisciplinary research. As LLMs continue to improve, their integration into scientific workflows presents both new opportunities and new challenges, requiring ongoing exploration, continued refinement, and further research to address reliability, interpretability, and reproducibility.
AB - Large language models (LLMs) are reshaping many aspects of materials science and chemistry research, enabling advances in molecular property prediction, materials design, scientific automation, knowledge extraction, and more. Recent developments demonstrate that the latest class of models are able to integrate structured and unstructured data, assist in hypothesis generation, and streamline research workflows. To explore the frontier of LLM capabilities across the research lifecycle, we review applications of LLMs through 32 total projects developed during the second annual LLM hackathon for applications in materials science and chemistry, a global hybrid event. These projects spanned seven key research areas: (1) molecular and material property prediction, (2) molecular and material design, (3) automation and novel interfaces, (4) scientific communication and education, (5) research data management and automation, (6) hypothesis generation and evaluation, and (7) knowledge extraction and reasoning from the scientific literature. Collectively, these applications illustrate how LLMs serve as versatile predictive models, platforms for rapid prototyping of domain-specific tools, and much more. In particular, improvements in both open source and proprietary LLM performance through the addition of reasoning, additional training data, and new techniques have expanded effectiveness, particularly in low-data environments and interdisciplinary research. As LLMs continue to improve, their integration into scientific workflows presents both new opportunities and new challenges, requiring ongoing exploration, continued refinement, and further research to address reliability, interpretability, and reproducibility.
KW - machine learning
KW - chemistry
KW - LLM
KW - materials science
KW - AI
U2 - 10.1088/2632-2153/ae011a
DO - 10.1088/2632-2153/ae011a
M3 - Article
SN - 2632-2153
VL - 6
JO - Machine Learning: Science and Technology
JF - Machine Learning: Science and Technology
IS - 3
M1 - 030701
ER -