Skip to content

The official repo for the paper "Leveraging large language models for nanaosynthesis mechanisms explanation"

Notifications You must be signed in to change notification settings

Dandelionym/llm_for_mechanisms

Repository files navigation

Leveraging Large Language Models for Explaining Material Synthesis Mechanisms: The Foundation of Materials Discovery

NeurIPS 2024 AI4Mat Workshop
Image-1

Figure 1: Semantic illustration of our proposed framework for large language model evaluation in nanomaterial synthesis prediction, highlighting concepts and workflow. a) nanosynthesis study loop: begins with basic conditions, leading to the discovery of novel synthesis rules through experiments involving variable adjustments. b) exemplifies the synthesis mechanism, dissected into causality and correlations, with an emphasis on correlations described through condition-observation pairs. c) outlines the process from sourcing relevant literature (using key area keywords) for benchmark construction and model evaluation.

Dataset

We manually created the dataset for evaluation. It is in the format of FastChat or ShareGPT, both are popular in LLMs area. If you are interested in using this dataset, please reference our paper (see below).

Image-2

Figure 2: Evaluation data set illustration. a) shows the distribution of collected evaluation sets containing 775 questions categorized by synthesis methods and structures, respectively. b) displays a jittered scatter plot of manually curated research papers with the counts of mechanism, conditions and observations, with mechanism relevance from low to high, indicated by varying colors to represent the frequency of observations and varying sizes to represent the biasing towards mechanism. c) showcases the multiple selection question considered in the evaluation. The model is instructed to give the correct option. d) illustration of the probing test in our evaluation study based on the proposed c-score.

How to run benchmarking

Before running, the deployment of 🚀FastChat is recommended for inferencing LLMs with OpenAI API fashion. It is an open platform for training, serving, and evaluating large language model based chatbots.

After deployed LLMs, you may start the inference API in LAN, e.g., http://10.11.50.197:7860, and IP address 10.11.50.197 should be your machine's actual IP. The port can also be configured in FastChat. Then you should change the configuration of address and port in eval_opensourced_llms.py, it is easy to config.

Finally, you could run each evaluation by simply run python eval_opensourced_llms.py

If you have OpenAI API, you could change the API Key in the code. And the Claude API configuration is in similar.

Notes

This study focus on the evaluation of LLMs in science mechanisms understanding, trying to open a new perspective for AI for Science Research. If anyone is interested, please see our manuscript.

Contact

If you have any questions, please feel free to email: yingmingpu@gmail.com

Citation

@inproceedings{
  pu2024leveraging,
  title={Leveraging Large Language Models for Explaining Material Synthesis Mechanisms: The Foundation of Materials Discovery},
  author={Yingming Pu and Liping Huang and Tao Lin and Hongyu Chen},
  booktitle={AI for Accelerated Materials Design - NeurIPS 2024},
  year={2024},
  url={https://openreview.net/forum?id=I6jYRbaai8}
}

About

The official repo for the paper "Leveraging large language models for nanaosynthesis mechanisms explanation"

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages