LacMaterial: Large Language Models as Analogical Chemists for Materials Discovery

📅 2025-10-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Human scientific discovery is often constrained by domain-specific knowledge limitations and surface-level similarity biases, hindering cross-disciplinary, structure-driven analogical reasoning and impeding innovation in novel materials. Method: We propose an explicit analogical reasoning framework powered by large language models (LLMs). It first performs cross-domain analogical retrieval to transcend disciplinary boundaries, then constructs in-domain structural analogy templates from few-shot labeled examples, integrating relation extraction, knowledge transfer, and prompt engineering to generate interpretable hypotheses. Contribution/Results: The framework shifts battery material design from empirical dopant substitution toward deep structural-guided innovation. In candidate material generation, it significantly outperforms standard prompting baselines and successfully identifies multiple electrochemically promising novel materials—providing the first empirical validation of LLMs as capable “analogical chemists.”

Technology Category

Application Category

📝 Abstract
Analogical reasoning, the transfer of relational structures across contexts (e.g., planet is to sun as electron is to nucleus), is fundamental to scientific discovery. Yet human insight is often constrained by domain expertise and surface-level biases, limiting access to deeper, structure-driven analogies both within and across disciplines. Large language models (LLMs), trained on vast cross-domain data, present a promising yet underexplored tool for analogical reasoning in science. Here, we demonstrate that LLMs can generate novel battery materials by (1) retrieving cross-domain analogs and analogy-guided exemplars to steer exploration beyond conventional dopant substitutions, and (2) constructing in-domain analogical templates from few labeled examples to guide targeted exploitation. These explicit analogical reasoning strategies yield candidates outside established compositional spaces and outperform standard prompting baselines. Our findings position LLMs as interpretable, expert-like hypothesis generators that leverage analogy-driven generalization for scientific innovation.
Problem

Research questions and friction points this paper is trying to address.

LLMs generate novel battery materials via cross-domain analogies
Overcome human expertise limitations in analogical reasoning
Construct in-domain analogical templates from few examples
Innovation

Methods, ideas, or system contributions that make the work stand out.

Retrieving cross-domain analogs to steer material exploration
Constructing in-domain analogical templates from few examples
Leveraging analogy-driven generalization for scientific innovation