Transforming Expert Knowledge into Scalable Ontology via Large Language Models

📅 2025-06-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In domain ontology construction, mapping multi-source terminology to foundational concepts faces three key challenges: high cost and subjectivity of manual approaches, shallow semantic modeling and poor cross-domain consistency of automated methods, and weak interpretability. To address these, this paper proposes the first LLM-driven framework integrating expert calibration with iterative prompt optimization. The framework combines expert-guided annotation, multi-stage prompt engineering, and a human-in-the-loop validation cycle to generate concept links with high confidence and full interpretability. Evaluated on the concept necessity mapping task, it achieves an F1-score of 0.97—substantially surpassing the human baseline (0.68)—and marks the first instance of scalable ontology alignment that simultaneously attains expert-level accuracy and transparent, auditable reasoning.

Technology Category

Application Category

📝 Abstract
Having a unified, coherent taxonomy is essential for effective knowledge representation in domain-specific applications as diverse terminologies need to be mapped to underlying concepts. Traditional manual approaches to taxonomy alignment rely on expert review of concept pairs, but this becomes prohibitively expensive and time-consuming at scale, while subjective interpretations often lead to expert disagreements. Existing automated methods for taxonomy alignment have shown promise but face limitations in handling nuanced semantic relationships and maintaining consistency across different domains. These approaches often struggle with context-dependent concept mappings and lack transparent reasoning processes. We propose a novel framework that combines large language models (LLMs) with expert calibration and iterative prompt optimization to automate taxonomy alignment. Our method integrates expert-labeled examples, multi-stage prompt engineering, and human validation to guide LLMs in generating both taxonomy linkages and supporting rationales. In evaluating our framework on a domain-specific mapping task of concept essentiality, we achieved an F1-score of 0.97, substantially exceeding the human benchmark of 0.68. These results demonstrate the effectiveness of our approach in scaling taxonomy alignment while maintaining high-quality mappings and preserving expert oversight for ambiguous cases.
Problem

Research questions and friction points this paper is trying to address.

Automating taxonomy alignment to replace costly manual expert review
Improving semantic relationship handling in cross-domain taxonomy mapping
Enhancing transparency and consistency in automated concept mapping processes
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLMs automate taxonomy alignment efficiently
Expert calibration enhances semantic relationship handling
Iterative prompt optimization ensures consistent mappings
🔎 Similar Papers
No similar papers found.
I
Ikkei Itoku
Amazon, New York, USA
D
David Theil
Amazon, Arlington, USA
E
Evelyn Eichelsdoerfer Uehara
Amazon, Seattle, USA
Sreyoshi Bhaduri
Sreyoshi Bhaduri
Amazon
Artificial IntelligenceNatural Language ProcessingEducation
J
Junnosuke Kuroda
Amazon, Seattle, USA
T
Toshi Yumoto
Amazon, Arlington, USA
A
Alex Gil
Amazon, New York, USA
Natalie Perez
Natalie Perez
Senior Research Scientist @ Amazon
qualitativemethodologygenAIlarge language modelslearning
R
R. Cherukuri
Amazon, Seattle, USA
N
Naumaan Nayyar
Amazon, Seattle, USA