A comprehensive benchmark for evaluating LLM-generated ontologies

Plu, Julien; Moreno Escobar, Oscar; Trouillez, Edouard; Gapin, Axelle; Troncy, Raphaël
ISWC 2024, 23rd International Semantic Web Conference, 11-15 November 2024, Baltimore, USA

This paper presents a methodology for evaluating ontologies that are automatically generated by Large Language Models (LLMs). Our approach combines quantitative metrics that compare generated ontologies with respect to a human-made reference and qualitative user assessments across diverse domains. We apply this methodology to evaluate the ontologies produced by various LLMs, including Claude 3.5 Sonnet, GPT-4o, and GPT-4o-mini. The results demonstrate the benchmark’s effectiveness in identifying strengths and weaknesses of LLM-generated ontologies, providing valuable insights for improving automated ontology generation techniques. 


DOI
Type:
Conference
City:
Baltimore
Date:
2024-11-11
Department:
Data Science
Eurecom Ref:
7945
Copyright:
CEUR

PERMALINK : https://www.eurecom.fr/publication/7945