LLM-Integrated Knowledge Graph Generation

Knowledge Graphs (KGs) are crucial for enhancing semantic web technologies, improving information retrieval, and bolstering data-driven AI systems. Despite their importance, constructing KGs from text corpora remains a significant challenge. Traditional methods, which rely on manually crafted rules and machine learning techniques, often struggle with domain-specific texts and cross-domain transferability. Recent advances in generative AI, particularly Large Language Models (LLMs) like GPT-4o, LLAMA-3.2, and the newer Qwen, promise to continue to revolutionize traditional text mining paradigms, including KG construction [1, 2, 3], due to their advanced capabilities in understanding and generating human-like text. Integrating LLMs into the KG construction pipeline can enable richer, more accurate extraction and inference of knowledge from unstructured text sources and provide solutions that are readily transferable across domains. Our proposed workshop, unlike others related, stands out for the breadth and uniqueness of its focus on generating KGs from diverse text domains—including research papers, legal documents, newswires, and social media. It emphasizes discussions on scalable strategies, maintaining consistency, and mitigating errors. Applications range from news analytics to scientific research.

Evolving from its previous four iterations, the TEXT2KG initiative, proposed as LLM-TEXT2KG in its 5th iteration, aims to explore the novel intersection of LLMs and KG generation, focusing on innovative approaches, best practices, and challenges. It will serve as a platform to discuss how LLMs can be utilized for improved knowledge extraction, context-aware entity disambiguation, named entity recognition, relation extraction, knowledge representation and seamless ontology alignment. The workshop solicits a broad range of papers including full research papers, negative results, position papers, and system demos examining the wide range of issues and processes related to knowledge graphs generation from text corpora. Papers on resources (methods, tools, benchmarks, libraries, datasets) are also welcomed.

The proceedings of the previous four editions of the workshops are in:
(2022) https://ceur-ws.org/Vol-3184, (2023) https://ceur-ws.org/Vol-3447, (2024) https://ceur-ws.org/Vol-3747, and (2025) https://ceur-ws.org/Vol-4020.

Why attend the LLM-Text2KG Workshop?

This workshop aims to bring together researchers from multiple areas such as Natural Language Processing (NLP), Entity Linking (EL), Relation Extraction (RE), Knowledge Representation & Reasoning (KRR), Deep Learning (DL), Knowledge Base Construction (KBC), Large Language Models (LLMs), Retrieval-Augmented Generation (RAG), Knowledge Sharing between Agent, Semantic Web, Linked Data, & other related fields to foster a discussion and enhance the state-of-the-art in knowledge graph generation from text.
The participants will find opportunities to present and hear about other emerging research and applications, to exchange ideas and experiences, and to identify new opportunities for collaborations across disciplines. We plan to involve the many prominent research groups in the Semantic Web community which in the last years focused on the generation of knowledge graphs from textual sources in different fields, such as research data (ORKG, AI-KG, Nanopublications), question answering (ParaQA, NSQA), common sense (CSKG), automotive (CoSI, ASKG), biomedical (Hetionet), and many others.


Themes & Topics

We are interested in (including but not limited to) the following themes and topics that study the generation of Knowledge Graphs from text with LLMs,
based on quantitative, qualitative, and mixed research methods.

Themes

  • LLM-based Entity Recognition and Relation Extraction from Complex, Unstructured Text
  • LLM-driven Inference of Implicit Relationships and Knowledge Discovery
  • Addressing and Mitigating Hallucinations and Biases in LLM outputs
  • Advances in Fine-tuning and Customizing LLMs for KG Generation Tasks
  • Industrial Applications Involving KGs Generation from Text

Topics

  • Open Information Extraction
  • Deep Learning and Generative approaches
  • Human-in-the-loop methods
  • Large Language Models and Knowledge Graphs
  • RAG-Driven Knowledge Extraction
  • Intelligent Agents for Text to Knowledge Graphs
  • LLM-KG Integration
  • Benchmarks for KG Generation from Text
  • Evaluation Methods for KGs Generated from Text

Important Dates


TBD

Submission Instructions


We invite full research papers, negative results, position papers, dataset and system demo papers.
The page limit for the full research papers, negative results and dataset papers is 16 pages excluding references and for the short papers and demos it is 7 pages excluding references.
Submissions must be original and should not have been published previously or be under consideration for publication while being evaluated for this workshop. Submissions will be evaluated by the program committee based on the quality of the work and its fit to the workshop themes. All submissions are double-blind and a high-resolution PDF of the paper should be uploaded to the EasyChair submission site before the paper submission deadline.
The accepted papers will be presented at the Text2KG workshop integrated with the conference, and they will be published as CEUR proceedings.
All must be submitted and formatted in the style of the CEUR proceedings format.
For details on CEUR style, see CEUR's Author Instruction.
Also see Overleaf Template.

Workshop Schedule

TBD

Co-located Events:


TBD


(Previous) 2025 TEXT2SPARQL Challenge Collocated Event: https://ceur-ws.org/Vol-4094



Organizing Committee

Sanju
Tiwari

Sharda University, Delhi-NCR, India &
TIB Hannover Germany

tiwarisanju18@ieee.org

Nandana Mihindukulasooriya

IBM Research, Dublin, Ireland

nandana.m@ibm.com

Jennifer
D’Souza

TIB, Germany

jennifer.dsouza@tib.eu

Francesco
Osborne

KMi, The Open University

francesco.osborne@open.ac.uk



Steering Committee
& Publicity Chair

Amit
Sheth

AIISC, University of South Carolina

amit@sc.edu

Joey
Yip

AIISC, University of South Carolina

hyip@email.sc.edu



Advisory Committee

Program Committee

  • Tommaso Soru, University of Leipzig, Germany
  • Nandana Mihindukulasooriya, IBM Research AI
  • Muhammad Raza Naqvi, Universite de Toulouse Ecole Nationale d’Ing ́enieurs de Tarbes ENIT, Tarbes
  • Fatima Zahra Amara, University of Abbes Laghrour Khenchela, Algeria
  • Maosheng Guo, Harbin Institute of Technology
  • Tek Raj Chhetri, Massachusetts Institute of Technology, USA
  • Jennifer D'Souza, TIB Leibniz Information Centre for Science and Technology University Library, Hannover, Germany
  • Dimitris Kontokostas, Medidata, Greece
  • Mauro Dragoni, Fondazione Bruno Kessler
  • Serge Sonfack Sounchio, INP-Toulouse, France
  • Hong Yung Yip, Artificial Intelligence Institute, University of South Carolina
  • Edgard Marx, Leipzig University of Applied Sciences (HTWK), Germany
  • Daniil Dobriy, Vienna University of Economics and Business, Vienna, Austria
  • Patience Usoro Usip, University of Uyo, Uyo, Nigeria
  • Hamed Babaei Giglou, TIB Leibniz Information Centre for Science and Technology, Germany
  • Francesco Osborne, KMi, The Open University, UK
  • Sven Groppe, University of Luebeck, Germany
  • Hanieh Khorashadizedeh, University of Luebeck, Germany
  • Mayank Kejriwal, University of Southern California, USA
  • Davide Buscaldi, LIPN, Université Paris 13, Sorbonne Paris Cité, Paris
  • Marlene Goncalves, Universidad Simón Bolívar, Venezuela

Previous Workshops

References

  • [1] Khorashadizadeh, H., Amara, F.Z., Ezzabady, M., Ieng, F., Tiwari, S., Mihindukulasooriya, N., Groppe, J., Sahri, S., Benamara, F., Groppe, S.: Research trends for the interplay between large language models and knowledge graphs. arXiv preprint arXiv:2406.08223 (2024)
  • [2] Liang, X., Wang, Z., Li, M., Yan, Z.: A survey of llm-augmented knowledge graph construction and application in complex product design. Procedia CIRP 128, 870–875 (2024)
  • [3] Zhu, Y., Wang, X., Chen, J., Qiao, S., Ou, Y., Yao, Y., Deng, S., Chen, H., Zhang, N.: Llms for knowledge graph construction and reasoning: Recent capabilities and future opportunities. World Wide Web 27(5), 58 (2024)