Enhancing Knowledge Graph Consistency through Open Large Language Models: A Case Study

dc.contributor.authorPadia, Ankur
dc.contributor.authorFerraro, Francis
dc.contributor.authorFinin, Tim
dc.date.accessioned2024-02-29T17:57:44Z
dc.date.available2024-02-29T17:57:44Z
dc.date.issued2024-05-20
dc.descriptionAAAI-MAKE: Empowering Machine Learning and Large Language Models with Domain and Commonsense Knowledge, 25 March, 2024
dc.description.abstractHigh-quality knowledge graphs (KGs) play a crucial role in many applications. However, KGs created by automated information extraction systems can suffer from erroneous extractions or be inconsistent with provenance/source text. It is important to identify and correct such problems. In this paper, we study leveraging the emergent reasoning capabilities of large language models (LLMs) to detect inconsistencies between extracted facts and their provenance. With a focus on ?open? LLMs that can be run and trained locally, we find that few-shot approaches can yield an absolute performance gain of 2.5-3.4% over the state-of-the-art method with only 9% of training data. We examine the LLM architectures? effect and show that Decoder-Only models underperform Encoder-Decoder approaches. We also explore how model size impacts performance and counterintuitively find that larger models do not result in consistent performance gains. Our detailed analyses suggest that while LLMs can improve KG consistency, the different LLM models learn different aspects of KG consistency and are sensitive to the number of entities involved.
dc.description.sponsorshipWe thank the anonymous reviewers for their comments,questions, and suggestions. This material is partly based onwork supported by the National Science Foundation underGrant Nos. IIS-2024878 and DGE-2114892. This material isalso based on research that is in part supported by the ArmyResearch Laboratory, Grant No. W911NF2120076, and bythe Air Force Research Laboratory (AFRL), DARPA, for theKAIROS program under agreement number FA8750-19-2-1003. The U.S. Government is authorized to reproduce anddistribute reprints for Governmental purposes, notwithstand-ing any copyright notation thereon. The views and conclu-sions contained herein are those of the authors and shouldnot be interpreted as necessarily representing the officialpolicies or endorsements, either express or implied, of theAir Force Research Laboratory (AFRL), DARPA, or theU.S. Government.
dc.description.urihttps://ojs.aaai.org/index.php/AAAI-SS/article/view/31201
dc.format.extent6 pages
dc.genreconference papers and proceedings
dc.genrepreprints
dc.identifierdoi:10.13016/m2pdvz-nyxj
dc.identifier.citationPadia, Ankur, Francis Ferraro, and Tim Finin. “Enhancing Knowledge Graph Consistency through Open Large Language Models: A Case Study.” Proceedings of the AAAI Symposium Series 3, no. 1 (May 20, 2024): 203–8. https://doi.org/10.1609/aaaiss.v3i1.31201.
dc.identifier.urihttps://doi.org/10.1609/aaaiss.v3i1.31201
dc.language.isoen_US
dc.publisherAAAI
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Faculty Collection
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department
dc.rightsThis item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.This item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.
dc.subjectUMBC Ebiquity Research Group
dc.titleEnhancing Knowledge Graph Consistency through Open Large Language Models: A Case Study
dc.typeText
dcterms.creatorhttps://orcid.org/0000-0002-6593-1792

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
1384.pdf
Size:
308.92 KB
Format:
Adobe Portable Document Format