Enhancing Knowledge Graph Consistency through Open Large Language Models: A Case Study
| dc.contributor.author | Padia, Ankur | |
| dc.contributor.author | Ferraro, Francis | |
| dc.contributor.author | Finin, Tim | |
| dc.date.accessioned | 2024-02-29T17:57:44Z | |
| dc.date.available | 2024-02-29T17:57:44Z | |
| dc.date.issued | 2024-05-20 | |
| dc.description | AAAI-MAKE: Empowering Machine Learning and Large Language Models with Domain and Commonsense Knowledge, 25 March, 2024 | |
| dc.description.abstract | High-quality knowledge graphs (KGs) play a crucial role in many applications. However, KGs created by automated information extraction systems can suffer from erroneous extractions or be inconsistent with provenance/source text. It is important to identify and correct such problems. In this paper, we study leveraging the emergent reasoning capabilities of large language models (LLMs) to detect inconsistencies between extracted facts and their provenance. With a focus on ?open? LLMs that can be run and trained locally, we find that few-shot approaches can yield an absolute performance gain of 2.5-3.4% over the state-of-the-art method with only 9% of training data. We examine the LLM architectures? effect and show that Decoder-Only models underperform Encoder-Decoder approaches. We also explore how model size impacts performance and counterintuitively find that larger models do not result in consistent performance gains. Our detailed analyses suggest that while LLMs can improve KG consistency, the different LLM models learn different aspects of KG consistency and are sensitive to the number of entities involved. | |
| dc.description.sponsorship | We thank the anonymous reviewers for their comments,questions, and suggestions. This material is partly based onwork supported by the National Science Foundation underGrant Nos. IIS-2024878 and DGE-2114892. This material isalso based on research that is in part supported by the ArmyResearch Laboratory, Grant No. W911NF2120076, and bythe Air Force Research Laboratory (AFRL), DARPA, for theKAIROS program under agreement number FA8750-19-2-1003. The U.S. Government is authorized to reproduce anddistribute reprints for Governmental purposes, notwithstand-ing any copyright notation thereon. The views and conclu-sions contained herein are those of the authors and shouldnot be interpreted as necessarily representing the officialpolicies or endorsements, either express or implied, of theAir Force Research Laboratory (AFRL), DARPA, or theU.S. Government. | |
| dc.description.uri | https://ojs.aaai.org/index.php/AAAI-SS/article/view/31201 | |
| dc.format.extent | 6 pages | |
| dc.genre | conference papers and proceedings | |
| dc.genre | preprints | |
| dc.identifier | doi:10.13016/m2pdvz-nyxj | |
| dc.identifier.citation | Padia, Ankur, Francis Ferraro, and Tim Finin. “Enhancing Knowledge Graph Consistency through Open Large Language Models: A Case Study.” Proceedings of the AAAI Symposium Series 3, no. 1 (May 20, 2024): 203–8. https://doi.org/10.1609/aaaiss.v3i1.31201. | |
| dc.identifier.uri | https://doi.org/10.1609/aaaiss.v3i1.31201 | |
| dc.language.iso | en_US | |
| dc.publisher | AAAI | |
| dc.relation.isAvailableAt | The University of Maryland, Baltimore County (UMBC) | |
| dc.relation.ispartof | UMBC Faculty Collection | |
| dc.relation.ispartof | UMBC Computer Science and Electrical Engineering Department | |
| dc.rights | This item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.This item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author. | |
| dc.subject | UMBC Ebiquity Research Group | |
| dc.title | Enhancing Knowledge Graph Consistency through Open Large Language Models: A Case Study | |
| dc.type | Text | |
| dcterms.creator | https://orcid.org/0000-0002-6593-1792 |
Files
Original bundle
1 - 1 of 1
