Jointly Identifying and Fixing Inconsistent Readings from Information Extraction Systems

dc.contributor.authorPadia, Ankur
dc.contributor.authorFerraro, Francis
dc.contributor.authorFinin, Tim
dc.date.accessioned2022-05-31T16:33:59Z
dc.date.available2022-05-31T16:33:59Z
dc.date.issued2022-05-27
dc.descriptionProceedings of Deep Learning Inside Out (DeeLIO 2022): The 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, pages 42–52, Dublin, Ireland and Online.
dc.description.abstractInformation extraction systems analyze text to produce entities and beliefs, but their output often has errors. In this paper, we analyze the reading consistency of the extracted facts with respect to the text from which they were derived and show how to detect and correct errors. We consider both the scenario when the provenance text is automatically found by an information extraction system and when it is curated by humans. We contrast consistency with credibility; define and explore consistency and repair tasks; and demonstrate a simple yet effective and generalizable model. We analyze these tasks and evaluate this approach on three datasets. Against a strong baseline model, we consistently improve both consistency and repair across three datasets using a simple MLP model with attention and lexical features.en_US
dc.description.sponsorshipWe would also like to thank the anonymous reviewers for their comments, questions, and suggestions. This material is based in part upon work supported by the National Science Foundation under Grant Nos. IIS-1940931, IIS-2024878, and DGE-2114892. Some experiments were conducted on the UMBC HPCF, supported by the National Science Foundation under Grant No. CNS1920079.This material is also based on research that is in part supported by the Army Research Laboratory, Grant No. W911NF2120076, and by the Air Force Research Laboratory (AFRL), DARPA, for the KAIROS program under agreement number FA8750-19-2-1003. The U.S.Government is authorized to reproduce and distribute reprints for Governmental purposes notwithstanding any copyright notation thereon. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either express or implied, of the Air Force Research Laboratory (AFRL), DARPA, or the U.S. Government.en_US
dc.description.urihttps://aclanthology.org/2022.deelio-1.5/en_US
dc.format.extent11 pagesen_US
dc.genreconference papers and preceedingsen_US
dc.identifierdoi:10.13016/m2twfe-z8xc
dc.identifier.citationAnkur Padia, Francis Ferraro, and Tim Finin, Jointly Identifying and Fixing Inconsistent Readings from Information Extraction Systems, 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, 60th Annual Meeting of the Association for Computational Linguistics, May 2022. http://dx.doi.org/10.18653/v1/2022.deelio-1.5en_US
dc.identifier.urihttp://hdl.handle.net/11603/24762
dc.identifier.urihttp://dx.doi.org/10.18653/v1/2022.deelio-1.5
dc.language.isoen_USen_US
dc.publisherAssociation for Computational Linguisticsen_US
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department Collection
dc.relation.ispartofUMBC Faculty Collection
dc.rightsThis item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.en_US
dc.subjectUMBC Ebiquity Research Group
dc.titleJointly Identifying and Fixing Inconsistent Readings from Information Extraction Systemsen_US
dc.typeTexten_US
dcterms.creatorhttps://orcid.org/0000-0002-6593-1792en_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
2022.deelio-1.5.pdf
Size:
470.36 KB
Format:
Adobe Portable Document Format
Description:

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
2.56 KB
Format:
Item-specific license agreed upon to submission
Description: