Leveraging Explainable AI for LLM Text Attribution: Differentiating Human-Written and Multiple LLM-Generated Text

dc.contributor.authorNajjar, Ayat A.
dc.contributor.authorAshqar, Huthaifa
dc.contributor.authorDarwish, Omar
dc.contributor.authorHammad, Eman
dc.date.accessioned2025-10-16T15:27:07Z
dc.date.issued2025-09-04
dc.description.abstractThe development of generative AI Large Language Models (LLMs) raised the alarm regarding the identification of content produced by generative AI vs. humans. In one case, issues arise when students heavily rely on such tools in a manner that can affect the development of their writing or coding skills. Other issues of plagiarism also apply. This study aims to support efforts to detect and identify textual content generated using LLM tools. We hypothesize that LLM-generated text is detectable by machine learning (ML) and investigate ML models that can recognize and differentiate between texts generated by humans and multiple LLM tools. We used a dataset of student-written text in comparison with LLM-written text. We leveraged several ML and Deep Learning (DL) algorithms, such as Random Forest (RF) and Recurrent Neural Networks (RNNs) and utilized Explainable Artificial Intelligence (XAI) to understand the important features in attribution. Our method is divided into (1) binary classification to differentiate between human-written and AI-generated text and (2) multi-classification to differentiate between human-written text and text generated by five different LLM tools (ChatGPT, LLaMA, Google Bard, Claude, and Perplexity). Results show high accuracy in multi- and binary classification. Our model outperformed GPTZero (78.3%), with an accuracy of 98.5%. Notably, GPTZero was unable to recognize about 4.2% of the observations, but our model was able to recognize the complete test dataset. XAI results showed that understanding feature importance across different classes enables detailed author/source profiles, aiding in attribution and supporting plagiarism detection by highlighting unique stylistic and structural elements, thereby ensuring robust verification of content originality.
dc.description.sponsorshipPublication made possible in part by support from Eastern Michigan University’s Faculty Open Access Publishing Fund, administered by the Associate Provost and Vice President for Graduate Studies and Research, with assistance from the EMU Library
dc.description.urihttps://www.mdpi.com/2078-2489/16/9/767
dc.format.extent16 pages
dc.genrejournal articles
dc.identifierdoi:10.13016/m2ikuj-ao0b
dc.identifier.citationNajjar, Ayat A., Huthaifa I. Ashqar, Omar Darwish, and Eman Hammad. “Leveraging Explainable AI for LLM Text Attribution: Differentiating Human-Written and Multiple LLM-Generated Text.” Information 16, no. 9 (2025): 767. https://doi.org/10.3390/info16090767.
dc.identifier.urihttps://doi.org/10.3390/info16090767
dc.identifier.urihttp://hdl.handle.net/11603/40434
dc.language.isoen
dc.publisherMDPI
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Data Science
dc.rightsAttribution 4.0 International
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.subjectattribution
dc.subjectChatGPT
dc.subjectGoogle Bard
dc.subjectLLaMA
dc.subjectexplainable AI
dc.subjectplagiarism
dc.subjectLLMs
dc.subjectClaude
dc.titleLeveraging Explainable AI for LLM Text Attribution: Differentiating Human-Written and Multiple LLM-Generated Text
dc.typeText
dcterms.creatorhttps://orcid.org/0000-0002-6835-8338

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
information1600767v2.pdf
Size:
1.5 MB
Format:
Adobe Portable Document Format