Evaluating Causal AI Techniques for Health Misinformation Detection
Permanent Link
Author/Creator
Author/Creator ORCID
Date
Type of Work
Department
Program
Citation of Original Publication
Clark, Ommo, and Karuna Pande Joshi. "Evaluating Causal AI Techniques for Health Misinformation Detection." Causal AI for Robust Decision Making (CARD 2025) Workshop, Held in Conjunction with 23rd International Conference on Pervasive Computing and Communications (PerCom 2025), March 17, 2025. https://ebiquity.umbc.edu/paper/html/id/1187/Evaluating-Causal-AI-Techniques-for-Health-Misinformation-Detection.
Rights
This item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.
Subjects
natural language processing
social media analysis
UMBC Cybersecurity Institute
causal ai
misinformation detection
topic modeling
digital health
UMBC Ebiquity Research Group
UMBC Knowledge, Analytics, Cognitive and Cloud Computing (KnACC) lab
UMBC Cybersecurity Institute
UMBC Knowledge, Analytics, Cognitive and Cloud Computing (KnACC) lab
health misinformation
social media analysis
UMBC Cybersecurity Institute
causal ai
misinformation detection
topic modeling
digital health
UMBC Ebiquity Research Group
UMBC Knowledge, Analytics, Cognitive and Cloud Computing (KnACC) lab
UMBC Cybersecurity Institute
UMBC Knowledge, Analytics, Cognitive and Cloud Computing (KnACC) lab
health misinformation
Abstract
The proliferation of health misinformation on social media, particularly regarding chronic conditions such as diabetes, hypertension, and obesity, poses significant public health risks. This study evaluates the feasibility of leveraging Natural Language Processing (NLP) techniques for real-time misinformation detection and classification, focusing on Reddit discussions. Using logistic regression as a baseline model, supplemented by Latent Dirichlet Allocation (LDA) for topic modeling and K-Means clustering, we identify clusters prone to misinformation. While the model achieved a 73% accuracy rate, its recall for misinformation was limited to 12%, reflecting challenges such as class imbalance and linguistic nuances. The findings underscore the importance of advanced NLP models, such as transformer based architectures like BERT, and propose the integration of causal reasoning to enhance the interpretability and robustness of AI systems for public health interventions.
