Do LLMs Adhere to Label Definitions? Examining Their Receptivity to External Label Definitions
Links to Files
Date
Type of Work
Department
Program
Citation of Original Publication
Mohammadi, Seyedali, Bhaskara Hanuma Vedula, Hemank Lamba, Edward Raff, Ponnurangam Kumaraguru, Francis Ferraro, and Manas Gaur. 2025. “Do Llms Adhere to Label Definitions? Examining Their Receptivity to External Label Definitions.” ACL Anthology. November 2025. https://aclanthology.org/2025.emnlp-main.1648/.
Rights
This item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.
Subjects
UMBC Interactive Robotics and Language Lab
UMBC Discovery, Research, and Experimental Analysis of Malware Lab (DREAM Lab)
UMBC Interactive Robotics and Language Lab (IRAL Lab)
Computer Science - Machine Learning
Computer Science - Computation and Language
UMBC Ebiquity Research Group
Computer Science - Artificial Intelligence
UMBC KAI2 Knowledge-infused AI and Inference lab
UMBC Discovery, Research, and Experimental Analysis of Malware Lab (DREAM Lab)
UMBC Interactive Robotics and Language Lab (IRAL Lab)
Computer Science - Machine Learning
Computer Science - Computation and Language
UMBC Ebiquity Research Group
Computer Science - Artificial Intelligence
UMBC KAI2 Knowledge-infused AI and Inference lab
Abstract
Do LLMs genuinely incorporate external definitions, or do they primarily rely on their parametric knowledge? To address these questions, we conduct controlled experiments across multiple explanation benchmark datasets (general and domain-specific) and label definition conditions, including expert-curated, LLM-generated, perturbed, and swapped definitions. Our results reveal that while explicit label definitions can enhance accuracy and explainability, their integration into an LLM's task-solving processes is neither guaranteed nor consistent, suggesting reliance on internalized representations in many cases. Models often default to their internal representations, particularly in general tasks, whereas domain-specific tasks benefit more from explicit definitions. These findings underscore the need for a deeper understanding of how LLMs process external knowledge alongside their pre-existing capabilities.
