Do LLMs Adhere to Label Definitions? Examining Their Receptivity to External Label Definitions
Files
Links to Files
Date
Type of Work
Department
Program
Citation of Original Publication
Rights
Attribution 4.0 International
Subjects
UMBC Interactive Robotics and Language Lab
UMBC Discovery, Research, and Experimental Analysis of Malware Lab (DREAM Lab)
UMBC Interactive Robotics and Language Lab (IRAL Lab)
Computer Science - Machine Learning
Computer Science - Computation and Language
UMBC Ebiquity Research Group
Computer Science - Artificial Intelligence
UMBC Discovery, Research, and Experimental Analysis of Malware Lab (DREAM Lab)
UMBC Interactive Robotics and Language Lab (IRAL Lab)
Computer Science - Machine Learning
Computer Science - Computation and Language
UMBC Ebiquity Research Group
Computer Science - Artificial Intelligence
Abstract
Do LLMs genuinely incorporate external definitions, or do they primarily rely on their parametric knowledge? To address these questions, we conduct controlled experiments across multiple explanation benchmark datasets (general and domain-specific) and label definition conditions, including expert-curated, LLM-generated, perturbed, and swapped definitions. Our results reveal that while explicit label definitions can enhance accuracy and explainability, their integration into an LLM's task-solving processes is neither guaranteed nor consistent, suggesting reliance on internalized representations in many cases. Models often default to their internal representations, particularly in general tasks, whereas domain-specific tasks benefit more from explicit definitions. These findings underscore the need for a deeper understanding of how LLMs process external knowledge alongside their pre-existing capabilities.
