Oates, TimLagnese, Joseph Anthony2021-09-012021-09-012020-01-2012199http://hdl.handle.net/11603/22789As the applications and number of production-level machine learning models continue to increase, so too does the need for appropriate monitoring frameworks for these models. Models applied to ever-changing real world data will inevitably experience a shift in their distribution of incoming data referred to as concept drift. The quick and accurate detection of concept drift is critical to the efficient and effective use of these models. While previous approaches to solving this problem have required partially or fully labeled testing data or have focused on monitoring a single metric, we propose a model- and metric-independent approach which is able to detect concept drift in unlabeled data streams. We utilize symmetrized Kullback-Leibler divergence in combination with statistical randomization testing to provide an approach which is able to detect drift with tunable sensitivity. To demonstrate the utility of our approach, we apply to logistic regression models tested on a variety of problems using the reduced-resolution MNIST dataset from UCI [1], the National Weather Service's CF6 climate dataset [2], and Blitzer et al.'s multi-domain sentiment analysis dataset [3]. Our results show that our approach is able to reliably detect sudden drift as well as gradual drift using a sliding window approach.application:pdfconcept driftdetectionKullback-Leibler divergencemachine learningrandomization testingunlabeledStatistical Methods for Detecting Anomalous Model Behavior with Unlabeled DataText