Enhanced self-directed training (ESDT): dynamic data balancing in regression models
Author/Creator
Author/Creator ORCID
Date
Department
Program
Citation of Original Publication
Kacenjar, Steve, Ronald Neely, Philip Feldman, and Aaron Dant. “Enhanced Self-Directed Training (ESDT): Dynamic Data Balancing in Regression Models.” Synthetic Data for Artificial Intelligence and Machine Learning: Tools, Techniques, and Applications III (May 2025): 236–55. https://doi.org/10.1117/12.3066302.
Rights
©2025 Society of Photo-Optical Instrumentation Engineers (SPIE). One print or electronic copy may be made for personal use only. Systematic reproduction and distribution, duplication of any material in this paper for a fee or for commercial purposes, or modification of the content of the paper are prohibited.
Subjects
Abstract
Enhanced Self-Directed Training (ESDT) is a novel framework designed to enhance regression-based Multilayer Perceptron (MLP) models through strategic synthetic data augmentation. Unlike traditional methods that rely on random data generation, ESDT focuses on high-loss validation cases, applying Synthetic Data Extension (SDE) techniques, such as SMOTE-based interpolation for continuous data and time-based translation for discontinuous events. This strategy significantly improves model generalization, reduces overfitting, and boosts performance in unbalanced datasets. By tailoring data augmentation efforts to specific areas where the model encounters challenges, ESDT effectively increases data diversity and resilience, ensuring robust predictive modeling across various complex domains. This research establishes ESDT as a robust, adaptive training methodology, promising advancements in scalability and efficiency for future machine learning applications.
