Evaluating Classifiers During Dataset Shift

dc.contributor.advisorRazia Azen
dc.contributor.committeememberBrian Patterson
dc.contributor.committeememberBo Zhang
dc.contributor.committeememberJake Luo
dc.creatorFritsch, Corey
dc.date.accessioned2025-01-16T18:58:53Z
dc.date.available2025-01-16T18:58:53Z
dc.date.issued2023-05-01
dc.description.abstractDeployment of a classifier into a machine learning application likely begins with training different types of algorithms on a subset of the available historical data and then evaluating them on datasets that are drawn from identical distributions. The goal of this evaluation process is to select the classifier that is believed to be most robust in maintaining good future performance, and then deploy that classifier to end-users who use it to make predictions on new data. Often times, predictive models are deployed in conditions that differ from those used in training, meaning that dataset shift occurred. In these situations, there are no guarantees that predictions made by the predictive model in deployment will still be as reliable and accurate as they were during the training of the model. This study demonstrated a technique that can be utilized by others when selecting a classifier for deployment, as well as the first comparative study that evaluates machine learning classifier performance on synthetic datasets with different levels of prior-probability, covariate, and concept dataset shifts. The results from this study showed the impact of dataset shift on the performance of different classifiers for two real-world datasets related to teacher retention in Wisconsin and detecting fraud in testing, as well as demonstrated a framework that can be used by others when selecting a classifier for deployment. By using the methods from this study as a proactive approach to evaluate classifiers on synthetic dataset shift, different classifiers would have been considered for deployment of both predictive models, compared to only using evaluation datasets that were drawn from identical distributions. The results from both real-world datasets also showed that there was no classifier that dealt well with prior-probability shift and that classifiers were affected less by covariate and concept shift than was expected. Two supplemental demonstrations of the methodology showed that it can be extended for additional purposes of evaluating classifiers on dataset shift. Results from analyzing the effects of hyperparameter choices on classifier performance under dataset shift, as well as the effects of actual dataset shift on classifier performance, showed that different hyperparameter configurations have an impact on the performance of a classifier in general, but can also have an impact on how robust that classifier might be to dataset shift.
dc.identifier.urihttp://digital.library.wisc.edu/1793/87653
dc.relation.replaceshttps://dc.uwm.edu/etd/3147
dc.subjectBinary Classification
dc.subjectDataset Shift
dc.subjectMachine Learning
dc.titleEvaluating Classifiers During Dataset Shift
dc.typedissertation
thesis.degree.disciplineEducational Psychology
thesis.degree.grantorUniversity of Wisconsin-Milwaukee
thesis.degree.nameDoctor of Philosophy

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Fritsch_uwm_0263D_13541.pdf
Size:
7.83 MB
Format:
Adobe Portable Document Format
Description:
Main File