Show simple item record

Exploring Training Provenance for Clues of Data Poisoning in Machine Learning

dc.contributor.authorJackson, Jon-Nicklaus Z.
dc.contributor.advisorBirhanu Eshete
dc.date.accessioned2023-05-02T14:27:55Z
dc.date.available2023-05-02T14:27:55Z
dc.date.issued2023-04-30
dc.identifier.urihttps://hdl.handle.net/2027.42/176346
dc.description.abstractMachine Learning today plays a vital role in a wide range of critical applications. To ensure ML models are consistently able to produce correct output, such models must be retrained as new input samples become available to avoid performance degradation as a result of data drift. During retraining, such models are left vulnerable to possible injection of poisonous samples via data poisoning attacks, where adversaries manipulate the training data to have the ML model misbehave to serve adversarial goals. Inspired by previous works that leverage data provenance for detecting and filtering out poisonous data from the training set, we adapt the definition of provenance for what we define as training provenance, the history of training metrics captured over training time-frame. We then build a framework that allows us to capture both key performance metrics for the overall dataset and per-class metrics for each label at every epoch as training provenance. Through exploratory analysis of captured training provenance we aim to find clues that standout to serve as strong signals for making a call on training data poisoning. We evaluate our proposed framework on two benchmark image classification datasets: MNIST and CIFAR-10. For MNIST, we observed promising signal(s) for establishing a per-epoch poisoning detection threshold based on captured metrics at the overall dataset level. For CIFAR-10, captured overall dataset metrics as training provenance for clean and poisoned training data were not as effective as compared to our observations for MNIST experiments. As for captured per-class metrics, we discovered that these metrics provided little insight as a result of the nondeterministic nature of machine learning. Overall, we observe that this is a promising direction that invites further exploration with more poisoning attacks and diverse datasets.
dc.languageEnglish
dc.subjectTraining data poisoning
dc.subjectData provenance
dc.subjectMachine learning
dc.titleExploring Training Provenance for Clues of Data Poisoning in Machine Learning
dc.typeThesis
dc.description.thesisdegreenameMaster of Science (MS)en_US
dc.description.thesisdegreedisciplineCybersecurity and Information Assurance, College of Engineering & Computer Science
dc.description.thesisdegreegrantorUniversity of Michigan-Dearborn
dc.contributor.committeememberJin Lu
dc.contributor.committeememberProbir Roy
dc.subject.hlbtoplevelComputer Engineering
dc.description.bitstreamurlhttp://deepblue.lib.umich.edu/bitstream/2027.42/176346/1/Jon-Nicklaus Jackson Final Thesis.pdf
dc.identifier.doihttps://dx.doi.org/10.7302/7196
dc.identifier.orcid0000-0001-6948-4662
dc.identifier.name-orcidJackson, Jon-Nicklaus; 0000-0001-6948-4662en_US
dc.working.doi10.7302/7196en
dc.owningcollnameDissertations and Theses (Ph.D. and Master's)


Files in this item

Show simple item record

Remediation of Harmful Language

The University of Michigan Library aims to describe library materials in a way that respects the people and communities who create, use, and are represented in our collections. Report harmful or offensive language in catalog records, finding aids, or elsewhere in our collections anonymously through our metadata feedback form. More information at Remediation of Harmful Language.

Accessibility

If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.