Semi-supervised Expectation Maximization with Contrastive Outlier Removal.

dc.contributor.advisorChapman, David
dc.contributor.authorMenon, Sumeet
dc.contributor.departmentComputer Science and Electrical Engineering
dc.contributor.programComputer Science
dc.date.accessioned2022-09-29T15:37:59Z
dc.date.available2022-09-29T15:37:59Z
dc.date.issued2022-01-01
dc.description.abstractSemi-supervised learning has proven to be one of the most widely used techniques to overcome the concern of limited labels. One of the concerns while using neural networks for semi-supervised learning in presence of an extremely small labeled dataset is the occurrence of confidently predicted incorrect labels. This phenomenon of confidently predicting incorrect labels for unsupervised data is called confounding bias. Even though pseudo-labeling and consistency regularization are among the state-of-the-art techniques for semi-supervised learning, these techniques are susceptible to the problem of confounding bias while using neural networks. We propose a methodology that could help neural networks overcome this problem by leveraging information from unlabeled images using cluster generating techniques and smoothness generating techniques in a tightly-coupled way to overcome the fundamental problem of outliers. These techniques could help the model to learn certain attributes from the image which could not be learned from the original resolution of the unlabeled images. We argue both theoretically and empirically that contrastive outlier suppression is a necessary yet overlooked criteria in the application of EM-derived latent bootstrapping, because discrimination models such as neural networks have the potential to make erronous predictions with high confidence if these datasets are far from the decision boundary, whereas generative methods for which Expectation Maximization (EM) was originally designed have no such issue. Contrastive outlier suppression is derived under the assumption that the latent feature vector predictions should follow a multivariate gaussian mixture distribution. Our results show that contrastive latent bootstrapping greatly improves semi-supervised classification accuracy over a baseline, and furthermore when combined with a state-of-the-art consistency regularization method, our results achieve the highest reported semi-supervised accuracy for the CIFAR-10 classification using only 250 labeled sample images.
dc.formatapplication:pdf
dc.genredissertations
dc.identifierdoi:10.13016/m2iuae-ynix
dc.identifier.other12528
dc.identifier.urihttp://hdl.handle.net/11603/25989
dc.languageen
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department Collection
dc.relation.ispartofUMBC Theses and Dissertations Collection
dc.relation.ispartofUMBC Graduate School Collection
dc.relation.ispartofUMBC Student Collection
dc.rightsThis item may be protected under Title 17 of the U.S. Copyright Law. It is made available by UMBC for non-commercial research and education. For permission to publish or reproduce, please see http://aok.lib.umbc.edu/specoll/repro.php or contact Special Collections at speccoll(at)umbc.edu
dc.sourceOriginal File Name: Menon_umbc_0434D_12528.pdf
dc.subjectConsistency Regularization
dc.subjectOutlier Removal
dc.subjectProxy-Label
dc.subjectSemi-Supervised Learning
dc.titleSemi-supervised Expectation Maximization with Contrastive Outlier Removal.
dc.typeText
dcterms.accessRightsDistribution Rights granted to UMBC by the author.
dcterms.accessRightsAccess limited to the UMBC community. Item may possibly be obtained via Interlibrary Loan thorugh a local library, pending author/copyright holder's permission.

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Menon_umbc_0434D_12528.pdf
Size:
1.39 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
Menon-Sumeet_Open.pdf
Size:
265.38 KB
Format:
Adobe Portable Document Format
Description: