SELF-SUPERVISED LEARNING BY COMPRESSING REPRESENTATIONS FOR LIGHTWEIGHT MODELS
dc.contributor.advisor | Pirsiavash, Hamed | |
dc.contributor.author | Abbasi Koohpayegani, Soroush | |
dc.contributor.department | Computer Science and Electrical Engineering | |
dc.contributor.program | Computer Science | |
dc.date.accessioned | 2023-07-07T16:02:12Z | |
dc.date.available | 2023-07-07T16:02:12Z | |
dc.date.issued | 2022-01-01 | |
dc.description.abstract | Self-supervised learning aims to learn good representations with unlabeled data. Recent works have shown that larger models benefit more from self-supervised learning than smaller models. As a result, the gap between supervised and self-supervised learning has been greatly reduced for larger models. In this work, instead of designing a new pseudo task for self-supervised learning, we develop a model compression method to compress an already learned, deep self-supervised model (teacher) to a smaller one (student). We train the student model so that it mimics the relative similarity between the datapoints in the teacher's embedding space. For AlexNet, our method outperforms all previous methods including the fully supervised model on ImageNet linear evaluation (59.0% compared to 56.5%) and on nearest neighbor evaluation (50.7% compared to 41.4%). To the best of our knowledge, this is the first time a self-supervised AlexNet has outperformed supervised one on ImageNet classification. Moreover, we show that our method is effective in a few other applications: reducing the computation precision rather than the model depth only, learning small models for video representations, learning across modalities, and self-distillation. | |
dc.format | application:pdf | |
dc.genre | thesis | |
dc.identifier | doi:10.13016/m2ecxs-6dgw | |
dc.identifier.other | 12572 | |
dc.identifier.uri | http://hdl.handle.net/11603/28463 | |
dc.language | en | |
dc.relation.isAvailableAt | The University of Maryland, Baltimore County (UMBC) | |
dc.relation.ispartof | UMBC Computer Science and Electrical Engineering Collection | |
dc.relation.ispartof | UMBC Theses and Dissertations Collection | |
dc.relation.ispartof | UMBC Graduate School Collection | |
dc.relation.ispartof | UMBC Student Collection | |
dc.rights | This item may be protected under Title 17 of the U.S. Copyright Law. It is made available by UMBC for non-commercial research and education. For permission to publish or reproduce, please see http://aok.lib.umbc.edu/specoll/repro.php or contact Special Collections at speccoll(at)umbc.edu | |
dc.source | Original File Name: AbbasiKoohpayegani_umbc_0434M_12572.pdf | |
dc.subject | Computer Vision | |
dc.subject | Distillation | |
dc.subject | Efficient Model | |
dc.subject | Machine Learning | |
dc.subject | Representation Learning | |
dc.subject | Self-Supervised Learning | |
dc.title | SELF-SUPERVISED LEARNING BY COMPRESSING REPRESENTATIONS FOR LIGHTWEIGHT MODELS | |
dc.type | Text | |
dcterms.accessRights | Access limited to the UMBC community. Item may possibly be obtained via Interlibrary Loan through a local library, pending author/copyright holder's permission. | |
dcterms.accessRights | Access limited to the UMBC community. Item may possibly be obtained via Interlibrary Loan thorugh a local library, pending author/copyright holder's permission. |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- AbbasiKoohpayegani_umbc_0434M_12572.pdf
- Size:
- 3.29 MB
- Format:
- Adobe Portable Document Format