Sparsity and heterogeneous dropout for continual learning in the null space of neural activations

dc.contributor.authorAbbasi, Ali
dc.contributor.authorNooralinejad, Parsa
dc.contributor.authorBraverman, Vladimir
dc.contributor.authorPirsiavash, Hamed
dc.contributor.authorKolouri, Soheil
dc.date.accessioned2022-11-14T15:46:18Z
dc.date.available2022-11-14T15:46:18Z
dc.date.issued2022
dc.description1st Conference on Lifelong Learning Agents; Montreal, Canada; August 22 - 24, 2022
dc.description.abstractContinual/lifelong learning from a non-stationary input data stream is a cornerstone of intelligence. Despite their phenomenal performance in a wide variety of applications, deep neural networks are prone to forgetting their previously learned information upon learning new ones. This phenomenon is called “catastrophic forgetting” and is deeply rooted in the stability-plasticity dilemma. Overcoming catastrophic forgetting in deep neural networks has become an active field of research in recent years. In particular, gradient projection-based methods have recently shown exceptional performance at overcoming catastrophic forgetting. This paper proposes two biologically-inspired mechanisms based on sparsity and heterogeneous dropout that significantly increase a continual learner’s performance over a long sequence of tasks. Our proposed approach builds on the Gradient Projection Memory (GPM) framework. We leverage k-winner activations in each layer of a neural network to enforce layer-wise sparse activations for each task, together with a between-task heterogeneous dropout that encourages the network to use non-overlapping activation patterns between different tasks. In addition, we introduce two new benchmarks for continual learning under distributional shift, namely Continual Swiss Roll and ImageNet SuperDog-40. Lastly, we provide an in-depth analysis of our proposed method and demonstrate a significant performance boost on various benchmark continual learning problems.en_US
dc.description.sponsorshipSK and HP were supported by the Defense Advanced Research Projects Agency (DARPA) under Contract No. HR00112190135. VB was supported by DARPA under Contract No. HR00112190130.en_US
dc.description.urihttps://proceedings.mlr.press/v199/abbasi22aen_US
dc.format.extent12 pagesen_US
dc.genreconference papers and proceedingsen_US
dc.genrepreprintsen_US
dc.identifierdoi:10.13016/m2wgwn-iyjz
dc.identifier.citationAbbasi, Ali, Parsa Nooralinejad, Vladimir Braverman, Hamed Pirsiavash, and Soheil Kolouri. “Sparsity and Heterogeneous Dropout for Continual Learning in the Null Space of Neural Activations.” In Proceedings of The 1st Conference on Lifelong Learning Agents, 617–28. PMLR, 2022. https://proceedings.mlr.press/v199/abbasi22a.html.
dc.identifier.urihttp://hdl.handle.net/11603/26318
dc.language.isoen_USen_US
dc.publisherMLResearchPress
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department Collection
dc.relation.ispartofUMBC Faculty Collection
dc.rightsThis item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.en_US
dc.titleSparsity and heterogeneous dropout for continual learning in the null space of neural activationsen_US
dc.typeTexten_US

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
abbasi22a.pdf
Size:
1.3 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.56 KB
Format:
Item-specific license agreed upon to submission
Description: