A Fast Method to Fine-Tune Neural Networks for the Least Energy Consumption on FPGAs

dc.contributor.authorHosseini, Morteza
dc.contributor.authorEbrahimabadi, Mohammad
dc.contributor.authorMazumder, Arnab Neelim
dc.contributor.authorHomayoun, Houman
dc.contributor.authorMohsenin, Tinoosh
dc.date.accessioned2021-07-26T16:32:06Z
dc.date.available2021-07-26T16:32:06Z
dc.date.issued2021
dc.descriptionHAET workshop of ICLR 2021en_US
dc.description.abstractBecause of their simple hardware requirements, low bitwidth neural networks (NNs) have gained significant attention over the recent years, and have been extensively employed in electronic devices that seek efficiency and performance. Research has shown that scaled-up low bitwidth NNs can have accuracy levels on par with their full-precision counterparts. As a result, there seems to be a tradeoff between quantization (q) and scaling (s) of NNs to maintain the accuracy. In this paper, we propose QS-NAS which is a systematic approach to explore the best quantization and scaling factors for a NN architecture that satisfies a targeted accuracy level and results in the least energy consumption per inference when deployed to a hardware–FPGA in this work. Compared to the literature using the same VGG-like NN with different q and s over the same datasets, our selected optimal NNs deployed to a low-cost tiny Xilinx FPGA from the ZedBoard resulted in accuracy levels higher or on par with those of the related work, while giving the least power dissipation and the highest inference/Joule.en_US
dc.description.urihttps://eehpc.csee.umbc.edu/publications/pdf/2021/A_fast_method.pdfen_US
dc.format.extent5 pagesen_US
dc.genreconference papers and proceedingsen_US
dc.identifierdoi:10.13016/m2m8bc-5lxc
dc.identifier.citationHosseini, Morteza et al.; A Fast Method to Fine-Tune Neural Networks for the Least Energy Consumption on FPGAs; HAET workshop of ICLR 2021; https://eehpc.csee.umbc.edu/publications/pdf/2021/A_fast_method.pdfen_US
dc.identifier.urihttp://hdl.handle.net/11603/22103
dc.language.isoen_USen_US
dc.publisherUMBC Energy Efficient High Performance Computing Laben_US
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department Collection
dc.relation.ispartofUMBC Faculty Collection
dc.relation.ispartofUMBC Student Collection
dc.rightsThis item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.
dc.titleA Fast Method to Fine-Tune Neural Networks for the Least Energy Consumption on FPGAsen_US
dc.typeTexten_US

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
A_fast_method.pdf
Size:
1.32 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.56 KB
Format:
Item-specific license agreed upon to submission
Description: