EA: An Event Autoencoder for High-Speed Vision Sensing

dc.contributor.authorIslam, Riadul
dc.contributor.authorMule, Joey
dc.contributor.authorChallagundla, Dhandeep
dc.contributor.authorRizvi, Shahmir
dc.contributor.authorCarson, Sean
dc.date.accessioned2025-07-30T19:22:17Z
dc.date.issued2025-07-09
dc.description.abstractHigh-speed vision sensing is essential for real-time perception in applications such as robotics, autonomous vehicles, and industrial automation. Traditional frame-based vision systems suffer from motion blur, high latency, and redundant data processing, limiting their performance in dynamic environments. Event cameras, which capture asynchronous brightness changes at the pixel level, offer a promising alternative but pose challenges in object detection due to sparse and noisy event streams. To address this, we propose an event autoencoder architecture that efficiently compresses and reconstructs event data while preserving critical spatial and temporal features. The proposed model employs convolutional encoding and incorporates adaptive threshold selection and a lightweight classifier to enhance recognition accuracy while reducing computational complexity. Experimental results on the existing Smart Event Face Dataset (SEFD) demonstrate that our approach achieves comparable accuracy to the YOLO-v4 model while utilizing up to 35.5 X fewer parameters. Implementations on embedded platforms, including Raspberry Pi 4B and NVIDIA Jetson Nano, show high frame rates ranging from 8 FPS up to 44.8 FPS. The proposed classifier exhibits up to 87.84x better FPS than the state-of-the-art and significantly improves event-based vision performance, making it ideal for low-power, high-speed applications in real-time edge computing.
dc.description.sponsorshipThis work was supported in part by the National Science Foundation (NSF) award number: 2138253, the Maryland Industrial Partnerships (MIPS) program under award number MIPS0012, and the UMBC Startup grant.
dc.description.urihttp://arxiv.org/abs/2507.06459
dc.format.extent7 pages
dc.genrejournal articles
dc.genrepreprints
dc.identifierdoi:10.13016/m2xnx9-wpnz
dc.identifier.urihttps://doi.org/10.48550/arXiv.2507.06459
dc.identifier.urihttp://hdl.handle.net/11603/39521
dc.language.isoen_US
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department
dc.relation.ispartofUMBC Student Collection
dc.relation.ispartofUMBC Faculty Collection
dc.rightsAttribution 4.0 International
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/deed.en
dc.subjectComputer Science - Artificial Intelligence
dc.subjectUMBC Cybersecurity Institute
dc.subjectUMBC Multi-Scale Thermal Transport Research Lab
dc.subjectComputer Science - Computer Vision and Pattern Recognition
dc.subjectUMBC VLSI-SOC GROUP
dc.titleEA: An Event Autoencoder for High-Speed Vision Sensing
dc.typeText
dcterms.creatorhttps://orcid.org/0000-0002-4649-3467
dcterms.creatorhttps://orcid.org/0009-0002-8522-0872
dcterms.creatorhttps://orcid.org/0000-0001-7491-1710

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
2507.06459v1.pdf
Size:
3.54 MB
Format:
Adobe Portable Document Format