Vision-Language Models for Autonomous Driving: CLIP-Based Dynamic Scene Understanding

dc.contributor.authorElhenawy, Mohammed
dc.contributor.authorAshqar, Huthaifa
dc.contributor.authorRakotonirainy, Andry
dc.contributor.authorAlhadidi, Taqwa I.
dc.contributor.authorJaber, Ahmed
dc.contributor.authorTami, Mohammad Abu
dc.date.accessioned2025-10-16T15:27:11Z
dc.date.issued2025-03-24
dc.description.abstractScene understanding is essential for enhancing driver safety, generating human-centric explanations for Automated Vehicle (AV) decisions, and leveraging Artificial Intelligence (AI) for retrospective driving video analysis. This study developed a dynamic scene retrieval system using Contrastive Language–Image Pretraining (CLIP) models, which can be optimized for real-time deployment on edge devices. The proposed system outperforms state-of-the-art in-context learning methods, including the zero-shot capabilities of GPT-4o, particularly in complex scenarios. By conducting frame-level analyses on the Honda Scenes Dataset, which contains a collection of about 80 h of annotated driving videos capturing diverse real-world road and weather conditions, our study highlights the robustness of CLIP models in learning visual concepts from natural language supervision. The results also showed that fine-tuning the CLIP models, such as ViT-L/14 (Vision Transformer) and ViT-B/32, significantly improved scene classification, achieving a top F1-score of 91.1%. These results demonstrate the ability of the system to deliver rapid and precise scene recognition, which can be used to meet the critical requirements of advanced driver assistance systems (ADASs). This study shows the potential of CLIP models to provide scalable and efficient frameworks for dynamic scene understanding and classification. Furthermore, this work lays the groundwork for advanced autonomous vehicle technologies by fostering a deeper understanding of driver behavior, road conditions, and safety-critical scenarios, marking a significant step toward smarter, safer, and more context-aware autonomous driving systems.
dc.description.sponsorshipThis research was funded partially by the Australian Government through the Australian Research Council Discovery Project DP220102598.
dc.description.urihttps://www.mdpi.com/2079-9292/14/7/1282
dc.format.extent27 pages
dc.genrejournal articles
dc.identifierdoi:10.13016/m2zpbl-om08
dc.identifier.citationElhenawy, Mohammed, Huthaifa I. Ashqar, Andry Rakotonirainy, Taqwa I. Alhadidi, Ahmed Jaber, and Mohammad Abu Tami. “Vision-Language Models for Autonomous Driving: CLIP-Based Dynamic Scene Understanding.” Electronics 14, no. 7 (2025): 1282. https://doi.org/10.3390/electronics14071282.
dc.identifier.urihttps://doi.org/10.3390/electronics14071282
dc.identifier.urihttp://hdl.handle.net/11603/40449
dc.language.isoen
dc.publisherMDPI
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Data Science
dc.rightsAttribution 4.0 International
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.subjectadvanced driver assistance systems (ADASs)
dc.subjectscene understanding
dc.subjectautomated vehicle (AV)
dc.subjectcontrastive language–image pretraining (CLIP)
dc.subjectfine-tuning
dc.titleVision-Language Models for Autonomous Driving: CLIP-Based Dynamic Scene Understanding
dc.typeText
dcterms.creatorhttps://orcid.org/0000-0002-6835-8338

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
electronics1401282.pdf
Size:
9.88 MB
Format:
Adobe Portable Document Format