Floorplan2Guide: LLM-Guided Floorplan Parsing for BLV Indoor Navigation
| dc.contributor.author | Ayanzadeh, Aydin | |
| dc.contributor.author | Oates, Tim | |
| dc.date.accessioned | 2026-01-22T16:19:04Z | |
| dc.date.issued | 2025-12-13 | |
| dc.description | IEEE International Conference on Big Data (IEEE BigData 2025), December 8-11, 2025, Macau, China | |
| dc.description.abstract | Indoor navigation remains a critical challenge for people with visual impairments. The current solutions mainly rely on infrastructure-based systems, which limit their ability to navigate safely in dynamic environments. We propose a novel navigation approach that utilizes a foundation model to transform floor plans into navigable knowledge graphs and generate human-readable navigation instructions. Floorplan2Guide integrates a large language model (LLM) to extract spatial information from architectural layouts, reducing the manual preprocessing required by earlier floorplan parsing methods. Experimental results indicate that few-shot learning improves navigation accuracy in comparison to zero-shot learning on simulated and real-world evaluations. Claude 3.7 Sonnet achieves the highest accuracy among the evaluated models, with 92.31%, 76.92%, and 61.54% on the short, medium, and long routes, respectively, under 5-shot prompting of the MP-1 floor plan. The success rate of graph-based spatial structure is 15.4% higher than that of direct visual reasoning among all models, which confirms that graphical representation and in-context learning enhance navigation performance and make our solution more precise for indoor navigation of Blind and Low Vision (BLV) users. | |
| dc.description.uri | http://arxiv.org/abs/2512.12177 | |
| dc.format.extent | 10 pages | |
| dc.genre | conference papers and proceedings | |
| dc.genre | postprints | |
| dc.identifier | doi:10.13016/m2dovk-8i0d | |
| dc.identifier.uri | https://doi.org/10.48550/arXiv.2512.12177 | |
| dc.identifier.uri | http://hdl.handle.net/11603/41540 | |
| dc.language.iso | en | |
| dc.relation.isAvailableAt | The University of Maryland, Baltimore County (UMBC) | |
| dc.relation.ispartof | UMBC Faculty Collection | |
| dc.relation.ispartof | UMBC Computer Science and Electrical Engineering Department | |
| dc.relation.ispartof | UMBC Student Collection | |
| dc.rights | Attribution 4.0 International | |
| dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | |
| dc.subject | UMBC Accelerated Cognitive Cybersecurity Laboratory | |
| dc.subject | UMBC Ebiquity Research Group | |
| dc.subject | Computer Science - Artificial Intelligence | |
| dc.subject | UMBC Cognition, Robotics, and Learning (CoRaL) Lab | |
| dc.title | Floorplan2Guide: LLM-Guided Floorplan Parsing for BLV Indoor Navigation | |
| dc.type | Text | |
| dcterms.creator | https://orcid.org/0000-0002-8816-3204 |
Files
Original bundle
1 - 1 of 1
