International Journal of Science and Research (IJSR)

International Journal of Science and Research (IJSR)
Call for Papers | Fully Refereed | Open Access | Double Blind Peer Reviewed

ISSN: 2319-7064


Downloads: 1

India | Computer Technology | Volume 14 Issue 4, April 2025 | Pages: 1310 - 1315


Indoor Object Detection for Blind

Akash P S, Preethi Thomas

Abstract: "Indoor Object Detection for Blind" is a dedicated assistive technology project designed to enhance indoor navigation for visually impaired individuals. Utilizing advanced computer vision techniques, the system detects and identifies objects in real-time using wearable devices like smart glasses or smartphones. It integrates YOLO-based deep learning models, fine-tuned with a custom indoor dataset, to ensure high accuracy in recognizing objects such as furniture, doors, and appliances. Additionally, natural language processing (NLP) enables context-aware descriptions, providing adaptive feedback. The project is designed to be flexible and seamlessly integrate with wearable technology, ensuring a smooth and efficient user experience. The system evaluates various aspects of objects, including their spatial positioning and contextual information, to assist users in real-time. By concentrating solely on object detection within indoor environments, this system significantly improves spatial awareness and navigation for visually impaired individuals. Extensive testing demonstrates the system's performance in detecting and identifying objects while minimizing false positives.

Keywords: Indoor navigation, object detection, visually impaired, assistive technology, YOLO, deep learning

How to Cite?: Akash P S, Preethi Thomas, "Indoor Object Detection for Blind", Volume 14 Issue 4, April 2025, International Journal of Science and Research (IJSR), Pages: 1310-1315, https://www.ijsr.net/getabstract.php?paperid=SR25417001126, DOI: https://dx.doi.org/10.21275/SR25417001126


Download Article PDF


Rate This Article!

Received Comments

No approved comments available.


Top