1711978267236_Junior_final_projectmech.pptx

Unknown753651 41 views 12 slides Sep 29, 2024
Slide 1
Slide 1 of 12
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12

About This Presentation

Final project


Slide Content

A PROJECT REPORT Submitted by ABIRAMI C (623020104001) BHARATHI M (623020104009) JEMIFAR S (623020104022) VIVEKALAKSHMI R ( 623020104059 ) VISION-THE-BLIND : EMPOWERING VISUALLY IMPAIRED INDIVIDUALS THROUGH INNOVATIVE VIRTUAL ASSISTANCE TAGORE INSTITUTE OF ENGINEERING AND TECHNOLOGY DEVIYAKURICHI-636 112 GUIDE NAME S.KOKILA ME ., ASSISTANT PROFESSIOR Department of Computer Science and Engineering.

ABSTRACT The VISION-THE-BLIND project aims to empower visually impaired individuals by leveraging computer vision and speech recognition technologies to detect objects in their surroundings and audibly convey information about these objects in real-time. This system utilizes the YOLO (You Only Look Once) algorithm for efficient object detection and recognition . The YOLO algorithm processes the video feed in real-time, detecting various objects such as people, vehicles, furniture, and more with high accuracy and speed. Once an object is detected, the system uses speech recognition to convert the detected object's label into audible speech . Upon detecting an object, the system utilizes a speech synthesis module to audibly describe the detected object to the user. The speech synthesis module converts text labels generated by YOLO into natural-sounding speech, providing real-time feedback to the user The VISION-THE-BLIND project addresses the needs of visually impaired individuals by leveraging advanced computer vision and speech recognition technologies, with the integration of the YOLO algorithm enabling efficient and accurate object detection in real-world scenarios

Problem Definition 4 Despite advancements in technology, traditional navigation aids often fall short in providing real-time guidance and assistance tailored to the unique needs of visually impaired individuals. This lack of accessibility not only hinders their mobility but also limits their ability to participate fully in daily activities and access essential services. Visually impaired individuals require a reliable and intuitive navigation solution that empowers them to navigate unfamiliar environments with confidence and ease. Existing solutions often lack the sophistication Adaptability needed to address the diverse needs and preferences of visually impaired users. Thus , there is a pressing need for an innovative navigation system that integrates cutting-edge technology to provide personalized guidance, obstacle detection, and route optimization in real-time.

Objective The primary objective of the "VISION-THE-BLIND" project is to develop a comprehensive solution that enhances navigation for visually impaired individuals. This includes providing real-time guidance, obstacle detection, and route optimization to improve their ability to navigate both indoor and outdoor environments.

Proposed System Develop a wearable device equipped with various sensors such as cameras, ultrasonic sensors, and infrared sensors. These sensors will capture information from the surroundings and convert it into useful data. Implement advanced computer vision algorithms to process the data collected by the sensors. These algorithms will interpret the visual information and extract relevant features such as objects, obstacles, people, and text.. Train the system to recognize and classify objects in the environment. This includes common objects like furniture, doors, stairs, as well as potential hazards like obstacles or uneven terrain.. Continuously improve the system's performance and accuracy through machine learning  techniques. .

Existing System Object Detection: The system should be able to detect objects in the environment using computer vision techniques. This could involve the use of pre-trained deep learning models such as YOLO (You Only Look Once), SSD (Single Shot Multi Box Detector), or Faster R-CNN (Region-based Convolutional Neural Network). Camera Input: The system would likely require a camera to capture the live feed of the environment. This camera feed would be processed by the object detection model to identify objects. Speech Synthesis: Once objects are detected, the system should convert the detected objects into spoken words. This can be achieved using text-to-speech (TTS) synthesis techniques. There are several TTS libraries available in Python, such as pyttsx3 or GTTS (Google Text-to-Speech). Real-time Processing: The system should be capable of real-time processing to provide immediate feedback to the user as they navigate their surroundings.

Virtual Assistance Methodology 6 Develop a sophisticated natural language understanding system capable of processing and interpreting user queries and commands in a conversational manner. Train the system to recognize a wide range of spoken commands and questions related to navigation, object identification, task assistance, and general  inquiries. Implement a robust dialogue management system to maintain context and coherence throughout interactions with the virtual  assistant. Employ techniques such as reinforcement learning or rule-based approaches to optimize dialogue flow and handle multi-turn conversations  effectively. Implement integrations with databases, web services, and third-party APIs to access information related to navigation routes, points of interest, product details, news, weather updates, and more.

Dis-Advantages : Technological Limitations : Despite advancements in technology, there may still be limitations in the accuracy and reliability of the system. Computer vision algorithms may struggle with certain environmental conditions, such as low lighting or complex backgrounds, leading to inaccuracies in object recognition and navigation. 2 . Accessibility Challenges : Depending on the implementation and technology used, the project may not be accessible to all visually impaired individuals. Some may face difficulties in using the virtual assistant due to limitations in technology or user interface design. 3. Privacy and Security Risks: Collecting and processing sensitive user data, such as location information and personal preferences, raises concerns about privacy and data security.

Software Specification Server Side : Python 3.7.4(64-bit) or (32-bit) Client Side : HTML, CSS, Bootstrap IDE : Django Back end : MySQL 6.

System Architecture

Conclusion In conclusion, the VISION-THE-BLIND project represents a significant step forward in leveraging technology to assist visually impaired individuals in their daily lives. Through the implementation of our virtual assistant methodology, we have developed a robust system aimed at providing enhanced accessibility and support to users with visual impairments . Moving forward, we envision further enhancements and developments for the VISION-THE-BLIND project. This includes expanding the functionality of the virtual assistant, incorporating additional features based on user needs and feedback, and collaborating with relevant stakeholders to promote accessibility and inclusivity..

References Smith, J., & Johnson, A. (2022). "Empowering the Visually Impaired: A Review of Assistive Technology Solutions." Journal of Accessibility Studies, 8(2), 112-126. Brown, K., et al. (2023). "Development of a Virtual Assistant for Visually Impaired Individuals." Proceedings of the International Conference on Human-Computer Interaction. National Federation of the Blind. (2021). "Technology Resource Guide for Individuals with Visual Impairments." Retrieved from https://www.nfb.org/programs-services/technology-resource-list.