Beyond Vision Logo
Hacettepe University Logo
Project Logo

Vision Beyond Sight

Introduction

Today, over 285 million people worldwide live with visual impairments—39 million of them are completely blind World Health Organization.While assistive technology has come a long way, traditional tools often fall short in adapting to real-world environments. That's where AI and Vision-Language Models (VLMs) step in—offering the ability to understand both images and language for smarter, more responsive assistance.At Beyond Vision, we're building modular, AI-powered systems that help visually impaired individuals navigate safely, interact independently, and access information in real time—bringing us closer to a more inclusive and accessible world.

Beyond Vision Logo

Core Modules

Attention

Apply

Communicate

Hover over a module to see its details

Key Results & Achievements

Object Detection Excellence

Real-time Processing Power

Advanced Text Recognition

Intelligent Visual Understanding

Our system achieved remarkable accuracy with 91.3% in campus obstacle detection and 89.4% for ATM interface elements, setting new standards in assistive technology.

Object detection visualization

Analysis

Overall Performance

Object Class Detection Accuracy

Conclusion

Impact & Innovation

Our project represents a transformative approach to how visually impaired individuals interact with their environment, particularly focusing on two critical areas: ATM usage and campus navigation. By enabling independent access to these essential services, we're taking significant steps toward greater autonomy.

Key Features

ATM Module

Features finger-tracking technology with real-time spoken feedback, powered by YOLOv8 for precise finger and button detection, complemented by EasyOCR for accurate text recognition.

Navigation Module

Utilizes a custom-trained YOLOv8 model on thousands of campus images to identify obstacles and provide timely safety alerts, enhancing independent mobility.

Future Directions

  • Enhanced model accuracy through expanded dataset collection
  • Implementation of lightweight AI models for mobile devices and smart glasses
  • Integration of user feedback for improved accessibility
  • Real-world testing in banking and educational environments

While currently in its conceptual phase, this project demonstrates significant potential for real-world impact. With continued development and support, it could evolve into a powerful tool that meaningfully enhances the independence and quality of life for visually impaired individuals.

References

Implementation Studies

  • Johnson, M., et al. (2021)Computer vision-based ATM access system
  • Park, H., et al. (2020)Text recognition on ATM screens using OCR
  • Wang, C., et al. (2022)Domain-specific fine-tuning of YOLO