Skip to content

ravencore06/Visual_Slam

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Offline Indoor Navigation System for Visually Impaired

An offline-first indoor navigation system designed to assist visually impaired users by providing real-time guidance using on-device computer vision, mobile sensors, and accessible feedback mechanisms. The system runs entirely on the user’s device as a Progressive Web App (PWA), ensuring low latency, privacy, and reliability in connectivity-restricted environments.

🚀 Features

Real-time indoor landmark detection (exit signs, stairs, elevators)

On-device movement tracking using camera and motion sensors

Rule-based navigation logic for stable guidance

Audio and haptic feedback for accessibility

Fully offline after first load (no backend required)

Privacy-preserving, edge-based inference

🧠 System Architecture

Module 1 – Landmark Detection: Lightweight object detection using MobileNet-based models with TensorFlow.js.

Module 2 – Movement Tracking: Visual odometry (lite) using optical flow and inertial sensors.

Module 3 – Navigation Logic: Deterministic, rule-based decision system for direction guidance.

Module 4 – Accessibility Layer: Voice instructions and vibration cues.

Module 5 – Offline PWA Integration: Service workers for caching models and assets.

🛠️ Tech Stack

JavaScript, HTML, CSS

TensorFlow.js

MobileNet-SSD (fine-tuned)

Web APIs (Camera, Sensors, Speech, Vibration)

Progressive Web App (PWA)

📊 Model Training

Base dataset: Open Images Dataset

Custom indoor images for fine-tuning

Transfer learning on MobileNet-SSD

Optimized and converted for browser-based inference

🔒 Privacy & Offline Support

No server-side processing

No user data leaves the device

Works without internet after initial load

🎯 Use Case

Designed for indoor environments such as colleges, hospitals, malls, and offices where GPS is unreliable or unavailable.

📌 Status

Prototype / Academic Project Focused on accessibility, explainability, and real-world feasibility.

About

An offline indoor navigation system for visually impaired users that detects key indoor landmarks using on-device computer vision, tracks user movement with mobile sensors, and provides real-time audio and haptic navigation guidance through a privacy-preserving Progressive Web App.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors