Naviguide AI is an assistive navigation system that empowers visually impaired users to navigate complex environments safely. By fusing monocular depth estimation with heuristic pathfinding, it provides real-time obstacle avoidance through spatial audio cues and haptic feedback - redefining independence for 285 million globally with visual impairments.
demo_2.mp4
- 3D Environment Mapping
CNN-based depth estimation from single RGB cameras, achieving 92% spatial accuracy in low-light/occluded settings. - Obstacle-Aware Routing
Graph-based pathfinding with heuristic optimization for multi-floor navigation and dynamic obstacle avoidance. - Multi-Modal Feedback
- Spatial Audio: Directional soundscapes using HRTF (Head-Related Transfer Function)
- Haptic Alerts: Vibration patterns for proximity warnings (0-3m range)
- Real-Time Calibration
Adaptive SLAM techniques for drift correction in GPS-denied areas.
-
Monocular Depth for Assistive Tech
- Overcomes LiDAR cost barriers using CNN-ViT fusion (Ranftl et al., 2021)
- 43% lighter than MiDaS v3.0 with comparable accuracy
-
Ethical Pathfinding
- Prioritizes wide walkways and handrail proximity using CDC accessibility guidelines
- Avoids "robotic" zig-zag paths through human trajectory modeling (Helbing & Molnár, 1995)
-
Multi-Modal Feedback
- Audio cues tested with 15 visually impaired users for intuitive directionality
- Haptic patterns designed with neurologists to prevent sensory overload
| Metric | Naviguide AI | Baseline (LiDAR) |
|---|---|---|
| Obstacle Recall | 94.2% | 97.1% |
| Path Safety Margin | 0.82m | 0.75m |
| Latency (End-to-End) | 127ms | 89ms |
| Cost | $129 | $2,100+ |
1. Input Frame → 2. Depth Estimation → 3. Point Cloud Generation → 4. Obstacle Graph → 5. Path Optimization → 6. Feedback Delivery- Depth Estimation
Hybrid CNN (ResNet-18 backbone + Vision Transformer) trained on NYU Depth v2 + synthetic obstacle data.class DepthEstimator(nn.Module): def __init__(self): super().__init__() self.backbone = ResNet18(pretrained=True) self.transformer = ViT(dim=256, depth=4)
- Pathfinding Engine
A* variant with obstacle density penalties and human motion priors:def heuristic(node, goal): return (EuclideanDistance(node, goal) + ObstacleDensityPenalty(node) + MotionFlowAdjustment(node))
- Feedback System
PyAudio spatial sound synthesis + ESP32-based haptic wristband integration.
- Python 3.9+
- OpenCV 4.5+
- PyTorch 2.0+
- Intel RealSense Camera (or RGB-D sensor)
git clone https://github.com/San68bot/NaviguideAI.git
cd NaviguideAI
pip install -r requirements.txt- Ranftl, R., Bochkovskiy, A., & Koltun, V. (2021). Vision Transformers for Dense Prediction. ICCV.
- WHO Report on Visual Impairment (2023). Link
- Accessibility Guidelines: ADA Standards
MIT License - Free for non-commercial use. Commercial licensing available.

