--- title: Sentinel Vision Navigator emoji: 👁️ colorFrom: green colorTo: blue sdk: static pinned: false license: apache-2.0 short_description: "AI camera guidance for blind users." --- # Sentinel Vision Navigator Sentinel Vision Navigator is a web + Android assistive AI prototype for blind and partially sighted people. It uses live camera frames, voice commands, Akash-hosted Qwen multimodal inference, and a small vision-navigation RAG layer to provide short, calm, physical guidance such as: - Stop. Chair directly ahead, one step away. - Doorway ahead, slightly left. Take two small steps forward. - I cannot see enough. Hold still and slowly pan left, then right. ## Demo - Web app: https://amdvision.qubitpage.com/ - Android APK: https://amdvision.qubitpage.com/downloads/sentinel-vision.apk - Whitepaper: https://amdvision.qubitpage.com/downloads/sentinel-vision-whitepaper.pdf - Pitch deck: https://amdvision.qubitpage.com/downloads/sentinel-vision-pitch-deck.pdf - Model/progress page: https://huggingface.co/lablab-ai-amd-developer-hackathon/SentinelBrain-14B-MoE-v0.1 ## Hackathon Fit - Primary track: Vision & Multimodal AI - Secondary track: AI Agents & Agentic Workflows - Supporting track: Fine-tuning / AMD MI300X progress via SentinelBrain model artifacts - Hugging Face category: Space deployment for public testing ## Safety This is an assistive prototype, not a medical device and not a replacement for a cane, guide dog, trained mobility support, or human assistance. It is designed to be conservative: if it is uncertain, it should ask the user to stop and pan slowly rather than invent a route.