qubitpage commited on
Commit
888c6bf
·
verified ·
1 Parent(s): 10643b7

Add Sentinel Vision Navigator progress log

Browse files
Files changed (1) hide show
  1. README.md +34 -1
README.md CHANGED
@@ -84,4 +84,37 @@ model.eval()
84
 
85
  ## Next Phase Direction
86
 
87
- The recommended next phase is a controlled SFT/auto-critic cycle: train from the pretrain anchor plus selected v2 weights only after passing format probes, prioritize executable Python/TypeScript/code-repair datasets, reject non-compiling generations, and benchmark every 250-500 steps before continuing.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
84
 
85
  ## Next Phase Direction
86
 
87
+ The recommended next phase is a controlled SFT/auto-critic cycle: train from the pretrain anchor plus selected v2 weights only after passing format probes, prioritize executable Python/TypeScript/code-repair datasets, reject non-compiling generations, and benchmark every 250-500 steps before continuing.
88
+
89
+ <!-- sentinel-vision-navigator-progress-2026-05-07 -->
90
+
91
+ ## Sentinel Vision Navigator Progress Log
92
+
93
+ **Date:** 2026-05-07
94
+ **Hackathon:** AMD Developer Hackathon
95
+ **Tracks:** Vision & Multimodal AI, AI Agents & Agentic Workflows, Fine-Tuning on AMD GPUs, Hugging Face Space
96
+
97
+ Sentinel Vision Navigator is now connected to this model page as the applied assistive-AI product layer for the SentinelBrain roadmap. The public prototype focuses on blind and partially sighted users who need immediate, camera-first understanding of their surroundings.
98
+
99
+ ### Current build
100
+
101
+ - Public web demo: https://amdvision.qubitpage.com/
102
+ - Hugging Face Space: https://huggingface.co/spaces/lablab-ai-amd-developer-hackathon/sentinel-vision-navigator
103
+ - Whitepaper PDF: https://amdvision.qubitpage.com/downloads/sentinel-vision-whitepaper.pdf
104
+ - Pitch deck PDF: https://amdvision.qubitpage.com/downloads/sentinel-vision-pitch-deck.pdf
105
+ - Android APK distribution path: https://amdvision.qubitpage.com/downloads/sentinel-vision.apk
106
+
107
+ ### Implemented application features
108
+
109
+ - Live camera scene analysis for blind mobility prompts.
110
+ - Camera-first local navigation behavior for commands such as `guide me`, `get out`, `avoid obstacles`, and `which way`.
111
+ - Embedded vision-direction RAG knowledge for indoor paths, outdoor sidewalks, stairs, curbs, text reading, and object finding.
112
+ - Fluent English speech output with calmer, shorter instructions.
113
+ - Android first-run setup for camera, mic/speech, contacts, location, email handoff, WhatsApp handoff, SMS, and calls.
114
+ - Multitasking agent routing for navigation, calls, SMS, WhatsApp, email, maps, app opening, battery, time, text reading, object finding, and conversation.
115
+ - 3D Terminator-style camera HUD with reticle, scan line, perspective grid, connector status, and detected-object chips.
116
+
117
+ ### Model status note
118
+
119
+ The production demo currently uses Akash-hosted Qwen multimodal inference for visual reasoning while SentinelBrain-14B-MoE-v0.1 remains the project model artifact and training/progress page for the specialized assistant roadmap. The next step is to collect interaction traces and evaluate a fine-tuned SentinelBrain navigation/router layer on AMD GPU infrastructure.
120
+