Add Sentinel Vision Navigator progress log
Browse files
README.md
CHANGED
|
@@ -84,4 +84,37 @@ model.eval()
|
|
| 84 |
|
| 85 |
## Next Phase Direction
|
| 86 |
|
| 87 |
-
The recommended next phase is a controlled SFT/auto-critic cycle: train from the pretrain anchor plus selected v2 weights only after passing format probes, prioritize executable Python/TypeScript/code-repair datasets, reject non-compiling generations, and benchmark every 250-500 steps before continuing.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 84 |
|
| 85 |
## Next Phase Direction
|
| 86 |
|
| 87 |
+
The recommended next phase is a controlled SFT/auto-critic cycle: train from the pretrain anchor plus selected v2 weights only after passing format probes, prioritize executable Python/TypeScript/code-repair datasets, reject non-compiling generations, and benchmark every 250-500 steps before continuing.
|
| 88 |
+
|
| 89 |
+
<!-- sentinel-vision-navigator-progress-2026-05-07 -->
|
| 90 |
+
|
| 91 |
+
## Sentinel Vision Navigator Progress Log
|
| 92 |
+
|
| 93 |
+
**Date:** 2026-05-07
|
| 94 |
+
**Hackathon:** AMD Developer Hackathon
|
| 95 |
+
**Tracks:** Vision & Multimodal AI, AI Agents & Agentic Workflows, Fine-Tuning on AMD GPUs, Hugging Face Space
|
| 96 |
+
|
| 97 |
+
Sentinel Vision Navigator is now connected to this model page as the applied assistive-AI product layer for the SentinelBrain roadmap. The public prototype focuses on blind and partially sighted users who need immediate, camera-first understanding of their surroundings.
|
| 98 |
+
|
| 99 |
+
### Current build
|
| 100 |
+
|
| 101 |
+
- Public web demo: https://amdvision.qubitpage.com/
|
| 102 |
+
- Hugging Face Space: https://huggingface.co/spaces/lablab-ai-amd-developer-hackathon/sentinel-vision-navigator
|
| 103 |
+
- Whitepaper PDF: https://amdvision.qubitpage.com/downloads/sentinel-vision-whitepaper.pdf
|
| 104 |
+
- Pitch deck PDF: https://amdvision.qubitpage.com/downloads/sentinel-vision-pitch-deck.pdf
|
| 105 |
+
- Android APK distribution path: https://amdvision.qubitpage.com/downloads/sentinel-vision.apk
|
| 106 |
+
|
| 107 |
+
### Implemented application features
|
| 108 |
+
|
| 109 |
+
- Live camera scene analysis for blind mobility prompts.
|
| 110 |
+
- Camera-first local navigation behavior for commands such as `guide me`, `get out`, `avoid obstacles`, and `which way`.
|
| 111 |
+
- Embedded vision-direction RAG knowledge for indoor paths, outdoor sidewalks, stairs, curbs, text reading, and object finding.
|
| 112 |
+
- Fluent English speech output with calmer, shorter instructions.
|
| 113 |
+
- Android first-run setup for camera, mic/speech, contacts, location, email handoff, WhatsApp handoff, SMS, and calls.
|
| 114 |
+
- Multitasking agent routing for navigation, calls, SMS, WhatsApp, email, maps, app opening, battery, time, text reading, object finding, and conversation.
|
| 115 |
+
- 3D Terminator-style camera HUD with reticle, scan line, perspective grid, connector status, and detected-object chips.
|
| 116 |
+
|
| 117 |
+
### Model status note
|
| 118 |
+
|
| 119 |
+
The production demo currently uses Akash-hosted Qwen multimodal inference for visual reasoning while SentinelBrain-14B-MoE-v0.1 remains the project model artifact and training/progress page for the specialized assistant roadmap. The next step is to collect interaction traces and evaluate a fine-tuned SentinelBrain navigation/router layer on AMD GPU infrastructure.
|
| 120 |
+
|