qubitpage's picture
Add Sentinel Vision Navigator progress log
888c6bf verified
---
license: apache-2.0
tags:
- sentinelbrain
- moe
- rocm
- mi300x
- pytorch
- checkpoint
- realignment
library_name: pytorch
pipeline_tag: text-generation
---
# SentinelBrain 14B MoE v0.1 - Frankenstein Realignment v2
This repository now includes SentinelBrain Frankenstein realignment v2 artifacts from the AMD MI300X run completed on 2026-05-03.
## v2 Training Update
- Architecture: custom SentinelBrain sparse MoE decoder, approximately 14.4B stored parameters, 4 experts, top-2 routing, 24 layers, d_model 4096, seq_len 4096.
- Hardware: AMD Instinct MI300X via ROCm/HIP.
- Run: Frankenstein realignment v2 from raw merged checkpoint.
- Completed steps: 5,000.
- Total training tokens during realignment: approximately 0.98B.
- Best validation loss observed: 5.1359.
- Final checkpoint: `checkpoints/frankenstein_v2_final.pt`.
- Best checkpoint: `checkpoints/frankenstein_v2_best.pt`.
- EMA best checkpoint: `checkpoints/frankenstein_v2_ema_best.pt`.
- Previous Hugging Face version preserved on branch: `previous-before-v2-realign-5000-20260503-103121`.
## Included Files
- `checkpoints/frankenstein_v2_final.pt`: full final checkpoint at step 5000, including optimizer/progress state.
- `checkpoints/frankenstein_v2_best.pt`: best model-only checkpoint by validation loss.
- `checkpoints/frankenstein_v2_ema_best.pt`: EMA best checkpoint from the v2 run.
- `checkpoints/sentinelbrain_pretrain_step2471_hf.pt`: pretrain anchor used for comparison.
- `logs/realign_v2.log`: full realignment console log.
- `logs/realign_v2_metrics.jsonl`: step metrics emitted during training.
- `reports/train_metrics_final.json`: final dashboard training metrics snapshot.
- `reports/conductor_state_final.json`: final dashboard/conductor state.
- `reports/sft_combined_ready_report.*`: cleaned SFT dataset preflight report.
- `reports/sentinelbrain_quality_stub_full_fixed.json`: MI300X executable-code benchmark report.
A full progress archive containing all v2 milestones and optimizer-bearing checkpoints is backed up off-Hub on the Azure VM at `/home/msrusu/sentinelbrain_backups/v2_realign_5000/sentinelbrain_v2_realign_full_20260503.tar.zst`. A SHA256 sidecar is generated at archive completion.
## Current Evaluation Notes
MI300X executable-code tests show that v2 realignment is not yet ready as a coding assistant checkpoint:
| Checkpoint | Pass@1 | Syntax Rate | Notes |
| --- | ---: | ---: | --- |
| `frankenstein_v2_best.pt` | 0.0% | 62.5% | Failed all 8 HumanEval-style stub tasks. |
| `frankenstein_v2_final.pt` | 0.0% | 75.0% | Failed all 8 HumanEval-style stub tasks. |
| `sentinelbrain_pretrain_step2471_hf.pt` | 0.0% | 87.5% | Failed all 8 tasks but produced the most syntactically valid Python. |
Interpretation: v2 successfully completed corpus realignment and preserved all progress artifacts, but it needs a focused next phase of executable code SFT, function-call/chat formatting, and auto-critic rejection sampling before quality claims should be made.
## Dataset Preparation Status
The next SFT combined dataset was cleaned non-destructively on the MI300X server:
- Input rows: 42,138.
- Kept rows: 32,996 (78.3%).
- Removed rows: 9,142.
- Max estimated tokens: 3,072.
- Main removals: short assistant/user outputs, garbage responses, repetitive responses, and over-length samples.
## Loading
These are custom SentinelBrain PyTorch checkpoints, not standard Hugging Face `AutoModelForCausalLM` weights. Load with the SentinelBrain code from `/workspace/sentinelprime` or the matching source package.
```python
import torch
from config import ModelConfig
from model.sentinel import SentinelBrain
ckpt = torch.load("checkpoints/frankenstein_v2_best.pt", map_location="cpu", weights_only=False)
model = SentinelBrain(ModelConfig())
state = ckpt.get("model_state_dict") or ckpt.get("model") or ckpt
model.load_state_dict(state, strict=False)
model.eval()
```
## Next Phase Direction
The recommended next phase is a controlled SFT/auto-critic cycle: train from the pretrain anchor plus selected v2 weights only after passing format probes, prioritize executable Python/TypeScript/code-repair datasets, reject non-compiling generations, and benchmark every 250-500 steps before continuing.
<!-- sentinel-vision-navigator-progress-2026-05-07 -->
## Sentinel Vision Navigator Progress Log
**Date:** 2026-05-07
**Hackathon:** AMD Developer Hackathon
**Tracks:** Vision & Multimodal AI, AI Agents & Agentic Workflows, Fine-Tuning on AMD GPUs, Hugging Face Space
Sentinel Vision Navigator is now connected to this model page as the applied assistive-AI product layer for the SentinelBrain roadmap. The public prototype focuses on blind and partially sighted users who need immediate, camera-first understanding of their surroundings.
### Current build
- Public web demo: https://amdvision.qubitpage.com/
- Hugging Face Space: https://huggingface.co/spaces/lablab-ai-amd-developer-hackathon/sentinel-vision-navigator
- Whitepaper PDF: https://amdvision.qubitpage.com/downloads/sentinel-vision-whitepaper.pdf
- Pitch deck PDF: https://amdvision.qubitpage.com/downloads/sentinel-vision-pitch-deck.pdf
- Android APK distribution path: https://amdvision.qubitpage.com/downloads/sentinel-vision.apk
### Implemented application features
- Live camera scene analysis for blind mobility prompts.
- Camera-first local navigation behavior for commands such as `guide me`, `get out`, `avoid obstacles`, and `which way`.
- Embedded vision-direction RAG knowledge for indoor paths, outdoor sidewalks, stairs, curbs, text reading, and object finding.
- Fluent English speech output with calmer, shorter instructions.
- Android first-run setup for camera, mic/speech, contacts, location, email handoff, WhatsApp handoff, SMS, and calls.
- Multitasking agent routing for navigation, calls, SMS, WhatsApp, email, maps, app opening, battery, time, text reading, object finding, and conversation.
- 3D Terminator-style camera HUD with reticle, scan line, perspective grid, connector status, and detected-object chips.
### Model status note
The production demo currently uses Akash-hosted Qwen multimodal inference for visual reasoning while SentinelBrain-14B-MoE-v0.1 remains the project model artifact and training/progress page for the specialized assistant roadmap. The next step is to collect interaction traces and evaluate a fine-tuned SentinelBrain navigation/router layer on AMD GPU infrastructure.