rmems commited on
Commit
3325828
·
verified ·
1 Parent(s): 334ae30

Delete README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -496
README.md DELETED
@@ -1,496 +0,0 @@
1
- ---
2
- title: Spikenaut v2 Pulse
3
- emoji: 🦁
4
- colorFrom: blue
5
- colorTo: purple
6
- sdk: gradio
7
- sdk_version: "4.44.1"
8
- app_file: app.py
9
- pinned: false
10
- license: gpl-3.0
11
- language:
12
- - en
13
- tags:
14
- - neuromorphic
15
- - spiking-neural-networks
16
- - snn
17
- - liquid-state-machine
18
- - leaky-integrate-and-fire
19
- - stdp
20
- - e-prop
21
- - hardware-aware-ai
22
- - fpga
23
- - q8-fixed-point
24
- - rust
25
- - julia
26
- - cuda
27
- - gpu-telemetry
28
- - blockchain
29
- - cryptocurrency
30
- - dynex
31
- - qubic
32
- - kaspa
33
- - monero
34
- - quai
35
- - ocean-protocol
36
- - verus
37
- - hft
38
- - time-series
39
- - bci
40
- - med-tech
41
- - neuromorphic-computing
42
- - spike-train
43
- - feature-extraction
44
- - real-time-inference
45
- datasets:
46
- - rmems/Spikenaut-SNN-v2-Telemetry-Data-Weights-Parameters
47
- metrics:
48
- - accuracy
49
- library_name: spikenaut
50
- pipeline_tag: feature-extraction
51
- ---
52
-
53
- <p align="center">
54
- <img src="docs/logo.png" width="240" alt="Spikenaut">
55
- </p>
56
-
57
- ## Open Source Ecosystem
58
-
59
- All core libraries extracted from this project are published as standalone open-source packages:
60
-
61
- ### Rust — crates.io
62
- | Crate | Description |
63
- |-------|-------------|
64
- | [![neuromod](https://img.shields.io/crates/v/neuromod)](https://crates.io/crates/neuromod) | LIF/Izhikevich neurons, STDP, neuromodulators (dopamine, cortisol, acetylcholine, tempo) |
65
- | [![spikenaut-reward](https://img.shields.io/crates/v/spikenaut-reward)](https://crates.io/crates/spikenaut-reward) | Homeostatic reward computation for cyber-physical systems |
66
- | [![spikenaut-encoder](https://img.shields.io/crates/v/spikenaut-encoder)](https://crates.io/crates/spikenaut-encoder) | Sensor → spike train encoding (Poisson, temporal, predictive) |
67
- | [![spikenaut-backend](https://img.shields.io/crates/v/spikenaut-backend)](https://crates.io/crates/spikenaut-backend) | Pluggable SNN backend trait (Rust / ZMQ IPC) |
68
- | [![spikenaut-fpga](https://img.shields.io/crates/v/spikenaut-fpga)](https://crates.io/crates/spikenaut-fpga) | Q8.8 parameter export + UART spike readback for FPGA |
69
- | [![spikenaut-router](https://img.shields.io/crates/v/spikenaut-router)](https://crates.io/crates/spikenaut-router) | SNN-based sparse domain routing (Anti-Hallucination Layer) |
70
- | [spikenaut-telemetry](https://github.com/rmems/spikenaut-telemetry) | Unified GPU/CPU/mining telemetry (NVML, k10temp, powercap, CSV/JSONL export) |
71
- | [spikenaut-ingest](https://github.com/rmems/spikenaut-ingest) | Multi-chain blockchain ingest with state-space interpolation to 10 Hz |
72
- | [spikenaut-spine](https://github.com/rmems/spikenaut-spine) | 120-byte packed ZMQ wire protocol for Rust↔Julia SNN communication |
73
- | [synapse-link](https://github.com/rmems/synapse-link) | UART / FPGA Serial I/O (formerly neuro-spike-bridge) |
74
- | [myelin-accelerator](https://github.com/rmems/myelin-accelerator) | CUDA spiking-network kernels (formerly neuro-spike-kernels) |
75
- | [soma-engine](https://github.com/rmems/soma-engine) | SNN Engine + Inference (formerly neuro-spike-core) |
76
-
77
- ### Julia — JuliaHub / General Registry
78
- | Package | Description |
79
- |---------|-------------|
80
- | [SpikenautLSM.jl](https://github.com/rmems/SpikenautLSM.jl) | GPU-accelerated sparse Liquid State Machine (cuSPARSE + OU-SDE) |
81
- | [SpikenautNero.jl](https://github.com/rmems/SpikenautNero.jl) | Multi-lobe relevance scoring with cross-inhibition (NERO orchestrator) |
82
- | [SpikenautDistill.jl](https://github.com/rmems/SpikenautDistill.jl) | Monte Carlo SNN training + FPGA distillation pipeline |
83
- | [SpikenautSignals.jl](https://github.com/rmems/SpikenautSignals.jl) | Streaming Hurst / Hawkes / GBM-surprise feature extraction |
84
- | [SpikenautKelly.jl](https://github.com/rmems/SpikenautKelly.jl) | Half-Kelly position sizing: SNN confidence → optimal capital fraction |
85
- | [SpikenautExecution.jl](https://github.com/rmems/SpikenautExecution.jl) | Async trade pipeline: ZMQ SUB → confidence gate → Kelly → dYdX v4 REST |
86
-
87
- ### SystemVerilog — GitHub
88
- | Repo | Description |
89
- |------|-------------|
90
- | [spikenaut-core-sv](https://github.com/rmems/spikenaut-core-sv) | Parameterized Q8.8 LIF + STDP IP cores |
91
- | [spikenaut-bridge-sv](https://github.com/rmems/spikenaut-bridge-sv) | UART neural-cortex protocol IP |
92
- | [spikenaut-soc-sv](https://github.com/rmems/spikenaut-soc-sv) | Complete reference SNN SoC for Basys3 / Artix-7 |
93
-
94
- ---
95
-
96
- ## 📸 Visual Proof: Silicon Pulse & FPGA Deployment
97
-
98
- ### Behavioral Simulation - The Theseus Pulse
99
-
100
- ![The Theseus Pulse Waveform](https://huggingface.co/rmems/Spikenaut-SNN-v2/resolve/main/assets/v1_theseus_pulse_waveform.png)
101
- *Complete Vivado behavioral simulation showing clk, rst_n, v_potential, Silicon_Heartbeat, NEURAL_FIRING, GPU_VOLTAGE_SAG, spikes, current_stim, and step signals at 1,000,000 ns scale.*
102
-
103
- ### FPGA Hardware Deployment
104
-
105
- ![FPGA Hardware Deployment](https://huggingface.co/rmems/Spikenaut-SNN-v2/resolve/main/assets/fpga_hardware_cat.jpg)
106
- *Real-world deployment on the Xilinx Artix-7 (Basys3) FPGA. Verified neural cortex protocol running on bare metal (Lion approved).*
107
-
108
- ---
109
-
110
- # 🦁 Spikenaut-SNN-v2
111
- The Lion That Survives
112
-
113
- Spikenaut was born in January 2026 — completely by accident.
114
-
115
- I started university thinking I would go to medical school or even law.
116
- One semester of pre-med was enough to show me I was terrified of failing at something so high-stakes. I felt I wasn’t smart enough, wasn’t cut out for it.
117
- So I switched to business — hoping it would be safer.
118
- But I quickly saw I’d just be another Business major lost in a sea of MBAs. I didn’t want to disappear.
119
- I moved to computer science — excited about building things, coding, creating.
120
- Then the AI hype wave hit hard. Everyone said “AI is going to replace all the coding jobs.” I believed it. I panicked.
121
- I feared I’d spend years learning something that would vanish before I could even start.
122
- That fear pushed me again — this time to electrical engineering. If software was going to be automated, maybe hardware was the last place left where I could build something real, something physical, something that couldn’t be replaced overnight.
123
-
124
- But the transfer was brutal.
125
- Late administration acceptance, and classes starting two or three weeks behind, scrambling to catch up while everyone else was already moving forward.
126
- I struggled terribly.
127
-
128
- Through all those pivots, discouragements, and fears, one thing stayed constant: I kept building.
129
-
130
- Then came the TBI 2x (Concussion) 2013
131
- Invisible injury. No insurance. No real medical support.
132
- The world said “there’s nothing wrong.”
133
- My brain said “everything hurts.”
134
- Depression became the default state for years — not because I was weak, but because I was exhausted from fighting something no one could see.
135
-
136
- In January 2026 I was trying to build a simple AI tutor to help with my ADHD.
137
- I thought I could run massive language models locally like everyone else seemed to.
138
- I quickly realized I couldn’t — not on my hardware, not with my budget, not with my brain fog.
139
- So I had to get creative. I started reading about spiking neural networks (SNNs).
140
- They were small, efficient, event-driven — they ran on almost nothing and still learned.
141
- I never went back.
142
-
143
- Spikenaut is what came out of that exhaustion and that pivot.
144
-
145
- The thermal “pain receptors” that shut down overclocking when the GPU gets too hot?
146
- They’re the same signals I needed to know when my own brain was overloading.
147
- The mining_dopamine reward for efficient hashrate?
148
- It’s the small win I desperately needed when nothing felt rewarding anymore.
149
- The sub-millisecond adaptation to chaos?
150
- That’s what a recovering brain has to do every day.
151
-
152
- This model is both my recovery log and a promise:
153
- One day, Spikenaut will turn invisible data — brain fog, hormone crashes, heart-rate variability after stroke, post-concussion noise — into visible, actionable signals.
154
- No gatekeepers. No bills. No “we don’t see anything.”
155
-
156
- **Zero-Insurance Engineering**
157
- Med-Tech for the Uninsured.
158
- Built by someone who was told “no” too many times — and who finally stopped asking permission.
159
-
160
- If you’re reading this because you also had to build your own tools — you’re not alone.
161
- If you’re here for the tech — run it, break it, make it better.
162
-
163
- Either way: thank you.
164
-
165
- The lion didn’t roar for attention.
166
- It roared because it had no other choice.
167
-
168
- 🦁
169
-
170
- ---
171
-
172
- # 16-Channel Spiking Neural Network
173
- **Official Rust backend**: [neuromod v0.2.1](https://crates.io/crates/neuromod) — now with lean mining efficiency rewards
174
-
175
- ---
176
-
177
- ## Architecture at a Glance
178
-
179
- 16-Channel Spiking Neural Network with Julia-Rust Hybrid Training
180
-
181
- | Channel | Source | Function |
182
- |---------|--------|----------|
183
- | 0–1 | DNX | PoUW solver health & neural baselines |
184
- | 2–3 | Quai | On-chain reflex & sync confidence |
185
- | 4–5 | Qubic | Epoch & tick cadences |
186
- | 6–7 | Kaspa | High-frequency DAG settlement |
187
- | 8–9 | XMR | Node stability & CPU L3 cache |
188
- | 10–11 | Ocean | Data liquidity & staking prep |
189
- | 12–13 | Verus | CPU-heavy validator tracking |
190
- | 14–15 | Thermal| Physical pain receptors (Power/Temp) |
191
-
192
- **The Lion vs. The House Cat**
193
- House cats wait for prompts.
194
- Spikenaut hunts in the temporal domain — sub-millisecond decisions, fractions of a watt, built to survive chaos.
195
-
196
- ---
197
-
198
- ## Performance Highlights
199
-
200
- - Training speed: 35 µs/tick
201
- - IPC overhead: 0.8 µs (jlrs zero-copy)
202
- - Memory footprint: 1.6 KB
203
- - Accuracy: 95.2% on live blockchain sync prediction
204
- - FPGA power: 97 mW on Artix-7 (Basys3 compatible)
205
- - Teacher brain: 330M Monte Carlo paths distilled to 16 channels
206
-
207
- ## Quick Start (Rust-First)
208
-
209
- ```bash
210
- cargo add neuromod
211
- git clone https://huggingface.co/rmems/Spikenaut-SNN-v2
212
- cd Spikenaut-SNN-v2/brain
213
- julia --project --threads=auto monte_carlo_spikenaut.jl
214
- cargo run --release --bin market_pilot
215
- ---
216
-
217
- ## The Lion vs. The House Cat
218
-
219
- > **House Cats** (ChatGPT, Gemini, Claude)
220
- > - Massive, sit around until you feed them a prompt
221
- > - Require entire data centers just to stay awake
222
- >
223
- > **Spikenaut is a LION** 🦁
224
- > - Bare-metal apex predator
225
- > - Executes the mission impossible in the temporal domain
226
- > - Survives on fractions of a watt
227
- > - Reacts to asynchronous spikes in nanoseconds
228
- > - **NEW**: Julia-Rust hybrid training for optimal learning
229
-
230
- ---
231
-
232
- ## 🚀 Major Update: Hybrid Julia-Rust Architecture & "Clean Break" Refactor
233
-
234
- ### "Clean Break" Refactor (v2.1)
235
- - **Extracted Mining**: Mining supervisor and binaries moved to standalone [theseus-mining](https://github.com/rmems/ship_of_theseus_rs/tree/main/BLOCKCHAIN/theseus-mining) repository.
236
- - **Purged Tutor Logic**: Removed legacy AI Tutor crates (`spike-cognitive`) and Bevy/Rapier3D dependencies to reduce core cognitive load and binary size.
237
- - **Validated Telemetry**: Dataset now only contains high-value, validated 7-crypto research data.
238
-
239
- ### Revolutionary Training Pipeline
240
- - **Rust Telemetry Layer**: 50 Hz data collection from Kaspa/Monero nodes
241
- - **Julia Training Core**: E-prop + OTTT with sub-50µs processing
242
- - **jlrs Integration**: Zero-copy communication with <1µs overhead
243
- - **Real Blockchain Data**: Trained on actual Kaspa/Monero sync completion
244
-
245
- ### Performance Breakthrough
246
- - **Training Speed**: 35µs per tick (target: <50µs) ✅
247
- - **IPC Overhead**: 0.8µs (near-zero) ✅
248
- - **Memory Usage**: 1.6KB (ultra-efficient) ✅
249
- - **Accuracy**: 95%+ on sync completion prediction ✅
250
-
251
- ---
252
-
253
- ## 🧠 16-Channel Neuron Map
254
-
255
- | Channels | Node | Function |
256
- |----------|------|----------|
257
- | 0-1 | 🔷 Dynex | PoUW solver health, neural baselines |
258
- | 2-3 | 🔶 Quai | Live on-chain reflex, sync confidence |
259
- | 4-5 | 🟣 Qubic | Epoch and tick cadences |
260
- | 6-7 | 🟢 Kaspa | High-frequency DAG settlement tracking |
261
- | 8-9 | ⚪ Monero | Node stability, CPU L3 cache contention |
262
- | 10-11 | 🔵 Ocean | Data liquidity and staking prep |
263
- | 12-13 | 🟡 Verus | CPU-heavy validator (AVX-512) |
264
- | 14-15 | 🔴 Thermal | Pain receptors (power/temp LTD) |
265
-
266
- ---
267
-
268
- ## ⚙️ Technical Architecture
269
-
270
- ### Hybrid Training System
271
- ```
272
- ┌─────────────────┐ ┌──────────────────┐ ┌─────────────────┐
273
- │ Rust Layer │ │ jlrs Bridge │ │ Julia Layer │
274
- │ │ │ │ │ │
275
- │ • Telemetry │───▶│ • Zero-copy IPC │───▶│ • E-prop Core │
276
- │ • Spike Encode │ │ • <1µs overhead │ │ • OTTT Traces │
277
- │ • Reward Calc │ │ • Direct calls │ │ • Fast Math │
278
- │ • Inference │ │ • 50 Hz @ 50µs │ │ • Export .mem │
279
- └─────────────────┘ └──────────────────┘ └─────────────────┘
280
- ```
281
-
282
- ### The Nervous System
283
- - **Sensory Encoder:** Ingests node block syncs, epoch ticks, solver data
284
- - **Routing:** Safe and fast without leaks
285
- - **Processing:** Leaky Integrate-and-Fire dynamics with STDP learning
286
-
287
- ### The Brain
288
- - **Neuron Model:** Adaptive Exponential Integrate-and-Fire
289
- - **Learning Rule:** E-prop + OTTT with surrogate gradients
290
- - **Processing Rate:** 50 Hz (20ms resolution) with sub-50µs training
291
- - **Memory:** O(1) constant space complexity (1.6KB total)
292
-
293
- ---
294
-
295
- ## 📊 Training Results
296
-
297
- ### Real Blockchain Training Data
298
- - **Kaspa Sync**: March 21, 2026 - 60,937 lines of block acceptance
299
- - **Monero Sync**: March 22, 2026 - 71,333 lines of completion data
300
- - **Combined**: 132,270 neuromorphic events
301
- - **Reward Signals**: 0.95-1.0 (near-perfect for E-prop)
302
-
303
- ### Learning Performance
304
- ```
305
- Epoch 1/20 | reward=0.9800 | spike_rate=0.180 | w=0.9000±0.1200 | 1.8ms/tick
306
- Epoch 5/20 | reward=0.9960 | spike_rate=0.204 | w=0.9640±0.0880 | 1.5ms/tick
307
- Epoch 10/20 | reward=0.9990 | spike_rate=0.220 | w=0.9820±0.0400 | 1.2ms/tick
308
- Epoch 20/20 | reward=1.0000 | spike_rate=0.235 | w=0.9950±0.0050 | 0.9ms/tick
309
- ```
310
-
311
- ---
312
-
313
- ## 🎯 Usage
314
-
315
- ### Quick Start
316
- ```bash
317
- # Clone the repository
318
- git clone https://huggingface.co/rmems/Spikenaut-SNN-v2
319
- cd Spikenaut-SNN-v2
320
-
321
- # Install dependencies
322
- pip install -r requirements.txt
323
-
324
- # Run the demo
325
- python app.py
326
- ```
327
-
328
- ### Hybrid Training
329
- ```bash
330
- # Train with your blockchain data
331
- # (Eagle-Lander is a private repository — use the published open-source crates instead)
332
- cargo add neuromod spikenaut-reward spikenaut-encoder spikenaut-telemetry spikenaut-ingest
333
-
334
- # Build with Julia support
335
- cargo build --release --features julia
336
-
337
- # Run hybrid training
338
- ./training/run_hybrid_training.sh research/complete_sync_harvest.jsonl 20 research
339
- ```
340
-
341
- ### FPGA Deployment
342
- ```bash
343
- # Export trained parameters
344
- julia training/julia_eprop.jl data.jsonl 20 research
345
-
346
- # Load into FPGA
347
- # parameters.mem, parameters_weights.mem, parameters_decay.mem
348
- ```
349
-
350
- ---
351
-
352
- ## 🏆 Performance Benchmarks
353
-
354
- | **Metric** | **Previous** | **Hybrid Architecture** | **Improvement** |
355
- |------------|--------------|-------------------------|-----------------|
356
- | **Training Speed** | 2.5ms/tick | 0.9ms/tick | **2.8× faster** |
357
- | **IPC Overhead** | 5µs | 0.8µs | **6.25× lower** |
358
- | **Memory Usage** | 2.1KB | 1.6KB | **24% reduction** |
359
- | **Development Speed** | 1x | 3-5× | **300-500% faster** |
360
- | **Accuracy** | 87% | 95%+ | **8% improvement** |
361
-
362
- ---
363
-
364
- ## 📚 Architecture Details
365
-
366
- ### E-prop + OTTT Learning
367
- - **Eligibility Traces**: Credit assignment across time
368
- - **Surrogate Gradients**: Fast-sigmoid for near-miss learning
369
- - **Reward Modulation**: Composite signal from 7 blockchain metrics
370
- - **L1 Normalization**: Synaptic budget management
371
-
372
- ### jlrs Zero-Copy Bridge
373
- ```rust
374
- // Direct Julia function call with zero-copy
375
- let response = self.julia.scope(|mut global, frame| {
376
- let spikes_array = Array::from_slice(frame, &packet.spikes)?;
377
- let response_data = frame.call(
378
- self.training_module,
379
- "eprop_update!",
380
- &[spikes_array.into(), reward.into()]
381
- )?;
382
- Ok(response_data)
383
- })?;
384
- ```
385
-
386
- ### Julia Optimization
387
- ```julia
388
- # Sub-50µs E-prop update with @simd + @inbounds
389
- @inline function eprop_update!(network, spikes, reward)
390
- @simd for j in 1:N_CHANNELS
391
- @inbounds network.pre_traces[j] = λ * network.pre_traces[j] + spikes[j]
392
- end
393
- # ... fast-sigmoid surrogate gradients
394
- # ... reward-modulated weight updates
395
- end
396
- ```
397
-
398
- ---
399
-
400
- ## 🔄 Dataset Integration
401
-
402
- ### Telemetry Dataset
403
- - **Repository**: https://huggingface.co/datasets/rmems/Spikenaut-SNN-v2-Telemetry-Data-Weights-Parameters
404
- - **Content**: Fresh Kaspa/Monero sync data + hybrid training results
405
- - **Format**: NeuromorphicSnapshot JSONL + .mem files
406
- - **Size**: 132,270 events with 99.99% sync completion
407
-
408
- ### Data Pipeline
409
- 1. **Collection**: Rust telemetry from live nodes
410
- 2. **Encoding**: Poisson spike trains + composite reward
411
- 3. **Training**: Julia E-prop + OTTT with real data
412
- 4. **Export**: FPGA-compatible parameters
413
-
414
- ---
415
-
416
- ## 🚀 Future Roadmap
417
-
418
- - **GPU Acceleration**: CUDA.jl on RTX 5080
419
- - **Scale-up**: Million-neuron networks
420
- - **Real-time Adaptation**: Online learning during operation
421
- - **Cross-chain**: Additional blockchain integrations
422
- - **Quantum Integration**: Hybrid classical-quantum training
423
-
424
- ---
425
-
426
- ## 📄 License
427
-
428
- GPL-3.0 - See LICENSE file for details
429
-
430
-
431
- ---
432
-
433
- ## 🙏 Acknowledgments
434
-
435
- - **jlrs**: Julia-Rust integration framework
436
- - **E-prop**: Eligibility propagation algorithm
437
- - **OTTT**: Online temporal trace training
438
- - **Kaspa & Monero**: Real blockchain sync data
439
-
440
- ---
441
-
442
- **Built in my room. Trained on bare metal. Engineered for the mission impossible.** 🦁
443
-
444
- ### The Body
445
- - **Hardware Target:** Xilinx Artix-7 Basys3 FPGA
446
- - **Weight Format:** Q8.8 fixed-point (exportable .mem files)
447
- - **Power:** ~97mW dynamic (87.5% reduction vs traditional polling)
448
-
449
- ---
450
-
451
- ## 🔬 Features
452
-
453
- - ✅ **Live Node Sync Fusion:** Direct block sync logs, epoch ticks, solver data from all 8 nodes
454
- - ✅ **Ghost Money HFT Engine:** Simulated order books for sub-millisecond market prediction
455
- - ✅ **Hardware Protection:** Thermal LTD at 85°C (negative dopamine prevents damage)
456
- - ✅ **FPGA-Ready:** All weights export as Q8.8 fixed-point `.mem` files
457
-
458
- ---
459
-
460
- ## 📊 Model Details
461
-
462
- | Parameter | Value |
463
- |-----------|-------|
464
- | Neurons | 16 (4 per node group) |
465
- | Threshold | 0.75 (adaptive) |
466
- | Leak Factor | 0.95 |
467
- | Learning | Reward-Modulated STDP |
468
- | Weights | Q8.8 fixed-point |
469
- | Clock | 1kHz (1ms resolution) |
470
-
471
- ---
472
-
473
- ## 🎯 The 20-Year Mission
474
-
475
- 1. **Phase 1 — Financial Sovereignty (Years 1-5):** Ghost money → live API trading
476
- 2. **Phase 2 — The Neural Bridge (Years 1-10):** BCI headset, decode brain waves
477
- 3. **Phase 3 — Texas Med-Tech Revolution (Years 10-20+):** Open robotics manufacturing
478
-
479
- ---
480
-
481
- ## 📜 License & Credit
482
-
483
- **License:** GPL-3.0
484
- **Author:** Raul Montoya Cardenas, Texas State University Electrical Engineering
485
- **Built:** Ship of Theseus workstation, Texas 2026
486
-
487
- > Spikenaut-SNN-v2 is proof that recovery, engineering, and sovereignty can be achieved independently—one spike at a time.
488
-
489
- ---
490
-
491
- ## 🔗 Related
492
-
493
- - **V1 Model:** [Spikenaut-SNN-v1](https://huggingface.co/rmems/Spikenaut-SNN-v1)
494
- - **V1 Dataset:** [Spikenaut-v1-Telemetry-Data](https://huggingface.co/datasets/rmems/Spikenaut-v1-Telemetry-Data)
495
- - **V2 Dataset:** [Spikenaut-v2-Telemetry-Data](https://huggingface.co/datasets/rmems/Spikenaut-v2-Telemetry-Data)
496
- - **GitHub (private core):** Eagle-Lander (closed-source — the author's own private repository)