File size: 3,066 Bytes
1efb490
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
---
license: apache-2.0
tags:
  - robotics
  - haptics
  - spatial-understanding
  - touch-sensing
  - force-estimation
pipeline_tag: robotics
---

# Motoko Spatial 1B

Motoko Spatial 1B is a foundation model for 3D haptic spatial understanding in robotics. It takes raw sensor array input from distributed touch sensors across a robot surface and outputs spatial force maps, contact region predictions, and pressure distribution fields.

## Model Details

- **Model type:** 3D haptic spatial foundation model
- **Parameters:** 1B
- **Architecture:** Hybrid CNN + Transformer
- **Input:** 3D coordinate arrays and sensor pressure grids
- **Output:** Force field maps, contact region masks, and pressure heatmaps
- **License:** Apache-2.0

## Intended Use

Motoko Spatial 1B is designed for robotics systems that need dense touch and contact understanding from distributed tactile sensors.

Primary use cases include:

- Dexterous multi-finger manipulation
- Full-body robot touch sensing
- Terrain and surface contact mapping
- Collision detection
- Safe human-robot contact

## Inputs

The model expects structured haptic sensor input containing:

- 3D sensor coordinates
- Pressure grid values
- Optional force and torque channels
- Sensor timing or sampling metadata when available

Raw haptic arrays should be converted into model input tensors with `preprocessor/feature_extractor.py`.

## Outputs

The model produces spatial predictions for downstream robotics control and perception:

- Spatial force field maps
- Contact region masks
- Pressure distribution heatmaps

## Repository Files

| File | Description |
| --- | --- |
| `config.json` | Architecture definition, including layers, attention heads, hidden size, channel count, and spatial dimensions. |
| `configs/sensor_config.yaml` | Sensor array layout, sampling rate, axes, channel names, and physical units. |
| `preprocessor/preprocessor_config.json` | Signal normalization, channel statistics, windowing, and resampling configuration. |
| `model/model.safetensors` | Actual trained model weights. The current scaffold contains a placeholder until trained weights are added. |
| `model/model.safetensors.index.json` | Weight index used for loading sharded or indexed safetensors weights. |
| `preprocessor/feature_extractor.py` | Converts raw haptic arrays into normalized model input tensors. |
| `tokenizer_config.json` | Signal tokenizer metadata for quantized or discretized haptic tokens. |
| `tokenizer.json` | Minimal tokenizer vocabulary placeholder. |
| `configs/training_config.yaml` | Training hyperparameters and checkpoint cadence. |
| `examples/inference.py` | Basic inference preprocessing example. |
| `examples/spatial_map.py` | Spatial force map construction example. |

## Limitations

This repository is currently a minimal Hugging Face model scaffold. The included `model/model.safetensors` file is a placeholder and should be replaced with trained weights before production use.

## Citation

Citation information will be added when a technical report or paper is available.