File size: 4,765 Bytes
f6d1762
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
76d8203
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f6d1762
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c9ea9aa
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
---
license: cc-by-sa-4.0
base_model: mistralai/Devstral-Small-2-24B-Instruct-2512
library_name: peft
tags:
- mlx
- lora
- peft
- ailiance
- devstral
- platformio
language:
- en
- fr
pipeline_tag: text-generation
---

# Ailiance — Devstral-Small-2-24B-BF16 platformio LoRA

LoRA adapter fine-tuned on `mistralai/Devstral-Small-2-24B-Instruct-2512` for **platformio** tasks.

> Maintained by **Ailiance** — French AI org publishing EU AI Act aligned LoRA adapters and datasets.

## Quick start (MLX)

```python
from mlx_lm import load, generate

model, tokenizer = load(
    "mistralai/Devstral-Small-2-24B-Instruct-2512",
    adapter_path="Ailiance-fr/devstral-platformio-lora",
)

print(generate(model, tokenizer, prompt="..."))
```

## Training

| Hyperparameter   | Value                  |
|------------------|------------------------|
| Base model       | `mistralai/Devstral-Small-2-24B-Instruct-2512`     |
| Method           | LoRA via `mlx-lm`      |
| Rank             | 16            |
| Scale            | 2.0           |
| Alpha            | 32           |
| Max seq length   | 2048  |
| Iterations       | 500           |
| Optimizer        | Adam, LR 1e-5          |
| Hardware         | Apple M3 Ultra 512 GB  |

## Training data lineage

| Role            | Dataset                                                                                          | License        |
|-----------------|--------------------------------------------------------------------------------------------------|----------------|
| Primary corpus  | [`Ailiance-fr/mascarade-platformio-dataset`](https://huggingface.co/datasets/Ailiance-fr/mascarade-platformio-dataset)                          | cc-by-sa-4.0 |

For per-sample provenance and attribution status, consult the dataset card.

## Benchmark roadmap

This LoRA has **not yet been evaluated** through `electron-bench` (the current
pipeline supports `gemma-4-E4B` base only). Training was completed with the
standard `mlx-lm` LoRA trainer (rank 16, alpha 32, scale 2.0, AdamW
LR 1e-5, 500 iters) — full hyperparameters are in the `Training` table above.

Planned evaluations:

- Perplexity on the validation split of the training data
- Functional benchmark on **devstral**-specific tasks
- Comparison vs base `mistralai/Devstral-Small-2-24B-Instruct-2512`

Track progress: [ailiance-bench issues](https://github.com/ailiance/ailiance-bench/issues).

For reference benchmarks on the `gemma-4-E4B` base, see the
[base-vs-LoRA matrix](https://github.com/ailiance/ailiance-bench/blob/main/bench-results/compare_base_vs_lora.md).

## License chain

| Component                         | License           |
|-----------------------------------|-------------------|
| Base model (`mistralai/Devstral-Small-2-24B-Instruct-2512`)        | apache-2.0    |
| Training data ([`Ailiance-fr/mascarade-platformio-dataset`](https://huggingface.co/datasets/Ailiance-fr/mascarade-platformio-dataset))         | cc-by-sa-4.0      |
| **LoRA adapter (this repo)**      | **cc-by-sa-4.0**|

_Most restrictive license in the chain (CC-BY-SA-4.0 share-alike) propagates to derivatives._

## EU AI Act compliance

- **Article 53(1)(c)**: training data licenses preserved (per-dataset cards declare upstream licenses).
- **Article 53(1)(d)**: training data summary — see upstream dataset cards on Ailiance-fr.
- **GPAI Code of Practice (July 2025)**: base `mistralai/Devstral-Small-2-24B-Instruct-2512` released under apache-2.0.
- **No web scraping by Ailiance**, **no licensed data**, **no PII**.
- Upstream Stack Exchange content (where applicable) is CC-BY-SA-4.0 and propagates to this adapter.

## License

LoRA weights: **cc-by-sa-4.0** — see License chain table above for derivation rationale.

## Citation

```bibtex
@misc{ailiance_devstral_platformio_2026,
  author    = {Ailiance},
  title     = {Ailiance — Devstral-Small-2-24B-BF16 platformio LoRA},
  year      = {2026},
  publisher = {Hugging Face},
  url       = {https://huggingface.co/Ailiance-fr/devstral-platformio-lora}
}
```

## Related

See the full [Ailiance-fr LoRA collection](https://huggingface.co/Ailiance-fr).


## Bench comparison (2026-05-11)

### Base model (Devstral-Small-2-24B-MLX-4bit) capability

| Task | Score | Notes |
|---|---:|---|
| GSM8K-CoT flex EM | **0.96** | W3 lm-eval-harness (--limit 100) |
| ARC-Easy acc / acc_norm | **0.80 / 0.75** | |
| MMLU-Pro Computer Science | **0.64** | |

Source: <https://github.com/ailiance/ailiance/tree/main/output/lm-eval-base-2026-05-11>

### This LoRA (tuned) — bench PENDING

Will include kicad-sch / iact-bench validators + W3 lm-eval delta. See spec for
methodology:
<https://github.com/ailiance/ailiance-bench/blob/main/docs/superpowers/specs/2026-05-11-kicad-sch-gap-design.md>