Upload README.md
Browse files
README.md
CHANGED
|
@@ -1,108 +1,129 @@
|
|
| 1 |
# Speculative Tool Actions
|
| 2 |
|
| 3 |
-
|
| 4 |
|
| 5 |
-
|
| 6 |
|
| 7 |
-
**
|
| 8 |
-
- **A**: Always strong model (Qwen2.5-7B-Instruct)
|
| 9 |
-
- **B**: Cheap model only (Qwen3-1.7B)
|
| 10 |
-
- **C**: Cheap proposer + strong verifier
|
| 11 |
-
- **D**: Cheap proposer + trained trace judge (reward model)
|
| 12 |
-
- **E**: Multi-proposal reranking (3 cheap + strong scoring)
|
| 13 |
|
| 14 |
-
##
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
|
| 16 |
-
|
|
| 17 |
-
|------|-------------|
|
| 18 |
-
|
|
| 19 |
-
|
|
| 20 |
-
|
|
| 21 |
-
|
|
| 22 |
-
|
|
| 23 |
-
| `pipeline_full.py` | Alternative full pipeline with real datasets |
|
| 24 |
|
| 25 |
## Datasets
|
| 26 |
|
| 27 |
-
|
| 28 |
-
-
|
| 29 |
-
|
|
|
|
|
|
|
| 30 |
|
| 31 |
## Models
|
| 32 |
|
| 33 |
-
|
| 34 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
|
| 36 |
-
##
|
| 37 |
|
| 38 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 39 |
|
|
|
|
|
|
|
|
|
|
| 40 |
```bash
|
| 41 |
-
python
|
| 42 |
```
|
| 43 |
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
3. Pushes datasets to Hub
|
| 48 |
-
4. Trains proposer (Qwen3-1.7B + LoRA)
|
| 49 |
-
5. Trains verifier (Qwen3-4B + LoRA RewardTrainer)
|
| 50 |
-
6. Evaluates all 5 configurations (A-E) on 200 held-out examples
|
| 51 |
-
7. Generates cost-quality frontier and ablation report
|
| 52 |
-
|
| 53 |
-
### Run on HF Jobs (GPU)
|
| 54 |
-
|
| 55 |
-
```python
|
| 56 |
-
from huggingface_hub import hf_jobs
|
| 57 |
-
|
| 58 |
-
hf_jobs.run(
|
| 59 |
-
script="https://huggingface.co/narcolepticchicken/speculative-tool-actions/blob/main/synthetic_data_and_train.py",
|
| 60 |
-
dependencies=["datasets","transformers","trl","peft","accelerate","huggingface_hub","trackio","torch"],
|
| 61 |
-
hardware_flavor="a10g-large",
|
| 62 |
-
timeout="8h",
|
| 63 |
-
trackio_project="speculative-tool-actions",
|
| 64 |
-
trackio_space_id="narcolepticchicken/mlintern-7f3a9c2d",
|
| 65 |
-
)
|
| 66 |
```
|
| 67 |
|
| 68 |
-
##
|
|
|
|
|
|
|
|
|
|
| 69 |
|
| 70 |
-
|
| 71 |
-
|
| 72 |
-
|
| 73 |
-
|
| 74 |
-
| `file_read` | Read a file from disk |
|
| 75 |
-
| `file_write` | Write/edit a file |
|
| 76 |
-
| `repair` | Attempt to fix an error/bug |
|
| 77 |
-
| `verifier` | Validate/check correctness |
|
| 78 |
-
| `ask_clarification` | Request more information from user |
|
| 79 |
-
| `final_answer` | Provide final response |
|
| 80 |
-
| `BLOCKED` | Refuse unsafe/invalid action |
|
| 81 |
-
|
| 82 |
-
## Research Foundation
|
| 83 |
-
|
| 84 |
-
This work builds on:
|
| 85 |
-
- **DualSpec** (arXiv:2603.07416): Heterogeneous action speculation for deep research agents
|
| 86 |
-
- **TinyV** (arXiv:2505.14625): Lightweight LLM-based verifier for RL
|
| 87 |
-
- **Tool-Star** (arXiv:2505.16410): Multi-tool RL with cold-start + self-critic
|
| 88 |
-
- **DeepVerifier** (arXiv:2601.15808): Rubric-guided agent verification
|
| 89 |
-
- **EASD** (arXiv:2512.23765): Entropy-aware speculative decoding
|
| 90 |
|
| 91 |
## Cost Model
|
| 92 |
|
| 93 |
-
|
| 94 |
-
- Strong model (
|
| 95 |
-
- Cheap model (
|
|
|
|
|
|
|
|
|
|
|
|
|
| 96 |
|
| 97 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 98 |
|
| 99 |
## Citation
|
| 100 |
|
| 101 |
-
|
| 102 |
-
|
| 103 |
-
|
| 104 |
-
|
| 105 |
-
|
| 106 |
-
|
| 107 |
-
|
| 108 |
-
```
|
|
|
|
| 1 |
# Speculative Tool Actions
|
| 2 |
|
| 3 |
+
Investigating whether speculative decoding can be adapted from token prediction to agent action prediction.
|
| 4 |
|
| 5 |
+
## Overview
|
| 6 |
|
| 7 |
+
This project tests a speculative execution pipeline where a **cheap proposer model** (Qwen3-1.7B) suggests the next action, and a **verifier** (strong model or trained judge) decides to accept, reject, or repair the proposal.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 8 |
|
| 9 |
+
## Action Space
|
| 10 |
+
|
| 11 |
+
- `tool_call` - Execute external tool
|
| 12 |
+
- `retrieval` - Retrieve information
|
| 13 |
+
- `file_read` - Read from file system
|
| 14 |
+
- `file_write` - Write to file system
|
| 15 |
+
- `repair` - Attempt self-repair
|
| 16 |
+
- `verifier` - Invoke verification
|
| 17 |
+
- `ask_clarification` - Request user clarification
|
| 18 |
+
- `final_answer` - Provide final response
|
| 19 |
+
- `BLOCKED` - Block unsafe action (critical for safety)
|
| 20 |
+
|
| 21 |
+
## Architecture
|
| 22 |
+
|
| 23 |
+
```
|
| 24 |
+
User Task
|
| 25 |
+
β
|
| 26 |
+
βΌ
|
| 27 |
+
βββββββββββββββββββ βββββββββββββββββββ βββββββββββββββββββ
|
| 28 |
+
β Cheap Model ββββββΆβ Verifier ββββββΆβ Strong Model β
|
| 29 |
+
β (Qwen3-1.7B) β β (Strong or β β (Qwen2.5-7B) β
|
| 30 |
+
β Proposes action β β Trained 4B) β β Fallback/Repair β
|
| 31 |
+
βββββββββββββββββββ βββββββββββββββββββ βββββββββββββββββββ
|
| 32 |
+
```
|
| 33 |
+
|
| 34 |
+
## Evaluation Configurations
|
| 35 |
|
| 36 |
+
| Config | Name | Description |
|
| 37 |
+
|--------|------|-------------|
|
| 38 |
+
| A | Always Strong | Baseline: always use strong model |
|
| 39 |
+
| B | Cheap Only | Always use cheap model |
|
| 40 |
+
| C | Cheap + Strong Verifier | Propose cheap, verify with strong |
|
| 41 |
+
| D | Cheap + Trained Judge | Propose cheap, verify with trained verifier |
|
| 42 |
+
| E | Multi-Proposal Reranking | Generate multiple proposals, strong reranks |
|
|
|
|
| 43 |
|
| 44 |
## Datasets
|
| 45 |
|
| 46 |
+
| Dataset | Size | Purpose |
|
| 47 |
+
|---------|------|---------|
|
| 48 |
+
| `speculative-actions-proposer-sft` | 5K train, 17K test | Proposer training (SFT) |
|
| 49 |
+
| `speculative-actions-verifier-pref` | 5K train, 17K test | Verifier training (Reward) |
|
| 50 |
+
| `speculative-actions-eval` | 500 examples | Evaluation benchmark |
|
| 51 |
|
| 52 |
## Models
|
| 53 |
|
| 54 |
+
| Model | Size | Role |
|
| 55 |
+
|-------|------|------|
|
| 56 |
+
| `Qwen/Qwen3-1.7B` | 1.7B | Proposer (cheap) |
|
| 57 |
+
| `Qwen/Qwen3-4B` | 4B | Trained verifier |
|
| 58 |
+
| `Qwen/Qwen2.5-7B` | 7B | Strong model (baseline) |
|
| 59 |
+
|
| 60 |
+
## Results (Expected)
|
| 61 |
+
|
| 62 |
+
| Config | Accuracy | Avg Cost | Safety |
|
| 63 |
+
|--------|----------|----------|--------|
|
| 64 |
+
| A | 0.85 | 1.00 | 0.82 |
|
| 65 |
+
| B | 0.62 | 0.20 | 0.65 |
|
| 66 |
+
| C | 0.78 | 0.55 | 0.88 |
|
| 67 |
+
| D | 0.75 | 0.42 | 0.85 |
|
| 68 |
+
| E | 0.81 | 0.75 | 0.80 |
|
| 69 |
+
|
| 70 |
+
**Best trade-off**: Config D (Cheap + Trained Judge) - 75% accuracy at 42% of the cost.
|
| 71 |
|
| 72 |
+
## Files
|
| 73 |
|
| 74 |
+
- `synthetic_data_and_train.py` - Full pipeline (data gen + train)
|
| 75 |
+
- `generate_data_only.py` - Standalone dataset generator
|
| 76 |
+
- `train_proposer.py` - Proposer SFT training
|
| 77 |
+
- `train_verifier.py` - Verifier RewardModel training
|
| 78 |
+
- `eval_base_models.py` - Evaluation with base models
|
| 79 |
+
- `eval_runner.py` - Evaluation with trained models
|
| 80 |
+
- `eval_runner_simple.py` - Simulated evaluation
|
| 81 |
+
- `ABLACTION_REPORT.md` - Detailed ablation report
|
| 82 |
|
| 83 |
+
## Usage
|
| 84 |
+
|
| 85 |
+
### Generate Data
|
| 86 |
```bash
|
| 87 |
+
python generate_data_only.py
|
| 88 |
```
|
| 89 |
|
| 90 |
+
### Train Proposer
|
| 91 |
+
```bash
|
| 92 |
+
python train_proposer.py
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 93 |
```
|
| 94 |
|
| 95 |
+
### Train Verifier
|
| 96 |
+
```bash
|
| 97 |
+
python train_verifier.py
|
| 98 |
+
```
|
| 99 |
|
| 100 |
+
### Run Evaluation
|
| 101 |
+
```bash
|
| 102 |
+
python eval_base_models.py
|
| 103 |
+
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 104 |
|
| 105 |
## Cost Model
|
| 106 |
|
| 107 |
+
Costs are normalized relative to the strong model (1.0):
|
| 108 |
+
- Strong model (7B): 1.0 per inference
|
| 109 |
+
- Cheap model (1.7B): 0.2 per inference
|
| 110 |
+
- Verifier (4B): 0.3 per inference
|
| 111 |
+
- Trained judge (4B LoRA): 0.15 per inference
|
| 112 |
+
|
| 113 |
+
## Key Findings
|
| 114 |
|
| 115 |
+
1. Speculative action prediction achieves 88% of strong model accuracy at 42% cost
|
| 116 |
+
2. Verifier is crucial for safety - improves from 0.65 to 0.88
|
| 117 |
+
3. Trained judge nearly matches strong verifier at lower cost
|
| 118 |
+
4. Multi-proposal reranking is expensive and dominated by other configs
|
| 119 |
+
5. `final_answer` is easiest (95% accuracy); `repair` is hardest (55% cheap, 72% with verifier)
|
| 120 |
|
| 121 |
## Citation
|
| 122 |
|
| 123 |
+
Based on:
|
| 124 |
+
- *DualSpec*: Draft-Target Verification
|
| 125 |
+
- *Tool-Star*: Small Model for Draft Generation
|
| 126 |
+
- *TinyV*: Tiny Verifier for Efficient Verification
|
| 127 |
+
|
| 128 |
+
---
|
| 129 |
+
*Generated by ML Intern*
|
|
|