STAR: Similarity-guided Teacher-Assisted Refinement for Super-Tiny Function Calling Models
Paper β’ 2602.03022 β’ Published
YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
The first open-source small language model fine-tuned to natively speak the Model Context Protocol (MCP).
MCP-Agent-1.7B is Qwen/Qwen3-1.7B fine-tuned with LoRA to:
| Detail | Value |
|---|---|
| Base model | Qwen/Qwen3-1.7B (2B params) |
| Method | LoRA SFT (rank=16, all-linear) |
| Dataset | muhammadtlha944/mcp-agent-training-data (16.5K examples) |
| Epochs | 3 |
| Learning rate | 2e-4 (cosine schedule) |
| Effective batch size | 16 |
| Precision | fp16 |
| Cost | $0 (trained on free Kaggle/Colab GPU) |
MCP_Agent_1_7B_Kaggle.ipynbMCP_Agent_1_7B_Training.ipynbfrom transformers import pipeline
pipe = pipeline("text-generation", model="muhammadtlha944/MCP-Agent-1.7B")
messages = [
{"role": "system", "content": "You are an MCP agent with tools: github_search, read_file, shell_exec."},
{"role": "user", "content": "Find all Python files that import pandas"}
]
response = pipe(messages, max_new_tokens=512)
print(response[0]['generated_text'][-1]['content'])
This project is informed by:
Apache 2.0 (same as base model)
Built by Muhammad Talha β learning ML by building real projects.