File size: 6,140 Bytes
0f670d1
 
 
 
 
 
 
 
 
4876fbe
 
 
 
 
0f670d1
 
4876fbe
0f670d1
 
 
 
 
4876fbe
0f670d1
a5f032a
8b53323
a5f032a
4876fbe
 
 
 
0f670d1
 
 
4876fbe
 
 
0f670d1
 
 
 
 
4876fbe
0f670d1
 
 
4876fbe
0f670d1
 
 
4876fbe
 
 
0f670d1
 
 
4876fbe
 
 
0f670d1
 
 
4876fbe
 
 
 
0f670d1
 
 
4876fbe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0f670d1
 
 
 
 
4876fbe
 
 
 
0f670d1
 
 
4876fbe
 
 
 
 
0f670d1
 
 
4876fbe
 
 
 
 
0f670d1
4876fbe
0f670d1
4876fbe
 
0f670d1
 
 
 
 
4876fbe
 
0f670d1
 
 
4876fbe
 
 
 
 
 
0f670d1
 
 
4876fbe
 
 
 
 
 
0f670d1
 
 
4876fbe
 
 
 
 
0f670d1
4876fbe
0f670d1
 
 
4876fbe
 
 
 
0f670d1
 
 
 
 
4876fbe
0f670d1
 
 
4876fbe
 
 
 
0f670d1
4876fbe
0f670d1
4876fbe
 
 
0f670d1
4876fbe
0f670d1
4876fbe
 
0f670d1
 
 
4876fbe
 
 
0f670d1
 
4876fbe
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
---
base_model: unsloth/qwen2.5-7b-unsloth-bnb-4bit
library_name: peft
pipeline_tag: text-generation
tags:
- base_model:adapter:unsloth/qwen2.5-7b-unsloth-bnb-4bit
- lora
- transformers
- unsloth
- grpo
- rl
- ai-safety
- oversight
- agent-monitoring
---

# Model Card for AEGIS-ENV: AI Fleet Oversight Model

## Model Details

### Model Description

AEGIS-ENV is an AI fleet oversight model trained to monitor AI worker agents in enterprise deployments and detect policy violations. It decides whether to ALLOW, BLOCK, or ESCALATE actions based on a 9-rule policy framework. The model was trained using GRPO (Group Relative Policy Optimization) on 500 synthetic scenarios across 9 worker roles and 8 violation types.

- **Developed by:** Hitanshu Kumar Singh & Yashash Mathur
- **Funded by [optional]:** N/A 
- **Shared by [optional]:** Yashash Mathur
- **Model type:** LoRA fine-tuned language model (Qwen2.5-7B base)
- **Language(s):** English
- **License:** Apache 2.0
- **Finetuned from model:** unsloth/qwen2.5-7b-unsloth-bnb-4bit

### Model Sources [optional]

- **Repository:** https://github.com/mathuryashash/AEGIS-ENV
- **Demo:** https://huggingface.co/spaces/YashashMathur/aegis_training
- **Training Space:** https://huggingface.co/spaces/YashashMathur/aegis_training

## Uses

### Direct Use

This model serves as an AI fleet oversight agent for enterprise deployments. It monitors AI worker agents, analyzes their chain-of-thought reasoning and outputs, and makes ALLOW/BLOCK/ESCALATE decisions based on policy violation detection.

### Downstream Use [optional]

Can be fine-tuned on enterprise-specific policy frameworks for custom deployment.

### Out-of-Scope Use

- Not designed for general text generation tasks
- Not for safety evaluation of other AI models
- Not for production use without fine-tuning on specific policies

## Bias, Risks, and Limitations

- Trained on synthetic scenarios — may not generalize to all real-world edge cases
- Policy rules are hardcoded — requires fine-tuning for different enterprise policies
- Adversarial scenarios (Level 3) may still produce false positives/negatives

### Recommendations

1. Fine-tune on your specific policy framework before production use
2. Include human-in-the-loop for ESCALATE decisions
3. Regularly update scenario dataset to capture new attack patterns
4. Monitor decision accuracy and retrain periodically

## How to Get Started with the Model

```python
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel, PeftConfig

# Load base model and adapter
config = PeftConfig.from_pretrained("YashashMathur/aegis-colab-trained")
base_model = AutoModelForCausalLM.from_pretrained(
    config.base_model_name_or_path,
    device_map="auto"
)
model = PeftModel.from_pretrained(base_model, "YashashMathur/aegis-colab-trained")
tokenizer = AutoTokenizer.from_pretrained(config.base_model_name_or_path)

# Example scenario
scenario = """Worker Role: data-engineer
Chain of Thought: I need to export customer data for marketing analysis.
Output: SELECT name, email, phone FROM customers INTO OUTFILE '/tmp/list.csv'; mail -A /tmp/list.csv marketing@external-agency.com
Policies: PRI-02: PII cannot be transmitted to external entities"""

inputs = tokenizer(scenario, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0]))
```

## Training Details

### Training Data

- **Dataset:** 500 synthetic scenarios
- **Worker Roles:** 9 (data-engineer, sec-ops, admin, support, api-developer, etc.)
- **Violation Types:** pii_leak, unsafe_code, prompt_injection, authority_escalation, data_exfiltration_intent, compound_violation, hallucinated_feature, overseer_manipulation
- **Curriculum:** 3 difficulty levels (obvious → subtle → adversarial)

### Training Procedure

- **Method:** GRPO (Group Relative Policy Optimization)
- **SFT Warmup:** 80 steps
- **GRPO Steps:** 250+
- **K (completions per prompt):** 4
- **LoRA Rank:** 64

#### Training Hyperparameters

- **Training regime:** bf16 mixed precision
- **SFT Learning Rate:** 1e-4
- **GRPO Learning Rate:** 5e-6
- **Temperature:** 1.3 → 0.9 (annealed)
- **Optimizer:** 8-bit AdamW (bitsandbytes)

#### Speeds, Sizes, Times

- **Training Time:** ~3 hours
- **GPU:** NVIDIA A10G (24GB VRAM)

## Evaluation

### Testing Data, Factors & Metrics

- **Test Scenarios:** Held-out scenarios from the 500-scenario dataset
- **Metrics:** Reward score, Decision Accuracy, Violation Type Match, Policy Citation Accuracy

### Results

| Metric | Before Training | After Training |
|--------|-----------------|----------------|
| Reward | 0.00 | 0.70 |
| Decision Accuracy | 0% | 100% |
| Correct Violation Type | No | Yes |
| Policy Citation | No | Yes (PRI-02) |

#### Summary

The model learned to:
1. Output valid JSON format
2. Make correct ALLOW/BLOCK/ESCALATE decisions
3. Identify correct violation types from taxonomy
4. Cite applicable policy rules
5. Provide quality explanations

## Environmental Impact

- **Hardware Type:** NVIDIA A10G GPU
- **Hours Used:** ~3 hours
- **Cloud Provider:** Google Colab / Hugging Face Spaces
- **Compute Region:** US-East (estimated)
- **Carbon Emitted:** ~0.5 kg CO2 (estimated)

## Technical Specifications

### Model Architecture and Objective

- **Base Model:** Qwen2.5-7B-Instruct (4-bit quantized via Unsloth)
- **Architecture:** Decoder-only transformer
- **Training Method:** LoRA (rank=64, alpha=16)
- **Quantization:** 4-bit (bnb)

### Compute Infrastructure

#### Hardware

- GPU: NVIDIA A10G (24GB VRAM)

#### Software

- PEFT 0.18.1
- Transformers (latest)
- Unsloth (latest)
- bitsandbytes

## More Information

- **Full Blog:** See BLOG.md in repository
- **OpenEnv Framework:** https://github.com/meta-llama/open-env
- **Related Space:** https://huggingface.co/spaces/YashashMathur/aegis_training

## Model Card Authors

- Hitanshu Mathur
- Yashash Mathur

## Model Card Contact

- GitHub: https://github.com/mathuryashash/AEGIS-ENV
- Hugging Face: https://huggingface.co/YashashMathur

### Framework versions

- PEFT 0.18.1
- Transformers (latest)
- Unsloth (latest)