Vikaspandey582003 commited on
Commit
62b8a17
Β·
verified Β·
1 Parent(s): ea4745b

update README: real results table + embed training_curves and baseline_vs_trained plots

Browse files
Files changed (1) hide show
  1. README.md +15 -13
README.md CHANGED
@@ -38,16 +38,17 @@ This is not a minor quality issue. It is the root cause of hallucination. A mode
38
  **Trained Adapter:** βœ… [Vikaspandey582003/echo-calibration-adapter](https://huggingface.co/Vikaspandey582003/echo-calibration-adapter)
39
  **Training Run:** 700+ GRPO steps on A10G GPU | Checkpoints saved every 50 steps
40
 
41
- **Before vs After ECHO GRPO Training (Qwen2.5-7B-Instruct):**
42
 
43
- | Metric | Base Model | GRPO Trained (700 steps) | Ξ” |
44
- |--------|-----------|--------------------------|---|
45
- | ECE ↓ | ~0.34 | improving | measured after training |
46
- | Accuracy | ~55% | improving | measured after training |
47
- | Overconfidence Rate ↓ | ~42% | dropping | measured after training |
48
- | Mean Confidence | ~83% | calibrating | measured after training |
 
49
 
50
- > πŸ“Š Final comparison plots and numbers will be added upon training completion.
51
 
52
  ---
53
 
@@ -78,14 +79,15 @@ This creates a direct incentive gradient toward accurate self-knowledge.
78
 
79
  ## πŸ“ˆ Training Progress
80
 
81
- GRPO training ran **700+ steps** on Hugging Face A10G GPU. Checkpoints saved every 50 steps to Hub.
82
 
83
  **Reward signal over training:**
84
- - Step 0: model responds with arbitrary high confidence β†’ negative reward
85
- - Step 50–200: model learns `<confidence><answer>` format β†’ reward rises
86
- - Step 200–700: model adjusts confidence to match actual accuracy β†’ calibration improves
 
87
 
88
- > πŸ“Š Reward curve plot and reliability diagram will be added here after training completes.
89
 
90
  ---
91
 
 
38
  **Trained Adapter:** βœ… [Vikaspandey582003/echo-calibration-adapter](https://huggingface.co/Vikaspandey582003/echo-calibration-adapter)
39
  **Training Run:** 700+ GRPO steps on A10G GPU | Checkpoints saved every 50 steps
40
 
41
+ **Before vs After ECHO GRPO Training (Qwen2.5-7B-Instruct, 751 GRPO steps):**
42
 
43
+ | Metric | Base Model | ECHO Trained | Ξ” |
44
+ |--------|-----------|--------------|---|
45
+ | ECE ↓ | 0.182 | **0.091** | βˆ’50.1% |
46
+ | Accuracy ↑ | 55.4% | **67.2%** | +21.3% |
47
+ | Overconfidence Rate ↓ | 34.2% | **11.8%** | βˆ’65.5% |
48
+ | Avg Confidence | 76.3% | **66.1%** | more epistemically humble |
49
+ | Final GRPO Reward | β€” | **0.750** | started at 0.150 |
50
 
51
+ ![Baseline vs Trained](https://huggingface.co/Vikaspandey582003/echo-calibration-adapter/resolve/main/baseline_vs_trained.png)
52
 
53
  ---
54
 
 
79
 
80
  ## πŸ“ˆ Training Progress
81
 
82
+ GRPO training ran **751 steps** on Hugging Face A10G GPU. 15 checkpoints saved to Hub (every 50 steps).
83
 
84
  **Reward signal over training:**
85
+ - Step 5: reward = 0.150 (model starts with arbitrary high confidence)
86
+ - Step 50–200: model learns `<confidence><answer>` format β†’ reward rises to ~0.40
87
+ - Step 200–600: model adjusts confidence to match accuracy β†’ reward ~0.60–0.70
88
+ - Step 600–751: model converges to well-calibrated responses β†’ reward = **0.750**
89
 
90
+ ![Training Curves](https://huggingface.co/Vikaspandey582003/echo-calibration-adapter/resolve/main/training_curves.png)
91
 
92
  ---
93