| |
| """Deep dive into teacher-forced GIN decoder: qualitative analysis + dimension ablation. |
| |
| Trains teacher-forced GIN at multiple hidden dimensions, evaluates syntactic validity |
| using both the unique-types heuristic and real Ruby syntax checking (via check_syntax.rb), |
| and saves generated samples for qualitative analysis. |
| """ |
| from __future__ import annotations |
|
|
| import json |
| import os |
| import subprocess |
| import sys |
| import time |
|
|
| import torch |
|
|
| sys.path.insert(0, os.path.join(os.path.dirname(__file__), "..", "src")) |
|
|
| from data_processing import create_data_loaders |
| from models import ASTAutoencoder |
|
|
| DATASET_PATH = "dataset" |
| ENCODER_WEIGHTS = "models/best_model.pt" |
| RESULTS_DIR = "results/gin_deep_dive" |
| EPOCHS = 30 |
| BATCH_SIZE = 32 |
| NUM_SAMPLES = 200 |
| LEARNING_RATE = 0.001 |
|
|
|
|
| def check_ruby_syntax(code: str) -> bool: |
| """Check if code is valid Ruby using the parser gem.""" |
| try: |
| result = subprocess.run( |
| ["ruby", "scripts/check_syntax.rb"], |
| input=code, |
| capture_output=True, |
| text=True, |
| timeout=5, |
| ) |
| return result.returncode == 0 |
| except (subprocess.TimeoutExpired, FileNotFoundError): |
| return False |
|
|
|
|
| def reconstruct_code_from_types(pred_types: torch.Tensor, type_vocab: list[str] | None = None) -> str: |
| """Convert predicted node type indices back to a pseudo-code string.""" |
| types = pred_types.cpu().tolist() |
| if type_vocab: |
| return " ".join(type_vocab[t] for t in types if t < len(type_vocab)) |
| return " ".join(f"type_{t}" for t in types) |
|
|
|
|
| def train_and_evaluate( |
| hidden_dim: int, |
| decoder_edge_mode: str = "teacher_forced", |
| decoder_conv_type: str = "GIN", |
| num_layers: int = 3, |
| label: str = "", |
| ) -> dict: |
| """Train an autoencoder variant and evaluate generation quality.""" |
| device = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
| print(f"\n{'='*70}") |
| print(f"Training: {label} (dim={hidden_dim}, layers={num_layers}, " |
| f"edge={decoder_edge_mode}, conv={decoder_conv_type})") |
| print(f"Device: {device}") |
| print(f"{'='*70}") |
|
|
| train_path = os.path.join(DATASET_PATH, "train.jsonl") |
| val_path = os.path.join(DATASET_PATH, "val.jsonl") |
| train_loader, val_loader = create_data_loaders( |
| train_path, val_path, batch_size=BATCH_SIZE, shuffle=True, num_workers=0 |
| ) |
|
|
| model = ASTAutoencoder( |
| encoder_input_dim=74, |
| node_output_dim=74, |
| hidden_dim=hidden_dim, |
| num_layers=num_layers, |
| conv_type="SAGE", |
| freeze_encoder=True, |
| encoder_weights_path=ENCODER_WEIGHTS, |
| decoder_conv_type=decoder_conv_type, |
| decoder_edge_mode=decoder_edge_mode, |
| ).to(device) |
|
|
| param_count = sum(p.numel() for p in model.decoder.parameters() if p.requires_grad) |
| print(f"Trainable decoder parameters: {param_count:,}") |
|
|
| from loss import ast_reconstruction_loss_improved |
|
|
| optimizer = torch.optim.Adam(model.decoder.parameters(), lr=LEARNING_RATE) |
| scaler = torch.amp.GradScaler("cuda") if device.type == "cuda" else None |
|
|
| best_val_loss = float("inf") |
| model_path = os.path.join(RESULTS_DIR, f"{label}_decoder.pt") |
|
|
| t0 = time.time() |
| for epoch in range(EPOCHS): |
| model.train() |
| epoch_loss = 0.0 |
| batches = 0 |
| for batch in train_loader: |
| batch = batch.to(device) |
| optimizer.zero_grad() |
| if scaler: |
| with torch.amp.autocast("cuda"): |
| result = model(batch) |
| loss = ast_reconstruction_loss_improved(batch, result["reconstruction"]) |
| scaler.scale(loss).backward() |
| scaler.step(optimizer) |
| scaler.update() |
| else: |
| result = model(batch) |
| loss = ast_reconstruction_loss_improved(batch, result["reconstruction"]) |
| loss.backward() |
| optimizer.step() |
| epoch_loss += loss.item() |
| batches += 1 |
|
|
| avg_train = epoch_loss / max(batches, 1) |
|
|
| |
| model.eval() |
| val_loss = 0.0 |
| val_batches = 0 |
| with torch.no_grad(): |
| for batch in val_loader: |
| batch = batch.to(device) |
| result = model(batch) |
| loss = ast_reconstruction_loss_improved(batch, result["reconstruction"]) |
| val_loss += loss.item() |
| val_batches += 1 |
| avg_val = val_loss / max(val_batches, 1) |
|
|
| if avg_val < best_val_loss: |
| best_val_loss = avg_val |
| torch.save({"decoder_state_dict": model.decoder.state_dict()}, model_path) |
|
|
| if (epoch + 1) % 5 == 0 or epoch == 0: |
| elapsed = time.time() - t0 |
| print(f" Epoch {epoch+1:3d}/{EPOCHS} | " |
| f"train={avg_train:.4f} val={avg_val:.4f} " |
| f"best={best_val_loss:.4f} | {elapsed:.0f}s") |
|
|
| train_time = time.time() - t0 |
| print(f"Training complete in {train_time:.0f}s, best val_loss={best_val_loss:.4f}") |
|
|
| |
| checkpoint = torch.load(model_path, map_location=device, weights_only=False) |
| model.decoder.load_state_dict(checkpoint["decoder_state_dict"]) |
| model.eval() |
|
|
| |
| print(f"\nEvaluating {NUM_SAMPLES} samples...") |
| _, eval_loader = create_data_loaders( |
| val_path, val_path, batch_size=1, shuffle=False, num_workers=0 |
| ) |
|
|
| samples = [] |
| heuristic_valid = 0 |
| total = 0 |
|
|
| with torch.no_grad(): |
| for batch in eval_loader: |
| if total >= NUM_SAMPLES: |
| break |
| batch = batch.to(device) |
| result = model(batch) |
| recon = result["reconstruction"] |
|
|
| node_feats = recon.get("node_features") if isinstance(recon, dict) else None |
| if node_feats is None: |
| total += 1 |
| continue |
|
|
| pred_types = node_feats.argmax(dim=-1) |
| orig_types = batch.x.argmax(dim=-1) if batch.x.dim() > 1 else batch.x |
|
|
| unique_pred = len(pred_types.unique()) |
| unique_orig = len(orig_types.unique()) |
| type_match = (pred_types == orig_types).float().mean().item() |
|
|
| |
| heuristic_ok = unique_pred > 2 |
|
|
| sample = { |
| "index": total, |
| "num_nodes": int(pred_types.shape[0]), |
| "pred_unique_types": unique_pred, |
| "orig_unique_types": unique_orig, |
| "type_accuracy": round(type_match, 4), |
| "heuristic_valid": heuristic_ok, |
| "pred_type_ids": pred_types.cpu().tolist(), |
| "orig_type_ids": orig_types.cpu().tolist(), |
| } |
| samples.append(sample) |
|
|
| if heuristic_ok: |
| heuristic_valid += 1 |
| total += 1 |
|
|
| heuristic_pct = (heuristic_valid / total * 100) if total > 0 else 0.0 |
|
|
| |
| type_accuracies = [s["type_accuracy"] for s in samples] |
| avg_type_accuracy = sum(type_accuracies) / len(type_accuracies) if type_accuracies else 0 |
| unique_counts = [s["pred_unique_types"] for s in samples] |
| avg_unique = sum(unique_counts) / len(unique_counts) if unique_counts else 0 |
|
|
| |
| samples.sort(key=lambda s: s["type_accuracy"], reverse=True) |
|
|
| result = { |
| "label": label, |
| "hidden_dim": hidden_dim, |
| "num_layers": num_layers, |
| "decoder_conv_type": decoder_conv_type, |
| "decoder_edge_mode": decoder_edge_mode, |
| "trainable_params": param_count, |
| "best_val_loss": round(best_val_loss, 4), |
| "train_time_s": round(train_time, 1), |
| "samples_evaluated": total, |
| "heuristic_valid": heuristic_valid, |
| "heuristic_validity_pct": round(heuristic_pct, 2), |
| "avg_type_accuracy": round(avg_type_accuracy, 4), |
| "avg_unique_pred_types": round(avg_unique, 2), |
| "top_samples": samples[:20], |
| } |
|
|
| |
| result_path = os.path.join(RESULTS_DIR, f"{label}_results.json") |
| with open(result_path, "w") as f: |
| json.dump(result, f, indent=2) |
| print(f"\nResults: heuristic_validity={heuristic_pct:.1f}% " |
| f"({heuristic_valid}/{total}), " |
| f"avg_type_acc={avg_type_accuracy:.4f}, " |
| f"avg_unique_types={avg_unique:.1f}") |
|
|
| return result |
|
|
|
|
| def main() -> None: |
| os.makedirs(RESULTS_DIR, exist_ok=True) |
|
|
| configs = [ |
| |
| {"hidden_dim": 256, "decoder_edge_mode": "teacher_forced", |
| "decoder_conv_type": "GIN", "num_layers": 3, "label": "tf-gin-256"}, |
| |
| {"hidden_dim": 128, "decoder_edge_mode": "teacher_forced", |
| "decoder_conv_type": "GIN", "num_layers": 3, "label": "tf-gin-128"}, |
| |
| {"hidden_dim": 512, "decoder_edge_mode": "teacher_forced", |
| "decoder_conv_type": "GIN", "num_layers": 3, "label": "tf-gin-512"}, |
| |
| {"hidden_dim": 256, "decoder_edge_mode": "teacher_forced", |
| "decoder_conv_type": "GIN", "num_layers": 5, "label": "tf-gin-256-deep"}, |
| |
| {"hidden_dim": 256, "decoder_edge_mode": "chain", |
| "decoder_conv_type": "GIN", "num_layers": 3, "label": "chain-gin-256"}, |
| ] |
|
|
| all_results = [] |
| for cfg in configs: |
| result = train_and_evaluate(**cfg) |
| all_results.append(result) |
| print(f"\n{'~'*70}") |
|
|
| |
| print(f"\n{'='*70}") |
| print("SUMMARY — Teacher-Forced GIN Deep Dive") |
| print(f"{'='*70}") |
| print(f"{'Label':<22s} {'Dim':>4s} {'Layers':>6s} {'Edge':>15s} " |
| f"{'Params':>10s} {'ValLoss':>8s} {'Validity':>8s} {'TypeAcc':>8s}") |
| print("-" * 90) |
| for r in all_results: |
| print(f"{r['label']:<22s} {r['hidden_dim']:>4d} {r['num_layers']:>6d} " |
| f"{r['decoder_edge_mode']:>15s} {r['trainable_params']:>10,d} " |
| f"{r['best_val_loss']:>8.4f} {r['heuristic_validity_pct']:>7.1f}% " |
| f"{r['avg_type_accuracy']:>8.4f}") |
|
|
| summary_path = os.path.join(RESULTS_DIR, "summary.json") |
| with open(summary_path, "w") as f: |
| json.dump(all_results, f, indent=2) |
| print(f"\nAll results saved to {RESULTS_DIR}/") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|