Copy the full content of the script to Colab. Run in order: 1. First cell: `!pip install torch transformers datasets accelerate autoawq huggingface_hub -q` 2. Then run all remaining cells in order. The script benchmarks GSM8K (20 samples) and code generation (10 problems), then drops into an interactive chat loop with ARIA monitoring. Model: Qwen/Qwen3-8B-AWQ (same Qwen3-8B architecture as prism-ml/Ternary-Bonsai-8B) GPU: T4 (16GB VRAM) — fits comfortably with AWQ 4-bit quantization (~5.7GB)