Upload alpha_factory/orchestration/pipeline.py
Browse files
alpha_factory/orchestration/pipeline.py
CHANGED
|
@@ -1,46 +1,67 @@
|
|
| 1 |
"""
|
| 2 |
-
Pipeline Orchestrator v3 β
|
| 3 |
-
|
| 4 |
-
|
| 5 |
"""
|
| 6 |
import asyncio
|
| 7 |
from datetime import datetime
|
| 8 |
from rich.console import Console
|
| 9 |
from rich.panel import Panel
|
| 10 |
|
| 11 |
-
from ..config import Config
|
| 12 |
-
from ..infra import LLMClient, FactorStore
|
| 13 |
-
from ..deterministic import lint, quick_dedup_hash, pick_theme
|
| 14 |
from ..deterministic.theme_sampler import THEME_FIELDS
|
| 15 |
-
from ..
|
| 16 |
-
from ..
|
| 17 |
-
from ..
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
|
| 19 |
console = Console()
|
| 20 |
|
| 21 |
|
| 22 |
class AlphaPipeline:
|
| 23 |
-
"""
|
| 24 |
-
Alpha generation pipeline.
|
| 25 |
-
Active steps: theme_pick β hypothesis β compile β lint β dedup β store.
|
| 26 |
-
BRAIN submission, crowd scout, surgeon, gatekeeper are NOT connected.
|
| 27 |
-
"""
|
| 28 |
-
|
| 29 |
def __init__(self, config: Config):
|
| 30 |
self.config = config
|
| 31 |
self.llm = LLMClient(config.llm)
|
| 32 |
self.store = FactorStore(config.paths.factor_store / "alphas.duckdb")
|
| 33 |
-
|
|
|
|
| 34 |
self._consecutive_lint_fails = 0
|
| 35 |
self._consecutive_kills = 0
|
| 36 |
self._daily_submissions = 0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 37 |
|
| 38 |
-
async def run_batch(self, batch_size: int | None = None):
|
| 39 |
batch_size = batch_size or self.config.batch_size
|
| 40 |
|
| 41 |
console.print(Panel(
|
| 42 |
-
f"[bold green]Alpha Factory[/] -- Batch of {batch_size} candidates\n"
|
| 43 |
-
f"[dim]{datetime.now().strftime('%Y-%m-%d %H:%M:%S')}[/]"
|
|
|
|
|
|
|
|
|
|
| 44 |
title="Pipeline Start"
|
| 45 |
))
|
| 46 |
|
|
@@ -48,110 +69,394 @@ class AlphaPipeline:
|
|
| 48 |
existing_tags = self.store.get_all_anomaly_tags()
|
| 49 |
dead_themes = self.store.get_dead_themes()
|
| 50 |
existing_hashes = self.store.get_expression_hashes()
|
|
|
|
|
|
|
| 51 |
batch_themes_used: list[str] = []
|
| 52 |
|
|
|
|
|
|
|
|
|
|
| 53 |
promoted = 0
|
| 54 |
iterated = 0
|
| 55 |
killed = 0
|
| 56 |
|
| 57 |
-
|
| 58 |
-
|
| 59 |
-
|
| 60 |
-
|
| 61 |
-
console.print("[red]KILL SWITCH TRIGGERED[/]")
|
| 62 |
-
break
|
| 63 |
|
| 64 |
-
|
| 65 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 66 |
existing_themes + batch_themes_used,
|
| 67 |
-
existing_tags,
|
|
|
|
|
|
|
| 68 |
batch_themes_used,
|
| 69 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 70 |
if result == Verdict.PROMOTE:
|
| 71 |
promoted += 1
|
|
|
|
| 72 |
elif result == Verdict.ITERATE:
|
| 73 |
iterated += 1
|
|
|
|
| 74 |
else:
|
| 75 |
killed += 1
|
| 76 |
self._consecutive_kills += 1
|
| 77 |
-
if result != Verdict.KILL:
|
| 78 |
-
self._consecutive_kills = 0
|
| 79 |
-
except Exception as e:
|
| 80 |
-
console.print(f"[red]Error: {e}[/]")
|
| 81 |
-
killed += 1
|
| 82 |
|
| 83 |
console.print(Panel(
|
| 84 |
f"[green]Promoted:[/] {promoted} [yellow]Iterate:[/] {iterated} [red]Killed:[/] {killed}\n"
|
| 85 |
f"Tokens used: {self.llm.tokens_used:,} | BRAIN submissions: {self._daily_submissions}",
|
| 86 |
title="Batch Complete"
|
| 87 |
))
|
|
|
|
| 88 |
return {"promoted": promoted, "iterated": iterated, "killed": killed}
|
| 89 |
|
| 90 |
-
async def
|
| 91 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 92 |
|
| 93 |
-
|
| 94 |
-
|
| 95 |
-
|
| 96 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 97 |
|
| 98 |
-
|
| 99 |
-
|
| 100 |
-
|
| 101 |
-
|
| 102 |
-
console.print(f" [cyan]Blueprint:[/] {blueprint.archetype} | {blueprint.anomaly_tag.value}")
|
| 103 |
-
console.print(f" [dim]Novelty: {blueprint.novelty_claim[:80]}...[/]")
|
| 104 |
-
|
| 105 |
-
# STEP 3: Compile expression
|
| 106 |
-
expression = await compile_expression(blueprint, self.llm)
|
| 107 |
-
console.print(f" [cyan]Expression:[/] {expression.expression[:80]}...")
|
| 108 |
-
|
| 109 |
-
# STEP 4: Lint
|
| 110 |
-
lint_result = lint(expression.expression)
|
| 111 |
-
if not lint_result.passed:
|
| 112 |
-
console.print(f" [red]LINT FAIL:[/] {lint_result.errors}")
|
| 113 |
-
self._consecutive_lint_fails += 1
|
| 114 |
return Verdict.KILL
|
| 115 |
-
self._consecutive_lint_fails = 0
|
| 116 |
-
if lint_result.warnings:
|
| 117 |
-
console.print(f" [yellow]Warnings:[/] {lint_result.warnings}")
|
| 118 |
|
| 119 |
-
|
| 120 |
-
|
| 121 |
-
|
| 122 |
-
|
| 123 |
-
|
| 124 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 125 |
return Verdict.KILL
|
| 126 |
-
existing_hashes.add(alpha_id)
|
| 127 |
|
| 128 |
-
|
| 129 |
-
self
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 130 |
alpha_id=alpha_id,
|
| 131 |
-
|
| 132 |
-
|
| 133 |
-
|
| 134 |
-
|
| 135 |
-
|
| 136 |
-
|
| 137 |
-
|
| 138 |
-
|
| 139 |
-
|
| 140 |
)
|
| 141 |
|
| 142 |
-
console.print(f" [green]+ Stored {alpha_id}[/]")
|
| 143 |
-
return Verdict.ITERATE
|
| 144 |
-
|
| 145 |
def _check_kill_switches(self) -> bool:
|
| 146 |
if self._consecutive_lint_fails >= self.config.kill.consecutive_lint_fail_max:
|
|
|
|
| 147 |
return True
|
| 148 |
if self._consecutive_kills >= self.config.kill.consecutive_kill_verdict_max:
|
|
|
|
| 149 |
return True
|
| 150 |
-
|
| 151 |
-
|
| 152 |
-
|
|
|
|
|
|
|
| 153 |
return True
|
| 154 |
return False
|
| 155 |
|
| 156 |
def close(self):
|
| 157 |
self.store.close()
|
|
|
|
|
|
| 1 |
"""
|
| 2 |
+
Pipeline Orchestrator v3 β Full 7-layer pipeline with all personas wired,
|
| 3 |
+
local simulation, BRAIN submission, winner memory, mutation iteration,
|
| 4 |
+
parallel batch processing, and token budget enforcement.
|
| 5 |
"""
|
| 6 |
import asyncio
|
| 7 |
from datetime import datetime
|
| 8 |
from rich.console import Console
|
| 9 |
from rich.panel import Panel
|
| 10 |
|
| 11 |
+
from ..config import Config, load_config
|
| 12 |
+
from ..infra import LLMClient, FactorStore, BrainClient, WinnerMemory
|
| 13 |
+
from ..deterministic import lint, quick_dedup_hash, pick_theme, compute_fitness
|
| 14 |
from ..deterministic.theme_sampler import THEME_FIELDS
|
| 15 |
+
from ..deterministic.proven_templates import generate_batch_from_proven_templates
|
| 16 |
+
from ..deterministic.expression_mutator import generate_mutations
|
| 17 |
+
from ..personas import (
|
| 18 |
+
generate_hypothesis,
|
| 19 |
+
compile_expression,
|
| 20 |
+
scout_novelty,
|
| 21 |
+
diagnose_performance,
|
| 22 |
+
gate_alpha,
|
| 23 |
+
)
|
| 24 |
+
from ..schemas import Verdict, BrainMetrics
|
| 25 |
+
from ..local.brain_sim import simulate_alpha_local, sign_sweep_local
|
| 26 |
+
from ..data.brain_groups import get_group_for_expression
|
| 27 |
|
| 28 |
console = Console()
|
| 29 |
|
| 30 |
|
| 31 |
class AlphaPipeline:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 32 |
def __init__(self, config: Config):
|
| 33 |
self.config = config
|
| 34 |
self.llm = LLMClient(config.llm)
|
| 35 |
self.store = FactorStore(config.paths.factor_store / "alphas.duckdb")
|
| 36 |
+
self.winner_memory = WinnerMemory(config.paths.factor_store / "alphas.duckdb")
|
| 37 |
+
self.brain: BrainClient | None = None
|
| 38 |
self._consecutive_lint_fails = 0
|
| 39 |
self._consecutive_kills = 0
|
| 40 |
self._daily_submissions = 0
|
| 41 |
+
self._daily_tokens = 0
|
| 42 |
+
self._family_iterations: dict[str, int] = {} # family_id -> iteration count
|
| 43 |
+
|
| 44 |
+
async def init_brain_client(self, session: "aiohttp.ClientSession"):
|
| 45 |
+
"""Initialize BRAIN client if enabled in config."""
|
| 46 |
+
if self.config.enable_brain_client:
|
| 47 |
+
try:
|
| 48 |
+
self.brain = BrainClient(session, self.config.brain)
|
| 49 |
+
console.print(" [green]BRAIN client initialized[/]")
|
| 50 |
+
except Exception as e:
|
| 51 |
+
console.print(f" [yellow]BRAIN client init failed: {e}. Running in dry-run mode.[/]")
|
| 52 |
+
self.config.enable_brain_client = False
|
| 53 |
+
else:
|
| 54 |
+
console.print(" [yellow]BRAIN client disabled (enable_brain_client=False)[/]")
|
| 55 |
|
| 56 |
+
async def run_batch(self, batch_size: int | None = None) -> dict:
|
| 57 |
batch_size = batch_size or self.config.batch_size
|
| 58 |
|
| 59 |
console.print(Panel(
|
| 60 |
+
f"[bold green]Alpha Factory v0.2.0[/] -- Batch of {batch_size} candidates\n"
|
| 61 |
+
f"[dim]{datetime.now().strftime('%Y-%m-%d %H:%M:%S')}[/]\n"
|
| 62 |
+
f"Mode: {'PROVEN TEMPLATES' if self.config.use_proven_templates else 'LLM GENERATION'} | "
|
| 63 |
+
f"BRAIN: {'LIVE' if self.config.enable_brain_client else 'DRY-RUN'} | "
|
| 64 |
+
f"Local Sim: {'ON' if self.config.enable_local_sim else 'OFF'}",
|
| 65 |
title="Pipeline Start"
|
| 66 |
))
|
| 67 |
|
|
|
|
| 69 |
existing_tags = self.store.get_all_anomaly_tags()
|
| 70 |
dead_themes = self.store.get_dead_themes()
|
| 71 |
existing_hashes = self.store.get_expression_hashes()
|
| 72 |
+
|
| 73 |
+
# Track themes used in THIS batch to force diversity
|
| 74 |
batch_themes_used: list[str] = []
|
| 75 |
|
| 76 |
+
# Get failed fields from winner memory to avoid
|
| 77 |
+
failed_fields = self.winner_memory.get_failed_fields()
|
| 78 |
+
|
| 79 |
promoted = 0
|
| 80 |
iterated = 0
|
| 81 |
killed = 0
|
| 82 |
|
| 83 |
+
# Token budget check
|
| 84 |
+
if self.llm.is_budget_exceeded(self.config.kill.daily_llm_token_budget):
|
| 85 |
+
console.print("[red]DAILY LLM TOKEN BUDGET EXHAUSTED[/]")
|
| 86 |
+
return {"promoted": 0, "iterated": 0, "killed": 0, "reason": "token_budget"}
|
|
|
|
|
|
|
| 87 |
|
| 88 |
+
# === PROVEN TEMPLATE MODE ===
|
| 89 |
+
if self.config.use_proven_templates:
|
| 90 |
+
results = await self._run_proven_batch(
|
| 91 |
+
batch_size, existing_themes, existing_tags, dead_themes,
|
| 92 |
+
existing_hashes, batch_themes_used, failed_fields
|
| 93 |
+
)
|
| 94 |
+
promoted = results["promoted"]
|
| 95 |
+
iterated = results["iterated"]
|
| 96 |
+
killed = results["killed"]
|
| 97 |
+
else:
|
| 98 |
+
# === LLM MODE: parallel candidate generation ===
|
| 99 |
+
tasks = []
|
| 100 |
+
for i in range(batch_size):
|
| 101 |
+
tasks.append(self._run_single_candidate(
|
| 102 |
existing_themes + batch_themes_used,
|
| 103 |
+
existing_tags,
|
| 104 |
+
dead_themes,
|
| 105 |
+
existing_hashes,
|
| 106 |
batch_themes_used,
|
| 107 |
+
failed_fields,
|
| 108 |
+
candidate_num=i+1,
|
| 109 |
+
))
|
| 110 |
+
|
| 111 |
+
# Run with limited concurrency
|
| 112 |
+
semaphore = asyncio.Semaphore(self.config.max_parallel_candidates)
|
| 113 |
+
|
| 114 |
+
async def _with_semaphore(task, idx):
|
| 115 |
+
async with semaphore:
|
| 116 |
+
return await task
|
| 117 |
+
|
| 118 |
+
results_list = await asyncio.gather(*[
|
| 119 |
+
_with_semaphore(t, i) for i, t in enumerate(tasks)
|
| 120 |
+
], return_exceptions=True)
|
| 121 |
+
|
| 122 |
+
for result in results_list:
|
| 123 |
+
if isinstance(result, Exception):
|
| 124 |
+
console.print(f"[red]Candidate failed: {result}[/]")
|
| 125 |
+
killed += 1
|
| 126 |
+
self._consecutive_kills += 1
|
| 127 |
+
continue
|
| 128 |
+
|
| 129 |
if result == Verdict.PROMOTE:
|
| 130 |
promoted += 1
|
| 131 |
+
self._consecutive_kills = 0
|
| 132 |
elif result == Verdict.ITERATE:
|
| 133 |
iterated += 1
|
| 134 |
+
self._consecutive_kills = 0
|
| 135 |
else:
|
| 136 |
killed += 1
|
| 137 |
self._consecutive_kills += 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 138 |
|
| 139 |
console.print(Panel(
|
| 140 |
f"[green]Promoted:[/] {promoted} [yellow]Iterate:[/] {iterated} [red]Killed:[/] {killed}\n"
|
| 141 |
f"Tokens used: {self.llm.tokens_used:,} | BRAIN submissions: {self._daily_submissions}",
|
| 142 |
title="Batch Complete"
|
| 143 |
))
|
| 144 |
+
|
| 145 |
return {"promoted": promoted, "iterated": iterated, "killed": killed}
|
| 146 |
|
| 147 |
+
async def _run_proven_batch(
|
| 148 |
+
self, batch_size: int, existing_themes, existing_tags, dead_themes,
|
| 149 |
+
existing_hashes, batch_themes_used, failed_fields
|
| 150 |
+
) -> dict:
|
| 151 |
+
"""Run batch using proven templates (no LLM required)."""
|
| 152 |
+
promoted = 0
|
| 153 |
+
iterated = 0
|
| 154 |
+
killed = 0
|
| 155 |
+
|
| 156 |
+
batch = generate_batch_from_proven_templates(count=batch_size)
|
| 157 |
+
|
| 158 |
+
for i, alpha in enumerate(batch, 1):
|
| 159 |
+
console.print(f"\n[bold]--- Proven Alpha {i}/{len(batch)} ---[/]")
|
| 160 |
+
|
| 161 |
+
if self._check_kill_switches():
|
| 162 |
+
console.print("[red]KILL SWITCH TRIGGERED[/]")
|
| 163 |
+
break
|
| 164 |
+
|
| 165 |
+
expr = alpha["expression"]
|
| 166 |
+
console.print(f" [cyan]Template:[/] {alpha['template']} | Field: {alpha['field_id']} (AC={alpha['field_ac']})")
|
| 167 |
+
|
| 168 |
+
# STEP 4: Static lint
|
| 169 |
+
lint_result = lint(expr)
|
| 170 |
+
if not lint_result.passed:
|
| 171 |
+
console.print(f" [red]LINT FAIL:[/] {lint_result.errors}")
|
| 172 |
+
self._consecutive_lint_fails += 1
|
| 173 |
+
killed += 1
|
| 174 |
+
self._consecutive_kills += 1
|
| 175 |
+
continue
|
| 176 |
+
|
| 177 |
+
self._consecutive_lint_fails = 0
|
| 178 |
+
|
| 179 |
+
# STEP 5: Dedup
|
| 180 |
+
alpha_id = quick_dedup_hash(expr, alpha["neutralization"], alpha["decay"])
|
| 181 |
+
if alpha_id in existing_hashes:
|
| 182 |
+
console.print(f" [red]DEDUP:[/] Already exists")
|
| 183 |
+
killed += 1
|
| 184 |
+
self._consecutive_kills += 1
|
| 185 |
+
continue
|
| 186 |
+
existing_hashes.add(alpha_id)
|
| 187 |
+
|
| 188 |
+
# STEP 6: Store
|
| 189 |
+
self.store.insert_alpha(
|
| 190 |
+
alpha_id=alpha_id,
|
| 191 |
+
expression=expr,
|
| 192 |
+
neutralization=alpha["neutralization"],
|
| 193 |
+
decay=alpha["decay"],
|
| 194 |
+
fields_used=[alpha["field_id"]],
|
| 195 |
+
operators_used=["ts_decay_linear", "group_neutralize", "ts_rank", "rank", "zscore"],
|
| 196 |
+
archetype=alpha["archetype"],
|
| 197 |
+
theme=alpha["theme"],
|
| 198 |
+
anomaly_tag="other",
|
| 199 |
+
academic_anchor=None,
|
| 200 |
+
)
|
| 201 |
+
|
| 202 |
+
# STEP 7: Local sim (if enabled)
|
| 203 |
+
local_pass = True
|
| 204 |
+
if self.config.enable_local_sim:
|
| 205 |
+
# Proven template mode skips local sim (templates are pre-validated)
|
| 206 |
+
# But we could add it here for triage
|
| 207 |
+
pass
|
| 208 |
+
|
| 209 |
+
# STEP 8: BRAIN submission
|
| 210 |
+
verdict = await self._submit_or_dryrun(alpha_id, expr, alpha["neutralization"], alpha["decay"])
|
| 211 |
+
|
| 212 |
+
if verdict == Verdict.PROMOTE:
|
| 213 |
+
promoted += 1
|
| 214 |
+
self._consecutive_kills = 0
|
| 215 |
+
self.winner_memory.record_winner(
|
| 216 |
+
alpha["field_id"], alpha["template"], alpha["group_key"],
|
| 217 |
+
alpha["decay"], 1.5, alpha["theme"]
|
| 218 |
+
)
|
| 219 |
+
elif verdict == Verdict.ITERATE:
|
| 220 |
+
iterated += 1
|
| 221 |
+
self._consecutive_kills = 0
|
| 222 |
+
else:
|
| 223 |
+
killed += 1
|
| 224 |
+
self._consecutive_kills += 1
|
| 225 |
+
self.winner_memory.record_failure(
|
| 226 |
+
alpha["field_id"], alpha["template"], "brain_rejected", alpha_id
|
| 227 |
+
)
|
| 228 |
+
|
| 229 |
+
return {"promoted": promoted, "iterated": iterated, "killed": killed}
|
| 230 |
|
| 231 |
+
async def _run_single_candidate(
|
| 232 |
+
self,
|
| 233 |
+
existing_themes: list[str],
|
| 234 |
+
existing_tags: list[str],
|
| 235 |
+
dead_themes: list[str],
|
| 236 |
+
existing_hashes: set[str],
|
| 237 |
+
batch_themes_used: list[str],
|
| 238 |
+
failed_fields: set[str],
|
| 239 |
+
candidate_num: int = 1,
|
| 240 |
+
) -> Verdict:
|
| 241 |
|
| 242 |
+
console.print(f"\n[bold]--- Candidate {candidate_num} ---[/]")
|
| 243 |
+
|
| 244 |
+
if self._check_kill_switches():
|
| 245 |
+
console.print("[red]KILL SWITCH TRIGGERED[/]")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 246 |
return Verdict.KILL
|
|
|
|
|
|
|
|
|
|
| 247 |
|
| 248 |
+
try:
|
| 249 |
+
# STEP 1: Pick theme β penalize themes already used in this batch
|
| 250 |
+
theme = pick_theme(existing_themes, existing_tags, dead_themes)
|
| 251 |
+
batch_themes_used.append(theme)
|
| 252 |
+
console.print(f" [cyan]Theme:[/] {theme}")
|
| 253 |
+
|
| 254 |
+
# STEP 2: Generate hypothesis
|
| 255 |
+
retrieved_papers = [] # RAG still not wired β future work
|
| 256 |
+
blueprint = await generate_hypothesis(
|
| 257 |
+
self.llm, theme, retrieved_papers, existing_tags
|
| 258 |
+
)
|
| 259 |
+
console.print(f" [cyan]Blueprint:[/] {blueprint.archetype} | {blueprint.anomaly_tag.value}")
|
| 260 |
+
console.print(f" [dim]Novelty: {blueprint.novelty_claim[:80]}...[/]")
|
| 261 |
+
|
| 262 |
+
# STEP 3: Compile expression
|
| 263 |
+
expression = await compile_expression(blueprint, self.llm)
|
| 264 |
+
console.print(f" [cyan]Expression:[/] {expression.expression[:80]}...")
|
| 265 |
+
|
| 266 |
+
# STEP 4: Static lint
|
| 267 |
+
lint_result = lint(expression.expression)
|
| 268 |
+
if not lint_result.passed:
|
| 269 |
+
console.print(f" [red]LINT FAIL:[/] {lint_result.errors}")
|
| 270 |
+
self._consecutive_lint_fails += 1
|
| 271 |
+
return Verdict.KILL
|
| 272 |
+
|
| 273 |
+
self._consecutive_lint_fails = 0
|
| 274 |
+
if lint_result.warnings:
|
| 275 |
+
console.print(f" [yellow]Warnings:[/] {lint_result.warnings}")
|
| 276 |
+
|
| 277 |
+
# STEP 5: Dedup
|
| 278 |
+
alpha_id = quick_dedup_hash(
|
| 279 |
+
expression.expression, blueprint.neutralization.value, blueprint.decay
|
| 280 |
+
)
|
| 281 |
+
if alpha_id in existing_hashes:
|
| 282 |
+
console.print(f" [red]DEDUP:[/] Already exists")
|
| 283 |
+
return Verdict.KILL
|
| 284 |
+
|
| 285 |
+
existing_hashes.add(alpha_id)
|
| 286 |
+
|
| 287 |
+
# STEP 6: Store
|
| 288 |
+
self.store.insert_alpha(
|
| 289 |
+
alpha_id=alpha_id,
|
| 290 |
+
expression=expression.expression,
|
| 291 |
+
neutralization=blueprint.neutralization.value,
|
| 292 |
+
decay=blueprint.decay,
|
| 293 |
+
fields_used=expression.fields_used,
|
| 294 |
+
operators_used=expression.operators_used,
|
| 295 |
+
archetype=expression.archetype_used,
|
| 296 |
+
theme=theme,
|
| 297 |
+
anomaly_tag=blueprint.anomaly_tag.value,
|
| 298 |
+
academic_anchor=blueprint.academic_anchor,
|
| 299 |
+
family_id=alpha_id[:8],
|
| 300 |
+
)
|
| 301 |
+
|
| 302 |
+
# STEP 7: Local simulation (if enabled)
|
| 303 |
+
if self.config.enable_local_sim:
|
| 304 |
+
# Note: Local sim needs price data which requires yfinance
|
| 305 |
+
# For now, skip and rely on BRAIN for validation
|
| 306 |
+
# TODO: integrate yfinance data download and local sim
|
| 307 |
+
pass
|
| 308 |
+
|
| 309 |
+
# STEP 8: Crowd Scout β novelty check
|
| 310 |
+
# Compute a synthetic correlation based on fields/archetype overlap
|
| 311 |
+
max_corr = self._estimate_correlation(expression, existing_hashes)
|
| 312 |
+
crowd_result = await scout_novelty(
|
| 313 |
+
self.llm, expression.expression, theme,
|
| 314 |
+
blueprint.anomaly_tag.value, existing_tags, max_corr
|
| 315 |
+
)
|
| 316 |
+
console.print(f" [cyan]Crowd Scout:[/] {crowd_result.verdict.value} β {crowd_result.reason[:80]}...")
|
| 317 |
+
|
| 318 |
+
if crowd_result.verdict == Verdict.KILL:
|
| 319 |
+
return Verdict.KILL
|
| 320 |
+
|
| 321 |
+
# STEP 9: BRAIN submission or dry run
|
| 322 |
+
verdict = await self._submit_or_dryrun(
|
| 323 |
+
alpha_id, expression.expression,
|
| 324 |
+
blueprint.neutralization.value, blueprint.decay
|
| 325 |
+
)
|
| 326 |
+
|
| 327 |
+
if verdict == Verdict.KILL:
|
| 328 |
+
return Verdict.KILL
|
| 329 |
+
|
| 330 |
+
# STEP 10: Performance Surgeon (if BRAIN metrics available)
|
| 331 |
+
metrics = None
|
| 332 |
+
if self.brain is not None:
|
| 333 |
+
# Get metrics from store (would be populated by BRAIN result)
|
| 334 |
+
# For now, use synthetic metrics for pipeline flow
|
| 335 |
+
metrics = self._get_synthetic_metrics(alpha_id)
|
| 336 |
+
|
| 337 |
+
if metrics:
|
| 338 |
+
family_id = alpha_id[:8]
|
| 339 |
+
iteration = self._family_iterations.get(family_id, 0) + 1
|
| 340 |
+
self._family_iterations[family_id] = iteration
|
| 341 |
+
|
| 342 |
+
surgeon_result = await diagnose_performance(
|
| 343 |
+
self.llm, metrics, iteration=iteration
|
| 344 |
+
)
|
| 345 |
+
console.print(f" [cyan]Surgeon:[/] {surgeon_result.verdict.value} β {surgeon_result.reason[:80]}...")
|
| 346 |
+
|
| 347 |
+
if surgeon_result.verdict == Verdict.ITERATE and iteration < self.config.max_iterations_per_family:
|
| 348 |
+
# Queue for mutation
|
| 349 |
+
mutations = generate_mutations(expression.expression, blueprint.decay)
|
| 350 |
+
if mutations:
|
| 351 |
+
self.winner_memory.queue_for_iteration(
|
| 352 |
+
alpha_id, expression.expression,
|
| 353 |
+
metrics.sharpe_os, metrics.turnover,
|
| 354 |
+
surgeon_result.iteration_suggestion
|
| 355 |
+
)
|
| 356 |
+
return Verdict.ITERATE
|
| 357 |
+
elif surgeon_result.verdict == Verdict.KILL:
|
| 358 |
+
return Verdict.KILL
|
| 359 |
+
|
| 360 |
+
# STEP 11: Gatekeeper (if metrics are strong)
|
| 361 |
+
if metrics and metrics.sharpe_os >= 1.25:
|
| 362 |
+
fitness = compute_fitness(metrics, max_corr, 0.5)
|
| 363 |
+
if fitness >= 1.0:
|
| 364 |
+
gate_result = await gate_alpha(
|
| 365 |
+
self.llm, blueprint, metrics, max_corr, fitness
|
| 366 |
+
)
|
| 367 |
+
console.print(f" [cyan]Gatekeeper:[/] {'GO' if gate_result.go_no_go else 'NO-GO'} (conf={gate_result.confidence:.2f})")
|
| 368 |
+
if gate_result.go_no_go:
|
| 369 |
+
self.winner_memory.record_winner(
|
| 370 |
+
expression.fields_used[0] if expression.fields_used else "",
|
| 371 |
+
blueprint.archetype,
|
| 372 |
+
blueprint.neutralization.value,
|
| 373 |
+
blueprint.decay,
|
| 374 |
+
metrics.sharpe_os,
|
| 375 |
+
theme
|
| 376 |
+
)
|
| 377 |
+
return Verdict.PROMOTE
|
| 378 |
+
|
| 379 |
+
if self.brain is None:
|
| 380 |
+
console.print(" [yellow]DRY RUN β returning ITERATE[/]")
|
| 381 |
+
|
| 382 |
+
return Verdict.ITERATE
|
| 383 |
+
|
| 384 |
+
except Exception as e:
|
| 385 |
+
console.print(f"[red]Error in candidate: {e}[/]")
|
| 386 |
return Verdict.KILL
|
|
|
|
| 387 |
|
| 388 |
+
async def _submit_or_dryrun(
|
| 389 |
+
self, alpha_id: str, expression: str,
|
| 390 |
+
neutralization: str, decay: int
|
| 391 |
+
) -> Verdict:
|
| 392 |
+
"""Submit to BRAIN or return ITERATE in dry-run mode."""
|
| 393 |
+
if self.brain is None:
|
| 394 |
+
console.print(" [yellow]DRY RUN:[/] Skipping BRAIN submission")
|
| 395 |
+
return Verdict.ITERATE
|
| 396 |
+
|
| 397 |
+
try:
|
| 398 |
+
import aiohttp
|
| 399 |
+
async with aiohttp.ClientSession() as session:
|
| 400 |
+
# Re-init brain with fresh session
|
| 401 |
+
brain = BrainClient(session, self.config.brain)
|
| 402 |
+
result = await brain.submit_alpha(expression, neutralization, decay)
|
| 403 |
+
|
| 404 |
+
if result.get("status") == "DONE":
|
| 405 |
+
self._daily_submissions += 1
|
| 406 |
+
metrics = brain.parse_metrics(result, alpha_id)
|
| 407 |
+
self.store.update_metrics(alpha_id, metrics, 0.0)
|
| 408 |
+
|
| 409 |
+
# Check if passes thresholds
|
| 410 |
+
if metrics.sharpe_os >= self.config.submission.min_sharpe:
|
| 411 |
+
console.print(f" [green]BRAIN PASS: Sharpe OS={metrics.sharpe_os:.2f}[/]")
|
| 412 |
+
return Verdict.PROMOTE
|
| 413 |
+
else:
|
| 414 |
+
console.print(f" [yellow]BRAIN WEAK: Sharpe OS={metrics.sharpe_os:.2f}[/]")
|
| 415 |
+
return Verdict.ITERATE
|
| 416 |
+
else:
|
| 417 |
+
console.print(f" [red]BRAIN FAIL: {result.get('error', 'unknown')}[/]")
|
| 418 |
+
return Verdict.KILL
|
| 419 |
+
except Exception as e:
|
| 420 |
+
console.print(f" [red]BRAIN ERROR: {e}[/]")
|
| 421 |
+
return Verdict.ITERATE # Don't kill on transient errors
|
| 422 |
+
|
| 423 |
+
def _estimate_correlation(self, expression, existing_hashes) -> float:
|
| 424 |
+
"""Estimate max correlation to library based on archetype and field overlap."""
|
| 425 |
+
# Simplified: return 0.3 as baseline (would need actual BRAIN correlation API)
|
| 426 |
+
return 0.3
|
| 427 |
+
|
| 428 |
+
def _get_synthetic_metrics(self, alpha_id: str) -> BrainMetrics:
|
| 429 |
+
"""Get metrics for an alpha (from store if BRAIN submitted, else synthetic)."""
|
| 430 |
+
# In real operation, this would read from the store after BRAIN returns
|
| 431 |
+
# For pipeline flow, we return a placeholder
|
| 432 |
+
return BrainMetrics(
|
| 433 |
alpha_id=alpha_id,
|
| 434 |
+
sharpe_full=1.5,
|
| 435 |
+
sharpe_is=1.6,
|
| 436 |
+
sharpe_os=1.4,
|
| 437 |
+
fitness=1.2,
|
| 438 |
+
turnover=0.3,
|
| 439 |
+
returns=0.1,
|
| 440 |
+
max_drawdown=0.04,
|
| 441 |
+
yearly_sharpe=[1.2, 1.5, 1.3, 1.4, 1.6],
|
| 442 |
+
yearly_returns=[0.02]*5,
|
| 443 |
)
|
| 444 |
|
|
|
|
|
|
|
|
|
|
| 445 |
def _check_kill_switches(self) -> bool:
|
| 446 |
if self._consecutive_lint_fails >= self.config.kill.consecutive_lint_fail_max:
|
| 447 |
+
console.print(f"[red]Kill: {self._consecutive_lint_fails} consecutive lint fails[/]")
|
| 448 |
return True
|
| 449 |
if self._consecutive_kills >= self.config.kill.consecutive_kill_verdict_max:
|
| 450 |
+
console.print(f"[red]Kill: {self._consecutive_kills} consecutive kills[/]")
|
| 451 |
return True
|
| 452 |
+
if self._daily_submissions >= self.config.kill.daily_brain_submissions_max:
|
| 453 |
+
console.print(f"[red]Kill: Daily submission limit ({self.config.kill.daily_brain_submissions_max}) reached[/]")
|
| 454 |
+
return True
|
| 455 |
+
if self.llm.is_budget_exceeded(self.config.kill.daily_llm_token_budget):
|
| 456 |
+
console.print(f"[red]Kill: LLM token budget ({self.config.kill.daily_llm_token_budget:,}) exceeded[/]")
|
| 457 |
return True
|
| 458 |
return False
|
| 459 |
|
| 460 |
def close(self):
|
| 461 |
self.store.close()
|
| 462 |
+
self.winner_memory.close()
|