Rohan03's picture
Add purpose_agent package init
12c2cae verified
raw
history blame
1.63 kB
"""
purpose_agent — A Self-Improving Agentic Framework via State-Value Evaluation
Architecture based on:
- MUSE (arxiv:2510.08002): 3-tier hierarchical memory (strategic/procedural/tool)
- LATS (arxiv:2310.04406): LLM-as-value-function V(s) = λ·LM_score + (1-λ)·SC_score
- REMEMBERER (arxiv:2306.07929): Q-value experience replay with Bellman updates
- Reflexion (arxiv:2303.11366): Verbal reinforcement via episodic self-reflection
- SPC (arxiv:2504.19162): Anti-reward-hacking via adversarial critic patterns
Core philosophy: The agent improves via a "Purpose Function" Φ(s) that evaluates
intermediate state improvements (distance to goal) rather than binary outcome success.
No real-time backprop — improvement comes from expanding external memory with
learned heuristics extracted from high-reward trajectories.
"""
__version__ = "0.1.0"
from purpose_agent.types import (
State,
Action,
Trajectory,
TrajectoryStep,
Heuristic,
PurposeScore,
MemoryRecord,
)
from purpose_agent.llm_backend import LLMBackend, MockLLMBackend
from purpose_agent.actor import Actor
from purpose_agent.purpose_function import PurposeFunction
from purpose_agent.experience_replay import ExperienceReplay
from purpose_agent.optimizer import HeuristicOptimizer
from purpose_agent.orchestrator import Orchestrator
__all__ = [
"State",
"Action",
"Trajectory",
"TrajectoryStep",
"Heuristic",
"PurposeScore",
"MemoryRecord",
"LLMBackend",
"MockLLMBackend",
"Actor",
"PurposeFunction",
"ExperienceReplay",
"HeuristicOptimizer",
"Orchestrator",
]