asdf98's picture
Upload liqmamba/__init__.py
a1e6c8b verified
"""
LiqMamba: Liquid-Mamba Image Generator
A novel lightweight architecture combining:
- Liquid Time-Constant (CfC) networks for adaptive continuous-time gating
- Mamba-2 State Space Duality (SSD) for linear-time sequence processing
- Flow Matching for stable image generation
- Multi-directional 2D scans for image understanding
- ConFIG gradient stabilization (from PINN research)
Key innovations:
1. CfC-Gated Mamba blocks: Replace static nonlinearities with learnable
continuous-time dynamics that adapt computation depth per-token
2. Liquid State Modulation: The SSM state transition is modulated by CfC
dynamics, giving the model ODE-inspired expressivity
3. Physics-informed training: ConFIG gradient composition prevents
competing loss terms from destabilizing training
4. Extremely lightweight: ~25M params, trainable on Colab free tier
Paper References:
- CfC: "Closed-form Continuous-time Neural Networks" (Hasani et al., 2021)
- Mamba-2: "Transformers are SSMs" (Dao & Gu, 2024)
- DiM: "Diffusion Mamba" (Teng et al., 2024)
- ConFIG: "Towards Conflict-free Training of PINNs" (Liu et al., 2024)
"""
__version__ = "0.1.0"