moving-plantain / README.md
phanerozoic's picture
Initial placeholder README
fb28cee verified
metadata
language: en
license: apache-2.0
base_model: black-forest-labs/FLUX.2-klein-base-4B
library_name: diffusers
tags:
  - video-prediction
  - dynamics
  - lora
  - flux2
  - vision-banana
  - arxiv:2604.20329
pipeline_tag: image-to-image

moving-plantain

A LoRA adapter on FLUX.2 Klein (4B) for single-step future-frame prediction. Tests whether the latent physics priors of an image generator can be surfaced through the instruction-tuning recipe of Image Generators are Generalist Vision Learners (Gabeur et al., 2026; arXiv:2604.20329).

Thesis

Vision Banana argues that image generation pretraining produces a generalist vision learner. moving-plantain extends that argument to dynamics. A model that can render a physically coherent t+1 frame conditioned on a t=0 frame and a free-form intervention prompt — "the ball rolls left", "the cup tips over", "the cloth falls" — implicitly carries a forward physics simulator in its weights. Recovering that simulator under parameter-efficient adaptation is the empirical test of whether generative vision pretraining encodes object permanence, gravity, contact dynamics, and other physical structure beyond static appearance.

Method

Input: a single RGB frame at t=0 and an intervention prompt describing the change. Output: the predicted RGB frame at t=1. Training pairs are drawn from natural video datasets, with intervention prompts derived from the optical flow / motion description between consecutive frames. The loss is the diffusion objective on the t=1 target.

Status

Placeholder. Weights and training data forthcoming.

License

Apache 2.0 — matches base FLUX.2 Klein 4B.

References