DreamTransformer

A language model that generates mid-scene with no start token and no narrative entry point.

Architecture

  • 4 Transformer blocks, d_model=128, 4 heads
  • RoPE positional encoding โ€” local order awareness, no global anchor
  • Random latent injection replaces BOS token
  • Local window attention (64 tokens) โ€” simulates suppressed prefrontal coherence
  • Trained on DreamBank (~21K dream reports)
  • Parameters: 7.2M | Val loss: 4.2456 | Val PPL: 69.8

Quick start

import torch
ckpt = torch.load('dream_transformer.pt')
model.load_state_dict(ckpt['model_state'])
text = dream(model, temperature=1.3)
Downloads last month
2,440
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support