Flux Kontext GGUF β€” 4GB VRAM, 2-Pass + LoRA

By: The_frizzy1 Hardware target: 4 GB VRAM (laptop) CivitAI: https://civitai.com/models/1311703/flux-kontext-gguf-workflow-lowvram-realistic-cinematic-2pass-lora YouTube: https://www.youtube.com/@the_frizzy1

πŸŽ₯ Video Explainer: https://www.youtube.com/watch?v=4C0RJ01yRok

FLUX.1 [dev] Non-Commercial License β€” Black Forest Labs, Inc.


What This Is

Realistic, cinematic image generation using Flux Dev / Schnell / Kontext on a 4 GB VRAM laptop. Two samplers + Torch TeaCache for speed. Semi-realistic results in ~30 steps.


Model Downloads

Flux models (not VAE/CLIP):

VAE: https://huggingface.co/ffxvs/vae-flux/blob/main/ae.safetensors

CLIP:


Model Guide

Model Best for
Kontext Quality + image editing
Dev Quality generation
Schnell Speed, fewer steps
PixelWave Best realism

Quantisation

Quant Notes
q8 Best quality
q5 Best quality/speed trade-off
Below q5 Less VRAM, lower quality

Changelog

Version Notes
v2.2 Fixed settings, Torch Compile
v2.0 Two-pass setup
v1.0 Initial release
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support