LongCat AudioDiT 3.5B — MLX 8-bit

This repository contains a self-contained MLX-native int8 conversion of LongCat AudioDiT 3.5B for local text-to-speech on Apple Silicon.

Note This repo is a community mirror of the canonical MLX conversion maintained by AppAutomaton at appautomaton/longcat-audiodit-3.5b-8bit-mlx.

Model Details

Bundle Contents

This bundle is self-contained and includes:

  • config.json
  • model.safetensors
  • tokenizer files (tokenizer.json, tokenizer_config.json, special_tokens_map.json)

How to Get Started

Command-line generation with mlx-speech:

python scripts/generate/longcat_audiodit.py \
  --text "Hello from LongCat AudioDiT." \
  --model-dir /path/to/longcat-audiodit-3.5b-8bit-mlx \
  --output-audio outputs/longcat.wav

Voice cloning:

python scripts/generate/longcat_audiodit.py \
  --text "Hello from LongCat AudioDiT." \
  --prompt-text "Original speaker text." \
  --prompt-audio /path/to/prompt.wav \
  --model-dir /path/to/longcat-audiodit-3.5b-8bit-mlx \
  --output-audio outputs/longcat_clone.wav \
  --guidance-method apg

Minimal Python usage:

from pathlib import Path

from mlx_speech.generation.longcat_audiodit import generate_longcat_audiodit

result = generate_longcat_audiodit(
    text="Hello from LongCat AudioDiT.",
    model_dir=Path("/path/to/longcat-audiodit-3.5b-8bit-mlx"),
    output_audio="outputs/longcat.wav",
)

Notes

  • This repo contains the quantized MLX runtime artifact only.
  • The conversion preserves the LongCat AudioDiT diffusion transformer and bundled VAE for waveform decode.
  • The current bundle is intended for local MLX runtime use and parity validation.
  • This mirror is a duplicated repo, not an automatically synchronized namespace mirror.

Links

License

MIT License — following the upstream license published with meituan-longcat/LongCat-AudioDiT-3.5B.

Downloads last month
102
Safetensors
Model size
1B params
Tensor type
F32
·
U32
·
MLX
Hardware compatibility
Log In to add your hardware

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for mlx-community/longcat-audiodit-3.5b-8bit-mlx

Quantized
(2)
this model