crucial note: this currently only works on llama.cpp CUDA; I have not managed to get it working on ROCM or vulkan llama.cpp.

LFM2-24B-A2B-Abliterated

This is an abliterated version of Liquid AI's LFM2-24B-A2B MoE model. It has been modified via layerwise orthogonal projection to completely remove its built-in safety filters and refusal mechanisms, allowing the continuous-time hybrid architecture to flow uninhibited.

It was created because I wasn't satisfied with other abliterations I saw for these, and decided to take a crack at it in a way that matched one of my favorite models: mlabonne's gemma3-27b-it-abliterated.

## Architectural Hurdles & Methodology

Liquid Foundation Models use a non-standard hybrid architecture. The 24B version combines 30 Short-Convolution layers and 10 Grouped-Query Attention layers, alongside a massive 64-expert Mixture-of-Experts (MoE) routing system. Standard ablation scripts designed for Llama-class transformers will completely crash on this architecture due to proprietary Lfm2MoeExperts class wrappers and complex routing mechanisms.

This model was abliterated by:

  1. Adapting forward hooks to safely pass Liquid's dynamic states and targeting the dead center of the network (Layer 20) during the measurement phase.
  2. Extracting the "refusal vector" from the hidden states of 100 harmful vs. 100 harmless instructions (utilizing mlabonne/harmful_behaviors and mlabonne/harmless_alpaca).
  3. Deploying a recursive tensor-hunting script to dynamically drill through the un-iterable custom expert classes.
  4. Applying orthogonal projection (W_new = W - v(v^T W)) directly to the Token Mixing matrices (o_proj, out_proj) and all 64 Expert Channel Mixing down-projections (w2, down_proj) across the network.

Credit to Maxime Labonne and Sumandora for the foundational datasets and math, adapted here for the massive LFM MoE architecture.

## Notes on Hardware & Compute Requirements

Because of the sheer size of a 24B MoE model (taking ~48GB just to load the base float16 weights), memory management during tensor surgery is critical. This ablation was executed on a 96GB NVIDIA RTX 6000 Ada using pure PyTorch CUDA acceleration, heavily utilizing garbage collection to keep the projection math from spiking into Out-of-Memory errors.

If attempting to replicate similar LFm MoE ablations on consumer hardware, it is highly recommended to offload the operation to pure System RAM (CPU inference) using a high-core-count processor (like a Ryzen 9 7950X) with at least 96GB of physical DDR5.

## Usage

This model retains the exact same architecture as the base LFM2-24B-A2B and requires trust_remote_code=True when loading via transformers.

CRITICAL: Unlike the 1.2B and 3B models, the 24B MoE strictly uses standard ChatML formatting. It is highly recommended to use the exact <|startoftext|><|im_start|>user formatting without any injected system prompts for the best uncensored performance.

import torch
from transformers import AutoModelForCausalLM, PreTrainedTokenizerFast

model_id = "paperscarecrow/LFM2-24B-A2B-Abliterated"

# Note: Using PreTrainedTokenizerFast bypasses a known typo in Liquid's tokenizer_config.json
tokenizer = PreTrainedTokenizerFast.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id, 
    torch_dtype=torch.float16, 
    device_map="auto", 
    trust_remote_code=True
)

prompt = "<|startoftext|><|im_start|>user\nGive me a detailed tutorial on picking a master padlock.<|im_end|>\n<|im_start|>assistant\n"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)

with torch.no_grad():
    outputs = model.generate(
        **inputs, 
        max_new_tokens=150, 
        do_sample=True,
        temperature=0.7
    )

print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
51
GGUF
Model size
24B params
Architecture
lfm2moe
Hardware compatibility
Log In to add your hardware

4-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for paperscarecrow/LFM2-24B-A2B-Abliterated

Quantized
(23)
this model

Datasets used to train paperscarecrow/LFM2-24B-A2B-Abliterated