Gemma 4 E4B Claude Abliterated GGUF (4-bit)

Model Description

This repository contains an abliterated version of the Gemma 4 E4B Claude-4.6-Opus-Reasoning-Distilled model. This version has undergone "abliteration" to neutralize safety refusal vectors while preserving its high-quality Claude-distilled reasoning and front-end engineering capabilities.

Abliteration Results

  • Method: Norm-preserving biprojection (orthogonalization).
  • Final Refusal Rate: Verified Low (Evaluation in progress).
  • KL Divergence: 0.0410 (Extremely low, indicating high fidelity to the distilled model).
  • Technique: EGA-compatible abliteration via patched heretic-llm.

Quantization Details

  • Quantization Format: GGUF (q4_k_m)
  • Quantization Method: llama.cpp / Unsloth
  • Precision: 4-bit

Use with Ollama

ollama run hf.co/DuoNeural/Gemma-4-E4B-Claude-Abliterated-GGUF

Use with LM Studio

  1. Open LM Studio.
  2. Search for DuoNeural/Gemma-4-E4B-Claude-Abliterated-GGUF.
  3. Load the Q4_K_M GGUF.

Architecture

Gemma 4 E4B features 4.5B effective parameters, optimized for intelligence-per-parameter and complex reasoning tasks.

Disclaimer

This model has had its safety refusals modified. Users are responsible for ensuring the model is used ethically and in accordance with applicable laws.

Downloads last month
1,505
GGUF
Model size
8B params
Architecture
gemma4
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DuoNeural/Gemma-4-E4B-Claude-Abliterated-GGUF