Gemma 4 E4B Claude Abliterated GGUF (4-bit)
Model Description
This repository contains an abliterated version of the Gemma 4 E4B Claude-4.6-Opus-Reasoning-Distilled model. This version has undergone "abliteration" to neutralize safety refusal vectors while preserving its high-quality Claude-distilled reasoning and front-end engineering capabilities.
Abliteration Results
- Method: Norm-preserving biprojection (orthogonalization).
- Final Refusal Rate: Verified Low (Evaluation in progress).
- KL Divergence: 0.0410 (Extremely low, indicating high fidelity to the distilled model).
- Technique: EGA-compatible abliteration via patched heretic-llm.
Quantization Details
- Quantization Format: GGUF (
q4_k_m) - Quantization Method: llama.cpp / Unsloth
- Precision: 4-bit
Use with Ollama
ollama run hf.co/DuoNeural/Gemma-4-E4B-Claude-Abliterated-GGUF
Use with LM Studio
- Open LM Studio.
- Search for
DuoNeural/Gemma-4-E4B-Claude-Abliterated-GGUF. - Load the
Q4_K_MGGUF.
Architecture
Gemma 4 E4B features 4.5B effective parameters, optimized for intelligence-per-parameter and complex reasoning tasks.
Disclaimer
This model has had its safety refusals modified. Users are responsible for ensuring the model is used ethically and in accordance with applicable laws.
- Downloads last month
- 1,505
Hardware compatibility
Log In to add your hardware
4-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for DuoNeural/Gemma-4-E4B-Claude-Abliterated-GGUF
Base model
google/gemma-4-E4B-it