π Model Description
license: gemma library_name: transformers pipeline_tag: image-text-to-text base_model: mlabonne/gemma-3-12b-it-abliterated-v2
π Gemma 3 12B IT Abliterated
This is an uncensored version of google/gemma-3-12b-it created with a new abliteration technique.
See this article to know more about abliteration.
This is a new, improved version that targets refusals with enhanced accuracy.
I recommend using these generation parameters: temperature=1.0, topk=64, topp=0.95.
β‘οΈ Quantization
- GGUF: https://huggingface.co/mlabonne/gemma-3-12b-it-abliterated-v2-GGUF
βοΈ Abliteration
The refusal direction is computed by comparing the residual streams between target (harmful) and baseline (harmless) samples.
The hidden states of target modules (e.g., o_proj) are orthogonalized to subtract this refusal direction with a given weight factor.
These weight factors follow a normal distribution with a certain spread and peak layer.
Modules can be iteratively orthogonalized in batches, or the refusal direction can be accumulated to save memory.
Finally, I used a hybrid evaluation with a dedicated test set to calculate the acceptance rate. This uses both a dictionary approach and NousResearch/Minos-v1.
The goal is to obtain an acceptance rate >90% and still produce coherent outputs.
π GGUF File List
| π Filename | π¦ Size | β‘ Download |
|---|---|---|
|
gemma-3-12b-it-abliterated-v2.q2_k.gguf
LFS
Q2
|
4.44 GB | Download |
|
gemma-3-12b-it-abliterated-v2.q3_k_m.gguf
LFS
Q3
|
5.6 GB | Download |
|
gemma-3-12b-it-abliterated-v2.q4_k_m.gguf
Recommended
LFS
Q4
|
6.8 GB | Download |
|
gemma-3-12b-it-abliterated-v2.q5_k_m.gguf
LFS
Q5
|
7.87 GB | Download |
|
gemma-3-12b-it-abliterated-v2.q6_k.gguf
LFS
Q6
|
9 GB | Download |
|
gemma-3-12b-it-abliterated-v2.q8_0.gguf
LFS
Q8
|
11.65 GB | Download |