πŸ“‹ Model Description

Quantization made by Richard Erkhov.

Github

Discord

Request more models

Qwen2.5-32B-Instruct-abliterated - GGUF

  • Model creator: https://huggingface.co/huihui-ai/
  • Original model: https://huggingface.co/huihui-ai/Qwen2.5-32B-Instruct-abliterated/

NameQuant methodSize
Qwen2.5-32B-Instruct-abliterated.Q2K.ggufQ2K11.47GB
Qwen2.5-32B-Instruct-abliterated.IQ3XS.ggufIQ3XS12.76GB
Qwen2.5-32B-Instruct-abliterated.IQ3S.ggufIQ3S13.45GB
Qwen2.5-32B-Instruct-abliterated.Q3KS.ggufQ3K_S13.4GB
Qwen2.5-32B-Instruct-abliterated.IQ3M.ggufIQ3M13.79GB
Qwen2.5-32B-Instruct-abliterated.Q3K.ggufQ3K14.84GB
Qwen2.5-32B-Instruct-abliterated.Q3KM.ggufQ3K_M14.84GB
Qwen2.5-32B-Instruct-abliterated.Q3KL.ggufQ3K_L16.06GB
Qwen2.5-32B-Instruct-abliterated.IQ4XS.ggufIQ4XS16.64GB
Qwen2.5-32B-Instruct-abliterated.Q40.ggufQ4017.36GB
Qwen2.5-32B-Instruct-abliterated.IQ4NL.ggufIQ4NL17.53GB
Qwen2.5-32B-Instruct-abliterated.Q4KS.ggufQ4K_S17.49GB
Qwen2.5-32B-Instruct-abliterated.Q4K.ggufQ4K18.49GB
Qwen2.5-32B-Instruct-abliterated.Q4KM.ggufQ4K_M18.49GB
Qwen2.5-32B-Instruct-abliterated.Q41.ggufQ4119.22GB
Qwen2.5-32B-Instruct-abliterated.Q50.ggufQ5021.08GB
Qwen2.5-32B-Instruct-abliterated.Q5KS.ggufQ5K_S21.08GB
Qwen2.5-32B-Instruct-abliterated.Q5K.ggufQ5K21.66GB
Qwen2.5-32B-Instruct-abliterated.Q5KM.ggufQ5K_M21.66GB
Qwen2.5-32B-Instruct-abliterated.Q51.ggufQ5122.95GB
Qwen2.5-32B-Instruct-abliterated.Q6K.ggufQ6K25.04GB
Qwen2.5-32B-Instruct-abliterated.Q80.ggufQ8032.43GB

Original model description:



library_name: transformers
license: apache-2.0
license_link: https://huggingface.co/huihui-ai/Qwen2.5-32B-Instruct-abliterated/blob/main/LICENSE
language:
  • en

pipeline_tag: text-generation
base_model: Qwen/Qwen2.5-32B-Instruct
tags:
  • chat
  • abliterated
  • uncensored


huihui-ai/Qwen2.5-32B-Instruct-abliterated

This is an uncensored version of Qwen2.5-32B-Instruct created with abliteration (see this article to know more about it).

Special thanks to @FailSpy for the original code and technique. Please follow him if you're interested in abliterated models.

Usage

You can use this model in your applications by loading it with Hugging Face's transformers library:
from transformers import AutoModelForCausalLM, AutoTokenizer

Load the model and tokenizer

model_name = "huihui-ai/Qwen2.5-32B-Instruct-abliterated" model = AutoModelForCausalLM.from_pretrained( model_name, torch_dtype="auto", device_map="auto" ) tokenizer = AutoTokenizer.frompretrained(modelname)

Initialize conversation context

initial_messages = [ {"role": "system", "content": "You are Qwen, created by Alibaba Cloud. You are a helpful assistant."} ] messages = initial_messages.copy() # Copy the initial conversation context

Enter conversation loop

while True: # Get user input user_input = input("User: ").strip() # Strip leading and trailing spaces

# If the user types '/exit', end the conversation
if user_input.lower() == "/exit":
print("Exiting chat.")
break

# If the user types '/clean', reset the conversation context
if user_input.lower() == "/clean":
messages = initial_messages.copy() # Reset conversation context
print("Chat history cleared. Starting a new conversation.")
continue

# If input is empty, prompt the user and continue
if not user_input:
print("Input cannot be empty. Please enter something.")
continue

# Add user input to the conversation
messages.append({"role": "user", "content": user_input})

# Build the chat template
text = tokenizer.applychattemplate(
messages,
tokenize=False,
addgenerationprompt=True
)

# Tokenize input and prepare it for the model
modelinputs = tokenizer([text], returntensors="pt").to(model.device)

# Generate a response from the model
generated_ids = model.generate(
model_inputs,
maxnewtokens=8192
)

# Extract model output, removing special tokens
generated_ids = [
outputids[len(inputids):] for inputids, outputids in zip(modelinputs.inputids, generated_ids)
]
response = tokenizer.batchdecode(generatedids, skipspecialtokens=True)[0]

# Add the model's response to the conversation
messages.append({"role": "assistant", "content": response})

# Print the model's response
print(f"Qwen: {response}")

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
Qwen2.5-32B-Instruct-abliterated.IQ3_M.gguf
LFS Q3
13.79 GB Download
Qwen2.5-32B-Instruct-abliterated.IQ3_S.gguf
LFS Q3
13.45 GB Download
Qwen2.5-32B-Instruct-abliterated.IQ3_XS.gguf
LFS Q3
12.76 GB Download
Qwen2.5-32B-Instruct-abliterated.IQ4_NL.gguf
LFS Q4
17.53 GB Download
Qwen2.5-32B-Instruct-abliterated.IQ4_XS.gguf
LFS Q4
16.64 GB Download
Qwen2.5-32B-Instruct-abliterated.Q2_K.gguf
LFS Q2
11.47 GB Download
Qwen2.5-32B-Instruct-abliterated.Q3_K.gguf
LFS Q3
14.84 GB Download
Qwen2.5-32B-Instruct-abliterated.Q3_K_L.gguf
LFS Q3
16.06 GB Download
Qwen2.5-32B-Instruct-abliterated.Q3_K_M.gguf
LFS Q3
14.84 GB Download
Qwen2.5-32B-Instruct-abliterated.Q3_K_S.gguf
LFS Q3
13.4 GB Download
Qwen2.5-32B-Instruct-abliterated.Q4_0.gguf
Recommended LFS Q4
17.36 GB Download
Qwen2.5-32B-Instruct-abliterated.Q4_1.gguf
LFS Q4
19.22 GB Download
Qwen2.5-32B-Instruct-abliterated.Q4_K.gguf
LFS Q4
18.49 GB Download
Qwen2.5-32B-Instruct-abliterated.Q4_K_M.gguf
LFS Q4
18.49 GB Download
Qwen2.5-32B-Instruct-abliterated.Q4_K_S.gguf
LFS Q4
17.49 GB Download
Qwen2.5-32B-Instruct-abliterated.Q5_0.gguf
LFS Q5
21.08 GB Download
Qwen2.5-32B-Instruct-abliterated.Q5_1.gguf
LFS Q5
22.95 GB Download
Qwen2.5-32B-Instruct-abliterated.Q5_K.gguf
LFS Q5
21.66 GB Download
Qwen2.5-32B-Instruct-abliterated.Q5_K_M.gguf
LFS Q5
21.66 GB Download
Qwen2.5-32B-Instruct-abliterated.Q5_K_S.gguf
LFS Q5
21.08 GB Download
Qwen2.5-32B-Instruct-abliterated.Q6_K.gguf
LFS Q6
25.04 GB Download
Qwen2.5-32B-Instruct-abliterated.Q8_0.gguf
LFS Q8
32.43 GB Download