📋 Model Description
license: gemma language:
- en
- google/gemma-3-1b-it
- uncensored
- text-generation
- reasoning
- instruction-tuned
- lightweight
Gemma 3 – 1B IT GLM-4.7 Flash Heretic Uncensored Thinking
This repository hosts Gemma 3 – 1B IT GLM-4.7 Flash Heretic Uncensored Thinking, a lightweight 1 billion–parameter instruction-tuned model derived from Google’s Gemma 3 1B IT base.
This variant is optimized for fast inference, structured reasoning behavior, and minimal refusal patterns, while maintaining compatibility with Gemma’s native instruction format.
Model Overview
- Model Name: Gemma 3 – 1B IT GLM-4.7 Flash Heretic Uncensored Thinking
- Parameter Count: 1 Billion (1B)
- Base Architecture: Gemma 3
- Base Model: google/gemma-3-1b-it
- Model Type: Instruction-Tuned Causal Language Model
- Context Length: Inherits base model context window
- Primary Language: English
- License: Gemma License (inherits from base model)
- Maintainer / Publisher: DavidAU
What Is This Model?
This model is a modified derivative of Gemma 3 – 1B IT, configured for:
- Reduced refusal bias compared to default IT alignment
- Enhanced direct-answer behavior
- Stronger short-form reasoning output
- Faster response latency due to compact parameter size
- “Flash”-style concise and rapid generation
The “Heretic Uncensored Thinking” configuration emphasizes:
- Minimal conversational filtering
- Direct completion behavior
- Structured reasoning patterns when prompted
No additional safety layers beyond those present in the base architecture are intentionally introduced.
Key Features & Capabilities
Core Strengths
- Fast inference on consumer GPUs and CPUs
- Low VRAM requirements
- Instruction-following compatibility
- Concise reasoning outputs
- Suitable for lightweight agent pipelines
Performance Characteristics
- Optimized for short-to-medium generation tasks
- Responsive in real-time assistant applications
- Works well in tool-driven or chain-of-thought–style prompts
- Practical for edge deployments and experimentation
Intended Use Cases
- Lightweight AI assistant
- Prompt engineering experimentation
- Tool-augmented agents
- Rapid-response chat systems
- Local inference environments
- Educational or research workflows
- Controlled “uncensored” deployment environments
Chat Template & Prompt Format
This model follows the Gemma instruction format.
For best results:
- Provide explicit system instructions
- Use structured reasoning prompts when needed
- Avoid mixing non-Gemma chat formats
Hardware & Deployment Notes
Due to its 1B parameter size:
- Runs efficiently on 8GB GPUs
- Suitable for CPU inference with quantization
- Ideal for edge devices and low-resource setups
- Compatible with common inference engines supporting Gemma architecture
Quantized versions (GGUF, GPTQ, AWQ, etc.) may be used depending on deployment stack.
Alignment & Safety Notice
This is an “uncensored” derivative configuration.
- Reduced refusal behavior compared to standard IT
- Users are responsible for system prompt controls
- Deployment should follow local laws and ethical guidelines
- No additional alignment layers are added by this repository
Use responsibly.
License & Usage Notes
This model inherits the Gemma License from its base model (google/gemma-3-1b-it).
- The Gemma License is a custom license provided by Google
- You must review and comply with the Gemma License terms
- This repository does not change or replace the original licensing terms
Users are responsible for ensuring compliance with all applicable regulations.
Acknowledgements
- Google for the Gemma 3 architecture and base model
- The Hugging Face ecosystem
- Open-source tooling communities supporting lightweight deployment
Community & Support
- Use the Hugging Face Discussions tab for issues and questions
- Community experimentation and benchmarking feedback is welcome
📂 GGUF File List
| 📁 Filename | 📦 Size | ⚡ Download |
|---|---|---|
|
Gemma-3-1B-it-GLM-4.7-Flash-Heretic-Uncensored-Thinking_F16.gguf
LFS
FP16
|
1.87 GB | Download |
|
Gemma-3-1B-it-GLM-4.7-Flash-Heretic-Uncensored-Thinking_Q2_k.gguf
LFS
Q2
|
657.86 MB | Download |
|
Gemma-3-1B-it-GLM-4.7-Flash-Heretic-Uncensored-Thinking_Q3_k_m.gguf
LFS
Q3
|
688.95 MB | Download |
|
Gemma-3-1B-it-GLM-4.7-Flash-Heretic-Uncensored-Thinking_Q4_k_m.gguf
Recommended
LFS
Q4
|
768.72 MB | Download |
|
Gemma-3-1B-it-GLM-4.7-Flash-Heretic-Uncensored-Thinking_Q5_k_m.gguf
LFS
Q5
|
811.91 MB | Download |
|
Gemma-3-1B-it-GLM-4.7-Flash-Heretic-Uncensored-Thinking_Q6_k.gguf
LFS
Q6
|
964.87 MB | Download |
|
Gemma-3-1B-it-GLM-4.7-Flash-Heretic-Uncensored-Thinking_Q8_0.gguf
LFS
Q8
|
1019.77 MB | Download |