π Model Description
license: apache-2.0 tags: - gguf - qwen - qwen3 - qwen3-coder - qwen3-coder-30B - qwen3-coder-30B-gguf - llama.cpp - quantized - text-generation - reasoning - agent - multilingual base_model: Qwen/Qwen3-Coder-30B-A3B-Instruct author: geoffmunn pipeline_tag: text-generation language: - en - zh - es - fr - de - ru - ar - ja - ko - hi
Qwen3-Coder-30B-A3B-Instruct-f16-GGUF
This is a GGUF-quantized version of the Qwen/Qwen3-Coder-30B-A3B-Instruct language model.
Converted for use with llama.cpp, LM Studio, OpenWebUI, GPT4All, and more.
π‘ Key Features of Qwen3-Coder-30B-A3B-Instruct:
Available Quantizations (from f16)
| Level | Quality | Speed | Size | Recommendation |
|---|---|---|---|---|
| Q2_K | Minimal | β‘ Fast | 11.30 GB | Only on severely memory-constrained systems. |
>π‘ Recommendations by Use Case
- π» Standard Laptop (i5/M1 Mac): Q5KM (optimal quality)
- π§ Reasoning, Coding, Math: Q5KM or Q6K
- π RAG, Retrieval, Precision Tasks: Q6K or Q80
- π€ Agent & Tool Integration: Q5KM
- π οΈ Development & Testing: Test from Q4KM up to Q80
Usage
Load this model using:
- OpenWebUI β self-hosted AI interface with RAG & tools
- LM Studio β desktop app with GPU support
- GPT4All β private, offline AI chatbot
- Or directly via
llama.cpp
Each quantized model includes its own README.md and shares a common MODELFILE.
Author
π€ Geoff Munn (@geoffmunn)
π Hugging Face Profile
Disclaimer
This is a community conversion for local inference. Not affiliated with Alibaba Cloud or the Qwen team.
π GGUF File List
| π Filename | π¦ Size | β‘ Download |
|---|---|---|
|
Qwen3-Coder-30B-A3B-Instruct-f16-imatrix-4697-coder.gguf
LFS
FP16
|
116.38 MB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16-imatrix-4697-generic.gguf
LFS
FP16
|
116.38 MB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16-imatrix:Q3_K_HIFI.gguf
LFS
Q3
|
19.05 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16-imatrix:Q3_K_M.gguf
LFS
Q3
|
17.28 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16-imatrix:Q3_K_S.gguf
LFS
Q3
|
16.26 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16:Q2_K.gguf
LFS
Q2
|
10.49 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16:Q3_K_HIFI.gguf
LFS
Q3
|
15.69 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16:Q3_K_M.gguf
LFS
Q3
|
13.7 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16:Q3_K_S.gguf
LFS
Q3
|
12.38 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16:Q4_K_HIFI.gguf
LFS
Q4
|
19.05 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16:Q4_K_M.gguf
Recommended
LFS
Q4
|
17.28 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16:Q4_K_S.gguf
LFS
Q4
|
16.26 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16:Q5_K_M.gguf
LFS
Q5
|
20.23 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16:Q5_K_S.gguf
LFS
Q5
|
19.63 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16:Q6_K.gguf
LFS
Q6
|
23.37 GB | Download |
|
Qwen3-Coder-30B-A3B-Instruct-f16:Q8_0.gguf
LFS
Q8
|
30.25 GB | Download |