πŸ“‹ Model Description


base_model: nex-agi/DeepSeek-V3.1-Nex-N1.1 language:
  • en
library_name: transformers license: apache-2.0 mradermacher: readme_rev: 1 quantized_by: mradermacher

About









weighted/imatrix quants of https://huggingface.co/nex-agi/DeepSeek-V3.1-Nex-N1.1

For a convenient overview and download list, visit our model page for this model.

static quants are available at https://huggingface.co/mradermacher/DeepSeek-V3.1-Nex-N1.1-GGUF

Usage

If you are unsure how to use GGUF files, refer to one of TheBloke's
READMEs
for
more details, including on how to concatenate multi-part files.

Provided Quants

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

LinkTypeSize/GBNotes
GGUFimatrix1.1imatrix file (for creating your own quants)
GGUFi1-IQ1_S133.8for the desperate
GGUFi1-IQ1_M149.2mostly desperate
GGUFi1-IQ2_XXS174.7
GGUFi1-IQ2_XS195.3
GGUFi1-IQ2_S197.2
P1 P2 P3 P4 P5i1-IQ2_M217.7
P1 P2 P3 P4 P5i1-Q2K_S224.9very low quality
P1 P2 P3 P4 P5 P6i1-Q2K244.2IQ3XXS probably better
P1 P2 P3 P4 P5 P6i1-IQ3_XXS258.1lower quality
P1 P2 P3 P4 P5 P6i1-IQ3_XS273.0
P1 P2 P3 P4 P5 P6i1-IQ3S289.3beats Q3K*
P1 P2 P3 P4 P5 P6i1-Q3KS289.3IQ3_XS probably better
P1 P2 P3 P4 P5 P6 P7i1-IQ3_M292.3
P1 P2 P3 P4 P5 P6 P7i1-Q3KM319.4IQ3S probably better
P1 P2 P3 P4 P5 P6 P7 P8i1-Q3KL347.6IQ3_M probably better
P1 P2 P3 P4 P5 P6 P7 P8i1-IQ4_XS357.2
P1 P2 P3 P4 P5 P6 P7 P8i1-Q4_0379.1fast, low quality
P1 P2 P3 P4 P5 P6 P7 P8i1-Q4KS380.2optimal size/speed/quality
P1 P2 P3 P4 P5 P6 P7 P8 P9i1-Q4K_M404.6fast, recommended
P1 P2 P3 P4 P5 P6 P7 P8 P9i1-Q4_1420.0
P1 P2 P3 P4 P5 P6 P7 P8 P9 P10i1-Q5KS461.9
P1 P2 P3 P4 P5 P6 P7 P8 P9 P10i1-Q5KM475.5
P1 P2 P3 P4 P5 P6 P7 P8 P9 P10 P11 P12i1-Q6K551.0practically like static Q6K
Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

!image.png

And here are Artefact2's thoughts on the matter:
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9

FAQ / Model Request

See https://huggingface.co/mradermacher/model_requests for some answers to
questions you might have and/or if you want some other model quantized.

Thanks

I thank my company, nethype GmbH, for letting
me use its servers and providing upgrades to my workstation to enable
this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
DeepSeek-V3.1-Nex-N1.1.i1-IQ1_M.gguf
Recommended LFS
138.82 GB Download
DeepSeek-V3.1-Nex-N1.1.i1-IQ1_S.gguf
LFS
124.56 GB Download
DeepSeek-V3.1-Nex-N1.1.i1-IQ2_S.gguf
LFS Q2
183.6 GB Download
DeepSeek-V3.1-Nex-N1.1.i1-IQ2_XS.gguf
LFS Q2
181.82 GB Download
DeepSeek-V3.1-Nex-N1.1.i1-IQ2_XXS.gguf
LFS Q2
162.59 GB Download
DeepSeek-V3.1-Nex-N1.1.imatrix.gguf
LFS
960.44 MB Download