πŸ“‹ Model Description

Quantization made by Richard Erkhov.

Github

Discord

Request more models

Llama-3.1-BaseNSFW-pretrainede-0.5 - GGUF

  • Model creator: https://huggingface.co/athirdpath/
  • Original model: https://huggingface.co/athirdpath/Llama-3.1-BaseNSFW-pretrainede-0.5/

NameQuant methodSize
Llama-3.1-BaseNSFW-pretrainede-0.5.Q2K.ggufQ2K2.96GB
Llama-3.1-BaseNSFW-pretrainede-0.5.IQ3XS.ggufIQ3XS3.28GB
Llama-3.1-BaseNSFW-pretrainede-0.5.IQ3S.ggufIQ3S3.43GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q3KS.ggufQ3K_S3.41GB
Llama-3.1-BaseNSFW-pretrainede-0.5.IQ3M.ggufIQ3M3.52GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q3K.ggufQ3K3.74GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q3KM.ggufQ3K_M3.74GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q3KL.ggufQ3K_L4.03GB
Llama-3.1-BaseNSFW-pretrainede-0.5.IQ4XS.ggufIQ4XS4.18GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q40.ggufQ404.34GB
Llama-3.1-BaseNSFW-pretrainede-0.5.IQ4NL.ggufIQ4NL4.38GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q4KS.ggufQ4K_S4.37GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q4K.ggufQ4K4.58GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q4KM.ggufQ4K_M4.58GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q41.ggufQ414.78GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q50.ggufQ505.21GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q5KS.ggufQ5K_S5.21GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q5K.ggufQ5K5.34GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q5KM.ggufQ5K_M5.34GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q51.ggufQ515.65GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q6K.ggufQ6K6.14GB
Llama-3.1-BaseNSFW-pretrainede-0.5.Q80.ggufQ807.95GB

Original model description:



license: llama3.1

Llama 3.1 Base, continually pretrained with 0.5 Epochs (2100 steps @ total batch 64) of the same 1.5gb private dataset that underpins Iambe

Mostly a proof of concept, but outputs are better than expected. It'd likely be quite good with some instruction tuning.

-----

Why do this? I have a niche use case where I cannot increase compute over 8b, and L3/3.1 are the only models in this size category that meet my needs for logic. However, both versions of L3/3.1 have the damn repetition/token overconfidence problem, and this is meant to disrupt that certainty without disrupting the model's ability to function.

By the way, I think it's the lm_head that is causing the looping, but it might be the embeddings being too separated. I'm not going to pay two more times to test them separately, however :p

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.IQ3_M.gguf
LFS Q3
3.52 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.IQ3_S.gguf
LFS Q3
3.43 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.IQ3_XS.gguf
LFS Q3
3.28 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.IQ4_NL.gguf
LFS Q4
4.38 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.IQ4_XS.gguf
LFS Q4
4.18 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q2_K.gguf
LFS Q2
2.96 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q3_K.gguf
LFS Q3
3.74 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q3_K_L.gguf
LFS Q3
4.03 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q3_K_M.gguf
LFS Q3
3.74 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q3_K_S.gguf
LFS Q3
3.41 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q4_0.gguf
Recommended LFS Q4
4.34 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q4_1.gguf
LFS Q4
4.78 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q4_K.gguf
LFS Q4
4.58 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q4_K_M.gguf
LFS Q4
4.58 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q4_K_S.gguf
LFS Q4
4.37 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q5_0.gguf
LFS Q5
5.21 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q5_1.gguf
LFS Q5
5.65 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q5_K.gguf
LFS Q5
5.34 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q5_K_M.gguf
LFS Q5
5.34 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q5_K_S.gguf
LFS Q5
5.21 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q6_K.gguf
LFS Q6
6.14 GB Download
Llama-3.1-Base_NSFW-pretrained_e-0.5.Q8_0.gguf
LFS Q8
7.95 GB Download