πŸ“‹ Model Description

Quantization made by Richard Erkhov.

Github

Discord

Request more models

llama-3.1-minitron-6b-width-base-chatml - GGUF

  • Model creator: https://huggingface.co/dumping-grounds/
  • Original model: https://huggingface.co/dumping-grounds/llama-3.1-minitron-6b-width-base-chatml/

NameQuant methodSize
llama-3.1-minitron-6b-width-base-chatml.Q2K.ggufQ2K2.36GB
llama-3.1-minitron-6b-width-base-chatml.IQ3XS.ggufIQ3XS2.6GB
llama-3.1-minitron-6b-width-base-chatml.IQ3S.ggufIQ3S2.72GB
llama-3.1-minitron-6b-width-base-chatml.Q3KS.ggufQ3K_S2.7GB
llama-3.1-minitron-6b-width-base-chatml.IQ3M.ggufIQ3M2.82GB
llama-3.1-minitron-6b-width-base-chatml.Q3K.ggufQ3K2.97GB
llama-3.1-minitron-6b-width-base-chatml.Q3KM.ggufQ3K_M2.97GB
llama-3.1-minitron-6b-width-base-chatml.Q3KL.ggufQ3K_L3.21GB
llama-3.1-minitron-6b-width-base-chatml.IQ4XS.ggufIQ4XS3.32GB
llama-3.1-minitron-6b-width-base-chatml.Q40.ggufQ403.44GB
llama-3.1-minitron-6b-width-base-chatml.IQ4NL.ggufIQ4NL3.48GB
llama-3.1-minitron-6b-width-base-chatml.Q4KS.ggufQ4K_S3.46GB
llama-3.1-minitron-6b-width-base-chatml.Q4K.ggufQ4K3.62GB
llama-3.1-minitron-6b-width-base-chatml.Q4KM.ggufQ4K_M3.62GB
llama-3.1-minitron-6b-width-base-chatml.Q41.ggufQ413.79GB
llama-3.1-minitron-6b-width-base-chatml.Q50.ggufQ504.14GB
llama-3.1-minitron-6b-width-base-chatml.Q5KS.ggufQ5K_S4.14GB
llama-3.1-minitron-6b-width-base-chatml.Q5K.ggufQ5K4.23GB
llama-3.1-minitron-6b-width-base-chatml.Q5KM.ggufQ5K_M4.23GB
llama-3.1-minitron-6b-width-base-chatml.Q51.ggufQ514.49GB
llama-3.1-minitron-6b-width-base-chatml.Q6K.ggufQ6K4.88GB
llama-3.1-minitron-6b-width-base-chatml.Q80.ggufQ806.32GB

Original model description:



base_model:
  • IntervitensInc/Llama-3.1-Minitron-4B-Width-Base-chatml

library_name: transformers
tags:
  • mergekit
  • merge


merge

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the passthrough merge method.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

slices:
  - sources:
      - model: IntervitensInc/Llama-3.1-Minitron-4B-Width-Base-chatml
        layer_range: [0, 24]
  - sources: # add middle layers with residuals scaled to zero
      - model: IntervitensInc/Llama-3.1-Minitron-4B-Width-Base-chatml
        layer_range: [8, 24]
        parameters:
          scale:
            - filter: o_proj
              value: 0.0
            - filter: down_proj
              value: 0.0
            - value: 1.0
  - sources:
      - model: IntervitensInc/Llama-3.1-Minitron-4B-Width-Base-chatml
        layer_range: [24, 32]
merge_method: passthrough
dtype: bfloat16

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
llama-3.1-minitron-6b-width-base-chatml.IQ3_M.gguf
LFS Q3
2.82 GB Download
llama-3.1-minitron-6b-width-base-chatml.IQ3_S.gguf
LFS Q3
2.72 GB Download
llama-3.1-minitron-6b-width-base-chatml.IQ3_XS.gguf
LFS Q3
2.6 GB Download
llama-3.1-minitron-6b-width-base-chatml.IQ4_NL.gguf
LFS Q4
3.48 GB Download
llama-3.1-minitron-6b-width-base-chatml.IQ4_XS.gguf
LFS Q4
3.32 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q2_K.gguf
LFS Q2
2.36 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q3_K.gguf
LFS Q3
2.97 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q3_K_L.gguf
LFS Q3
3.21 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q3_K_M.gguf
LFS Q3
2.97 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q3_K_S.gguf
LFS Q3
2.7 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q4_0.gguf
Recommended LFS Q4
3.44 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q4_1.gguf
LFS Q4
3.79 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q4_K.gguf
LFS Q4
3.62 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q4_K_M.gguf
LFS Q4
3.62 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q4_K_S.gguf
LFS Q4
3.46 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q5_0.gguf
LFS Q5
4.14 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q5_1.gguf
LFS Q5
4.49 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q5_K.gguf
LFS Q5
4.23 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q5_K_M.gguf
LFS Q5
4.23 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q5_K_S.gguf
LFS Q5
4.14 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q6_K.gguf
LFS Q6
4.88 GB Download
llama-3.1-minitron-6b-width-base-chatml.Q8_0.gguf
LFS Q8
6.32 GB Download