πŸ“‹ Model Description


license: apache-2.0 library_name: transformers tags:
  • language
  • granite-4.0
  • gguf
base_model:
  • ibm-granite/granite-4.0-1b

Granite 4.0 1b (GGUF)

[!NOTE]

This repository contains models that have been converted to the GGUF format with various quantizations from an IBM Granite base model.

>

Please reference the base model's full model card here:

https://huggingface.co/ibm-granite/granite-4.0-1b

Known Issues

This model often uses the full numerical range of a 32-bit float (f32), so variants with smaller numerical ranges may run into precision errors at inference. The F16 variant is known to fail on many hardware combinations.

The recommended full-precision variant is bf16.

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
granite-4.0-1b-F16.gguf
LFS FP16
3.04 GB Download
granite-4.0-1b-Q2_K.gguf
LFS Q2
669.19 MB Download
granite-4.0-1b-Q3_K_L.gguf
LFS Q3
883.26 MB Download
granite-4.0-1b-Q3_K_M.gguf
LFS Q3
820.51 MB Download
granite-4.0-1b-Q3_K_S.gguf
LFS Q3
749.19 MB Download
granite-4.0-1b-Q4_0.gguf
Recommended LFS Q4
929.82 MB Download
granite-4.0-1b-Q4_1.gguf
LFS Q4
1014.82 MB Download
granite-4.0-1b-Q4_K_M.gguf
LFS Q4
976.22 MB Download
granite-4.0-1b-Q4_K_S.gguf
LFS Q4
935.32 MB Download
granite-4.0-1b-Q5_0.gguf
LFS Q5
1.07 GB Download
granite-4.0-1b-Q5_1.gguf
LFS Q5
1.16 GB Download
granite-4.0-1b-Q5_K_M.gguf
LFS Q5
1.1 GB Download
granite-4.0-1b-Q5_K_S.gguf
LFS Q5
1.07 GB Download
granite-4.0-1b-Q6_K.gguf
LFS Q6
1.25 GB Download
granite-4.0-1b-Q8_0.gguf
LFS Q8
1.62 GB Download
granite-4.0-1b-bf16.gguf
LFS FP16
3.04 GB Download