πŸ“‹ Model Description


base_model: google/umt5-xxl library_name: gguf license: apache-2.0 quantized_by: city96 language: en

This is a GGUF conversion of Google's UMT5 xxl model, specifically the encoder part.

The weights can be used with ./llama-embedding or with the ComfyUI-GGUF custom node together with image/video generation models.

This is a non imatrix quant as llama.cpp doesn't support imatrix creation for T5 models at the time of writing. It's therefore recommended to use Q5KM or larger for the best results, although smaller models may also still provide decent results in resource constrained scenarios.

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
umt5-xxl-encoder-F16.gguf
LFS FP16
10.59 GB Download
umt5-xxl-encoder-F32.gguf
LFS
21.17 GB Download
umt5-xxl-encoder-Q3_K_M.gguf
LFS Q3
2.85 GB Download
umt5-xxl-encoder-Q3_K_S.gguf
LFS Q3
2.66 GB Download
umt5-xxl-encoder-Q4_K_M.gguf
Recommended LFS Q4
3.4 GB Download
umt5-xxl-encoder-Q4_K_S.gguf
LFS Q4
3.26 GB Download
umt5-xxl-encoder-Q5_K_M.gguf
LFS Q5
3.86 GB Download
umt5-xxl-encoder-Q5_K_S.gguf
LFS Q5
3.77 GB Download
umt5-xxl-encoder-Q6_K.gguf
LFS Q6
4.35 GB Download
umt5-xxl-encoder-Q8_0.gguf
LFS Q8
5.63 GB Download