π Model Description
base_model: google/umt5-xxl library_name: gguf license: apache-2.0 quantized_by: city96 language: en
This is a GGUF conversion of Google's UMT5 xxl model, specifically the encoder part.
The weights can be used with ./llama-embedding
or with the ComfyUI-GGUF custom node together with image/video generation models.
This is a non imatrix quant as llama.cpp doesn't support imatrix creation for T5 models at the time of writing. It's therefore recommended to use Q5KM or larger for the best results, although smaller models may also still provide decent results in resource constrained scenarios.
π GGUF File List
π Filename | π¦ Size | β‘ Download |
---|---|---|
umt5-xxl-encoder-F16.gguf
LFS
FP16
|
10.59 GB | Download |
umt5-xxl-encoder-F32.gguf
LFS
|
21.17 GB | Download |
umt5-xxl-encoder-Q3_K_M.gguf
LFS
Q3
|
2.85 GB | Download |
umt5-xxl-encoder-Q3_K_S.gguf
LFS
Q3
|
2.66 GB | Download |
umt5-xxl-encoder-Q4_K_M.gguf
Recommended
LFS
Q4
|
3.4 GB | Download |
umt5-xxl-encoder-Q4_K_S.gguf
LFS
Q4
|
3.26 GB | Download |
umt5-xxl-encoder-Q5_K_M.gguf
LFS
Q5
|
3.86 GB | Download |
umt5-xxl-encoder-Q5_K_S.gguf
LFS
Q5
|
3.77 GB | Download |
umt5-xxl-encoder-Q6_K.gguf
LFS
Q6
|
4.35 GB | Download |
umt5-xxl-encoder-Q8_0.gguf
LFS
Q8
|
5.63 GB | Download |