πŸ“‹ Model Description


license: apache-2.0 model_name: Mistral-Nemo-Instruct-2407 base_model: mistralai/Mistral-Nemo-Instruct-2407 inference: false model_creator: mistralai quantized_by: Second State Inc. language: - en - fr - de - es - it - pt - ru - zh - ja








Mistral-Nemo-Instruct-2407-GGUF

Original Model

mistralai/Mistral-Nemo-Instruct-2407

Run with LlamaEdge

  • LlamaEdge version: v0.12.4
  • Prompt template

- Prompt type: mistral-instruct

- Prompt string

<s>[INST] {usermessage1} [/INST]{assistantmessage1}</s>[INST] {usermessage2} [/INST]{assistantmessage2}</s>
  • Context size: 128000
  • Run as LlamaEdge service
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Mistral-Nemo-Instruct-2407-Q5KM.gguf \
    llama-api-server.wasm \
    --prompt-template mistral-instruct \
    --ctx-size 128000 \
    --model-name Mistral-Nemo-Instruct-2407
  • Run as LlamaEdge command app
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Mistral-Nemo-Instruct-2407-Q5KM.gguf \
    llama-chat.wasm \
    --prompt-template mistral-instruct \
    --ctx-size 128000

Quantized GGUF Models

NameQuant methodBitsSizeUse case
Mistral-Nemo-Instruct-2407-Q2K.ggufQ2K24.79 GBsmallest, significant quality loss - not recommended for most purposes
Mistral-Nemo-Instruct-2407-Q3KL.ggufQ3K_L36.56 GBsmall, substantial quality loss
Mistral-Nemo-Instruct-2407-Q3KM.ggufQ3K_M36.08 GBvery small, high quality loss
Mistral-Nemo-Instruct-2407-Q3KS.ggufQ3K_S35.53 GBvery small, high quality loss
Mistral-Nemo-Instruct-2407-Q40.ggufQ4047.07 GBlegacy; small, very high quality loss - prefer using Q3KM
Mistral-Nemo-Instruct-2407-Q4KM.ggufQ4K_M47.48 GBmedium, balanced quality - recommended
Mistral-Nemo-Instruct-2407-Q4KS.ggufQ4K_S47.12 GBsmall, greater quality loss
Mistral-Nemo-Instruct-2407-Q50.ggufQ5058.52 GBlegacy; medium, balanced quality - prefer using Q4KM
Mistral-Nemo-Instruct-2407-Q5KM.ggufQ5K_M58.73 GBlarge, very low quality loss - recommended
Mistral-Nemo-Instruct-2407-Q5KS.ggufQ5K_S58.52 GBlarge, low quality loss - recommended
Mistral-Nemo-Instruct-2407-Q6K.ggufQ6K610.1 GBvery large, extremely low quality loss
Mistral-Nemo-Instruct-2407-Q80.ggufQ80813.0 GBvery large, extremely low quality loss - not recommended
Mistral-Nemo-Instruct-2407-f16.gguff161624.5 GB
Quantized with llama.cpp b3438.

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
Mistral-Nemo-Instruct-2407-Q2_K.gguf
LFS Q2
4.46 GB Download
Mistral-Nemo-Instruct-2407-Q3_K_L.gguf
LFS Q3
6.11 GB Download
Mistral-Nemo-Instruct-2407-Q3_K_M.gguf
LFS Q3
5.67 GB Download
Mistral-Nemo-Instruct-2407-Q3_K_S.gguf
LFS Q3
5.15 GB Download
Mistral-Nemo-Instruct-2407-Q4_0.gguf
Recommended LFS Q4
6.59 GB Download
Mistral-Nemo-Instruct-2407-Q4_K_M.gguf
LFS Q4
6.96 GB Download
Mistral-Nemo-Instruct-2407-Q4_K_S.gguf
LFS Q4
6.63 GB Download
Mistral-Nemo-Instruct-2407-Q5_0.gguf
LFS Q5
7.93 GB Download
Mistral-Nemo-Instruct-2407-Q5_K_M.gguf
LFS Q5
8.13 GB Download
Mistral-Nemo-Instruct-2407-Q5_K_S.gguf
LFS Q5
7.93 GB Download
Mistral-Nemo-Instruct-2407-Q6_K.gguf
LFS Q6
9.37 GB Download
Mistral-Nemo-Instruct-2407-Q8_0.gguf
LFS Q8
12.13 GB Download
Mistral-Nemo-Instruct-2407-f16.gguf
LFS FP16
22.82 GB Download