πŸ“‹ Model Description

Quantization made by Richard Erkhov.

Github

Discord

Request more models

llama-3-8b-samantha - GGUF

  • Model creator: https://huggingface.co/ruslandev/
  • Original model: https://huggingface.co/ruslandev/llama-3-8b-samantha/

NameQuant methodSize
llama-3-8b-samantha.Q2K.ggufQ2K2.96GB
llama-3-8b-samantha.IQ3XS.ggufIQ3XS3.28GB
llama-3-8b-samantha.IQ3S.ggufIQ3S3.43GB
llama-3-8b-samantha.Q3KS.ggufQ3K_S3.41GB
llama-3-8b-samantha.IQ3M.ggufIQ3M3.52GB
llama-3-8b-samantha.Q3K.ggufQ3K3.74GB
llama-3-8b-samantha.Q3KM.ggufQ3K_M3.74GB
llama-3-8b-samantha.Q3KL.ggufQ3K_L4.03GB
llama-3-8b-samantha.IQ4XS.ggufIQ4XS4.18GB
llama-3-8b-samantha.Q40.ggufQ404.34GB
llama-3-8b-samantha.IQ4NL.ggufIQ4NL4.38GB
llama-3-8b-samantha.Q4KS.ggufQ4K_S4.37GB
llama-3-8b-samantha.Q4K.ggufQ4K4.58GB
llama-3-8b-samantha.Q4KM.ggufQ4K_M4.58GB
llama-3-8b-samantha.Q41.ggufQ414.78GB
llama-3-8b-samantha.Q50.ggufQ505.21GB
llama-3-8b-samantha.Q5KS.ggufQ5K_S5.21GB
llama-3-8b-samantha.Q5K.ggufQ5K5.34GB
llama-3-8b-samantha.Q5KM.ggufQ5K_M5.34GB
llama-3-8b-samantha.Q51.ggufQ515.65GB
llama-3-8b-samantha.Q6K.ggufQ6K6.14GB
llama-3-8b-samantha.Q80.ggufQ807.95GB

Original model description:



language:
  • en

license: apache-2.0
tags:
  • text-generation-inference
  • transformers
  • unsloth
  • llama
  • trl

base_model: unsloth/llama-3-8b-bnb-4bit
datasets:
  • cognitivecomputations/samantha-data


Uploaded model

  • Developed by: ruslandev
  • License: apache-2.0
  • Finetuned from model : unsloth/llama-3-8b-bnb-4bit

This model is finetuned on the data of Samantha.
Prompt format is Alpaca. I used the same system prompt as the original Samantha.

"""Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.

Instruction:

{SYSTEM_PROMPT}

Input:

{QUESTION}

Response:

"""

Training

gptchain framework has been used for training.

Training hyperparameters

  • learningrate: 2e-4
  • seed: 3407
  • gradientaccumulationsteps: 4
  • perdevicetrainbatchsize: 2
  • optimizer: adamw8bit
  • lrschedulertype: linear
  • warmupsteps: 5
  • numepochs: 2
  • weight_decay: 0.01

Training results

Training LossEpochStep
2.07780.01
0.62550.18120
0.62080.94620
0.62442.01306
2 epoch finetuning from llama-3-8b took 1 hour on a single A100 with Unsloth and Huggingface's TRL library.

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
llama-3-8b-samantha.IQ3_M.gguf
LFS Q3
3.52 GB Download
llama-3-8b-samantha.IQ3_S.gguf
LFS Q3
3.43 GB Download
llama-3-8b-samantha.IQ3_XS.gguf
LFS Q3
3.28 GB Download
llama-3-8b-samantha.IQ4_NL.gguf
LFS Q4
4.38 GB Download
llama-3-8b-samantha.IQ4_XS.gguf
LFS Q4
4.18 GB Download
llama-3-8b-samantha.Q2_K.gguf
LFS Q2
2.96 GB Download
llama-3-8b-samantha.Q3_K.gguf
LFS Q3
3.74 GB Download
llama-3-8b-samantha.Q3_K_L.gguf
LFS Q3
4.03 GB Download
llama-3-8b-samantha.Q3_K_M.gguf
LFS Q3
3.74 GB Download
llama-3-8b-samantha.Q3_K_S.gguf
LFS Q3
3.41 GB Download
llama-3-8b-samantha.Q4_0.gguf
Recommended LFS Q4
4.34 GB Download
llama-3-8b-samantha.Q4_1.gguf
LFS Q4
4.78 GB Download
llama-3-8b-samantha.Q4_K.gguf
LFS Q4
4.58 GB Download
llama-3-8b-samantha.Q4_K_M.gguf
LFS Q4
4.58 GB Download
llama-3-8b-samantha.Q4_K_S.gguf
LFS Q4
4.37 GB Download
llama-3-8b-samantha.Q5_0.gguf
LFS Q5
5.21 GB Download
llama-3-8b-samantha.Q5_1.gguf
LFS Q5
5.65 GB Download
llama-3-8b-samantha.Q5_K.gguf
LFS Q5
5.34 GB Download
llama-3-8b-samantha.Q5_K_M.gguf
LFS Q5
5.34 GB Download
llama-3-8b-samantha.Q5_K_S.gguf
LFS Q5
5.21 GB Download
llama-3-8b-samantha.Q6_K.gguf
LFS Q6
6.14 GB Download
llama-3-8b-samantha.Q8_0.gguf
LFS Q8
7.95 GB Download