๐Ÿ“‹ Model Description

Quantization made by Richard Erkhov.

Github

Discord

Request more models

Llama-3-ELYZA-JP-8B - GGUF

  • Model creator: https://huggingface.co/elyza/
  • Original model: https://huggingface.co/elyza/Llama-3-ELYZA-JP-8B/

NameQuant methodSize
Llama-3-ELYZA-JP-8B.Q2K.ggufQ2K2.96GB
Llama-3-ELYZA-JP-8B.IQ3XS.ggufIQ3XS3.28GB
Llama-3-ELYZA-JP-8B.IQ3S.ggufIQ3S3.43GB
Llama-3-ELYZA-JP-8B.Q3KS.ggufQ3K_S3.41GB
Llama-3-ELYZA-JP-8B.IQ3M.ggufIQ3M3.52GB
Llama-3-ELYZA-JP-8B.Q3K.ggufQ3K3.74GB
Llama-3-ELYZA-JP-8B.Q3KM.ggufQ3K_M3.74GB
Llama-3-ELYZA-JP-8B.Q3KL.ggufQ3K_L4.03GB
Llama-3-ELYZA-JP-8B.IQ4XS.ggufIQ4XS4.18GB
Llama-3-ELYZA-JP-8B.Q40.ggufQ404.34GB
Llama-3-ELYZA-JP-8B.IQ4NL.ggufIQ4NL4.38GB
Llama-3-ELYZA-JP-8B.Q4KS.ggufQ4K_S4.37GB
Llama-3-ELYZA-JP-8B.Q4K.ggufQ4K4.58GB
Llama-3-ELYZA-JP-8B.Q4KM.ggufQ4K_M4.58GB
Llama-3-ELYZA-JP-8B.Q41.ggufQ414.78GB
Llama-3-ELYZA-JP-8B.Q50.ggufQ505.21GB
Llama-3-ELYZA-JP-8B.Q5KS.ggufQ5K_S5.21GB
Llama-3-ELYZA-JP-8B.Q5K.ggufQ5K5.34GB
Llama-3-ELYZA-JP-8B.Q5KM.ggufQ5K_M5.34GB
Llama-3-ELYZA-JP-8B.Q51.ggufQ515.65GB
Llama-3-ELYZA-JP-8B.Q6K.ggufQ6K6.14GB
Llama-3-ELYZA-JP-8B.Q80.ggufQ807.95GB

Original model description:



library_name: transformers
license: llama3
language:
  • ja
  • en



Llama-3-ELYZA-JP-8B

!Llama-3-ELYZA-JP-8B-image

Model Description

Llama-3-ELYZA-JP-8B is a large language model trained by ELYZA, Inc.
Based on meta-llama/Meta-Llama-3-8B-Instruct, it has been enhanced for Japanese usage through additional pre-training and instruction tuning. (Built with Meta Llama3)

For more details, please refer to our blog post.

Usage

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

DEFAULTSYSTEMPROMPT = "ใ‚ใชใŸใฏ่ช ๅฎŸใงๅ„ช็ง€ใชๆ—ฅๆœฌไบบใฎใ‚ขใ‚ทใ‚นใ‚ฟใƒณใƒˆใงใ™ใ€‚็‰นใซๆŒ‡็คบใŒ็„กใ„ๅ ดๅˆใฏใ€ๅธธใซๆ—ฅๆœฌ่ชžใงๅ›ž็ญ”ใ—ใฆใใ ใ•ใ„ใ€‚"
text = "ไป•ไบ‹ใฎ็†ฑๆ„ใ‚’ๅ–ใ‚Šๆˆปใ™ใŸใ‚ใฎใ‚ขใ‚คใƒ‡ใ‚ขใ‚’5ใคๆŒ™ใ’ใฆใใ ใ•ใ„ใ€‚"

model_name = "elyza/Llama-3-ELYZA-JP-8B"

tokenizer = AutoTokenizer.frompretrained(modelname)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype="auto",
device_map="auto",
)
model.eval()

messages = [
{"role": "system", "content": DEFAULTSYSTEMPROMPT},
{"role": "user", "content": text},
]
prompt = tokenizer.applychattemplate(
messages,
tokenize=False,
addgenerationprompt=True
)
token_ids = tokenizer.encode(
prompt, addspecialtokens=False, return_tensors="pt"
)

with torch.no_grad():
output_ids = model.generate(
token_ids.to(model.device),
maxnewtokens=1200,
do_sample=True,
temperature=0.6,
top_p=0.9,
)
output = tokenizer.decode(
outputids.tolist()[0][tokenids.size(1):], skipspecialtokens=True
)
print(output)

Developers

Listed in alphabetical order.

License

Meta Llama 3 Community License

How to Cite

@misc{elyzallama2024,
      title={elyza/Llama-3-ELYZA-JP-8B},
      url={https://huggingface.co/elyza/Llama-3-ELYZA-JP-8B},
      author={Masato Hirakawa and Shintaro Horie and Tomoaki Nakamura and Daisuke Oba and Sam Passaglia and Akira Sasaki},
      year={2024},
}

Citations

@article{llama3modelcard,
    title={Llama 3 Model Card},
    author={AI@Meta},
    year={2024},
    url = {https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md}
}

๐Ÿ“‚ GGUF File List

๐Ÿ“ Filename ๐Ÿ“ฆ Size โšก Download
Llama-3-ELYZA-JP-8B.IQ3_M.gguf
LFS Q3
3.52 GB Download
Llama-3-ELYZA-JP-8B.IQ3_S.gguf
LFS Q3
3.43 GB Download
Llama-3-ELYZA-JP-8B.IQ3_XS.gguf
LFS Q3
3.28 GB Download
Llama-3-ELYZA-JP-8B.IQ4_NL.gguf
LFS Q4
4.38 GB Download
Llama-3-ELYZA-JP-8B.IQ4_XS.gguf
LFS Q4
4.18 GB Download
Llama-3-ELYZA-JP-8B.Q2_K.gguf
LFS Q2
2.96 GB Download
Llama-3-ELYZA-JP-8B.Q3_K.gguf
LFS Q3
3.74 GB Download
Llama-3-ELYZA-JP-8B.Q3_K_L.gguf
LFS Q3
4.03 GB Download
Llama-3-ELYZA-JP-8B.Q3_K_M.gguf
LFS Q3
3.74 GB Download
Llama-3-ELYZA-JP-8B.Q3_K_S.gguf
LFS Q3
3.41 GB Download
Llama-3-ELYZA-JP-8B.Q4_0.gguf
Recommended LFS Q4
4.34 GB Download
Llama-3-ELYZA-JP-8B.Q4_1.gguf
LFS Q4
4.78 GB Download
Llama-3-ELYZA-JP-8B.Q4_K.gguf
LFS Q4
4.58 GB Download
Llama-3-ELYZA-JP-8B.Q4_K_M.gguf
LFS Q4
4.58 GB Download
Llama-3-ELYZA-JP-8B.Q4_K_S.gguf
LFS Q4
4.37 GB Download
Llama-3-ELYZA-JP-8B.Q5_0.gguf
LFS Q5
5.21 GB Download
Llama-3-ELYZA-JP-8B.Q5_1.gguf
LFS Q5
5.65 GB Download
Llama-3-ELYZA-JP-8B.Q5_K.gguf
LFS Q5
5.34 GB Download
Llama-3-ELYZA-JP-8B.Q5_K_M.gguf
LFS Q5
5.34 GB Download
Llama-3-ELYZA-JP-8B.Q5_K_S.gguf
LFS Q5
5.21 GB Download
Llama-3-ELYZA-JP-8B.Q6_K.gguf
LFS Q6
6.14 GB Download
Llama-3-ELYZA-JP-8B.Q8_0.gguf
LFS Q8
7.95 GB Download