๐ Model Description
Quantization made by Richard Erkhov.
Llama-3-ELYZA-JP-8B - GGUF
- Model creator: https://huggingface.co/elyza/
- Original model: https://huggingface.co/elyza/Llama-3-ELYZA-JP-8B/
| Name | Quant method | Size |
|---|---|---|
| Llama-3-ELYZA-JP-8B.Q2K.gguf | Q2K | 2.96GB |
| Llama-3-ELYZA-JP-8B.IQ3XS.gguf | IQ3XS | 3.28GB |
| Llama-3-ELYZA-JP-8B.IQ3S.gguf | IQ3S | 3.43GB |
| Llama-3-ELYZA-JP-8B.Q3KS.gguf | Q3K_S | 3.41GB |
| Llama-3-ELYZA-JP-8B.IQ3M.gguf | IQ3M | 3.52GB |
| Llama-3-ELYZA-JP-8B.Q3K.gguf | Q3K | 3.74GB |
| Llama-3-ELYZA-JP-8B.Q3KM.gguf | Q3K_M | 3.74GB |
| Llama-3-ELYZA-JP-8B.Q3KL.gguf | Q3K_L | 4.03GB |
| Llama-3-ELYZA-JP-8B.IQ4XS.gguf | IQ4XS | 4.18GB |
| Llama-3-ELYZA-JP-8B.Q40.gguf | Q40 | 4.34GB |
| Llama-3-ELYZA-JP-8B.IQ4NL.gguf | IQ4NL | 4.38GB |
| Llama-3-ELYZA-JP-8B.Q4KS.gguf | Q4K_S | 4.37GB |
| Llama-3-ELYZA-JP-8B.Q4K.gguf | Q4K | 4.58GB |
| Llama-3-ELYZA-JP-8B.Q4KM.gguf | Q4K_M | 4.58GB |
| Llama-3-ELYZA-JP-8B.Q41.gguf | Q41 | 4.78GB |
| Llama-3-ELYZA-JP-8B.Q50.gguf | Q50 | 5.21GB |
| Llama-3-ELYZA-JP-8B.Q5KS.gguf | Q5K_S | 5.21GB |
| Llama-3-ELYZA-JP-8B.Q5K.gguf | Q5K | 5.34GB |
| Llama-3-ELYZA-JP-8B.Q5KM.gguf | Q5K_M | 5.34GB |
| Llama-3-ELYZA-JP-8B.Q51.gguf | Q51 | 5.65GB |
| Llama-3-ELYZA-JP-8B.Q6K.gguf | Q6K | 6.14GB |
| Llama-3-ELYZA-JP-8B.Q80.gguf | Q80 | 7.95GB |
Original model description:
library_name: transformers
license: llama3
language:
- ja
- en
Llama-3-ELYZA-JP-8B
Model Description
Llama-3-ELYZA-JP-8B is a large language model trained by ELYZA, Inc.
Based on meta-llama/Meta-Llama-3-8B-Instruct, it has been enhanced for Japanese usage through additional pre-training and instruction tuning. (Built with Meta Llama3)
For more details, please refer to our blog post.
Usage
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
DEFAULTSYSTEMPROMPT = "ใใชใใฏ่ช ๅฎใงๅช็งใชๆฅๆฌไบบใฎใขใทในใฟใณใใงใใ็นใซๆ็คบใ็กใๅ ดๅใฏใๅธธใซๆฅๆฌ่ชใงๅ็ญใใฆใใ ใใใ"
text = "ไปไบใฎ็ฑๆใๅใๆปใใใใฎใขใคใใขใ5ใคๆใใฆใใ ใใใ"
model_name = "elyza/Llama-3-ELYZA-JP-8B"
tokenizer = AutoTokenizer.frompretrained(modelname)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype="auto",
device_map="auto",
)
model.eval()
messages = [
{"role": "system", "content": DEFAULTSYSTEMPROMPT},
{"role": "user", "content": text},
]
prompt = tokenizer.applychattemplate(
messages,
tokenize=False,
addgenerationprompt=True
)
token_ids = tokenizer.encode(
prompt, addspecialtokens=False, return_tensors="pt"
)
with torch.no_grad():
output_ids = model.generate(
token_ids.to(model.device),
maxnewtokens=1200,
do_sample=True,
temperature=0.6,
top_p=0.9,
)
output = tokenizer.decode(
outputids.tolist()[0][tokenids.size(1):], skipspecialtokens=True
)
print(output)
Developers
Listed in alphabetical order.
License
Meta Llama 3 Community License
How to Cite
@misc{elyzallama2024,
title={elyza/Llama-3-ELYZA-JP-8B},
url={https://huggingface.co/elyza/Llama-3-ELYZA-JP-8B},
author={Masato Hirakawa and Shintaro Horie and Tomoaki Nakamura and Daisuke Oba and Sam Passaglia and Akira Sasaki},
year={2024},
}
Citations
@article{llama3modelcard,
title={Llama 3 Model Card},
author={AI@Meta},
year={2024},
url = {https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md}
}
๐ GGUF File List
| ๐ Filename | ๐ฆ Size | โก Download |
|---|---|---|
|
Llama-3-ELYZA-JP-8B.IQ3_M.gguf
LFS
Q3
|
3.52 GB | Download |
|
Llama-3-ELYZA-JP-8B.IQ3_S.gguf
LFS
Q3
|
3.43 GB | Download |
|
Llama-3-ELYZA-JP-8B.IQ3_XS.gguf
LFS
Q3
|
3.28 GB | Download |
|
Llama-3-ELYZA-JP-8B.IQ4_NL.gguf
LFS
Q4
|
4.38 GB | Download |
|
Llama-3-ELYZA-JP-8B.IQ4_XS.gguf
LFS
Q4
|
4.18 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q2_K.gguf
LFS
Q2
|
2.96 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q3_K.gguf
LFS
Q3
|
3.74 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q3_K_L.gguf
LFS
Q3
|
4.03 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q3_K_M.gguf
LFS
Q3
|
3.74 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q3_K_S.gguf
LFS
Q3
|
3.41 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q4_0.gguf
Recommended
LFS
Q4
|
4.34 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q4_1.gguf
LFS
Q4
|
4.78 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q4_K.gguf
LFS
Q4
|
4.58 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q4_K_M.gguf
LFS
Q4
|
4.58 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q4_K_S.gguf
LFS
Q4
|
4.37 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q5_0.gguf
LFS
Q5
|
5.21 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q5_1.gguf
LFS
Q5
|
5.65 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q5_K.gguf
LFS
Q5
|
5.34 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q5_K_M.gguf
LFS
Q5
|
5.34 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q5_K_S.gguf
LFS
Q5
|
5.21 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q6_K.gguf
LFS
Q6
|
6.14 GB | Download |
|
Llama-3-ELYZA-JP-8B.Q8_0.gguf
LFS
Q8
|
7.95 GB | Download |