πŸ“‹ Model Description

Quantization made by Richard Erkhov.

Github

Discord

Request more models

japanese-stablelm-base-beta-7b - GGUF

  • Model creator: https://huggingface.co/stabilityai/
  • Original model: https://huggingface.co/stabilityai/japanese-stablelm-base-beta-7b/

NameQuant methodSize
japanese-stablelm-base-beta-7b.Q2K.ggufQ2K2.36GB
japanese-stablelm-base-beta-7b.IQ3XS.ggufIQ3XS2.6GB
japanese-stablelm-base-beta-7b.IQ3S.ggufIQ3S2.75GB
japanese-stablelm-base-beta-7b.Q3KS.ggufQ3K_S2.75GB
japanese-stablelm-base-beta-7b.IQ3M.ggufIQ3M2.9GB
japanese-stablelm-base-beta-7b.Q3K.ggufQ3K3.07GB
japanese-stablelm-base-beta-7b.Q3KM.ggufQ3K_M3.07GB
japanese-stablelm-base-beta-7b.Q3KL.ggufQ3K_L3.35GB
japanese-stablelm-base-beta-7b.IQ4XS.ggufIQ4XS3.4GB
japanese-stablelm-base-beta-7b.Q40.ggufQ403.56GB
japanese-stablelm-base-beta-7b.IQ4NL.ggufIQ4NL3.58GB
japanese-stablelm-base-beta-7b.Q4KS.ggufQ4K_S3.59GB
japanese-stablelm-base-beta-7b.Q4K.ggufQ4K3.8GB
japanese-stablelm-base-beta-7b.Q4KM.ggufQ4K_M3.8GB
japanese-stablelm-base-beta-7b.Q41.ggufQ413.95GB
japanese-stablelm-base-beta-7b.Q50.ggufQ504.33GB
japanese-stablelm-base-beta-7b.Q5KS.ggufQ5K_S4.33GB
japanese-stablelm-base-beta-7b.Q5K.ggufQ5K4.45GB
japanese-stablelm-base-beta-7b.Q5KM.ggufQ5K_M4.45GB
japanese-stablelm-base-beta-7b.Q51.ggufQ514.72GB
japanese-stablelm-base-beta-7b.Q6K.ggufQ6K5.15GB
japanese-stablelm-base-beta-7b.Q80.ggufQ806.67GB

Original model description:



language:
  • ja

tags:
  • japanese-stablelm
  • causal-lm

pipeline_tag: text-generation
datasets:
  • wikipedia
  • mc4
  • cc100
  • oscar-corpus/OSCAR-2301
  • oscar-corpus/OSCAR-2201
  • cerebras/SlimPajama-627B

license:
  • llama2

extragatedfields:
Name: text
Email: text
Country: text
Organization or Affiliation: text
I allow Stability AI to contact me about information related to its models and research: checkbox

Japanese-StableLM-Base-Beta-7B

!A cute robot wearing a kimono writes calligraphy with one single brush

A cute robot wearing a kimono writes calligraphy with one single brush β€” Stable Diffusion XL

Model Description

japanese-stablelm-base-beta-7b is a 7B-parameter decoder-only language model based on Llama-2-7b that has been fine-tuned on a diverse collection of Japanese data, with the intent of maximizing downstream performance on Japanese language tasks.

For an instruction-following model, check Japanese-StableLM-Instruct-Beta-7B. The base and instruct models are also available in larger 70b sizes. For a model that has faster inference times, see Japanese-StableLM-Base-JAVocab-Beta-7B, or the instruction-following version.

Usage

First install additional dependencies in requirements.txt:

pip install -r requirements.txt

Then start generating text with japanese-stablelm-base-beta-7b by using the following code snippet:

import torch
from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "stabilityai/japanese-stablelm-base-beta-7b"
tokenizer = AutoTokenizer.frompretrained(modelname)

The next line may need to be modified depending on the environment

model = AutoModelForCausalLM.frompretrained(modelname, torchdtype=torch.float16, lowcpumemusage=True, device_map="auto")

prompt = """
AI γ§η§‘ε­¦η ”η©Άγ‚’εŠ ι€Ÿγ™γ‚‹γ«γ―γ€
""".strip()

input_ids = tokenizer.encode(
prompt,
addspecialtokens=True,
return_tensors="pt"
)

this is for reproducibility.

feel free to change to get different result

seed = 23 torch.manual_seed(seed)

tokens = model.generate(
input_ids.to(device=model.device),
maxnewtokens=128,
temperature=0.99,
top_p=0.95,
do_sample=True,
)

out = tokenizer.decode(tokens[0], skipspecialtokens=True)
print(out)

We suggest playing with different generation config (topp, repetitionpenalty etc) to find the best setup for your tasks. For example, use higher temperature for roleplay task, lower temperature for reasoning.

Model Details

  • Model type: japanese-stablelm-base-beta-7b model is an auto-regressive language model based on the Llama2 transformer architecture.
  • Language(s): Japanese
  • License: Llama2 Community License.
  • Contact: For questions and comments about the model, please join Stable Community Japan. For future announcements / information about Stability AI models, research, and events, please follow https://twitter.com/StabilityAI_JP.

Training Dataset

Roughly 100B tokens from a mixture of the following corpora were used for continued pre-training.

Use and Limitations

Intended Use

The model is intended to be used by all individuals as a foundation for application-specific fine-tuning without strict limitations on commercial use.

Limitations and bias

The pre-training dataset may have contained offensive or inappropriate content even after applying data cleansing filters which can be reflected in the model generated text. We recommend users exercise reasonable caution when using these models in production systems. Do not use the model for any applications that may cause harm or distress to individuals or groups.

Authors

This model was developed by the Research & Development team at Stability AI Japan, and the development was co-led by Takuya Akiba and Meng Lee. The members of the team are as follows:

Acknowledgements

We thank Meta Research for releasing Llama 2 under an open license for others to build on.

We are grateful for the contributions of the EleutherAI Polyglot-JA team in helping us to collect a large amount of pre-training data in Japanese. Polyglot-JA members includes Hyunwoong Ko (Project Lead), Fujiki Nakamura (originally started this project when he commited to the Polyglot team), Yunho Mo, Minji Jung, KeunSeok Im, and Su-Kyeong Jang.

We are also appreciative of AI Novelist/Sta (Bit192, Inc.) and the numerous contributors from Stable Community Japan for assisting us in gathering a large amount of high-quality Japanese textual data for model training.

How to cite

@misc{JapaneseStableLMBaseBeta7B, 
      url={https://huggingface.co/stabilityai/japanese-stablelm-base-beta-7b}, 
      title={Japanese StableLM Base Beta 7B}, 
      author={Lee, Meng and Nakamura, Fujiki and Shing, Makoto and McCann, Paul and Akiba, Takuya and Orii, Naoki}
}

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
japanese-stablelm-base-beta-7b.IQ3_M.gguf
LFS Q3
2.9 GB Download
japanese-stablelm-base-beta-7b.IQ3_S.gguf
LFS Q3
2.75 GB Download
japanese-stablelm-base-beta-7b.IQ3_XS.gguf
LFS Q3
2.6 GB Download
japanese-stablelm-base-beta-7b.IQ4_NL.gguf
LFS Q4
3.58 GB Download
japanese-stablelm-base-beta-7b.IQ4_XS.gguf
LFS Q4
3.4 GB Download
japanese-stablelm-base-beta-7b.Q2_K.gguf
LFS Q2
2.36 GB Download
japanese-stablelm-base-beta-7b.Q3_K.gguf
LFS Q3
3.07 GB Download
japanese-stablelm-base-beta-7b.Q3_K_L.gguf
LFS Q3
3.35 GB Download
japanese-stablelm-base-beta-7b.Q3_K_M.gguf
LFS Q3
3.07 GB Download
japanese-stablelm-base-beta-7b.Q3_K_S.gguf
LFS Q3
2.75 GB Download
japanese-stablelm-base-beta-7b.Q4_0.gguf
Recommended LFS Q4
3.56 GB Download
japanese-stablelm-base-beta-7b.Q4_1.gguf
LFS Q4
3.95 GB Download
japanese-stablelm-base-beta-7b.Q4_K.gguf
LFS Q4
3.8 GB Download
japanese-stablelm-base-beta-7b.Q4_K_M.gguf
LFS Q4
3.8 GB Download
japanese-stablelm-base-beta-7b.Q4_K_S.gguf
LFS Q4
3.59 GB Download
japanese-stablelm-base-beta-7b.Q5_0.gguf
LFS Q5
4.33 GB Download
japanese-stablelm-base-beta-7b.Q5_1.gguf
LFS Q5
4.72 GB Download
japanese-stablelm-base-beta-7b.Q5_K.gguf
LFS Q5
4.45 GB Download
japanese-stablelm-base-beta-7b.Q5_K_M.gguf
LFS Q5
4.45 GB Download
japanese-stablelm-base-beta-7b.Q5_K_S.gguf
LFS Q5
4.33 GB Download
japanese-stablelm-base-beta-7b.Q6_K.gguf
LFS Q6
5.15 GB Download
japanese-stablelm-base-beta-7b.Q8_0.gguf
LFS Q8
6.67 GB Download