đ Model Description
license: llama2 language:
- hi
This repository is the first model in the OpenHathi series of models that will be released by Sarvam AI. This is a 7B parameter, based on Llama2, trained on Hindi, English, and Hinglish. More details about the model, its training procedure, and evaluations can be found here.
Note: this is a base model and not meant to be used as is. We recommend first finetuning it on task(s) you are interested in.
# Usage
import torch
from transformers import LlamaTokenizer, LlamaForCausalLM
tokenizer = LlamaTokenizer.from_pretrained('sarvamai/OpenHathi-7B-Hi-v0.1-Base')
model = LlamaForCausalLM.frompretrained('sarvamai/OpenHathi-7B-Hi-v0.1-Base', torchdtype=torch.bfloat16)
prompt = "ā¤ŽāĨ⤠ā¤ā¤ ā¤
ā¤āĨā¤ā¤ž ā¤šā¤žā¤ĨāĨ ā¤šāĨā¤"
inputs = tokenizer(prompt, return_tensors="pt")
Generate
generateids = model.generate(inputs.inputids, max_length=30)
tokenizer.batchdecode(generateids, skipspecialtokens=True, cleanuptokenization_spaces=False)[0]