Rashik24 commited on
Commit
059c7e4
·
1 Parent(s): 901e071

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -7
README.md CHANGED
@@ -22,23 +22,33 @@ To start using the Rashik24/Mistral-Instruct-Bangla model, you can use the follo
22
 
23
  ```Python
24
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
 
 
 
25
 
26
  def load_model(model_name):
27
- tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
28
- model = AutoModelForCausalLM.from_pretrained(model_name)
 
 
 
 
 
29
  return model, tokenizer
30
 
31
  def generate_text(prompt, model, tokenizer):
32
- inputs = tokenizer.encode(prompt, return_tensors='pt')
33
- outputs = model.generate(inputs, max_length=256, num_return_sequences=1)
34
- return tokenizer.decode(outputs[0], skip_special_tokens=True)
 
 
35
 
36
  #Load the model
37
  model_name = 'Rashik24/Mistral-Instruct-Bangla'
38
  model, tokenizer = load_model(model_name)
39
 
40
- #Example usage
41
- prompt = "আমি কেমন আছি?" # Bengali for "How am I?"
42
  generated_text = generate_text(prompt, model, tokenizer)
43
  print(generated_text)
44
  ```
 
22
 
23
  ```Python
24
  from transformers import AutoModelForCausalLM, AutoTokenizer
25
+ import torch
26
+ from peft import PeftModel, PeftConfig
27
+ from transformers import AutoModelForCausalLM
28
+
29
 
30
  def load_model(model_name):
31
+ tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2", trust_remote_code=True)
32
+ tokenizer.pad_token = tokenizer.eos_token
33
+ tokenizer.padding_side = "right"
34
+ config = PeftConfig.from_pretrained("Rashik24/Mistral-Instruct-Bangla")
35
+ model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2").to("cuda")
36
+ model = PeftModel.from_pretrained(model, "Rashik24/Mistral-Instruct-Bangla").to("cuda")
37
+ model.eval()
38
  return model, tokenizer
39
 
40
  def generate_text(prompt, model, tokenizer):
41
+ inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
42
+ with torch.no_grad():
43
+ generated_code = tokenizer.decode(model.generate(**inputs, max_new_tokens=1024, pad_token_id=2)[0], skip_special_tokens=True)
44
+ print(generated_code)
45
+ return generated_code
46
 
47
  #Load the model
48
  model_name = 'Rashik24/Mistral-Instruct-Bangla'
49
  model, tokenizer = load_model(model_name)
50
 
51
+ prompt = "একটি গ্রামের বর্ণনা করুন।"
 
52
  generated_text = generate_text(prompt, model, tokenizer)
53
  print(generated_text)
54
  ```