# Usage ```python from unsloth import FastLanguageModel import torch import json HUGGINGFACE_TOKEN = "your token here" model, tokenizer = FastLanguageModel.from_pretrained( model_name = "solaoi/gemma2-9b-magpie-sft-v1-v11", max_seq_length = 1024, dtype = None, load_in_4bit = True, token = HUGGINGFACE_TOKEN, ) FastLanguageModel.for_inference(model) system_prompt = "あなたは知識豊富な人工知能アシスタントです。ユーザーは日本語であなたに質問をし、あなたは自身の知識に基づいて日本語で誠実に回答します。" prompt = tokenizer.apply_chat_template([ {"role": "system", "content": system_prompt}, {"role": "user", "content": input} ], tokenize=False, add_generation_prompt=True) inputs = tokenizer(prompt, return_tensors="pt").to(model.device) outputs = model.generate( **inputs, max_new_tokens = 304, use_cache = True, temperature=0.7, top_k=50, top_p=0.85, repetition_penalty = 1.15, eos_token_id = tokenizer.eos_token_id, pad_token_id = tokenizer.pad_token_id ) answer = tokenizer.decode(outputs[0], skip_special_tokens=True)[len(prompt):].strip() ```