Edit model card

Llama-2-7b-chat-finetune

This model is a fine-tuned version of NousResearch/Llama-2-7b-chat-hf using the mlabonne/guanaco-llama2-1k dataset. It has been fine-tuned using LoRA (Low-Rank Adaptation) with the PEFT library and the SFTTrainer from TRL.

Model Description

This model is intended for text generation and instruction following tasks. It has been fine-tuned on a dataset of 1,000 instruction-following examples.

Intended Uses & Limitations

This model can be used for a variety of text generation tasks, including:

  • Generating creative text formats, like poems, code, scripts, musical pieces, email, letters, etc.
  • Answering your questions in an informative way, even if they are open ended, challenging, or strange.
  • Following your instructions and completing your requests thoughtfully.

Limitations:

  • The model may generate biased or harmful content.
  • The model may not be able to follow all instructions perfectly.
  • The model may not be able to generate text that is factually accurate.

Training and Fine-tuning

This model was fine-tuned using the following parameters:

  • LoRA attention dimension (lora_r): 64
  • Alpha parameter for LoRA scaling (lora_alpha): 16
  • Dropout probability for LoRA layers (lora_dropout): 0.1
  • 4-bit precision base model loading (use_4bit): True
  • Number of training epochs (num_train_epochs): 1
  • Batch size per GPU for training (per_device_train_batch_size): 4
  • Learning rate (learning_rate): 2e-4

How to Use

You can use this model with the following code:

from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline

model_name = "chaitanya42/Llama-2-7b-chat-finetune"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

prompt = "What is a large language model?"
pipe = pipeline(task="text-generation", model=model, tokenizer=tokenizer, max_length=200)
result = pipe(f"[INST] {prompt} [/INST]")
print(result[0]['generated_text'])
Downloads last month
0
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.