metadata
license: mit
language:
- en
tags:
- IFT
Introduction
This model originate from "LLaMA 2-7b" we trained only response part with the "Alpaca-GPT-4" dataset, utilizing LoRA (Low-Rank Adaptation) training. The weights from LoRA are merged into the model.
Details
Used Datasets
- vicgalle/alpaca-gpt4