RichardErkhov's picture
uploaded readme
795fab9 verified

Quantization made by Richard Erkhov.

Github

Discord

Request more models

OPT-1.3B-SFT-DSChatLoRA - AWQ

Original model description:

license: cc-by-nc-sa-4.0 datasets: - Dahoas/rm-static - Dahoas/synthetic-instruct-gptj-pairwise - Anthropic/hh-rlhf language: - en

Model Card for Model ID

This a model is a fine-tuned with SFT using DeepSpeed Chat. It is based on OPT-1.3M.B

Model Details

Model Description

Model Sources

The model has been trained with the procedure described in this article:

Train Instruct LLMs On Your GPU with DeepSpeed Chat — Step #1: Supervised Fine-tuning