YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co./docs/hub/model-cards#model-card-metadata)
Quantization made by Richard Erkhov.
OPT-1.3B-SFT-DSChatLoRA - AWQ
- Model creator: https://huggingface.co./kaitchup/
- Original model: https://huggingface.co./kaitchup/OPT-1.3B-SFT-DSChatLoRA/
Original model description:
license: cc-by-nc-sa-4.0 datasets: - Dahoas/rm-static - Dahoas/synthetic-instruct-gptj-pairwise - Anthropic/hh-rlhf language: - en
Model Card for Model ID
This a model is a fine-tuned with SFT using DeepSpeed Chat. It is based on OPT-1.3M.B
Model Details
Model Description
- Developed by: The Kaitchup
- Model type: Causal
- Language(s) (NLP): English
- License: cc-by-nc-sa-4.0
- Finetuned from model: facebook/opt-1.3b
Model Sources
The model has been trained with the procedure described in this article:
Train Instruct LLMs On Your GPU with DeepSpeed Chat — Step #1: Supervised Fine-tuning
- Downloads last month
- 4