metadata
license: apache-2.0
base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
datasets:
- cerebras/SlimPajama-627B
- bigcode/starcoderdata
- HuggingFaceH4/ultrachat_200k
- HuggingFaceH4/ultrafeedback_binarized
language:
- en
widget:
- example_title: Fibonacci (Python)
messages:
- role: system
content: You are a chatbot who can help code!
- role: user
content: >-
Write me a function to calculate the first 10 digits of the fibonacci
sequence in Python and print it out to the CLI.
tags:
- transformers
- safetensors
- finetuned
- 4-bit
- AWQ
- text-generation
- text-generation-inference
- autotrain_compatible
- endpoints_compatible
- chatml
model_creator: TinyLlama
model_name: TinyLlama-1.1B-Chat-v1.0
inference: false
pipeline_tag: text-generation
quantized_by: Suparious
TinyLlama/TinyLlama-1.1B-Chat-v1.0 AWQ
- Model creator: TinyLlama
- Original model: TinyLlama-1.1B-Chat-v1.0
Model Summary
This is the chat model finetuned on top of TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T. We follow HF's Zephyr's training recipe. The model was " initially fine-tuned on a variant of the UltraChat
dataset, which contains a diverse range of synthetic dialogues generated by ChatGPT.
We then further aligned the model with 🤗 TRL's DPOTrainer
on the openbmb/UltraFeedback dataset, which contain 64k prompts and model completions that are ranked by GPT-4."