Edit model card

Built with Axolotl

Base model:

https://huggingface.co./TinyLlama/tinyLlama-intermediate-checkpoints/tree/step-720k-token-1510B This fine tune was done on the "early" version of tinyllama-1.5T which suffers from a bug in dataset processing. See https://github.com/jzhang38/TinyLlama/issues/67. Through it suffers from the glitch, its performance seems not being damaged and still showing improvement(metrics needed)

Dataset:

Fine tuned on OpenOrca GPT4 subset for 1 epoch,Using CHATML format

Model License:

Apache 2.0, following the TinyLlama base model.

Quantisation:

GGUF format:https://huggingface.co./s3nh/jeff31415-TinyLlama-1.1B-1.5T-OpenOrca-Alpha-GGUF

Hardware and training details:

Hardware: 1*RTX A5000, ~16 hours to complete 1 epoch. GPU from autodl.com, cost around $3 for this finetuning. https://wandb.ai/jeff200402/TinyLlama-1.5T-alpha-Orca?workspace= for more details.

Downloads last month
81
Safetensors
Model size
1.1B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train jeff31415/TinyLlama-1.1B-1.5T-OpenOrca-Alpha