Model Card for Model ID

This Model is Finetuned for Document Question and Answering purpose Trained on the yahma/alpaca-cleaned(https://huggingface.co./TheBloke/zephyr-7B-beta-GPTQ) dataset.

Model Details

Training hyperparameters

The following hyperparameters were used during training:

-gradient_accumulation_steps=1,

-warmup_steps=5,

-max_steps=20,

-learning_rate=2e-4,

-fp16=not torch.cuda.is_bf16_supported(),

-bf16=torch.cuda.is_bf16_supported(),

-logging_steps=1,

-optim="adamw_8bit",

-weight_decay=0.01,

-lr_scheduler_type="linear",

-seed=3407,

  • Framework versions

  • PEFT 0.7.1

  • Transformers 4.36.0

  • Pytorch 2.0.0

  • Datasets 2.16.1

  • Tokenizers 0.15.0

Downloads last month
12
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Dataset used to train Feluda/Zephyr-7b-QnA