--- library_name: peft base_model: unsloth/mistral-7b-bnb-4bit datasets: - yahma/alpaca-cleaned --- # Model Card for Model ID A 4bit Mistral 7B model finetuned using unsloth on T4 GPU ## Model Details ### Model Description - **Finetuned from model:** unsloth/mistral-7b-bnb-4bit - **Repository:** https://github.com/unslothai/unsloth ## Training Details ### Training Data https://huggingface.co./datasets/yahma/alpaca-cleaned ### Training Procedure #### Preprocessing Alpaca prompt template is used: ``` alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. ### Instruction: {} ### Input: {} ### Response: {}""" ``` #### Training Hyperparameters ``` per_device_train_batch_size = 2, gradient_accumulation_steps = 4, warmup_steps = 5, max_steps = 60, learning_rate = 2e-4, fp16 = not torch.cuda.is_bf16_supported(), bf16 = torch.cuda.is_bf16_supported(), logging_steps = 1, optim = "adamw_8bit", weight_decay = 0.01, lr_scheduler_type = "linear", seed = 3407 ``` - **Hardware Type:** T4 GPU - **Cloud Provider:** Google Colab ### Framework versions - PEFT 0.7.1