kirankunapuli commited on
Commit
01e53a2
1 Parent(s): 9e4c38d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +60 -1
README.md CHANGED
@@ -14,7 +14,6 @@ datasets:
14
  - yahma/alpaca-cleaned
15
  - ravithejads/samvaad-hi-filtered
16
  - HydraIndicLM/hindi_alpaca_dolly_67k
17
- pipeline_tag: text-generation
18
  ---
19
 
20
  # TinyLlama-1.1B-Hinglish-LORA-v1.0 model
@@ -22,6 +21,66 @@ pipeline_tag: text-generation
22
  - **Developed by:** [Kiran Kunapuli](https://www.linkedin.com/in/kirankunapuli/)
23
  - **License:** apache-2.0
24
  - **Finetuned from model :** unsloth/gemma-2b-bnb-4bit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
 
26
  This gemma model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
27
 
 
14
  - yahma/alpaca-cleaned
15
  - ravithejads/samvaad-hi-filtered
16
  - HydraIndicLM/hindi_alpaca_dolly_67k
 
17
  ---
18
 
19
  # TinyLlama-1.1B-Hinglish-LORA-v1.0 model
 
21
  - **Developed by:** [Kiran Kunapuli](https://www.linkedin.com/in/kirankunapuli/)
22
  - **License:** apache-2.0
23
  - **Finetuned from model :** unsloth/gemma-2b-bnb-4bit
24
+ - - **Model config:**
25
+ ```python
26
+ model = FastLanguageModel.get_peft_model(
27
+ model,
28
+ r = 16,
29
+ target_modules = ["q_proj", "k_proj", "v_proj", "o_proj",
30
+ "gate_proj", "up_proj", "down_proj",],
31
+ lora_alpha = 32,
32
+ lora_dropout = 0,
33
+ bias = "none",
34
+ use_gradient_checkpointing = True,
35
+ random_state = 42,
36
+ use_rslora = True,
37
+ loftq_config = None,
38
+ )
39
+ ```
40
+ - **Training parameters:**
41
+ ```python
42
+ trainer = SFTTrainer(
43
+ model = model,
44
+ tokenizer = tokenizer,
45
+ train_dataset = dataset,
46
+ dataset_text_field = "text",
47
+ max_seq_length = max_seq_length,
48
+ dataset_num_proc = 2,
49
+ packing = True,
50
+ args = TrainingArguments(
51
+ per_device_train_batch_size = 2,
52
+ gradient_accumulation_steps = 4,
53
+ warmup_steps = 5,
54
+ max_steps = 120,
55
+ learning_rate = 2e-4,
56
+ fp16 = not torch.cuda.is_bf16_supported(),
57
+ bf16 = torch.cuda.is_bf16_supported(),
58
+ logging_steps = 1,
59
+ optim = "adamw_8bit",
60
+ weight_decay = 0.01,
61
+ lr_scheduler_type = "linear",
62
+ seed = 42,
63
+ output_dir = "outputs",
64
+ report_to = "wandb",
65
+ ),
66
+ )
67
+ ```
68
+ - **Training details:**
69
+ ```
70
+ ==((====))== Unsloth - 2x faster free finetuning | Num GPUs = 1
71
+ \\ /| Num examples = 14,343 | Num Epochs = 1
72
+ O^O/ \_/ \ Batch size per device = 2 | Gradient Accumulation steps = 4
73
+ \ / Total batch size = 8 | Total steps = 120
74
+ "-____-" Number of trainable parameters = 19,611,648
75
+
76
+ GPU = Tesla T4. Max memory = 14.748 GB.
77
+ 2118.7553 seconds used for training.
78
+ 35.31 minutes used for training.
79
+ Peak reserved memory = 9.172 GB.
80
+ Peak reserved memory for training = 6.758 GB.
81
+ Peak reserved memory % of max memory = 62.191 %.
82
+ Peak reserved memory for training % of max memory = 45.823 %.
83
+ ```
84
 
85
  This gemma model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
86