kcsmta's picture
Upload config.yaml with huggingface_hub
bb7e2be verified
exp_name: 'vi-en-fix-v1'
# Training dataset (from Huggingface)
# data_source: "MedCat/MedCAT-SFT-v1"
data_source: "MedCat/MedCAT-SFT-v1.1"
# The base model (from HuggingFace model hub)
# model_name: "Qwen/Qwen2.5-0.5B"
model_name: "MedCat/MedCAT-PT-Qwen2.5-0.5B-v1-stream-data-v1-checkpoint-600000"
# model_name: "MedCat/MedCAT-PT-Apollo-0.5B-v1-stream-data-v1-checkpoint-600000"
# Tokenizer
tokenizer_batch_size: 1_000
max_length: 512
# Checkpoints configuration
output_folder: "./checkpoints/MedCAT-SFT" # Where to save checkpoints during the training
save_total_limit: 2 # Limit on number of checkpoints to keep
save_strategy: "steps" # Saving strategy (either 'steps' or 'epoch')
save_steps: 500 # Save model every ... steps
# LoRA
r: 8 # Rank of the low-rank matrices
lora_alpha: 32 # LoRA alpha
lora_dropout: 0.1 # Dropout rate
bias: "none" # Whether to train biases ("none", "all", or "lora_only")
task_type: "CAUSAL_LM" # Task type: casual language modeling
# Logging configuration
logging_dir: "./logs/MedCAT-SFT" # Directory for logs + base_model + data_version
logging_steps: 100 # Frequency of logging
# Training configuration
per_device_train_batch_size: 4 # Training batch size
per_device_eval_batch_size: 4 # Evaluation batch size
num_train_epochs: 2 # Number of epochs
# max_steps: 500 # Total training steps (or use num_train_epochs instead)
eval_steps: 500 # Frequency of evaluation. Should equal to logging_steps (can be different, but should be equal)
evaluation_strategy: "steps" # Evaluation strategy (either 'steps' or 'epoch')
seed: 3407 # Random seed for reproducibility
gradient_accumulation_steps: 8
learning_rate: 0.00001