{"base_model_dir": "/llm-downloader-destination/base/fireworks/llama-v3p3-70b-instruct/hf/", "output_model_dir": "gs://fireworks-artifacts-andreagonzalesmexico3-4fec3c-31c5b0/sftj-51c78962/b5777b/alheli-docs-experimental-70b/checkpoint", "checkpoint_dir": "/mnt/persistent/checkpoints", "max_checkpoints_to_keep": 1, "train": true, "learning_rate": 0.0001, "learning_rate_warmup_steps": 0, "grad_accum_steps": 1, "epochs": 3, "early_stop": true, "seed": 42, "dataset_dir": "/mnt/staging/dataset", "eval_dataset_dir": "/mnt/staging/evaluation-dataset", "train_limit": null, "max_context_len": 32000, "batch_size": null, "min_evals_per_epoch": 4, "base_model_weight_precision": "nf4", "status_file": "gs://fireworks-fine-tuning-job-status/sftj-andreagonzalesmexico3-4fec3c-51c78962", "billing_file": "gs://fireworks-fine-tuning-metadata/sftj-andreagonzalesmexico3-4fec3c-51c78962/billing", "wandb": false, "wandb_entity": null, "wandb_api_key": null, "wandb_project": null, "wandb_run_id": null, "profile": false, "peft_addon_dir": null, "lora_rank": 8, "template_kind": "conversation", "template": null, "eval_train_ratio": 0.02}