File size: 489 Bytes
9e60901
1
{"sae": {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "signed": false, "multi_topk": false}, "batch_size": 16, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["gpt_neox.layers.0", "gpt_neox.layers.1"], "layers": [0, 1], "layer_stride": 1, "distribute_modules": false, "save_every": 1000, "log_to_wandb": false, "run_name": null, "wandb_log_frequency": 1}