Epic3123 commited on
Commit
9e60901
1 Parent(s): 496ee0a

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sae": {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "signed": false, "multi_topk": false}, "batch_size": 16, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["gpt_neox.layers.0", "gpt_neox.layers.1"], "layers": [0, 1], "layer_stride": 1, "distribute_modules": false, "save_every": 1000, "log_to_wandb": false, "run_name": null, "wandb_log_frequency": 1}
gpt_neox.layers.0/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "signed": false, "multi_topk": false, "d_in": 768}
gpt_neox.layers.0/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81360d569933c436403a308e65fcc5282a5de39d16a7631d52514ccefca06234
3
+ size 151096648
gpt_neox.layers.1/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "signed": false, "multi_topk": false, "d_in": 768}
gpt_neox.layers.1/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb76ca23151a14013631d287695f6fffbd2d44ff28d1d314368711bed4b39c35
3
+ size 151096648
lr_scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f77592197726d86b927a990e9bf7ae447cd017a26dab0d9b5ff432de7d701bf
3
+ size 1076
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:011af16b7fba524aa1f49d8ff824962288fec861a180cf8fbf7a158ddb32a892
3
+ size 604392710
state.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:497cdc632be5523bed2a0f4d8c6e49b05cbdb0facc07c131f7a347c0fb7dbace
3
+ size 394636