ctigges commited on
Commit
03382b1
·
verified ·
1 Parent(s): e62cc2d

Upload folder using huggingface_hub

Browse files
0-mlp-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.0.hook_mlp_out", "hook_layer": 0, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
0-mlp-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fe81e861b3440573bedd6190017e8d212fc7dae49e2d25422602d9816ed4ea0
3
+ size 134351168
0-mlp-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:738cb2b3c4d1508e318e98c5fb810928acd53ae26439bd19573abc6781ea2a78
3
+ size 131152
1-mlp-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.1.hook_mlp_out", "hook_layer": 1, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
1-mlp-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbeda2b0da0b9094c2a9ffd44f2399e32c3ab2512f374c7afc68cbf49b4c4510
3
+ size 134351168
1-mlp-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d46b572dac5fb0d98bc1df38aad90af9dff22542d7839584b4afe8f17b435ffa
3
+ size 131152
2-mlp-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.2.hook_mlp_out", "hook_layer": 2, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
2-mlp-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ed59a82df0b25a95fc74b8b7bb18654e19f4dfda6bdf3a5f9d545aca87f27eb
3
+ size 134351168
2-mlp-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44337fd989163d5e270496f5b64ef755a3a0a5286bb3e93a7c7a05dbdb18d8cb
3
+ size 131152
3-mlp-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.3.hook_mlp_out", "hook_layer": 3, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
3-mlp-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f15c2d3efecdbd7f70fc54a2daec531cfe9151de0add623e0c6255f85dbf5219
3
+ size 134351168
3-mlp-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0681673fbad1aa499feeaa7c40d87c6ae3f0ae972f9c451c5db20362ccfdc9d0
3
+ size 131152
4-mlp-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.4.hook_mlp_out", "hook_layer": 4, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
4-mlp-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf8276d25a6d6626f2b12dbdc27dbd1490fda5d7fced4e232ab24cdc36b24e18
3
+ size 134351168
4-mlp-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ece271c263ab22e613344b529c8a0f4478e5c94d155c5b5af8b4020508c1ada2
3
+ size 131152
5-mlp-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.5.hook_mlp_out", "hook_layer": 5, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
5-mlp-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41cec5e820e4dfe8f538cac5e9b4e90d622240feee3b9715c3d013f1806f5753
3
+ size 134351168
5-mlp-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddfef726cea983d9e41024dcb1513dbe2b0b5bde0efded90d143277438320bd5
3
+ size 131152