100 million
Browse files
README.md
CHANGED
@@ -36,7 +36,7 @@ Training logs are available [via wandb](https://wandb.ai/lewington/ViT-L-14-laio
|
|
36 |
- Training Dataset: [Laion-2b](https://huggingface.co/datasets/laion/laion2B-multi-joined-translated-to-en)
|
37 |
- SAE Architecture: topk with k=32
|
38 |
- Layer Location: always the residual stream
|
39 |
-
- Training Checkpoints: every ~
|
40 |
- Number of features per autoencoder: 65536 (expansion factor 16)
|
41 |
|
42 |
## Usage
|
|
|
36 |
- Training Dataset: [Laion-2b](https://huggingface.co/datasets/laion/laion2B-multi-joined-translated-to-en)
|
37 |
- SAE Architecture: topk with k=32
|
38 |
- Layer Location: always the residual stream
|
39 |
+
- Training Checkpoints: every ~100 million tokens
|
40 |
- Number of features per autoencoder: 65536 (expansion factor 16)
|
41 |
|
42 |
## Usage
|