c4-10k-tokenized-gpt2 / sae_lens.json
chanind's picture
Add upload script
75cd4b5 verified
raw
history blame contribute delete
321 Bytes
{
"sae_lens_version": "2.1.2",
"tokenizer_name": "gpt2",
"original_dataset": "NeelNanda/c4-10k",
"original_split": "train",
"original_data_files": null,
"context_size": 128,
"shuffled": true,
"seed": null,
"begin_batch_token": "bos",
"begin_sequence_token": null,
"sequence_separator_token": "eos"
}