miguelvictor commited on
Commit
e6ef522
1 Parent(s): 61d23ae

added tokenizer config

Browse files
Files changed (4) hide show
  1. config.json +2 -8
  2. merges.txt +0 -0
  3. tokenizer.json +0 -0
  4. vocab.json +0 -0
config.json CHANGED
@@ -1,21 +1,16 @@
1
  {
2
- "_name_or_path": "gpt2-medium",
3
  "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPT2LMHeadModel"
6
- ],
7
  "attn_pdrop": 0.1,
8
  "bos_token_id": 50256,
9
  "embd_pdrop": 0.1,
10
  "eos_token_id": 50256,
11
- "gradient_checkpointing": true,
12
  "initializer_range": 0.02,
13
- "layer_norm_epsilon": 1e-05,
14
  "model_type": "gpt2",
15
  "n_ctx": 1024,
16
  "n_embd": 1024,
17
  "n_head": 16,
18
- "n_inner": null,
19
  "n_layer": 24,
20
  "n_positions": 1024,
21
  "n_special": 0,
@@ -33,6 +28,5 @@
33
  }
34
  },
35
  "transformers_version": "4.5.1",
36
- "use_cache": false,
37
  "vocab_size": 50257
38
  }
 
1
  {
 
2
  "activation_function": "gelu_new",
3
+ "architectures": ["GPT2LMHeadModel"],
 
 
4
  "attn_pdrop": 0.1,
5
  "bos_token_id": 50256,
6
  "embd_pdrop": 0.1,
7
  "eos_token_id": 50256,
 
8
  "initializer_range": 0.02,
9
+ "layer_norm_epsilon": 1e-5,
10
  "model_type": "gpt2",
11
  "n_ctx": 1024,
12
  "n_embd": 1024,
13
  "n_head": 16,
 
14
  "n_layer": 24,
15
  "n_positions": 1024,
16
  "n_special": 0,
 
28
  }
29
  },
30
  "transformers_version": "4.5.1",
 
31
  "vocab_size": 50257
32
  }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.json ADDED
The diff for this file is too large to render. See raw diff