corgicoder commited on
Commit
9686d28
1 Parent(s): 236f3d0

Delete codegen-350M-mono-mbpp

Browse files
codegen-350M-mono-mbpp/added_tokens.json DELETED
@@ -1,42 +0,0 @@
1
- {
2
- "\t\t": 50294,
3
- "\t\t\t": 50293,
4
- "\t\t\t\t": 50292,
5
- "\t\t\t\t\t": 50291,
6
- "\t\t\t\t\t\t": 50290,
7
- "\t\t\t\t\t\t\t": 50289,
8
- "\t\t\t\t\t\t\t\t": 50288,
9
- "\t\t\t\t\t\t\t\t\t": 50287,
10
- " ": 50286,
11
- " ": 50285,
12
- " ": 50284,
13
- " ": 50283,
14
- " ": 50282,
15
- " ": 50281,
16
- " ": 50280,
17
- " ": 50279,
18
- " ": 50278,
19
- " ": 50277,
20
- " ": 50276,
21
- " ": 50275,
22
- " ": 50274,
23
- " ": 50273,
24
- " ": 50272,
25
- " ": 50271,
26
- " ": 50270,
27
- " ": 50269,
28
- " ": 50268,
29
- " ": 50267,
30
- " ": 50266,
31
- " ": 50265,
32
- " ": 50264,
33
- " ": 50263,
34
- " ": 50262,
35
- " ": 50261,
36
- " ": 50260,
37
- " ": 50259,
38
- " ": 50258,
39
- " ": 50257,
40
- "<|beginoftext|>": 50295,
41
- "<|paddingoftext|>": 50296
42
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
codegen-350M-mono-mbpp/config.json DELETED
@@ -1,42 +0,0 @@
1
- {
2
- "_name_or_path": "Salesforce/codegen-350M-mono",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "CodeGenForCausalLM"
6
- ],
7
- "attn_pdrop": 0.0,
8
- "bos_token_id": 1,
9
- "embd_pdrop": 0.0,
10
- "eos_token_id": 50256,
11
- "gradient_checkpointing": false,
12
- "initializer_range": 0.02,
13
- "layer_norm_epsilon": 1e-05,
14
- "model_type": "codegen",
15
- "n_ctx": 2048,
16
- "n_embd": 1024,
17
- "n_head": 16,
18
- "n_inner": null,
19
- "n_layer": 20,
20
- "n_positions": 2048,
21
- "resid_pdrop": 0.0,
22
- "rotary_dim": 32,
23
- "scale_attn_weights": true,
24
- "summary_activation": null,
25
- "summary_first_dropout": 0.1,
26
- "summary_proj_to_labels": true,
27
- "summary_type": "cls_index",
28
- "summary_use_proj": true,
29
- "task_specific_params": {
30
- "text-generation": {
31
- "do_sample": true,
32
- "max_length": 50,
33
- "temperature": 1.0
34
- }
35
- },
36
- "tie_word_embeddings": false,
37
- "tokenizer_class": "GPT2Tokenizer",
38
- "torch_dtype": "float32",
39
- "transformers_version": "4.29.2",
40
- "use_cache": true,
41
- "vocab_size": 51200
42
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
codegen-350M-mono-mbpp/generation_config.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "bos_token_id": 1,
4
- "eos_token_id": 50256,
5
- "transformers_version": "4.29.2"
6
- }
 
 
 
 
 
 
 
codegen-350M-mono-mbpp/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
codegen-350M-mono-mbpp/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f7d4e128ba455d8bc223e882380e1d13bdb6f2353de5180388cf3607be07f377
3
- size 1510796785
 
 
 
 
codegen-350M-mono-mbpp/special_tokens_map.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "bos_token": "<|beginoftext|>",
3
- "eos_token": "<|endoftext|>",
4
- "pad_token": "<|paddingoftext|>",
5
- "unk_token": "<|endoftext|>"
6
- }
 
 
 
 
 
 
 
codegen-350M-mono-mbpp/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
codegen-350M-mono-mbpp/tokenizer_config.json DELETED
@@ -1,9 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "bos_token": "<|endoftext|>",
4
- "clean_up_tokenization_spaces": true,
5
- "eos_token": "<|endoftext|>",
6
- "model_max_length": 2048,
7
- "tokenizer_class": "CodeGenTokenizer",
8
- "unk_token": "<|endoftext|>"
9
- }
 
 
 
 
 
 
 
 
 
 
codegen-350M-mono-mbpp/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9c025f02861f65655bff2a36457b9195c56c225d5d89074e83bacfd66afecd9
3
- size 3899
 
 
 
 
codegen-350M-mono-mbpp/vocab.json DELETED
The diff for this file is too large to render. See raw diff