mtasic85 commited on
Commit
d4d6807
1 Parent(s): cb4823d
evaluate-0/config.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:94e73e086a5ed14149cee99a1aa3e2563ec7ab536c1653ff332999afa3520694
3
- size 546
 
 
 
 
evaluate-0/model_config.yaml DELETED
@@ -1,33 +0,0 @@
1
- attention_logit_softcapping: null
2
- attention_scores_scalar: null
3
- bias: false
4
- block_size: 32768
5
- final_logit_softcapping: null
6
- gelu_approximate: none
7
- head_size: 26
8
- hf_config: {}
9
- intermediate_size: 1092
10
- lm_head_bias: false
11
- mlp_class_name: LLaMAMLP
12
- n_embd: 312
13
- n_expert: 0
14
- n_expert_per_token: 0
15
- n_head: 12
16
- n_layer: 10
17
- n_query_groups: 4
18
- name: ''
19
- norm_class_name: RMSNorm
20
- norm_eps: 1.0e-05
21
- padded_vocab_size: 32768
22
- padding_multiple: 512
23
- parallel_residual: false
24
- post_attention_norm: false
25
- post_mlp_norm: false
26
- rope_base: 500000
27
- rope_condense_ratio: 1
28
- rotary_percentage: 1.0
29
- scale_embeddings: false
30
- shared_attention_norm: false
31
- sliding_window_layer_placing: null
32
- sliding_window_size: null
33
- vocab_size: 32768
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
evaluate-0/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:85703bfe7255763468bcc1a1d2ddb7e13c665cf78f12716ce9177915cec173ef
3
- size 66570501
 
 
 
 
evaluate-0/tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b496a30dc268bcb8adfd551f693e68e9eadd06b81cab385c088a61e7663649c
3
- size 1368561
 
 
 
 
evaluate-0/tokenizer_config.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6333d68c3280be6081b795cc160fd5872707562021f9889b2e2bd3ae508fa62
3
- size 23043
 
 
 
 
out/pretrain/final/evaluate/config.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:94e73e086a5ed14149cee99a1aa3e2563ec7ab536c1653ff332999afa3520694
3
- size 546
 
 
 
 
out/pretrain/final/evaluate/model_config.yaml DELETED
@@ -1,33 +0,0 @@
1
- attention_logit_softcapping: null
2
- attention_scores_scalar: null
3
- bias: false
4
- block_size: 32768
5
- final_logit_softcapping: null
6
- gelu_approximate: none
7
- head_size: 26
8
- hf_config: {}
9
- intermediate_size: 1092
10
- lm_head_bias: false
11
- mlp_class_name: LLaMAMLP
12
- n_embd: 312
13
- n_expert: 0
14
- n_expert_per_token: 0
15
- n_head: 12
16
- n_layer: 10
17
- n_query_groups: 4
18
- name: ''
19
- norm_class_name: RMSNorm
20
- norm_eps: 1.0e-05
21
- padded_vocab_size: 32768
22
- padding_multiple: 512
23
- parallel_residual: false
24
- post_attention_norm: false
25
- post_mlp_norm: false
26
- rope_base: 500000
27
- rope_condense_ratio: 1
28
- rotary_percentage: 1.0
29
- scale_embeddings: false
30
- shared_attention_norm: false
31
- sliding_window_layer_placing: null
32
- sliding_window_size: null
33
- vocab_size: 32768
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
out/pretrain/final/evaluate/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:85703bfe7255763468bcc1a1d2ddb7e13c665cf78f12716ce9177915cec173ef
3
- size 66570501
 
 
 
 
out/pretrain/final/evaluate/results.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fdfcb715382c318246f445c75353b1cc336cfe0e32db3a2d429c4ece6a94f22f
3
- size 139346753
 
 
 
 
out/pretrain/final/evaluate/tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b496a30dc268bcb8adfd551f693e68e9eadd06b81cab385c088a61e7663649c
3
- size 1368561
 
 
 
 
out/pretrain/final/evaluate/tokenizer_config.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6333d68c3280be6081b795cc160fd5872707562021f9889b2e2bd3ae508fa62
3
- size 23043