Omaratef3221 commited on
Commit
946cc19
1 Parent(s): fe4c2ad

basic-jais-13b-arabic-text-summarizer

Browse files
README.md CHANGED
@@ -1,5 +1,6 @@
1
  ---
2
- base_model: Omaratef3221/qwen-0.5-rbf-mlp
 
3
  tags:
4
  - trl
5
  - sft
@@ -14,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # tmp_trainer
16
 
17
- This model is a fine-tuned version of [Omaratef3221/qwen-0.5-rbf-mlp](https://huggingface.co/Omaratef3221/qwen-0.5-rbf-mlp) on an unknown dataset.
18
 
19
  ## Model description
20
 
@@ -48,6 +49,6 @@ The following hyperparameters were used during training:
48
  ### Framework versions
49
 
50
  - Transformers 4.37.2
51
- - Pytorch 2.3.1+cu121
52
- - Datasets 2.12.0
53
  - Tokenizers 0.15.2
 
1
  ---
2
+ license: apache-2.0
3
+ base_model: inceptionai/jais-family-590m-chat
4
  tags:
5
  - trl
6
  - sft
 
15
 
16
  # tmp_trainer
17
 
18
+ This model is a fine-tuned version of [inceptionai/jais-family-590m-chat](https://huggingface.co/inceptionai/jais-family-590m-chat) on the None dataset.
19
 
20
  ## Model description
21
 
 
49
  ### Framework versions
50
 
51
  - Transformers 4.37.2
52
+ - Pytorch 2.2.0
53
+ - Datasets 2.21.0
54
  - Tokenizers 0.15.2
config.json CHANGED
@@ -1,31 +1,43 @@
1
  {
2
- "_name_or_path": "Omaratef3221/qwen-0.5-rbf-mlp",
 
 
3
  "architectures": [
4
- "CustomQwen2ForCausalLM"
5
  ],
6
- "attention_dropout": 0.0,
7
- "bos_token_id": 151643,
8
- "eos_token_id": 151645,
9
- "hidden_act": "silu",
10
- "hidden_size": 896,
 
 
 
 
 
 
 
11
  "initializer_range": 0.02,
12
- "intermediate_size": 4864,
13
- "max_position_embeddings": 32768,
14
- "max_window_layers": 24,
15
- "model_type": "qwen2",
16
- "num_attention_heads": 14,
17
- "num_hidden_layers": 24,
18
- "num_key_value_heads": 2,
19
- "rbf_norm_function": "euclidean",
20
- "rbf_num_kernels": 100,
21
- "rbf_radial_function": "gaussian",
22
- "rms_norm_eps": 1e-06,
23
- "rope_theta": 1000000.0,
24
- "sliding_window": 32768,
25
- "tie_word_embeddings": true,
 
 
 
 
26
  "torch_dtype": "float32",
27
  "transformers_version": "4.37.2",
28
  "use_cache": true,
29
- "use_sliding_window": false,
30
- "vocab_size": 151936
31
  }
 
1
  {
2
+ "_name_or_path": "inceptionai/jais-family-590m-chat",
3
+ "activation_function": "swiglu",
4
+ "alibi_scaling": null,
5
  "architectures": [
6
+ "JAISLMHeadModel"
7
  ],
8
+ "attn_pdrop": 0.0,
9
+ "auto_map": {
10
+ "AutoConfig": "inceptionai/jais-family-590m-chat--configuration_jais.JAISConfig",
11
+ "AutoModel": "inceptionai/jais-family-590m-chat--modeling_jais.JAISModel",
12
+ "AutoModelForCausalLM": "inceptionai/jais-family-590m-chat--modeling_jais.JAISLMHeadModel",
13
+ "AutoModelForQuestionAnswering": "inceptionai/jais-family-590m-chat--modeling_jais.JAISForQuestionAnswering",
14
+ "AutoModelForSequenceClassification": "inceptionai/jais-family-590m-chat--modeling_jais.JAISForSequenceClassification",
15
+ "AutoModelForTokenClassification": "inceptionai/jais-family-590m-chat--modeling_jais.JAISForTokenClassification"
16
+ },
17
+ "bos_token_id": 0,
18
+ "embd_pdrop": 0.0,
19
+ "eos_token_id": 0,
20
  "initializer_range": 0.02,
21
+ "layer_norm_epsilon": 1e-05,
22
+ "model_type": "jais",
23
+ "mup_embeddings_scale": 9.1705785388303,
24
+ "mup_output_alpha": 1.09518349815769,
25
+ "mup_scale_qk_dot_by_d": true,
26
+ "mup_width_scale": 0.16666666666666666,
27
+ "n_embd": 1536,
28
+ "n_head": 12,
29
+ "n_inner": 4096,
30
+ "n_layer": 18,
31
+ "n_positions": 2048,
32
+ "pad_token_id": 0,
33
+ "position_embedding_type": "alibi",
34
+ "reorder_and_upcast_attn": false,
35
+ "resid_pdrop": 0.0,
36
+ "rotary_dim": null,
37
+ "scale_attn_by_inverse_layer_idx": false,
38
+ "scale_attn_weights": true,
39
  "torch_dtype": "float32",
40
  "transformers_version": "4.37.2",
41
  "use_cache": true,
42
+ "vocab_size": 84992
 
43
  }
generation_config.json CHANGED
@@ -1,14 +1,7 @@
1
  {
2
- "bos_token_id": 151643,
3
- "do_sample": true,
4
- "eos_token_id": [
5
- 151645,
6
- 151643
7
- ],
8
- "pad_token_id": 151643,
9
- "repetition_penalty": 1.1,
10
- "temperature": 0.7,
11
- "top_k": 20,
12
- "top_p": 0.8,
13
  "transformers_version": "4.37.2"
14
  }
 
1
  {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "eos_token_id": 0,
5
+ "pad_token_id": 0,
 
 
 
 
 
 
 
6
  "transformers_version": "4.37.2"
7
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e234b977e2f81ba25c3c05748b240988b6c741553c00551570fadf3d146d78e1
3
- size 724351912
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31c5ec5092be3872df4db8d595bd0be059ad35eaeb1891e376b77892e704033f
3
+ size 2562246784
special_tokens_map.json CHANGED
@@ -1,10 +1,13 @@
1
  {
2
- "additional_special_tokens": [
3
- "<|im_start|>",
4
- "<|im_end|>"
5
- ],
 
 
 
6
  "eos_token": {
7
- "content": "<|im_end|>",
8
  "lstrip": false,
9
  "normalized": false,
10
  "rstrip": false,
@@ -16,5 +19,12 @@
16
  "normalized": false,
17
  "rstrip": false,
18
  "single_word": false
 
 
 
 
 
 
 
19
  }
20
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
  "eos_token": {
10
+ "content": "<|endoftext|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
 
19
  "normalized": false,
20
  "rstrip": false,
21
  "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<|endoftext|>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
  }
30
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,43 +1,20 @@
1
  {
2
- "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
- "151643": {
5
  "content": "<|endoftext|>",
6
  "lstrip": false,
7
  "normalized": false,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
11
- },
12
- "151644": {
13
- "content": "<|im_start|>",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false,
18
- "special": true
19
- },
20
- "151645": {
21
- "content": "<|im_end|>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false,
26
- "special": true
27
  }
28
  },
29
- "additional_special_tokens": [
30
- "<|im_start|>",
31
- "<|im_end|>"
32
- ],
33
- "bos_token": null,
34
- "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
- "clean_up_tokenization_spaces": false,
36
- "eos_token": "<|im_end|>",
37
- "errors": "replace",
38
- "model_max_length": 32768,
39
  "pad_token": "<|endoftext|>",
40
- "split_special_tokens": false,
41
- "tokenizer_class": "Qwen2Tokenizer",
42
- "unk_token": null
43
  }
 
1
  {
 
2
  "added_tokens_decoder": {
3
+ "0": {
4
  "content": "<|endoftext|>",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  }
11
  },
12
+ "bos_token": "<|endoftext|>",
13
+ "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = '### Instruction: ' + messages[0]['content'] + '\nComplete the conversation below between [|Human|] and [|AI|]:\n### Input:'%}{% else %}{% set loop_messages = messages %}{% set system_message = '### Instruction: Your name is \\'Jais\\', and you are named after Jebel Jais, the highest mountain in UAE. You were made by \\'Inception\\' in the UAE. You are a helpful, respectful, and honest assistant. Always answer as helpfully as possible, while being safe. Complete the conversation below between [|Human|] and [|AI|]:\n### Input:' %}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if loop.index0 == 0 and system_message != false %}{% set content = system_message %}{% else %}{% set content = message['content'] %}{% endif %}{% if message['role'] == 'user' %}{% if loop.index0 == 0 %}{{ content + ' [|Human|] ' + message['content'] }}{% else %}{{ '\n[|Human|] ' + content.strip() }}{% endif %}{% elif message['role'] == 'assistant' %}{{ '\n[|AI|] ' + content.strip() }}{% endif %}{% endfor %}{% if add_generation_prompt and messages[-1]['role'] != 'assistant' %} {{'\n[|AI|]\n### Response:'}}{% endif %}",
14
+ "clean_up_tokenization_spaces": true,
15
+ "eos_token": "<|endoftext|>",
16
+ "model_max_length": 2048,
 
 
 
 
 
17
  "pad_token": "<|endoftext|>",
18
+ "tokenizer_class": "PreTrainedTokenizerFast",
19
+ "unk_token": "<|endoftext|>"
 
20
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a2437373338b57fc0a580fecc19cc6a5fcdfbc9d9f24192dd532af70314c6ff
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3c279336da0be4b2e329964f1f1c6d8321dc46c0e81ed5a4fc3d0b110ef8065
3
  size 4920