Gokalpozer commited on
Commit
a8c03f7
1 Parent(s): 7029f7b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -15
README.md CHANGED
@@ -19,21 +19,20 @@ model-index:
19
  num_few_shot: 25
20
  metrics:
21
  - type: acc
22
- value: 46.08
23
  name: accuracy
24
  - task:
25
  type: text-generation
26
  name: Text Generation
27
  dataset:
28
- name: MMLU TR v0.2
29
- type: cais/mmlu
30
- config: all
31
- split: test
32
  args:
33
- num_few_shot: 5
34
  metrics:
35
  - type: acc
36
- value: 47.02
37
  name: accuracy
38
  - task:
39
  type: text-generation
@@ -48,7 +47,7 @@ model-index:
48
  metrics:
49
  - type: acc
50
  name: accuracy
51
- value: 49.38
52
  - task:
53
  type: text-generation
54
  name: Text Generation
@@ -61,7 +60,7 @@ model-index:
61
  num_few_shot: 5
62
  metrics:
63
  - type: acc
64
- value: 53.71
65
  name: accuracy
66
  - task:
67
  type: text-generation
@@ -75,7 +74,7 @@ model-index:
75
  num_few_shot: 5
76
  metrics:
77
  - type: acc
78
- value: 53.08
79
  name: accuracy
80
  pipeline_tag: text-generation
81
  ---
@@ -89,17 +88,15 @@ pipeline_tag: text-generation
89
  <aside>by <a href="https://curiosity.tech">Curiosity Technology</a></aside>
90
  </div>
91
 
92
- MARS is the first iteration of Curiosity Technology models, based on Llama 3 8B.
93
 
94
- We have trained MARS on in-house Turkish dataset, as well as several open-source datasets and their Turkish
95
- translations.
96
- It is our intention to release Turkish translations in near future for community to have their go on them.
97
 
98
  MARS have been trained for 3 days on 4xA100.
99
 
100
  ## Model Details
101
 
102
- - **Base Model**: Meta Llama 3 8B Instruct
103
  - **Training Dataset**: In-house & Translated Open Source Turkish Datasets
104
  - **Training Method**: LoRA Fine Tuning
105
 
 
19
  num_few_shot: 25
20
  metrics:
21
  - type: acc
22
+ value: 43.85
23
  name: accuracy
24
  - task:
25
  type: text-generation
26
  name: Text Generation
27
  dataset:
28
+ name: HellaSwag TR
29
+ type: hellaswag
30
+ split: validation
 
31
  args:
32
+ num_few_shot: 10
33
  metrics:
34
  - type: acc
35
+ value: 46.64
36
  name: accuracy
37
  - task:
38
  type: text-generation
 
47
  metrics:
48
  - type: acc
49
  name: accuracy
50
+ value: 48.66
51
  - task:
52
  type: text-generation
53
  name: Text Generation
 
60
  num_few_shot: 5
61
  metrics:
62
  - type: acc
63
+ value: 52.84
64
  name: accuracy
65
  - task:
66
  type: text-generation
 
74
  num_few_shot: 5
75
  metrics:
76
  - type: acc
77
+ value: 59.30
78
  name: accuracy
79
  pipeline_tag: text-generation
80
  ---
 
88
  <aside>by <a href="https://curiosity.tech">Curiosity Technology</a></aside>
89
  </div>
90
 
91
+ MARS-v0.2 is the second iteration of Curiosity Technology models, built on the foundation of Llama 3.1 8B. This version expands upon the initial MARS model by fine-tuning it with a more comprehensive dataset, with an increased emphasis on mathematical data to enhance its reasoning and problem-solving capabilities.
92
 
93
+ We've continued our commitment to Turkish language processing, utilizing both in-house Turkish datasets and a broader selection of translated open-source datasets. We believe this version will serve the community with even more versatility and depth.
 
 
94
 
95
  MARS have been trained for 3 days on 4xA100.
96
 
97
  ## Model Details
98
 
99
+ - **Base Model**: Meta Llama 3.1 8B Instruct
100
  - **Training Dataset**: In-house & Translated Open Source Turkish Datasets
101
  - **Training Method**: LoRA Fine Tuning
102