JeongwonChoi commited on
Commit
0111b5c
1 Parent(s): 05b11c7

Initial commit

Browse files
README.md CHANGED
@@ -50,13 +50,13 @@ text = """\
50
 
51
  ### **[Ko LM Eval Harness](https://github.com/Beomi/ko-lm-evaluation-harness)**
52
 
53
- | Task | 0-shot | 5-shot | 10-shot | 50-shot |
54
- | :--------------- | -----------: | -------------: | ------------: | -----------: |
55
- | kobest_boolq | 0.375807 | 0.822623 | 0.828582 | 0.822529 |
56
- | kobest_copa | 0.539993 | 0.665979 | 0.67998 | 0.694997 |
57
- | kobest_hellaswag | 0.405785 | 0.401975 | 0.438219 | 0.402962 |
58
- | kobest_sentineg | 0.794083 | 0.85276 | 0.883509 | 0.880932 |
59
- | **Average** | **0.528917** | **0.68583425** | **0.7075725** | **0.700355** |
60
 
61
  ### **[Ko-LLM-Leaderboard](https://huggingface.co/spaces/upstage/open-ko-llm-leaderboard)**
62
 
 
50
 
51
  ### **[Ko LM Eval Harness](https://github.com/Beomi/ko-lm-evaluation-harness)**
52
 
53
+ | Task | 0-shot | 5-shot | 10-shot | 50-shot |
54
+ | :--------------- | -----------: | -----------: | -----------: | -----------: |
55
+ | kobest_boolq | 0.903441 | 0.922987 | 0.919466 | 0.923032 |
56
+ | kobest_copa | 0.734711 | 0.778697 | 0.773796 | 0.796829 |
57
+ | kobest_hellaswag | 0.473673 | 0.480091 | 0.491471 | 0.488234 |
58
+ | kobest_sentineg | 0.536605 | 0.93185 | 0.952136 | 0.949596 |
59
+ | **Average** | **0.662107** | **0.778406** | **0.784217** | **0.789423** |
60
 
61
  ### **[Ko-LLM-Leaderboard](https://huggingface.co/spaces/upstage/open-ko-llm-leaderboard)**
62
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "Edentns/DataVortexS-10.7B-dpo-v1.1",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -26,4 +26,4 @@
26
  "transformers_version": "4.36.2",
27
  "use_cache": false,
28
  "vocab_size": 32000
29
- }
 
1
  {
2
+ "_name_or_path": "Edentns/DataVortexS-10.7B-dpo-v1.8",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
26
  "transformers_version": "4.36.2",
27
  "use_cache": false,
28
  "vocab_size": 32000
29
+ }
generation_config.json CHANGED
@@ -5,4 +5,4 @@
5
  "pad_token_id": 2,
6
  "transformers_version": "4.36.2",
7
  "use_cache": false
8
- }
 
5
  "pad_token_id": 2,
6
  "transformers_version": "4.36.2",
7
  "use_cache": false
8
+ }
model-00001-of-00005.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74252227b9e6eef8683cae5a1a2ca08bc79700ec505a9f65d8637381924afcb5
3
  size 4943162240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73ecf068164752782e8d846e3ea90bbedb8531ca80ee3f4bb3361bb4a47590ff
3
  size 4943162240
model-00002-of-00005.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef7a5790cca4b941f2e94c9dd83ff6ce9e71f4511cd1b0f03ad295e1615ade50
3
  size 4999819232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e2c40c25f39bb414b81346214e174238ef954be69e05c317a197d43319a6d03
3
  size 4999819232
model-00003-of-00005.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c4c4a5227a1d7b56039c84b68e53951615d5a1d3423061369b3952e7c7a51f2
3
  size 4915916080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c939ec3977e87192d61752e46ca26a9e620f1c10a102e3c13f15e811d5cfea56
3
  size 4915916080
model-00004-of-00005.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e90034f6aebb0ec4fb00c7642137aa62aa4d6ce2787ec18eb62380d0323990d
3
  size 4915916080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66da474a3e5dcdc9665d92ec40b6275e1783e0e4dde2b88194ed919b89e1e520
3
  size 4915916080
model-00005-of-00005.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd75e803d2e65720ca000352b8b037fdd9eb2bb14f364dbe3c2ea92e9ddfd57e
3
  size 1688284744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bc0376bc503637c3f8cdbb95d49d14cc155dc39fdf238270973eebf262e60d0
3
  size 1688284744