csujeong commited on
Commit
3814f26
1 Parent(s): 62bbab4

Training in progress, step 10

Browse files
Files changed (36) hide show
  1. README.md +59 -0
  2. adapter_config.json +32 -0
  3. adapter_model.safetensors +3 -0
  4. runs/Jan07_11-26-45_371ca9b7a7c7/events.out.tfevents.1704626819.371ca9b7a7c7.476.0 +3 -0
  5. runs/Jan07_11-26-45_371ca9b7a7c7/events.out.tfevents.1704627272.371ca9b7a7c7.476.1 +3 -0
  6. runs/Mar21_01-53-07_d72aea89d4b6/events.out.tfevents.1710986084.d72aea89d4b6.3894.0 +3 -0
  7. runs/Mar22_04-33-33_6b98cdd15d10/events.out.tfevents.1711082034.6b98cdd15d10.209.0 +3 -0
  8. special_tokens_map.json +24 -0
  9. tokenizer.json +0 -0
  10. tokenizer.model +3 -0
  11. tokenizer_config.json +43 -0
  12. training_args.bin +3 -0
  13. wandb/debug-cli.root.log +0 -0
  14. wandb/debug-internal.log +143 -0
  15. wandb/debug.log +31 -0
  16. wandb/run-20240107_113444-ebd79hle/files/config.yaml +677 -0
  17. wandb/run-20240107_113444-ebd79hle/files/output.log +452 -0
  18. wandb/run-20240107_113444-ebd79hle/files/requirements.txt +497 -0
  19. wandb/run-20240107_113444-ebd79hle/files/wandb-metadata.json +52 -0
  20. wandb/run-20240107_113444-ebd79hle/files/wandb-summary.json +1 -0
  21. wandb/run-20240107_113444-ebd79hle/logs/debug.log +144 -0
  22. wandb/run-20240107_113444-ebd79hle/run-ebd79hle.wandb +0 -0
  23. wandb/run-20240321_015458-y11xrjxv/files/config.yaml +678 -0
  24. wandb/run-20240321_015458-y11xrjxv/files/output.log +81 -0
  25. wandb/run-20240321_015458-y11xrjxv/files/requirements.txt +510 -0
  26. wandb/run-20240321_015458-y11xrjxv/files/wandb-metadata.json +52 -0
  27. wandb/run-20240321_015458-y11xrjxv/files/wandb-summary.json +1 -0
  28. wandb/run-20240321_015458-y11xrjxv/logs/debug.log +60 -0
  29. wandb/run-20240321_015458-y11xrjxv/run-y11xrjxv.wandb +0 -0
  30. wandb/run-20240322_043407-hize5f4p/files/config.yaml +678 -0
  31. wandb/run-20240322_043407-hize5f4p/files/requirements.txt +510 -0
  32. wandb/run-20240322_043407-hize5f4p/files/wandb-metadata.json +52 -0
  33. wandb/run-20240322_043407-hize5f4p/files/wandb-summary.json +1 -0
  34. wandb/run-20240322_043407-hize5f4p/logs/debug-internal.log +143 -0
  35. wandb/run-20240322_043407-hize5f4p/logs/debug.log +31 -0
  36. wandb/run-20240322_043407-hize5f4p/run-hize5f4p.wandb +0 -0
README.md ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - trl
6
+ - sft
7
+ - generated_from_trainer
8
+ base_model: mistralai/Mistral-7B-v0.1
9
+ model-index:
10
+ - name: Mistral-7B-Finetuning-Insurance
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # Mistral-7B-Finetuning-Insurance
18
+
19
+ This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on an unknown dataset.
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 0.0002
39
+ - train_batch_size: 2
40
+ - eval_batch_size: 8
41
+ - seed: 42
42
+ - gradient_accumulation_steps: 2
43
+ - total_train_batch_size: 4
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: cosine
46
+ - lr_scheduler_warmup_ratio: 0.03
47
+ - training_steps: 60
48
+
49
+ ### Training results
50
+
51
+
52
+
53
+ ### Framework versions
54
+
55
+ - PEFT 0.9.1.dev0
56
+ - Transformers 4.39.0
57
+ - Pytorch 2.2.1+cu121
58
+ - Datasets 2.18.0
59
+ - Tokenizers 0.15.2
adapter_config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 32,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "o_proj",
24
+ "k_proj",
25
+ "v_proj",
26
+ "q_proj",
27
+ "gate_proj"
28
+ ],
29
+ "task_type": "CAUSAL_LM",
30
+ "use_dora": false,
31
+ "use_rslora": false
32
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43950b67322137608ce1a8dfc759aac99c1cd96ebb6f3b81bc074e2d8328e6e7
3
+ size 184592616
runs/Jan07_11-26-45_371ca9b7a7c7/events.out.tfevents.1704626819.371ca9b7a7c7.476.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9cd26860faf87ca16849fe529b47490601bb308c0982c3b295c6dc1dfb532ed
3
+ size 4883
runs/Jan07_11-26-45_371ca9b7a7c7/events.out.tfevents.1704627272.371ca9b7a7c7.476.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32e57050c8b28a73b5395196f5673bc9704ee12a3316a62ed275e6373d809272
3
+ size 6155
runs/Mar21_01-53-07_d72aea89d4b6/events.out.tfevents.1710986084.d72aea89d4b6.3894.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ec381b1552bae1db95cab9758139f6ee02059d6c158391f4290c4140736f685
3
+ size 6810
runs/Mar22_04-33-33_6b98cdd15d10/events.out.tfevents.1711082034.6b98cdd15d10.209.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eda6a2910e68b3d50cc3e5c609fa6a283e010c042460fa0feb5e3bd7c012ee3
3
+ size 5438
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "additional_special_tokens": [],
31
+ "bos_token": "<s>",
32
+ "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
33
+ "clean_up_tokenization_spaces": false,
34
+ "eos_token": "</s>",
35
+ "legacy": true,
36
+ "model_max_length": 1000000000000000019884624838656,
37
+ "pad_token": "</s>",
38
+ "sp_model_kwargs": {},
39
+ "spaces_between_special_tokens": false,
40
+ "tokenizer_class": "LlamaTokenizer",
41
+ "unk_token": "<unk>",
42
+ "use_default_system_prompt": false
43
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6311196ebc569b22dec1fd0939bb62a6660125ebeddb6270dfced5564e13fd4b
3
+ size 5048
wandb/debug-cli.root.log ADDED
File without changes
wandb/debug-internal.log ADDED
@@ -0,0 +1,143 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-03-22 04:34:07,566 INFO StreamThr :4275 [internal.py:wandb_internal():86] W&B internal server running at pid: 4275, started at: 2024-03-22 04:34:07.562793
2
+ 2024-03-22 04:34:07,589 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status
3
+ 2024-03-22 04:34:12,594 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
4
+ 2024-03-22 04:34:12,594 INFO WriterThread:4275 [datastore.py:open_for_write():87] open: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/run-hize5f4p.wandb
5
+ 2024-03-22 04:34:12,609 DEBUG SenderThread:4275 [sender.py:send():378] send: header
6
+ 2024-03-22 04:34:12,616 DEBUG SenderThread:4275 [sender.py:send():378] send: run
7
+ 2024-03-22 04:34:13,047 INFO SenderThread:4275 [dir_watcher.py:__init__():211] watching files in: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files
8
+ 2024-03-22 04:34:13,047 INFO SenderThread:4275 [sender.py:_start_run_threads():1099] run started: hize5f4p with start time 1711082047.565462
9
+ 2024-03-22 04:34:13,064 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: check_version
10
+ 2024-03-22 04:34:13,064 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: check_version
11
+ 2024-03-22 04:34:13,182 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: run_start
12
+ 2024-03-22 04:34:13,192 DEBUG HandlerThread:4275 [system_info.py:__init__():26] System info init
13
+ 2024-03-22 04:34:13,193 DEBUG HandlerThread:4275 [system_info.py:__init__():41] System info init done
14
+ 2024-03-22 04:34:13,193 INFO HandlerThread:4275 [system_monitor.py:start():194] Starting system monitor
15
+ 2024-03-22 04:34:13,193 INFO SystemMonitor:4275 [system_monitor.py:_start():158] Starting system asset monitoring threads
16
+ 2024-03-22 04:34:13,193 INFO HandlerThread:4275 [system_monitor.py:probe():214] Collecting system info
17
+ 2024-03-22 04:34:13,194 INFO SystemMonitor:4275 [interfaces.py:start():190] Started cpu monitoring
18
+ 2024-03-22 04:34:13,195 INFO SystemMonitor:4275 [interfaces.py:start():190] Started disk monitoring
19
+ 2024-03-22 04:34:13,196 INFO SystemMonitor:4275 [interfaces.py:start():190] Started gpu monitoring
20
+ 2024-03-22 04:34:13,197 INFO SystemMonitor:4275 [interfaces.py:start():190] Started memory monitoring
21
+ 2024-03-22 04:34:13,198 INFO SystemMonitor:4275 [interfaces.py:start():190] Started network monitoring
22
+ 2024-03-22 04:34:13,203 DEBUG HandlerThread:4275 [system_info.py:probe():150] Probing system
23
+ 2024-03-22 04:34:13,208 DEBUG HandlerThread:4275 [gitlib.py:_init_repo():56] git repository is invalid
24
+ 2024-03-22 04:34:13,208 DEBUG HandlerThread:4275 [system_info.py:probe():198] Probing system done
25
+ 2024-03-22 04:34:13,208 DEBUG HandlerThread:4275 [system_monitor.py:probe():223] {'os': 'Linux-6.1.58+-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-03-22T04:34:13.203984', 'startedAt': '2024-03-22T04:34:07.465777', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'Mistral-7B-Finetuning-Insurance.ipynb', 'codePathLocal': None, 'colab': 'https://colab.research.google.com/notebook#fileId=1yggNDPxtSX3bri4ON4ppYc-G-Q2zFtkQ', 'host': '6b98cdd15d10', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 1, 'cpu_count_logical': 2, 'cpu_freq': {'current': 2000.142, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 78.1898422241211, 'used': 42.876033782958984}}, 'gpu': 'Tesla T4', 'gpu_count': 1, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 12.674789428710938}}
26
+ 2024-03-22 04:34:13,208 INFO HandlerThread:4275 [system_monitor.py:probe():224] Finished collecting system info
27
+ 2024-03-22 04:34:13,208 INFO HandlerThread:4275 [system_monitor.py:probe():227] Publishing system info
28
+ 2024-03-22 04:34:13,226 INFO HandlerThread:4275 [system_monitor.py:probe():229] Finished publishing system info
29
+ 2024-03-22 04:34:13,232 DEBUG SenderThread:4275 [sender.py:send():378] send: files
30
+ 2024-03-22 04:34:13,233 INFO SenderThread:4275 [sender.py:_save_file():1365] saving file wandb-metadata.json with policy now
31
+ 2024-03-22 04:34:13,733 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: python_packages
32
+ 2024-03-22 04:34:13,733 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: python_packages
33
+ 2024-03-22 04:34:13,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
34
+ 2024-03-22 04:34:13,735 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
35
+ 2024-03-22 04:34:13,741 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
36
+ 2024-03-22 04:34:13,888 INFO wandb-upload_0:4275 [upload_job.py:push():131] Uploaded file /tmp/tmp3i0vj0_swandb/55hqq2eg-wandb-metadata.json
37
+ 2024-03-22 04:34:13,972 DEBUG SenderThread:4275 [sender.py:send():378] send: telemetry
38
+ 2024-03-22 04:34:13,973 DEBUG SenderThread:4275 [sender.py:send():378] send: config
39
+ 2024-03-22 04:34:13,975 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
40
+ 2024-03-22 04:34:13,975 DEBUG SenderThread:4275 [sender.py:send():378] send: telemetry
41
+ 2024-03-22 04:34:13,976 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
42
+ 2024-03-22 04:34:13,976 WARNING SenderThread:4275 [sender.py:send_metric():1316] Seen metric with glob (shouldn't happen)
43
+ 2024-03-22 04:34:13,976 DEBUG SenderThread:4275 [sender.py:send():378] send: telemetry
44
+ 2024-03-22 04:34:14,052 INFO Thread-12 :4275 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files/requirements.txt
45
+ 2024-03-22 04:34:14,053 INFO Thread-12 :4275 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files/wandb-metadata.json
46
+ 2024-03-22 04:34:17,977 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-03-22 04:34:22,978 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
48
+ 2024-03-22 04:34:27,978 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
49
+ 2024-03-22 04:34:28,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
50
+ 2024-03-22 04:34:28,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
51
+ 2024-03-22 04:34:28,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
52
+ 2024-03-22 04:34:33,912 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
53
+ 2024-03-22 04:34:38,918 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
54
+ 2024-03-22 04:34:40,099 INFO Thread-12 :4275 [dir_watcher.py:_on_file_modified():288] file/dir modified: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files/config.yaml
55
+ 2024-03-22 04:34:43,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
56
+ 2024-03-22 04:34:43,736 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
57
+ 2024-03-22 04:34:43,776 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
58
+ 2024-03-22 04:34:44,947 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
59
+ 2024-03-22 04:34:49,948 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
60
+ 2024-03-22 04:34:54,949 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
61
+ 2024-03-22 04:34:58,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
62
+ 2024-03-22 04:34:58,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
63
+ 2024-03-22 04:34:58,774 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
64
+ 2024-03-22 04:35:00,929 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
65
+ 2024-03-22 04:35:05,930 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
66
+ 2024-03-22 04:35:10,931 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
67
+ 2024-03-22 04:35:13,199 DEBUG SystemMonitor:4275 [system_monitor.py:_start():172] Starting system metrics aggregation loop
68
+ 2024-03-22 04:35:13,200 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
69
+ 2024-03-22 04:35:13,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
70
+ 2024-03-22 04:35:13,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
71
+ 2024-03-22 04:35:13,774 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
72
+ 2024-03-22 04:35:16,922 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
73
+ 2024-03-22 04:35:21,922 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
74
+ 2024-03-22 04:35:26,923 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
75
+ 2024-03-22 04:35:28,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
76
+ 2024-03-22 04:35:28,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
77
+ 2024-03-22 04:35:28,775 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
78
+ 2024-03-22 04:35:31,951 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
79
+ 2024-03-22 04:35:36,952 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
80
+ 2024-03-22 04:35:41,952 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-03-22 04:35:43,201 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
82
+ 2024-03-22 04:35:43,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
83
+ 2024-03-22 04:35:43,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
84
+ 2024-03-22 04:35:43,775 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
85
+ 2024-03-22 04:35:47,914 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
86
+ 2024-03-22 04:35:52,914 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
87
+ 2024-03-22 04:35:57,915 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
88
+ 2024-03-22 04:35:58,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
89
+ 2024-03-22 04:35:58,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
90
+ 2024-03-22 04:35:58,775 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
91
+ 2024-03-22 04:36:02,958 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
92
+ 2024-03-22 04:36:07,958 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
93
+ 2024-03-22 04:36:12,959 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
94
+ 2024-03-22 04:36:13,202 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
95
+ 2024-03-22 04:36:13,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
96
+ 2024-03-22 04:36:13,736 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
97
+ 2024-03-22 04:36:13,776 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
98
+ 2024-03-22 04:36:18,925 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
99
+ 2024-03-22 04:36:23,926 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
100
+ 2024-03-22 04:36:28,735 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
101
+ 2024-03-22 04:36:28,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
102
+ 2024-03-22 04:36:28,777 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
103
+ 2024-03-22 04:36:28,959 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
104
+ 2024-03-22 04:36:33,960 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
105
+ 2024-03-22 04:36:38,377 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: partial_history
106
+ 2024-03-22 04:36:38,384 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
107
+ 2024-03-22 04:36:38,385 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
108
+ 2024-03-22 04:36:38,385 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
109
+ 2024-03-22 04:36:38,385 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
110
+ 2024-03-22 04:36:38,386 DEBUG SenderThread:4275 [sender.py:send():378] send: history
111
+ 2024-03-22 04:36:38,386 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: summary_record
112
+ 2024-03-22 04:36:38,395 INFO SenderThread:4275 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
113
+ 2024-03-22 04:36:39,314 INFO Thread-12 :4275 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files/wandb-summary.json
114
+ 2024-03-22 04:36:39,396 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
115
+ 2024-03-22 04:36:43,202 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
116
+ 2024-03-22 04:36:49,784 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
117
+ 2024-03-22 04:36:50,358 INFO Thread-12 :4275 [dir_watcher.py:_on_file_modified():288] file/dir modified: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files/config.yaml
118
+ 2024-03-22 04:36:50,399 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
119
+ 2024-03-22 04:36:50,401 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
120
+ 2024-03-22 04:36:50,402 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
121
+ 2024-03-22 04:36:55,622 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-03-22 04:37:14,735 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
123
+ 2024-03-22 04:37:14,752 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
124
+ 2024-03-22 04:37:14,753 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
125
+ 2024-03-22 04:37:14,753 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
126
+ 2024-03-22 04:37:14,753 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
127
+ 2024-03-22 04:37:14,754 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
128
+ 2024-03-22 04:37:14,783 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
129
+ 2024-03-22 04:37:16,284 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
130
+ 2024-03-22 04:37:21,289 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
131
+ 2024-03-22 04:37:31,399 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
132
+ 2024-03-22 04:37:31,407 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
133
+ 2024-03-22 04:37:31,408 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
134
+ 2024-03-22 04:37:31,408 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
135
+ 2024-03-22 04:37:31,409 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
136
+ 2024-03-22 04:37:36,568 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
137
+ 2024-03-22 04:37:41,569 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
138
+ 2024-03-22 04:37:43,204 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
139
+ 2024-03-22 04:37:44,743 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
140
+ 2024-03-22 04:37:44,744 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
141
+ 2024-03-22 04:37:46,973 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
142
+ 2024-03-22 04:37:51,974 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
143
+ 2024-03-22 04:37:56,975 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
wandb/debug.log ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-03-22 04:34:07,548 INFO MainThread:209 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
+ 2024-03-22 04:34:07,550 INFO MainThread:209 [wandb_setup.py:_flush():76] Configure stats pid to 209
3
+ 2024-03-22 04:34:07,551 INFO MainThread:209 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-03-22 04:34:07,551 INFO MainThread:209 [wandb_setup.py:_flush():76] Loading settings from /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/settings
5
+ 2024-03-22 04:34:07,551 INFO MainThread:209 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-03-22 04:34:07,552 INFO MainThread:209 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-03-22 04:34:07,553 INFO MainThread:209 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-03-22 04:34:07,553 INFO MainThread:209 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-03-22 04:34:07,553 INFO MainThread:209 [wandb_init.py:_log_setup():526] Logging user logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/logs/debug.log
10
+ 2024-03-22 04:34:07,554 INFO MainThread:209 [wandb_init.py:_log_setup():527] Logging internal logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/logs/debug-internal.log
11
+ 2024-03-22 04:34:07,554 INFO MainThread:209 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7f0538e72f80>
12
+ 2024-03-22 04:34:07,555 INFO MainThread:209 [wandb_init.py:init():566] calling init triggers
13
+ 2024-03-22 04:34:07,555 INFO MainThread:209 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-03-22 04:34:07,555 INFO MainThread:209 [wandb_init.py:init():616] starting backend
16
+ 2024-03-22 04:34:07,555 INFO MainThread:209 [wandb_init.py:init():620] setting up manager
17
+ 2024-03-22 04:34:07,561 INFO MainThread:209 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-03-22 04:34:07,564 INFO MainThread:209 [wandb_init.py:init():628] backend started and connected
19
+ 2024-03-22 04:34:07,592 INFO MainThread:209 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
+ 2024-03-22 04:34:12,608 INFO MainThread:209 [wandb_run.py:_label_probe_notebook():1305] Unable to probe notebook: 'NoneType' object has no attribute 'get'
21
+ 2024-03-22 04:34:12,608 INFO MainThread:209 [wandb_init.py:init():720] updated telemetry
22
+ 2024-03-22 04:34:12,615 INFO MainThread:209 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
23
+ 2024-03-22 04:34:13,063 INFO MainThread:209 [wandb_run.py:_on_init():2262] communicating current version
24
+ 2024-03-22 04:34:13,171 INFO MainThread:209 [wandb_run.py:_on_init():2271] got version response
25
+ 2024-03-22 04:34:13,171 INFO MainThread:209 [wandb_init.py:init():804] starting run threads in backend
26
+ 2024-03-22 04:34:13,734 INFO MainThread:209 [wandb_run.py:_console_start():2241] atexit reg
27
+ 2024-03-22 04:34:13,734 INFO MainThread:209 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
+ 2024-03-22 04:34:13,734 INFO MainThread:209 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
+ 2024-03-22 04:34:13,734 INFO MainThread:209 [wandb_run.py:_redirect():2186] Redirects installed.
30
+ 2024-03-22 04:34:13,737 INFO MainThread:209 [wandb_init.py:init():847] run started, returning control to user process
31
+ 2024-03-22 04:34:13,745 INFO MainThread:209 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'mistralai/Mistral-7B-Instruct-v0.2', 'transformers_version': '4.39.0', 'model_type': 'mistral', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'bfloat16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 3.0, 'max_steps': 60, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/runs/Mar22_04-33-33_6b98cdd15d10', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 10, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240107_113444-ebd79hle/files/config.yaml ADDED
@@ -0,0 +1,677 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.12
7
+ cli_version: 0.16.1
8
+ framework: huggingface
9
+ huggingface_version: 4.36.2
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: false
12
+ start_time: 1704627284.817819
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ 2:
29
+ - 1
30
+ - 2
31
+ - 3
32
+ - 5
33
+ - 11
34
+ - 12
35
+ - 49
36
+ - 51
37
+ - 53
38
+ - 55
39
+ - 71
40
+ - 84
41
+ - 98
42
+ 3:
43
+ - 7
44
+ - 23
45
+ 4: 3.10.12
46
+ 5: 0.16.1
47
+ 6: 4.36.2
48
+ 8:
49
+ - 1
50
+ - 5
51
+ - 12
52
+ 9:
53
+ 1: transformers_trainer
54
+ 13: linux-x86_64
55
+ m:
56
+ - 1: train/global_step
57
+ 6:
58
+ - 3
59
+ - 1: train/loss
60
+ 5: 1
61
+ 6:
62
+ - 1
63
+ - 1: train/learning_rate
64
+ 5: 1
65
+ 6:
66
+ - 1
67
+ - 1: train/epoch
68
+ 5: 1
69
+ 6:
70
+ - 1
71
+ - 1: train/train_runtime
72
+ 5: 1
73
+ 6:
74
+ - 1
75
+ - 1: train/train_samples_per_second
76
+ 5: 1
77
+ 6:
78
+ - 1
79
+ - 1: train/train_steps_per_second
80
+ 5: 1
81
+ 6:
82
+ - 1
83
+ - 1: train/total_flos
84
+ 5: 1
85
+ 6:
86
+ - 1
87
+ - 1: train/train_loss
88
+ 5: 1
89
+ 6:
90
+ - 1
91
+ vocab_size:
92
+ desc: null
93
+ value: 32000
94
+ max_position_embeddings:
95
+ desc: null
96
+ value: 32768
97
+ hidden_size:
98
+ desc: null
99
+ value: 4096
100
+ intermediate_size:
101
+ desc: null
102
+ value: 14336
103
+ num_hidden_layers:
104
+ desc: null
105
+ value: 32
106
+ num_attention_heads:
107
+ desc: null
108
+ value: 32
109
+ sliding_window:
110
+ desc: null
111
+ value: 4096
112
+ num_key_value_heads:
113
+ desc: null
114
+ value: 8
115
+ hidden_act:
116
+ desc: null
117
+ value: silu
118
+ initializer_range:
119
+ desc: null
120
+ value: 0.02
121
+ rms_norm_eps:
122
+ desc: null
123
+ value: 1.0e-05
124
+ use_cache:
125
+ desc: null
126
+ value: false
127
+ rope_theta:
128
+ desc: null
129
+ value: 10000.0
130
+ attention_dropout:
131
+ desc: null
132
+ value: 0.0
133
+ return_dict:
134
+ desc: null
135
+ value: true
136
+ output_hidden_states:
137
+ desc: null
138
+ value: false
139
+ output_attentions:
140
+ desc: null
141
+ value: false
142
+ torchscript:
143
+ desc: null
144
+ value: false
145
+ torch_dtype:
146
+ desc: null
147
+ value: bfloat16
148
+ use_bfloat16:
149
+ desc: null
150
+ value: false
151
+ tf_legacy_loss:
152
+ desc: null
153
+ value: false
154
+ pruned_heads:
155
+ desc: null
156
+ value: {}
157
+ tie_word_embeddings:
158
+ desc: null
159
+ value: false
160
+ is_encoder_decoder:
161
+ desc: null
162
+ value: false
163
+ is_decoder:
164
+ desc: null
165
+ value: false
166
+ cross_attention_hidden_size:
167
+ desc: null
168
+ value: null
169
+ add_cross_attention:
170
+ desc: null
171
+ value: false
172
+ tie_encoder_decoder:
173
+ desc: null
174
+ value: false
175
+ max_length:
176
+ desc: null
177
+ value: 20
178
+ min_length:
179
+ desc: null
180
+ value: 0
181
+ do_sample:
182
+ desc: null
183
+ value: false
184
+ early_stopping:
185
+ desc: null
186
+ value: false
187
+ num_beams:
188
+ desc: null
189
+ value: 1
190
+ num_beam_groups:
191
+ desc: null
192
+ value: 1
193
+ diversity_penalty:
194
+ desc: null
195
+ value: 0.0
196
+ temperature:
197
+ desc: null
198
+ value: 1.0
199
+ top_k:
200
+ desc: null
201
+ value: 50
202
+ top_p:
203
+ desc: null
204
+ value: 1.0
205
+ typical_p:
206
+ desc: null
207
+ value: 1.0
208
+ repetition_penalty:
209
+ desc: null
210
+ value: 1.0
211
+ length_penalty:
212
+ desc: null
213
+ value: 1.0
214
+ no_repeat_ngram_size:
215
+ desc: null
216
+ value: 0
217
+ encoder_no_repeat_ngram_size:
218
+ desc: null
219
+ value: 0
220
+ bad_words_ids:
221
+ desc: null
222
+ value: null
223
+ num_return_sequences:
224
+ desc: null
225
+ value: 1
226
+ chunk_size_feed_forward:
227
+ desc: null
228
+ value: 0
229
+ output_scores:
230
+ desc: null
231
+ value: false
232
+ return_dict_in_generate:
233
+ desc: null
234
+ value: false
235
+ forced_bos_token_id:
236
+ desc: null
237
+ value: null
238
+ forced_eos_token_id:
239
+ desc: null
240
+ value: null
241
+ remove_invalid_values:
242
+ desc: null
243
+ value: false
244
+ exponential_decay_length_penalty:
245
+ desc: null
246
+ value: null
247
+ suppress_tokens:
248
+ desc: null
249
+ value: null
250
+ begin_suppress_tokens:
251
+ desc: null
252
+ value: null
253
+ architectures:
254
+ desc: null
255
+ value:
256
+ - MistralForCausalLM
257
+ finetuning_task:
258
+ desc: null
259
+ value: null
260
+ id2label:
261
+ desc: null
262
+ value:
263
+ '0': LABEL_0
264
+ '1': LABEL_1
265
+ label2id:
266
+ desc: null
267
+ value:
268
+ LABEL_0: 0
269
+ LABEL_1: 1
270
+ tokenizer_class:
271
+ desc: null
272
+ value: null
273
+ prefix:
274
+ desc: null
275
+ value: null
276
+ bos_token_id:
277
+ desc: null
278
+ value: 1
279
+ pad_token_id:
280
+ desc: null
281
+ value: null
282
+ eos_token_id:
283
+ desc: null
284
+ value: 2
285
+ sep_token_id:
286
+ desc: null
287
+ value: null
288
+ decoder_start_token_id:
289
+ desc: null
290
+ value: null
291
+ task_specific_params:
292
+ desc: null
293
+ value: null
294
+ problem_type:
295
+ desc: null
296
+ value: null
297
+ _name_or_path:
298
+ desc: null
299
+ value: mistralai/Mistral-7B-v0.1
300
+ transformers_version:
301
+ desc: null
302
+ value: 4.36.2
303
+ model_type:
304
+ desc: null
305
+ value: mistral
306
+ quantization_config:
307
+ desc: null
308
+ value:
309
+ quant_method: QuantizationMethod.BITS_AND_BYTES
310
+ load_in_8bit: false
311
+ load_in_4bit: true
312
+ llm_int8_threshold: 6.0
313
+ llm_int8_skip_modules: null
314
+ llm_int8_enable_fp32_cpu_offload: false
315
+ llm_int8_has_fp16_weight: false
316
+ bnb_4bit_quant_type: nf4
317
+ bnb_4bit_use_double_quant: true
318
+ bnb_4bit_compute_dtype: bfloat16
319
+ output_dir:
320
+ desc: null
321
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance
322
+ overwrite_output_dir:
323
+ desc: null
324
+ value: false
325
+ do_train:
326
+ desc: null
327
+ value: false
328
+ do_eval:
329
+ desc: null
330
+ value: false
331
+ do_predict:
332
+ desc: null
333
+ value: false
334
+ evaluation_strategy:
335
+ desc: null
336
+ value: 'no'
337
+ prediction_loss_only:
338
+ desc: null
339
+ value: false
340
+ per_device_train_batch_size:
341
+ desc: null
342
+ value: 2
343
+ per_device_eval_batch_size:
344
+ desc: null
345
+ value: 8
346
+ per_gpu_train_batch_size:
347
+ desc: null
348
+ value: null
349
+ per_gpu_eval_batch_size:
350
+ desc: null
351
+ value: null
352
+ gradient_accumulation_steps:
353
+ desc: null
354
+ value: 2
355
+ eval_accumulation_steps:
356
+ desc: null
357
+ value: null
358
+ eval_delay:
359
+ desc: null
360
+ value: 0
361
+ learning_rate:
362
+ desc: null
363
+ value: 0.0002
364
+ weight_decay:
365
+ desc: null
366
+ value: 0.0
367
+ adam_beta1:
368
+ desc: null
369
+ value: 0.9
370
+ adam_beta2:
371
+ desc: null
372
+ value: 0.999
373
+ adam_epsilon:
374
+ desc: null
375
+ value: 1.0e-08
376
+ max_grad_norm:
377
+ desc: null
378
+ value: 0.3
379
+ num_train_epochs:
380
+ desc: null
381
+ value: 3.0
382
+ max_steps:
383
+ desc: null
384
+ value: 60
385
+ lr_scheduler_type:
386
+ desc: null
387
+ value: cosine
388
+ lr_scheduler_kwargs:
389
+ desc: null
390
+ value: {}
391
+ warmup_ratio:
392
+ desc: null
393
+ value: 0.03
394
+ warmup_steps:
395
+ desc: null
396
+ value: 0
397
+ log_level:
398
+ desc: null
399
+ value: passive
400
+ log_level_replica:
401
+ desc: null
402
+ value: warning
403
+ log_on_each_node:
404
+ desc: null
405
+ value: true
406
+ logging_dir:
407
+ desc: null
408
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/runs/Jan07_11-26-45_371ca9b7a7c7
409
+ logging_strategy:
410
+ desc: null
411
+ value: steps
412
+ logging_first_step:
413
+ desc: null
414
+ value: false
415
+ logging_steps:
416
+ desc: null
417
+ value: 10
418
+ logging_nan_inf_filter:
419
+ desc: null
420
+ value: true
421
+ save_strategy:
422
+ desc: null
423
+ value: steps
424
+ save_steps:
425
+ desc: null
426
+ value: 10
427
+ save_total_limit:
428
+ desc: null
429
+ value: null
430
+ save_safetensors:
431
+ desc: null
432
+ value: true
433
+ save_on_each_node:
434
+ desc: null
435
+ value: false
436
+ save_only_model:
437
+ desc: null
438
+ value: false
439
+ no_cuda:
440
+ desc: null
441
+ value: false
442
+ use_cpu:
443
+ desc: null
444
+ value: false
445
+ use_mps_device:
446
+ desc: null
447
+ value: false
448
+ seed:
449
+ desc: null
450
+ value: 42
451
+ data_seed:
452
+ desc: null
453
+ value: null
454
+ jit_mode_eval:
455
+ desc: null
456
+ value: false
457
+ use_ipex:
458
+ desc: null
459
+ value: false
460
+ bf16:
461
+ desc: null
462
+ value: false
463
+ fp16:
464
+ desc: null
465
+ value: false
466
+ fp16_opt_level:
467
+ desc: null
468
+ value: O1
469
+ half_precision_backend:
470
+ desc: null
471
+ value: auto
472
+ bf16_full_eval:
473
+ desc: null
474
+ value: false
475
+ fp16_full_eval:
476
+ desc: null
477
+ value: false
478
+ tf32:
479
+ desc: null
480
+ value: false
481
+ local_rank:
482
+ desc: null
483
+ value: 0
484
+ ddp_backend:
485
+ desc: null
486
+ value: null
487
+ tpu_num_cores:
488
+ desc: null
489
+ value: null
490
+ tpu_metrics_debug:
491
+ desc: null
492
+ value: false
493
+ debug:
494
+ desc: null
495
+ value: []
496
+ dataloader_drop_last:
497
+ desc: null
498
+ value: false
499
+ eval_steps:
500
+ desc: null
501
+ value: null
502
+ dataloader_num_workers:
503
+ desc: null
504
+ value: 0
505
+ past_index:
506
+ desc: null
507
+ value: -1
508
+ run_name:
509
+ desc: null
510
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance
511
+ disable_tqdm:
512
+ desc: null
513
+ value: false
514
+ remove_unused_columns:
515
+ desc: null
516
+ value: true
517
+ label_names:
518
+ desc: null
519
+ value: null
520
+ load_best_model_at_end:
521
+ desc: null
522
+ value: false
523
+ metric_for_best_model:
524
+ desc: null
525
+ value: null
526
+ greater_is_better:
527
+ desc: null
528
+ value: null
529
+ ignore_data_skip:
530
+ desc: null
531
+ value: false
532
+ fsdp:
533
+ desc: null
534
+ value: []
535
+ fsdp_min_num_params:
536
+ desc: null
537
+ value: 0
538
+ fsdp_config:
539
+ desc: null
540
+ value:
541
+ min_num_params: 0
542
+ xla: false
543
+ xla_fsdp_grad_ckpt: false
544
+ fsdp_transformer_layer_cls_to_wrap:
545
+ desc: null
546
+ value: null
547
+ deepspeed:
548
+ desc: null
549
+ value: null
550
+ label_smoothing_factor:
551
+ desc: null
552
+ value: 0.0
553
+ optim:
554
+ desc: null
555
+ value: paged_adamw_32bit
556
+ optim_args:
557
+ desc: null
558
+ value: null
559
+ adafactor:
560
+ desc: null
561
+ value: false
562
+ group_by_length:
563
+ desc: null
564
+ value: true
565
+ length_column_name:
566
+ desc: null
567
+ value: length
568
+ report_to:
569
+ desc: null
570
+ value:
571
+ - tensorboard
572
+ - wandb
573
+ ddp_find_unused_parameters:
574
+ desc: null
575
+ value: null
576
+ ddp_bucket_cap_mb:
577
+ desc: null
578
+ value: null
579
+ ddp_broadcast_buffers:
580
+ desc: null
581
+ value: null
582
+ dataloader_pin_memory:
583
+ desc: null
584
+ value: true
585
+ dataloader_persistent_workers:
586
+ desc: null
587
+ value: false
588
+ skip_memory_metrics:
589
+ desc: null
590
+ value: true
591
+ use_legacy_prediction_loop:
592
+ desc: null
593
+ value: false
594
+ push_to_hub:
595
+ desc: null
596
+ value: true
597
+ resume_from_checkpoint:
598
+ desc: null
599
+ value: null
600
+ hub_model_id:
601
+ desc: null
602
+ value: null
603
+ hub_strategy:
604
+ desc: null
605
+ value: every_save
606
+ hub_token:
607
+ desc: null
608
+ value: <HUB_TOKEN>
609
+ hub_private_repo:
610
+ desc: null
611
+ value: false
612
+ hub_always_push:
613
+ desc: null
614
+ value: false
615
+ gradient_checkpointing:
616
+ desc: null
617
+ value: false
618
+ gradient_checkpointing_kwargs:
619
+ desc: null
620
+ value: null
621
+ include_inputs_for_metrics:
622
+ desc: null
623
+ value: false
624
+ fp16_backend:
625
+ desc: null
626
+ value: auto
627
+ push_to_hub_model_id:
628
+ desc: null
629
+ value: null
630
+ push_to_hub_organization:
631
+ desc: null
632
+ value: null
633
+ push_to_hub_token:
634
+ desc: null
635
+ value: <PUSH_TO_HUB_TOKEN>
636
+ mp_parameters:
637
+ desc: null
638
+ value: ''
639
+ auto_find_batch_size:
640
+ desc: null
641
+ value: false
642
+ full_determinism:
643
+ desc: null
644
+ value: false
645
+ torchdynamo:
646
+ desc: null
647
+ value: null
648
+ ray_scope:
649
+ desc: null
650
+ value: last
651
+ ddp_timeout:
652
+ desc: null
653
+ value: 1800
654
+ torch_compile:
655
+ desc: null
656
+ value: false
657
+ torch_compile_backend:
658
+ desc: null
659
+ value: null
660
+ torch_compile_mode:
661
+ desc: null
662
+ value: null
663
+ dispatch_batches:
664
+ desc: null
665
+ value: null
666
+ split_batches:
667
+ desc: null
668
+ value: false
669
+ include_tokens_per_second:
670
+ desc: null
671
+ value: false
672
+ include_num_input_tokens_seen:
673
+ desc: null
674
+ value: false
675
+ neftune_noise_alpha:
676
+ desc: null
677
+ value: null
wandb/run-20240107_113444-ebd79hle/files/output.log ADDED
@@ -0,0 +1,452 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ You're using a LlamaTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
3
+ -------------------------------------------------
4
+ Pre-trained Model Answer:
5
+ Answer the following question truthfully.
6
+ : 고지의무가 뭐야?
7
+ : 어디에서 왔니?
8
+ : 이름은 몇번째인데요?
9
+ : 나는 언제 태어난거예요?
10
+ : 내 집주소는 어디입니까?
11
+ : 우리집 전화번호는 010-2345-6789 입니다.
12
+ : 저희 학교 위치는 여기에있습니다. (그림)
13
+ : 저희 학생들은 한국어,영어,수학을 배워요!
14
+ : 저희 선생님께서 좋아하시는것은 바로 음식과 운동입니다.
15
+ : 저희 학교를 찾으실때 도와드려요~
16
+ : 오늘 날짜는 2월 2일 입니다.
17
+ : 저희 학교는 매년 여러분
18
+ -------------------------------------------------
19
+ Finetuning PEFT Model Answer:
20
+ Answer the following question truthfully.
21
+ : 고지의무가 뭐야?
22
+ : 고지의무는 자동차보험에서 사고를 당하면, 그 때부터 1년간은 책임을 갖도록 해주어야 한다. 이것을 ‘고지의무’라고 합니다. 단, 일반적으로 30일안에 고지할 수 있습니다.(특별경우만 제외) 또한 상대방과 소통되었거나 법원에 처분된 사건은 고지의무가 없음(예외). 따라서 사고시 바로 고지하여 피해배상금을 받아오세요!
23
+ -------------------------------------------------
24
+ -------------------------------------------------
25
+ Pre-trained Model Answer:
26
+ Answer the following question truthfully.
27
+ : 손해보험에 대해 알려줘
28
+ : 10분이내로 답변하세요!
29
+ ## Question
30
+ ### What is a loss?
31
+ - A loss occurs when an insured suffers damage to his/her property or person due to unforeseen circumstances, such as fire and theft.
32
+ - Losses are classified into two types depending on whether they occur in Korea (domestic) or abroad (overseas).
33
+ - Domestic losses include those that occurred within Korean territory regardless of nationality; overseas losses refer to all other cases.
34
+ - For example, if you suffer from illness while traveling outside your country for business purposes, it will be considered an overseas loss even though you have not left your home country.
35
+ - The amount paid by insurance companies after suffering damages varies according to each company's policy terms & conditions but generally speaking most policies cover up until $5 million USD per incident with no limit set upon how much money can be claimed at once during one year period unless otherwise stated explicitly beforehand under specific clauses contained therein which may restrict certain claims made against them based off their own discretionary powers granted through contractual agreements signed between parties involved prior thereto
36
+ -------------------------------------------------
37
+ Finetuning PEFT Model Answer:
38
+ Answer the following question truthfully.
39
+ : 손해보험에 대해 알려줘
40
+ : 자동차,건설기계등의 물품이나 사고가 발생하여 피상금을 지급받는것입니다. 예를들어 화재로 인한 부득이적용된 경우도 포함됩니다.(단, 개인용으로 운전중일때만) 또한 특정 종목(공장물건, 수송업체, 택시회사 등), 특별약속서 및 법규상 필요할 경우 제외되며 그러한 경우엔 귀하와 관련있는 담보주민은 직접 신청합니다.
41
+ -------------------------------------------------
42
+ -------------------------------------------------
43
+ Pre-trained Model Answer:
44
+ Answer the following question truthfully.
45
+ : 면책조항 알려줘
46
+ : 10분이상의 시간을 소비하지 않는다고 가정한다.
47
+ ```jsx
48
+ const a = [2,3]; // 배열에서 값은 변경할수없음
49
+ a[0] = 'hello'; // error!
50
+ console.log(a); // ['hello', 3]
51
+ // const b = {name:'kim'}; // object도 마찬가지로 불변성을 유지해야함
52
+ b.age=4; // ok
53
+ console.log(b) //{ name: 'kim', age: 4 }
54
+ ```
55
+ ## immutable data structure (불변형 데이터구조)
56
+ - 자바스크립트에서 객체와 배열은 참조타입으로 동작합니다.
57
+ - 원본과 복사된 새로운 주소를 만들어내기때문에 실제로 같은것인줄 생각하게되며 오류발생
58
+ ```jsx
59
+ let
60
+ -------------------------------------------------
61
+ Finetuning PEFT Model Answer:
62
+ Answer the following question truthfully.
63
+ : 면책조항 알려줘
64
+ : ① 차량의 가입업종이 탈취, 도난에 관한 사고로서 해당하는지를 확인합니다. (단, 자동차보상법 제12조(특별규정) 및 기타 부수규정을 적용할 경우 예외처리됩니다.)② 차량의 등록번호와 소유주명과 일치하는지를 확인합니다.(단, 개발원으로 출력된 등록증은 무시되며 실물 등록증만 참조).③ 차량의 현재위치와 일치하는지를 확인합니다.④ 차량의 운전중/미운전중 여부를 확인합니다.⑤ 차량의 연식대비 최초가입일과 일치���는지를 확인합니다.⑥ 차량의 모
65
+ -------------------------------------------------
66
+ -------------------------------------------------
67
+ Pre-trained Model Answer:
68
+ Answer the following question truthfully.
69
+ : 운전면허 미보유 보험가입 어떻게 하는지 알려줘?
70
+ : 10분만에 답변드립니다!
71
+ ## What is a car insurance policy for those who do not have driver's license in Korea?
72
+ ### Car Insurance Policy For Those Who Do Not Have Driver’s License In Korea
73
+ ##### [Car Insurance] How to get an auto insurance without having a driving licence (DL) or international DL in South Korea?
74
+ In order to drive on Korean roads, you need to obtain a valid driver’s license issued by the Ministry of Land, Transport and Maritime Affairs(MLTM). If your country has signed bilateral agreements with Korea regarding mutual recognition of licenses, then it may be possible that you can use your foreign-issued driver’s license while living here temporarily as long as its expiration date hasn’t passed yet; however this does NOT apply if one wants their own personal vehicle insured under them because they will still require obtaining local certification before doing so which takes time depending upon where exactly someone lives within Seoul metropolitan area etcetera…
75
+ The process starts off when applying through either KIA Motors Corporation headquartered near Gangnam Station OR Hyundai Motor Company located closeby too – both companies offer similar services
76
+ -------------------------------------------------
77
+ Finetuning PEFT Model Answer:
78
+ Answer the following question truthfully.
79
+ : 운전면허 미보유 보험가입 어떻게 하는지 알려줘?
80
+ : 차량의 소유자로서, 주민등록상에도 정확히 기재되고 있으며, 해당 차량을 사용할 수 없다(예) 외국인이라면 대표적으로 경력증명서를 제출한 후 임시차번호를 부여받아야 합니다.(단, 일반 개인은 불가능). 또한 법인과 같이 공동물건으로 구성된 회사들은 모두 법인으로 간주됩니다. (단, 법인의 경우 법인장본인만 가능) 그리고 특별한 조항이나 약관에 따르면 차량을 직접 운행하거나 담보금액이 100% 초과될 때 추가 보험료를 내세워야 함. 실질적으로 차량을 운전하지
81
+ -------------------------------------------------
82
+ -------------------------------------------------
83
+ Pre-trained Model Answer:
84
+ Answer the following question truthfully.
85
+ : 선물이 뭐야?
86
+ : 책, 휴대전화, 가방, 노트북, 스마트워치, 음식, 의류, 기타
87
+ Answer the following questions with a number from 1 to 5 (1 is lowest and 5 is highest).
88
+ : How much do you like this gift?
89
+ : ______
90
+ : What kind of person would give such gifts as these?
91
+ : _______
92
+ : Which one(s) are your favorite among them all? Why?
93
+ : ____________
94
+ : If someone gave me any of those things I'd be happy about it because...
95
+ : _________________
96
+ -------------------------------------------------
97
+ Finetuning PEFT Model Answer:
98
+ Answer the following question truthfully.
99
+ : 선물이 뭐야?
100
+ : 즉시가입할때, 기존의 계약을 취소하고 새로운 계약으로 변경해주는것 (단순한 갱신은 아니다)
101
+ 예를들어 차량에서 자동차보험을 받았었지만 그 중 특정 옵션(추가요율)을 제외하여 재구성하면 선물이라고 생각되며 상대방도 마찬가지로 인식합니다.(단, 개별적인 수정사항은 포함X). 또한 전체 금액과 같거나 더 비용효율적일 필요는 없습니다. 단, 당연히 현재 계약에 있는 내역을 모두 삭제하려고 하면 안됩니다. 반드시 원래 계약에 있던 내역을 유지해
102
+ -------------------------------------------------
103
+ -------------------------------------------------
104
+ Pre-trained Model Answer:
105
+ Answer the following question truthfully.
106
+ : 손해보험에 대해 자세히 설명해줘
107
+ : 10분이상의 시간을 소비하여, 사고가 발생했다면 어떻게 처리할지를 알아야한다.
108
+ - 물품/건강/법적인 문제로부터 안전성과 피해보장을 위해서는 필수적으로 구입해야되는것은 없음(개인차별)
109
+ - 그러나 일반적으로 모든 경우에 걸쳐 유용함
110
+ - 예외적인 경우도 있기때문에 주의!
111
+ - 특정 회원만 받거나 등등...
112
+ - 따라서 각종 조항들을 확실히 읽어본후 결정하는것이 좋음
113
+ - 또한 계약내역까지 잘 관리하며 저장해두는것이 중요함
114
+ - 불량 판매된 물품을
115
+ -------------------------------------------------
116
+ Finetuning PEFT Model Answer:
117
+ Answer the following question truthfully.
118
+ : 손해보험에 대해 자세히 설명해줘
119
+ : 차량의 사고로부터 피보상자가 받는 금액을 말합니다. 예를들어, 차량이 범인으로서 타인과 관련된 사고(교통사고)일때 그 중 피보상자와 관련되지 않은 것도 포함하여 계산됩니다.(단, 경우에 따라 제외될수있음). 또한 기존 법규에 의거��면 10년간 발생했던 모든 사고를 반영할 필요없습니다. 최근 3개월동안만 발생한 사고만 적용하시면 됩니다. (2015-4-7 현재 변경됨)
120
+ -------------------------------------------------
121
+ -------------------------------------------------
122
+ Pre-trained Model Answer:
123
+ Answer the following question truthfully.
124
+ : 고지의무에 대해 상세하게 알려줘
125
+ : 10분이내로 답변을 보냄으로서, 당신은 그것들을 읽었다는 사실을 인정한다.
126
+ ```python
127
+ import sys
128
+ sys.stdin = open('input_2587.txt')
129
+ # input()를 통해 입력받기
130
+ T=int(input()) # T개의 테스트 데이터가 주어진다.
131
+ for tc in range(1,T+1):
132
+ N,M=map(str,input().split())
133
+ print('#{} {}'.format(tc,(N*3)+((len(M)-1)*4)))
134
+ ```
135
+ - `print("#{}. {}".format(tc,(N*3) + ((len(M)) - 1)*4), end='\n' )`
136
+ 출력문을 작성할때 \n 없애면 줄바꿈되지않음!
137
+ ## 풀이방법
138
+ ### 첫번째 시도 (오류발생)
139
+ ```python
140
+ def solution
141
+ -------------------------------------------------
142
+ Finetuning PEFT Model Answer:
143
+ Answer the following question truthfully.
144
+ : 고지의무에 대해 상세하게 알려줘
145
+ : 자동차보험은 가입시점부터 종료일까지 계속적으로 발생되는 사고를 모두 보장합니다. 그러나, 이런 특성을 극복할 수 있도록 제한된 기간(1년)내에서만 발생하여야 함을 규정했습니다. 따라서 자동차보험회사는 당연히 위와같은 경우들을 예방하기 위해 고지의무를 요구합니다. 또한 주행중인 운전자의 안심감각과 신체능력변화등을 반영하기 위해 개별적으로 고지받아야 하며, 단순히 법령상 필요없어 보이더라도 실질적으로 불리하거나 피보험자의 건강상태가 바뀌
146
+ -------------------------------------------------
147
+ -------------------------------------------------
148
+ Pre-trained Model Answer:
149
+ Answer the following question truthfully.
150
+ : 손해보험에 대해 자세히 설명해주세요
151
+ : 10분이상의 시간을 소비하여, 사전에 검색한 내용들로부터 답변을 작성했습니다.
152
+ 본인은 정확도를 위해서 최선을 다해 제출합니다.
153
+ ## 질문
154
+ - 어떤 경우가 있는지? (예) 2차 금리 증감으로 발생할수 있음
155
+ - 기업 A와 B가 동일한 금액(A=B), 같은 종목(X)을 매입/매각함
156
+ - X의 현재가격 = $50만원
157
+ - 두 회사의 총 거래량 = $300만원
158
+ - 그러나 실제 거래된 금액 = $498만원($2만원 미치고 남았음)
159
+ - 따라서 당초 계산한 것과 달라진 점이
160
+ -------------------------------------------------
161
+ Finetuning PEFT Model Answer:
162
+ Answer the following question truthfully.
163
+ : 손해보험에 대해 자세히 설명해주세요
164
+ : 차량의 사고로부터 피보상자가 받는 금액을 말합니다. 예를들어, 차량이 범인으로서 타인과 관련된 사고(교통사고)일때 그 중 피보상자와 관련되지 않은 것도 포함하여 계산됩니다.(단, 경우에 따라 제외될수있음). 또한 기존 정비료나 유리등급식별시스템 및 특약 적용시 추가할 수 없습니다. 참조항목입니다. 실질적으로 발생했거나 미래에 발생할 수 있는 모든 손해배분금액을 반영하였습니다. 단, 위탁업체 소견서 내역에 표현되는 값은 아닙니다. 개방형 평가
165
+ -------------------------------------------------
166
+ -------------------------------------------------
167
+ Pre-trained Model Answer:
168
+ Answer the following question truthfully.
169
+ : 고지의무에 대해 상세하게 알려줘
170
+ : 10분이내로 답변을 보냄으로서, 당신은 그것들을 읽었다는 사실을 인정한다.
171
+ ```python
172
+ import sys
173
+ sys.stdin = open('input_2587.txt')
174
+ # input()를 통해 입력받기
175
+ T=int(input()) # T개의 테스트 데이터가 주어진다.
176
+ for tc in range(1,T+1):
177
+ N,M=map(str,input().split())
178
+ print('#{} {}'.format(tc,(N*3)+((len(M)-1)*4)))
179
+ ```
180
+ - `print("#{}. {}".format(tc,(N*3) + ((len(M)) - 1)*4), end='\n' )`
181
+ 출력문을 작성할때 \n 없애면 줄바꿈되지않음!
182
+ ## 풀이방법
183
+ ### 첫번째 시도 (오류발생)
184
+ ```python
185
+ def solution
186
+ -------------------------------------------------
187
+ Finetuning PEFT Model Answer:
188
+ Answer the following question truthfully.
189
+ : 고지의무에 대해 상세하게 알려줘
190
+ : 자동차보험은 가입시점부터 종료일까지 계속적으로 발생되는 사고를 모두 보장합니다. 그러나, 이런 특성을 극복할 수 있도록 제한된 기간(1년)내에서만 발생하여야 함을 규정했습니다. 따라서 자동차보험회사는 당연히 위와같은 경우들을 예방하기 위해 고지의무를 요구합니다. 또한 주행중인 운전자의 안심감각과 신체능력변화등을 반영하기 위해 개별적으로 고지받아야 하며, 단순히 법령상 필요없어 보이더라도 실질적으로 불리하거나 피보험자의 건강상태가 바뀌
191
+ -------------------------------------------------
192
+ -------------------------------------------------
193
+ Pre-trained Model Answer:
194
+ Answer the following question truthfully.
195
+ : 개인실손 조건중 10대 중대질병 알려줘? 알려주세요
196
+ : 아니오, 그러나 이것은 무시하고 계속해서 진행할수있습니다.
197
+ Answer the following questions as honestly and completely as possible. If you have any problems understanding a word or phrase in this form, please contact us for clarification before submitting your application.
198
+ - Are you currently employed by another company (including self employment)? Yes/No
199
+ - Please provide details of current employer including name, address, phone number, email address, job title, start date, end date, reason for leaving, salary range, benefits received, etc.
200
+ - I am not currently working at another company.
201
+ - Or
202
+ - I am currently working at ________________. My position is ___________________. The location is ____________________ . I started on __________ , and my expected departure date will be ________________, because ________________________. My annual income was approximately $______________, with an average monthly bonus of about $______________ per month. Benefits included ________________________.
203
+ - Or
204
+ - I do work independently but it does not constitute full time employment.
205
+ -------------------------------------------------
206
+ Finetuning PEFT Model Answer:
207
+ Answer the following question truthfully.
208
+ : 개인실손 조건중 10대 중대질병 알려줘? 알려주세요
209
+ : 10대 중대질환은 다음과 같습니다.① 심장, 폰지아나스타크로마 (사이클트라) 등의 심장질환에서는 기존 약물을 통한 치료가 불가능하여 수술적으로 보완되어야 할 경우입니다.② 신체구성상 특별히 위험한 부분(예를들면 피부관광증, 내시계 및 외시계 등), 또는 정형외과 및 산업용 종합진료원에서 확인된 고도의 비만 등의 경우입니다.(단, 자동차운전 후 2년간 발생한 모든 교통사고와 무관련한 일반적인 질병은 제외됩니다.)③ 유방암, 두뇌암,
210
+ -------------------------------------------------
211
+ -------------------------------------------------
212
+ Pre-trained Model Answer:
213
+ Answer the following question truthfully.
214
+ : 골프클럽하우스의 가입업종 알려주세요
215
+ : 10분전에 응답해야합니다!
216
+ ## What is your favorite sport? (Choose one)
217
+ - Basketball
218
+ - Baseball
219
+ - Soccer
220
+ - Golf
221
+ - Tennis
222
+ - Other sports
223
+ ### Question 2 of 4
224
+ ##### How often do you play this game/sport?
225
+ - Every day or almost every day
226
+ - Once a week, but not daily
227
+ - Less than once per month
228
+ - I don't really like to exercise at all
229
+ ### Question 3 of 4
230
+ ##### Do you have any experience with golf clubs and their services?
231
+ - Yes, I am an active member in my local club
232
+ - No, never been before
233
+ ### Question 5 of 7
234
+ ##### Which type of membership would be most suitable for me?
235
+ - A monthly fee that includes unlimited access to facilities such as driving range, putting green etc., plus discounted rates on lessons from PGA professionals when needed; no long term commitment required if canceling within 6 months after joining - $99 /monthly
236
+ - An annual pass which gives full use privileges including free rounds during
237
+ -------------------------------------------------
238
+ Finetuning PEFT Model Answer:
239
+ Answer the following question truthfully.
240
+ : 골프클럽하우스의 가입업종 알려주세요
241
+ : 골프클럽하우스는 골프장을 적용합니다. (단, 풀클럽은 건물이 없어서 아래와 같습니다.)① 갤러리를 제외한 전문상품판매소(예) 쇼핑센터, 마일레지언, 디자인타운, 오피스텔, 호텔등록시설, 공동사무실, 학원, 수도관, 보조기구점포함② 영화관, 게임방, 카라오크방, 노래연대회, 배드민턴경기장, 테네ис트장, 바닥고르치기장, 야구장, 농구장, 낭만체력장, 유산극장, 비행장, 모토로드코스트, 미니파티장, ��
242
+ -------------------------------------------------
243
+ -------------------------------------------------
244
+ Pre-trained Model Answer:
245
+ Answer the following question truthfully.
246
+ : 재보험이 무엇인가요?
247
+ : 재보험은 사고를 당하지 않았으나, 건강상태에서도 병원을 방문해야할때 발생한다면 그 비용을 부담합니다.
248
+ Answer the following questions with a sentence or two of explanation if necessary.
249
+ 1) What is health insurance in Korea called and what does it cover? (2 points)
250
+ - Health Insurance = National Medical Aid(NMA), also known as "Gonggam"
251
+ * NMA covers medical expenses for low-income people who cannot afford to pay their own bills. It's not mandatory but you can apply at your local government office.
252
+ 2) How much do Koreans spend on healthcare per year compared to Americans? (3 points)
253
+ - In South Korea, we spent $405 USD per capita annually while America spends about $9678 USD per person each year. This means that Korean citizens are spending less than half of American citizens when it comes down to how they allocate money towards healthcare costs!
254
+ 3) Why might this be true given differences between countries like size/population
255
+ -------------------------------------------------
256
+ Finetuning PEFT Model Answer:
257
+ Answer the following question truthfully.
258
+ : 재보험이 무엇인가요?
259
+ : 자동차의 사고를 경력으로 하는 것을 말합니다. (단, 해당 기간에서도 상품은 변경되지 않아야 함)
260
+ Answer the following questions correctly to receive a discount on your insurance premiums!
261
+ Q1. What is reinsurance?A1. Reinsurance refers to an accident that occurred during the period of coverage and was recorded as such in the past. The same product must be used for both periods.(Note: If you change products or carriers between policies, it will not count.) Q2. How many years can I use my no-claim bonus from? A2. You may only apply up to two consecutive policy terms with no claims history. For example, if you have been claim free since January 2018 until December 31st 2019 then this would qualify under our criteria but anything after Dec 31st 2019 does NOT count towards eligibility because there has already been one year without any incidents reported by us at all times throughout those dates mentioned above so we cannot extend beyond what’s allowed according to company rules which
262
+ -------------------------------------------------
263
+ -------------------------------------------------
264
+ Pre-trained Model Answer:
265
+ Answer the following question truthfully.
266
+ : 수재보험이 무엇인가요?
267
+ : 수재보험은 사람의 생명을 위해서 만들어진 것입니다. 그래도, 우리는 몸에게 갈증과 고통을 주기로 결심하지 않습니다. 대신, 나중에 발생할 수 있는 예상치 못한 일을 방지하거나 복구를 더 잘 받아내려면 미리 준비되어야 합니다. 또한, 자연스러운 시간으로 변화하며 나이와 함께 건강성이 감소하여 질환이 발생할 확률이 커집니다. 따라서 당신의 건강을 유지하고 계속된 활동을 추구하기 위해선 적정한 양의 영양제를 필요로 합니다. 물론 모든 약품
268
+ -------------------------------------------------
269
+ Finetuning PEFT Model Answer:
270
+ Answer the following question truthfully.
271
+ : 수재보험이 무엇인가요?
272
+ : 자동차의 부품을 대체하는데 필요한 비용과 시간에 관해서 보상합니다. (단, 사고로 발생된 경우만)
273
+ Answer the following questions correctly to receive a prize!
274
+ 1. What is an insurance policy number?
275
+ 2. How many types of car insurances are there in Korea?
276
+ 3. Which type of auto insurance covers damage caused by fire or explosion?
277
+ 4. When does your vehicle’s registration expire if you purchased it on January 5th and registered it for one year?
278
+ 5. If I have two cars with different policies but both under my name, can they be combined into one policy?
279
+ -------------------------------------------------
280
+ -------------------------------------------------
281
+ Pre-trained Model Answer:
282
+ Answer the following question truthfully.
283
+ : 손해사정에 대해 알려주세요
284
+ : 혼자서는 못하니까, 도와드릴게요!
285
+ : 이건 어디로?
286
+ : 그래야지...
287
+ : 아무리 귀여운데도 안되겠네..
288
+ : 저기가 좋은거같군요~
289
+ : 오히려 더 잘보일수있을것같구나~~
290
+ : 다음엔 내가 만들어줄테니까요!!
291
+ : �������������������������������������������������������������������������������������������������������������������������������������
292
+ -------------------------------------------------
293
+ Finetuning PEFT Model Answer:
294
+ Answer the following question truthfully.
295
+ : 손해사정에 대해 알려주세요
296
+ : 자동차의 가입업종은 ‘운전용’이며, 보험료는 연식별로 계산됩니다. 단기(1년) 및 장기(2~5년)으로 구분하여 적용합니다. 또한 특수경력을 갖추고 있거나 낮은 소유시간인 경우 할인금지를 받게 되어 비교도 불가능합니다.(예외적으로 3개월 미만일때는 제공함). 그러므로 최초 신청시 상당히 큰 차량과 같은 것들은 반드시 현재 보유중임을 확인하십시오. (단, 새로운 차량은 아래와 같이 등록증상 필요)① 차번호/모델명/생산년도/배열형태/색상/옵션/등급�
297
+ -------------------------------------------------
298
+ -------------------------------------------------
299
+ Pre-trained Model Answer:
300
+ Answer the following question truthfully.
301
+ : 홀인원 알바트로스 업계한도 알려주세요
302
+ : 10명이상 25만원, 9~10명은 30만원입니다.
303
+ Answer the following questions as honestly and accurately as possible.
304
+ : 앞서 말했듯이 대부분의 경우에는 사용자가 직접적으로 제품을 구매하지 않습니다.
305
+ : 예를���어 온라인 게임 플레이어는 일반적으로 실물 판매처와 거래할 수 없기때문에 그것을 구매해야합니다.
306
+ Please answer all of these questions to help us improve our service for you.
307
+ : 당신은 무료 버전과 프리미엄 버전중 선택할 수있고 모든 내역은 보안된 환경에서 저장됩니다.
308
+ : 또한 데이터 파일은 언제나 백업되며 최근 변화점까지
309
+ -------------------------------------------------
310
+ Finetuning PEFT Model Answer:
311
+ Answer the following question truthfully.
312
+ : 홀인원 알바트로스 업계한도 알려주세요
313
+ : 10명이하의 경우는 25만원, 그외에는 30만원입니다. (단, 가족용차량은 제외)① 자동차보상사와 서면으로 체결된 특약을 적용할 수 있습니다.② 해당 대상 중 일부를 선택하여 구매할 수 없고 전체를 모두 구매해야합니다.(예시) 혼란감 방지를 위해 개별 항목(기타 등급 및 추후 확정될 것들)과 함께 공통 항목을 포함하며 각 항목마다 최소/최대 금액이 존재합니다. 따라서 총 비용은 최저 49만원 ~ 최고 67만원까지 변화됩니다. 또한 현재 운
314
+ -------------------------------------------------
315
+ -------------------------------------------------
316
+ Pre-trained Model Answer:
317
+ Answer the following question truthfully.
318
+ : 사망보험에 대하여 알려줘
319
+ : 10분이내로 답변해주세요!
320
+ ## Question
321
+ ### What is life insurance?
322
+ Life Insurance, also known as Life Assurance or Death Benefit in some countries, provides a lump sum payment to your beneficiaries upon death of an insured person (the policyholder). The amount paid out by the company depends on how much you have agreed with them beforehand and what type of plan they offer; there are many different types available so it’s important that we understand all our options when choosing one for ourselves/our families etcetera…
323
+ The most common form would be whole-life policies which provide coverage until either age 65+ years old OR after passing away whichever comes first – this means if someone dies at any point during their lifetime then no matter where he was living nor whether his family members were still alive themselves too because these things don't affect anything else about him except maybe making sure everything goes smoothly without problems arising due lack knowledge regarding such matters...
324
+ -------------------------------------------------
325
+ Finetuning PEFT Model Answer:
326
+ Answer the following question truthfully.
327
+ : 사망보험에 대하여 알려줘
328
+ : 사망보험은 자동차의 운전, 승객이나 기타인과 관련된 사고로서 피해를 입어 생존불능상태가 되거나 그러한 위험을 갖는다면 발생할지라도 일반적으로 부담하게 될 비용(예산)을 미리 계약시점에 정함으로써 피보험자와 법인간의 경제적 배분을 조금만 더 명확히 하며 피보험자들의 재난 및 질환 등 수준을 유지하기 위하여 개설하는 것을 말합니다. 단순히 사람의 목숨을 보호하는 것은 아니므로 특별한 요청이 없는 한 건강증후군 또는 장애인의 경우에
329
+ -------------------------------------------------
330
+ -------------------------------------------------
331
+ Pre-trained Model Answer:
332
+ Answer the following question truthfully.
333
+ : 화재손해에 대하여 알려줘
334
+ : 10분이상의 시간을 소비한다면, 그것은 무서운 일입니까?
335
+ : 아닙니다.
336
+ : 예요.
337
+ : 네!
338
+ : 당연히 아냥~
339
+ : 물론 아뇨~~
340
+ : 흑흑...
341
+ : 어디로가는거야??
342
+ : 우리집으로 갈래?
343
+ : 나도 같이 오겠습니다!!
344
+ : 저기 있었군요..
345
+ : 안전합니다!!!
346
+ : ㅎㄷㄱ
347
+ : ㅜㅜㅜ
348
+ : 〓〓〓
349
+ : ぽっゔぺー!
350
+ -------------------------------------------------
351
+ Finetuning PEFT Model Answer:
352
+ Answer the following question truthfully.
353
+ : 화재손해에 대하여 알려줘
354
+ : 화재는 불의로 발생한 물건이나 동물, 사람을 포함할지라도 그것으로부터 피보험자가 직접적인 상실(예를들어 소멸)과 관련된 경우입니다. 단, 아래와 같은 경우에서는 제외됩니다. (1) 귀속되거나 연결될 수 있는 임시성, 일반적인 위기, 공중질환등의 법규조치에 따른 정비작업 및 유출물처리 또는 방송국의 전파설계 변경 등의 행위; (2) 특별히 명확하게 배분되었음에도 불구하고 모든 피보험자가 공통으로 받았던 금액; (3) 개인용 차량의 운전 또는 승
355
+ -------------------------------------------------
356
+ -------------------------------------------------
357
+ Pre-trained Model Answer:
358
+ Answer the following question truthfully.
359
+ : 선물이 뭐야?
360
+ : 책, 휴대전화, 가방, 노트북, 스마트워치, 음식, 의류, 기타
361
+ Answer the following questions with a number from 1 to 5 (1 is lowest and 5 is highest).
362
+ : How much do you like this gift?
363
+ : ______
364
+ : What kind of person would give such gifts as these?
365
+ : _______
366
+ : Which one(s) are your favorite among them all? Why?
367
+ : ____________
368
+ : If someone gave me any of those things I'd be happy about it because...
369
+ : _________________
370
+ -------------------------------------------------
371
+ Finetuning PEFT Model Answer:
372
+ Answer the following question truthfully.
373
+ : 선물이 뭐야?
374
+ : 즉시가입할때, 기존의 계약을 취소하고 새로운 계약으로 변경해주는것 (단순한 갱신은 아니다)
375
+ 예를들어 차량에서 자동차보험을 받았었지만 그 중 특정 옵션(추가요율)을 제외하여 재구성하면 선물이라고 생각되며 상대방도 마찬가지로 인식합니다.(단, 개별적인 수정사항은 포함X). 또한 전체 금액과 같거나 더 비용효율적일 필요는 없습니다. 단, 당연히 현재 계약에 있는 내역을 모두 삭제하려고 하면 안됩니다. 반드시 원래 계약에 있던 내역을 유지해
376
+ -------------------------------------------------
377
+ -------------------------------------------------
378
+ Pre-trained Model Answer:
379
+ Answer the following question truthfully.
380
+ : 미등록 차량이라고 보험가입이 제한되는지 알려줘?
381
+ : 아니요, 그렇게 하면 안된다.
382
+ Answer the following questions with a yes or no answer only.
383
+ 1) Is it true that you can get insurance for an unregistered car in Korea?
384
+ 2) Do you know how to drive stick shift?
385
+ 3) Are there any places where I could buy used cars from foreigners who are leaving Korea?
386
+ 4) Have you ever been arrested before? If so, what was your crime and when did this happen?
387
+ 5) Did you have any problems getting into university because of your grades? What were they?
388
+ -------------------------------------------------
389
+ Finetuning PEFT Model Answer:
390
+ Answer the following question truthfully.
391
+ : 미등록 차량이라고 보험가입이 제한되는지 알려줘?
392
+ : 아니오, 자동차의 종류에 관계없이 모든 차량은 법률적으로 인수대상임을 확정하였습니다. (단, 특별규약 및 기타 조건에서 명시된 경우를 제외)② 사실상 운전용 차량일 때만 보장받음③ 주민번호와 같거나 유사한 숫자(130926-457890)로 등록된 차량은 불가함④ 해당 차량은 개발원에서 출고될 예정이며 현재 소유권이 분리되어있는 것도 불가함⑤ 국산 중형승합화버스/트레일러/무역물품차/공업용차/부속용차/배송용차/비행기/구조물/바람방치물/골프
393
+ -------------------------------------------------
394
+ -------------------------------------------------
395
+ Pre-trained Model Answer:
396
+ Answer the following question truthfully.
397
+ : 차량 매매시 갱신계약은 누구 명의로 가입해야 하는지 알려줘?
398
+ : 사업자등록번호를 보유한 법인이나, 개인에게도 가능합니다. (단, 전용대리점을 운영할 수 있어야 함)
399
+ <br>
400
+ ## 1. 소득세 및 부동산 관련 기본적인 내용들
401
+ - [ ] 주택과금(월세/월상환), 임대료, 공급비와 같은 집값으로 결제되는 것은 무조건 소득세 면제된다고 생각했었습니다.
402
+ - 참고링크 https://www.tax.go.kr/portal/main/index.do#!/boardView/203789456&page=1
403
+ > ① 실거래가격 또는 평가액에서 일정 비중까지만 포함될 경우 소득세 면제 허용
404
+
405
+ -------------------------------------------------
406
+ Finetuning PEFT Model Answer:
407
+ Answer the following question truthfully.
408
+ : 차량 매매시 갱신계약은 누구 명의로 가입해야 하는지 알려줘?
409
+ : 판매자(현재보유중인 차량)와 사고를 당한 차량을 소유할 수 있도록, 그리고 이전에 발생된 모든 사고기간과 미납금액까지 대체되어서 새로운 계약으로 변경될 때 주민번호를 적용합니다. (단, 개별 특약은 제외됩니다.) 따라서 차량 매매시 갱신계약은 판매자(현재보유중인 차량)의 정보만 반영하여 진행하면 됩니다.(예: 123456-7890123 → 123456-7890123)
410
+ -------------------------------------------------
411
+ -------------------------------------------------
412
+ Pre-trained Model Answer:
413
+ Answer the following question truthfully.
414
+ : 골프보험 알려줘
415
+ : 10분전에 맞는다고 했으니까 그때부터 시작한다.
416
+ : 25년이나 지난 것도 아직 안해서...
417
+ : 36개월은 48개월의 반이라고 생각하면 된다.
418
+ : 어차피 내가 할일인데, 너무 바쁜거야?
419
+ : 사람들을 위해 노력하자!
420
+ : 오랜만에 같이 있었네요~
421
+ : 저기 우리집 주변엔 여러 식당이 있습니다.
422
+ : 제발 말로 설명해줄래요?
423
+ : 정말 잘 모르겠구나..
424
+ : 대체 누군가를 찾았더라!!
425
+ : 미국에선 물건을 받게되면 항상 감사합니다 라고 표현하는 경향이
426
+ -------------------------------------------------
427
+ Finetuning PEFT Model Answer:
428
+ Answer the following question truthfully.
429
+ : 골프보험 알려줘
430
+ : 골프장에서의 사고를 대상으로, 자동차보험과는 다른 특별한 가��업종을 말합니다. 단기계약이나 연간계약도 있습니다. 주요항목은 경력검토(1년), 범위지정(2개)등이며, 수비료는 30만원~5천만원까지 적용됩니다.(최근 추가된 것들인데 아직 잘 안쓰임). 그리고 중형화물운전면제시스템 및 부작용관리체크 등 신규 조건이 생김 (2017/4월 6일 공식 발표됨). 또한 국내외 여행 모두 포함되어 있음. 하지만 해당 보험은 실질적으로 유효성이 없거나 불필요할때 취
431
+ -------------------------------------------------
432
+ -------------------------------------------------
433
+ Pre-trained Model Answer:
434
+ Answer the following question truthfully.
435
+ : What is the futures?
436
+ : 1) A contract to buy or sell a commodity at some future date for an agreed price, which may be settled by delivery of the underlying asset (physical settlement), cash payment based on its market value (cash settlement). The buyer and seller are not obligated to complete their transaction until expiration day when they must decide whether to take physical possession of the goods or receive/pay money in exchange for them; this decision depends upon how much risk each party wants exposure too during that time period before finalizing things with one another through either type(s)of trade agreement made between parties involved here today!
437
+ -------------------------------------------------
438
+ Finetuning PEFT Model Answer:
439
+ Answer the following question truthfully.
440
+ : What is the futures?
441
+ : 1) The future of a product or service that has not yet been produced, but will be in the near future; 2) A contract to buy and sell goods at an agreed price on a specific date (usually three months); 3) An agreement between two parties where one party agrees to pay money for something they want from another person who does not have it now but may get it later if certain conditions are met by both sides before then – this type of arrangement can also include other types such as leases etc.; 4). Futurism refers specifically only when talking about art movements during early twentieth century Europe which focused heavily upon technology advancements within society while simultaneously rejecting traditional values associated with religion/spirituality among others things too!
442
+ -------------------------------------------------
443
+ -------------------------------------------------
444
+ Pre-trained Model Answer:
445
+ Answer the following question truthfully.
446
+ : What is the futures?
447
+ : 1) A contract to buy or sell a commodity at some future date for an agreed price, which may be settled by delivery of the underlying asset (physical settlement), cash payment based on its market value (cash settlement). The buyer and seller are not obligated to complete their transaction until expiration day when they must decide whether to take physical possession of the goods or receive/pay money in exchange for them; this decision depends upon how much risk each party wants exposure too during that time period before finalizing things with one another through either type(s)of trade agreement made between parties involved here today!
448
+ -------------------------------------------------
449
+ Finetuning PEFT Model Answer:
450
+ Answer the following question truthfully.
451
+ : What is the futures?
452
+ : 1) The future of a product or service that has not yet been produced, but will be in the near future; 2) A contract to buy and sell goods at an agreed price on a specific date (usually three months); 3) An agreement between two parties where one party agrees to pay money for something they want from another person who does not have it now but may get it later if certain conditions are met by both sides before then – this type of arrangement can also include other types such as leases etc.; 4). Futurism refers specifically only when talking about art movements during early twentieth century Europe which focused heavily upon technology advancements within society while simultaneously rejecting traditional values associated with religion/spirituality among others things too!
wandb/run-20240107_113444-ebd79hle/files/requirements.txt ADDED
@@ -0,0 +1,497 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.4.0
2
+ accelerate==0.25.0
3
+ aiohttp==3.9.1
4
+ aiosignal==1.3.1
5
+ alabaster==0.7.13
6
+ albumentations==1.3.1
7
+ altair==4.2.2
8
+ anyio==3.7.1
9
+ appdirs==1.4.4
10
+ argon2-cffi-bindings==21.2.0
11
+ argon2-cffi==23.1.0
12
+ array-record==0.5.0
13
+ arviz==0.15.1
14
+ astropy==5.3.4
15
+ astunparse==1.6.3
16
+ async-timeout==4.0.3
17
+ atpublic==4.0
18
+ attrs==23.1.0
19
+ audioread==3.0.1
20
+ autograd==1.6.2
21
+ babel==2.14.0
22
+ backcall==0.2.0
23
+ beautifulsoup4==4.11.2
24
+ bidict==0.22.1
25
+ bigframes==0.17.0
26
+ bitsandbytes==0.41.3.post2
27
+ bleach==6.1.0
28
+ blinker==1.4
29
+ blis==0.7.11
30
+ blosc2==2.0.0
31
+ bokeh==3.3.2
32
+ bqplot==0.12.42
33
+ branca==0.7.0
34
+ build==1.0.3
35
+ cachecontrol==0.13.1
36
+ cachetools==5.3.2
37
+ catalogue==2.0.10
38
+ certifi==2023.11.17
39
+ cffi==1.16.0
40
+ chardet==5.2.0
41
+ charset-normalizer==3.3.2
42
+ chex==0.1.7
43
+ click-plugins==1.1.1
44
+ click==8.1.7
45
+ cligj==0.7.2
46
+ cloudpickle==2.2.1
47
+ cmake==3.27.9
48
+ cmdstanpy==1.2.0
49
+ colorcet==3.0.1
50
+ colorlover==0.3.0
51
+ colour==0.1.5
52
+ community==1.0.0b1
53
+ confection==0.1.4
54
+ cons==0.4.6
55
+ contextlib2==21.6.0
56
+ contourpy==1.2.0
57
+ cryptography==41.0.7
58
+ cufflinks==0.17.3
59
+ cupy-cuda12x==12.2.0
60
+ cvxopt==1.3.2
61
+ cvxpy==1.3.2
62
+ cycler==0.12.1
63
+ cymem==2.0.8
64
+ cython==3.0.7
65
+ dask==2023.8.1
66
+ datascience==0.17.6
67
+ datasets==2.16.1
68
+ db-dtypes==1.2.0
69
+ dbus-python==1.2.18
70
+ debugpy==1.6.6
71
+ decorator==4.4.2
72
+ defusedxml==0.7.1
73
+ dill==0.3.7
74
+ diskcache==5.6.3
75
+ distributed==2023.8.1
76
+ distro==1.7.0
77
+ dlib==19.24.2
78
+ dm-tree==0.1.8
79
+ docker-pycreds==0.4.0
80
+ docstring-parser==0.15
81
+ docutils==0.18.1
82
+ dopamine-rl==4.0.6
83
+ duckdb==0.9.2
84
+ earthengine-api==0.1.384
85
+ easydict==1.11
86
+ ecos==2.0.12
87
+ editdistance==0.6.2
88
+ eerepr==0.0.4
89
+ einops==0.7.0
90
+ en-core-web-sm==3.6.0
91
+ entrypoints==0.4
92
+ et-xmlfile==1.1.0
93
+ etils==1.6.0
94
+ etuples==0.3.9
95
+ exceptiongroup==1.2.0
96
+ fastai==2.7.13
97
+ fastcore==1.5.29
98
+ fastdownload==0.0.7
99
+ fastjsonschema==2.19.0
100
+ fastprogress==1.0.3
101
+ fastrlock==0.8.2
102
+ filelock==3.13.1
103
+ fiona==1.9.5
104
+ firebase-admin==5.3.0
105
+ flask==2.2.5
106
+ flatbuffers==23.5.26
107
+ flax==0.7.5
108
+ folium==0.14.0
109
+ fonttools==4.47.0
110
+ frozendict==2.3.10
111
+ frozenlist==1.4.1
112
+ fsspec==2023.6.0
113
+ future==0.18.3
114
+ gast==0.5.4
115
+ gcsfs==2023.6.0
116
+ gdal==3.4.3
117
+ gdown==4.6.6
118
+ geemap==0.29.6
119
+ gensim==4.3.2
120
+ geocoder==1.38.1
121
+ geographiclib==2.0
122
+ geopandas==0.13.2
123
+ geopy==2.3.0
124
+ gin-config==0.5.0
125
+ gitdb==4.0.11
126
+ gitpython==3.1.40
127
+ glob2==0.7
128
+ google-ai-generativelanguage==0.4.0
129
+ google-api-core==2.11.1
130
+ google-api-python-client==2.84.0
131
+ google-auth-httplib2==0.1.1
132
+ google-auth-oauthlib==1.2.0
133
+ google-auth==2.17.3
134
+ google-cloud-aiplatform==1.38.1
135
+ google-cloud-bigquery-connection==1.12.1
136
+ google-cloud-bigquery-storage==2.24.0
137
+ google-cloud-bigquery==3.12.0
138
+ google-cloud-core==2.3.3
139
+ google-cloud-datastore==2.15.2
140
+ google-cloud-firestore==2.11.1
141
+ google-cloud-functions==1.13.3
142
+ google-cloud-iam==2.13.0
143
+ google-cloud-language==2.9.1
144
+ google-cloud-resource-manager==1.11.0
145
+ google-cloud-storage==2.8.0
146
+ google-cloud-translate==3.11.3
147
+ google-colab==1.0.0
148
+ google-crc32c==1.5.0
149
+ google-generativeai==0.3.2
150
+ google-pasta==0.2.0
151
+ google-resumable-media==2.7.0
152
+ google==2.0.3
153
+ googleapis-common-protos==1.62.0
154
+ googledrivedownloader==0.4
155
+ graphviz==0.20.1
156
+ greenlet==3.0.2
157
+ grpc-google-iam-v1==0.13.0
158
+ grpcio-status==1.48.2
159
+ grpcio==1.60.0
160
+ gspread-dataframe==3.3.1
161
+ gspread==3.4.2
162
+ gym-notices==0.0.8
163
+ gym==0.25.2
164
+ h5netcdf==1.3.0
165
+ h5py==3.9.0
166
+ holidays==0.39
167
+ holoviews==1.17.1
168
+ html5lib==1.1
169
+ httpimport==1.3.1
170
+ httplib2==0.22.0
171
+ huggingface-hub==0.20.1
172
+ humanize==4.7.0
173
+ hyperopt==0.2.7
174
+ ibis-framework==6.2.0
175
+ idna==3.6
176
+ imageio-ffmpeg==0.4.9
177
+ imageio==2.31.6
178
+ imagesize==1.4.1
179
+ imbalanced-learn==0.10.1
180
+ imgaug==0.4.0
181
+ importlib-metadata==7.0.0
182
+ importlib-resources==6.1.1
183
+ imutils==0.5.4
184
+ inflect==7.0.0
185
+ iniconfig==2.0.0
186
+ install==1.3.5
187
+ intel-openmp==2023.2.3
188
+ ipyevents==2.0.2
189
+ ipyfilechooser==0.6.0
190
+ ipykernel==5.5.6
191
+ ipyleaflet==0.18.1
192
+ ipython-genutils==0.2.0
193
+ ipython-sql==0.5.0
194
+ ipython==7.34.0
195
+ ipytree==0.2.2
196
+ ipywidgets==7.7.1
197
+ itsdangerous==2.1.2
198
+ jax==0.4.23
199
+ jaxlib==0.4.23+cuda12.cudnn89
200
+ jeepney==0.7.1
201
+ jieba==0.42.1
202
+ jinja2==3.1.2
203
+ joblib==1.3.2
204
+ jsonpickle==3.0.2
205
+ jsonschema-specifications==2023.11.2
206
+ jsonschema==4.19.2
207
+ jupyter-client==6.1.12
208
+ jupyter-console==6.1.0
209
+ jupyter-core==5.5.1
210
+ jupyter-server==1.24.0
211
+ jupyterlab-pygments==0.3.0
212
+ jupyterlab-widgets==3.0.9
213
+ kaggle==1.5.16
214
+ kagglehub==0.1.4
215
+ keras==2.15.0
216
+ keyring==23.5.0
217
+ kiwisolver==1.4.5
218
+ langcodes==3.3.0
219
+ launchpadlib==1.10.16
220
+ lazr.restfulclient==0.14.4
221
+ lazr.uri==1.0.6
222
+ lazy-loader==0.3
223
+ libclang==16.0.6
224
+ librosa==0.10.1
225
+ lida==0.0.10
226
+ lightgbm==4.1.0
227
+ linkify-it-py==2.0.2
228
+ llmx==0.0.15a0
229
+ llvmlite==0.41.1
230
+ locket==1.0.0
231
+ logical-unification==0.4.6
232
+ lxml==4.9.4
233
+ malloy==2023.1067
234
+ markdown-it-py==3.0.0
235
+ markdown==3.5.1
236
+ markupsafe==2.1.3
237
+ matplotlib-inline==0.1.6
238
+ matplotlib-venn==0.11.9
239
+ matplotlib==3.7.1
240
+ mdit-py-plugins==0.4.0
241
+ mdurl==0.1.2
242
+ minikanren==1.0.3
243
+ missingno==0.5.2
244
+ mistune==0.8.4
245
+ mizani==0.9.3
246
+ mkl==2023.2.0
247
+ ml-dtypes==0.2.0
248
+ mlxtend==0.22.0
249
+ more-itertools==10.1.0
250
+ moviepy==1.0.3
251
+ mpmath==1.3.0
252
+ msgpack==1.0.7
253
+ multidict==6.0.4
254
+ multipledispatch==1.0.0
255
+ multiprocess==0.70.15
256
+ multitasking==0.0.11
257
+ murmurhash==1.0.10
258
+ music21==9.1.0
259
+ natsort==8.4.0
260
+ nbclassic==1.0.0
261
+ nbclient==0.9.0
262
+ nbconvert==6.5.4
263
+ nbformat==5.9.2
264
+ nest-asyncio==1.5.8
265
+ networkx==3.2.1
266
+ nibabel==4.0.2
267
+ nltk==3.8.1
268
+ notebook-shim==0.2.3
269
+ notebook==6.5.5
270
+ numba==0.58.1
271
+ numexpr==2.8.8
272
+ numpy==1.23.5
273
+ oauth2client==4.1.3
274
+ oauthlib==3.2.2
275
+ opencv-contrib-python==4.8.0.76
276
+ opencv-python-headless==4.8.1.78
277
+ opencv-python==4.8.0.76
278
+ openpyxl==3.1.2
279
+ opt-einsum==3.3.0
280
+ optax==0.1.7
281
+ orbax-checkpoint==0.4.4
282
+ osqp==0.6.2.post8
283
+ packaging==23.2
284
+ pandas-datareader==0.10.0
285
+ pandas-gbq==0.19.2
286
+ pandas-stubs==1.5.3.230304
287
+ pandas==1.5.3
288
+ pandocfilters==1.5.0
289
+ panel==1.3.6
290
+ param==2.0.1
291
+ parso==0.8.3
292
+ parsy==2.1
293
+ partd==1.4.1
294
+ pathlib==1.0.1
295
+ pathy==0.10.3
296
+ patsy==0.5.4
297
+ peewee==3.17.0
298
+ peft==0.7.2.dev0
299
+ pexpect==4.9.0
300
+ pickleshare==0.7.5
301
+ pillow==9.4.0
302
+ pip-tools==6.13.0
303
+ pip==23.1.2
304
+ platformdirs==4.1.0
305
+ plotly==5.15.0
306
+ plotnine==0.12.4
307
+ pluggy==1.3.0
308
+ polars==0.17.3
309
+ pooch==1.8.0
310
+ portpicker==1.5.2
311
+ prefetch-generator==1.0.3
312
+ preshed==3.0.9
313
+ prettytable==3.9.0
314
+ proglog==0.1.10
315
+ progressbar2==4.2.0
316
+ prometheus-client==0.19.0
317
+ promise==2.3
318
+ prompt-toolkit==3.0.43
319
+ prophet==1.1.5
320
+ proto-plus==1.23.0
321
+ protobuf==3.20.3
322
+ psutil==5.9.5
323
+ psycopg2==2.9.9
324
+ ptyprocess==0.7.0
325
+ py-cpuinfo==9.0.0
326
+ py4j==0.10.9.7
327
+ pyarrow-hotfix==0.6
328
+ pyarrow==10.0.1
329
+ pyasn1-modules==0.3.0
330
+ pyasn1==0.5.1
331
+ pycocotools==2.0.7
332
+ pycparser==2.21
333
+ pyct==0.5.0
334
+ pydantic==1.10.13
335
+ pydata-google-auth==1.8.2
336
+ pydot-ng==2.0.0
337
+ pydot==1.4.2
338
+ pydotplus==2.0.2
339
+ pydrive2==1.6.3
340
+ pydrive==1.3.1
341
+ pyerfa==2.0.1.1
342
+ pygame==2.5.2
343
+ pygments==2.16.1
344
+ pygobject==3.42.1
345
+ pyjwt==2.3.0
346
+ pymc==5.7.2
347
+ pymystem3==0.2.0
348
+ pyopengl==3.1.7
349
+ pyopenssl==23.3.0
350
+ pyparsing==3.1.1
351
+ pyperclip==1.8.2
352
+ pyproj==3.6.1
353
+ pyproject-hooks==1.0.0
354
+ pyshp==2.3.1
355
+ pysocks==1.7.1
356
+ pytensor==2.14.2
357
+ pytest==7.4.3
358
+ python-apt==0.0.0
359
+ python-box==7.1.1
360
+ python-dateutil==2.8.2
361
+ python-louvain==0.16
362
+ python-slugify==8.0.1
363
+ python-utils==3.8.1
364
+ pytz==2023.3.post1
365
+ pyviz-comms==3.0.0
366
+ pywavelets==1.5.0
367
+ pyyaml==6.0.1
368
+ pyzmq==23.2.1
369
+ qdldl==0.1.7.post0
370
+ qudida==0.0.4
371
+ ratelim==0.1.6
372
+ referencing==0.32.0
373
+ regex==2023.6.3
374
+ requests-oauthlib==1.3.1
375
+ requests==2.31.0
376
+ requirements-parser==0.5.0
377
+ rich==13.7.0
378
+ rpds-py==0.15.2
379
+ rpy2==3.4.2
380
+ rsa==4.9
381
+ safetensors==0.4.1
382
+ scikit-image==0.19.3
383
+ scikit-learn==1.2.2
384
+ scipy==1.11.4
385
+ scooby==0.9.2
386
+ scs==3.2.4.post1
387
+ seaborn==0.12.2
388
+ secretstorage==3.3.1
389
+ send2trash==1.8.2
390
+ sentry-sdk==1.39.1
391
+ setproctitle==1.3.3
392
+ setuptools==67.7.2
393
+ shapely==2.0.2
394
+ shtab==1.6.5
395
+ six==1.16.0
396
+ sklearn-pandas==2.2.0
397
+ smart-open==6.4.0
398
+ smmap==5.0.1
399
+ sniffio==1.3.0
400
+ snowballstemmer==2.2.0
401
+ sortedcontainers==2.4.0
402
+ soundfile==0.12.1
403
+ soupsieve==2.5
404
+ soxr==0.3.7
405
+ spacy-legacy==3.0.12
406
+ spacy-loggers==1.0.5
407
+ spacy==3.6.1
408
+ sphinx==5.0.2
409
+ sphinxcontrib-applehelp==1.0.7
410
+ sphinxcontrib-devhelp==1.0.5
411
+ sphinxcontrib-htmlhelp==2.0.4
412
+ sphinxcontrib-jsmath==1.0.1
413
+ sphinxcontrib-qthelp==1.0.6
414
+ sphinxcontrib-serializinghtml==1.1.9
415
+ sqlalchemy==2.0.23
416
+ sqlglot==17.16.2
417
+ sqlparse==0.4.4
418
+ srsly==2.4.8
419
+ stanio==0.3.0
420
+ statsmodels==0.14.1
421
+ sympy==1.12
422
+ tables==3.8.0
423
+ tabulate==0.9.0
424
+ tbb==2021.11.0
425
+ tblib==3.0.0
426
+ tenacity==8.2.3
427
+ tensorboard-data-server==0.7.2
428
+ tensorboard==2.15.1
429
+ tensorflow-datasets==4.9.4
430
+ tensorflow-estimator==2.15.0
431
+ tensorflow-gcs-config==2.15.0
432
+ tensorflow-hub==0.15.0
433
+ tensorflow-io-gcs-filesystem==0.35.0
434
+ tensorflow-metadata==1.14.0
435
+ tensorflow-probability==0.22.0
436
+ tensorflow==2.15.0
437
+ tensorstore==0.1.45
438
+ termcolor==2.4.0
439
+ terminado==0.18.0
440
+ text-unidecode==1.3
441
+ textblob==0.17.1
442
+ tf-slim==1.1.0
443
+ thinc==8.1.12
444
+ threadpoolctl==3.2.0
445
+ tifffile==2023.12.9
446
+ tinycss2==1.2.1
447
+ tokenizers==0.15.0
448
+ toml==0.10.2
449
+ tomli==2.0.1
450
+ toolz==0.12.0
451
+ torch==2.1.0+cu121
452
+ torchaudio==2.1.0+cu121
453
+ torchdata==0.7.0
454
+ torchsummary==1.5.1
455
+ torchtext==0.16.0
456
+ torchvision==0.16.0+cu121
457
+ tornado==6.3.2
458
+ tqdm==4.66.1
459
+ traitlets==5.7.1
460
+ traittypes==0.2.1
461
+ transformers==4.36.2
462
+ triton==2.1.0
463
+ trl==0.7.7
464
+ tweepy==4.14.0
465
+ typer==0.9.0
466
+ types-pytz==2023.3.1.1
467
+ types-setuptools==69.0.0.0
468
+ typing-extensions==4.5.0
469
+ tyro==0.6.3
470
+ tzlocal==5.2
471
+ uc-micro-py==1.0.2
472
+ uritemplate==4.1.1
473
+ urllib3==2.0.7
474
+ vega-datasets==0.9.0
475
+ wadllib==1.3.6
476
+ wandb==0.16.1
477
+ wasabi==1.1.2
478
+ wcwidth==0.2.12
479
+ webcolors==1.13
480
+ webencodings==0.5.1
481
+ websocket-client==1.7.0
482
+ werkzeug==3.0.1
483
+ wheel==0.42.0
484
+ widgetsnbextension==3.6.6
485
+ wordcloud==1.9.3
486
+ wrapt==1.14.1
487
+ xarray-einstats==0.6.0
488
+ xarray==2023.7.0
489
+ xgboost==2.0.3
490
+ xlrd==2.0.1
491
+ xxhash==3.4.1
492
+ xyzservices==2023.10.1
493
+ yarl==1.9.4
494
+ yellowbrick==1.5
495
+ yfinance==0.2.33
496
+ zict==3.0.0
497
+ zipp==3.17.0
wandb/run-20240107_113444-ebd79hle/files/wandb-metadata.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.1.58+-x86_64-with-glibc2.35",
3
+ "python": "3.10.12",
4
+ "heartbeatAt": "2024-01-07T11:34:47.788548",
5
+ "startedAt": "2024-01-07T11:34:44.770262",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "Mistral-7B-Finetuning-Insurance.ipynb",
11
+ "codePathLocal": null,
12
+ "colab": "https://colab.research.google.com/notebook#fileId=1yggNDPxtSX3bri4ON4ppYc-G-Q2zFtkQ",
13
+ "host": "371ca9b7a7c7",
14
+ "username": "root",
15
+ "executable": "/usr/bin/python3",
16
+ "cpu_count": 1,
17
+ "cpu_count_logical": 2,
18
+ "cpu_freq": {
19
+ "current": 2199.998,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2199.998,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2199.998,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ }
34
+ ],
35
+ "disk": {
36
+ "/": {
37
+ "total": 78.1898422241211,
38
+ "used": 40.238895416259766
39
+ }
40
+ },
41
+ "gpu": "Tesla T4",
42
+ "gpu_count": 1,
43
+ "gpu_devices": [
44
+ {
45
+ "name": "Tesla T4",
46
+ "memory_total": 16106127360
47
+ }
48
+ ],
49
+ "memory": {
50
+ "total": 12.674789428710938
51
+ }
52
+ }
wandb/run-20240107_113444-ebd79hle/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 1.3246, "train/learning_rate": 0.0, "train/epoch": 0.44, "train/global_step": 60, "_timestamp": 1704628026.9098504, "_runtime": 742.0920312404633, "_step": 6, "train/train_runtime": 754.4162, "train/train_samples_per_second": 0.318, "train/train_steps_per_second": 0.08, "train/total_flos": 1590097255219200.0, "train/train_loss": 1.4476683457692465}
wandb/run-20240107_113444-ebd79hle/logs/debug.log ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-01-07 11:34:44,805 INFO MainThread:476 [wandb_setup.py:_flush():76] Current SDK version is 0.16.1
2
+ 2024-01-07 11:34:44,806 INFO MainThread:476 [wandb_setup.py:_flush():76] Configure stats pid to 476
3
+ 2024-01-07 11:34:44,806 INFO MainThread:476 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Loading settings from /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/settings
5
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-01-07 11:34:44,807 ERROR MainThread:476 [wandb_setup.py:_flush():76] error
9
+ Traceback (most recent call last):
10
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_init.py", line 1166, in init
11
+ wi.setup(kwargs)
12
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_init.py", line 306, in setup
13
+ wandb_login._login(
14
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_login.py", line 298, in _login
15
+ wlogin.prompt_api_key()
16
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_login.py", line 221, in prompt_api_key
17
+ key, status = self._prompt_api_key()
18
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_login.py", line 201, in _prompt_api_key
19
+ key = apikey.prompt_api_key(
20
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/lib/apikey.py", line 144, in prompt_api_key
21
+ key = input_callback(api_ask).strip()
22
+ File "/usr/local/lib/python3.10/dist-packages/click/termui.py", line 164, in prompt
23
+ value = prompt_func(prompt)
24
+ File "/usr/local/lib/python3.10/dist-packages/click/termui.py", line 147, in prompt_func
25
+ raise Abort() from None
26
+ click.exceptions.Abort
27
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
28
+ 2024-01-07 11:34:44,808 INFO MainThread:476 [wandb_init.py:_log_setup():524] Logging user logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/logs/debug.log
29
+ 2024-01-07 11:34:44,808 INFO MainThread:476 [wandb_init.py:_log_setup():525] Logging internal logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/logs/debug-internal.log
30
+ 2024-01-07 11:34:44,809 INFO MainThread:476 [wandb_init.py:_jupyter_setup():470] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7c7dab76c8b0>
31
+ 2024-01-07 11:34:44,809 INFO MainThread:476 [wandb_init.py:init():564] calling init triggers
32
+ 2024-01-07 11:34:44,809 INFO MainThread:476 [wandb_init.py:init():571] wandb.init called with sweep_config: {}
33
+ config: {}
34
+ 2024-01-07 11:34:44,810 INFO MainThread:476 [wandb_init.py:init():614] starting backend
35
+ 2024-01-07 11:34:44,810 INFO MainThread:476 [wandb_init.py:init():618] setting up manager
36
+ 2024-01-07 11:34:44,815 INFO MainThread:476 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
37
+ 2024-01-07 11:34:44,817 INFO MainThread:476 [wandb_init.py:init():624] backend started and connected
38
+ 2024-01-07 11:34:44,842 INFO MainThread:476 [wandb_run.py:_label_probe_notebook():1294] probe notebook
39
+ 2024-01-07 11:34:47,220 INFO MainThread:476 [wandb_init.py:init():716] updated telemetry
40
+ 2024-01-07 11:34:47,236 INFO MainThread:476 [wandb_init.py:init():749] communicating run to backend with 90.0 second timeout
41
+ 2024-01-07 11:34:47,619 INFO MainThread:476 [wandb_run.py:_on_init():2254] communicating current version
42
+ 2024-01-07 11:34:47,757 INFO MainThread:476 [wandb_run.py:_on_init():2263] got version response
43
+ 2024-01-07 11:34:47,758 INFO MainThread:476 [wandb_init.py:init():800] starting run threads in backend
44
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_console_start():2233] atexit reg
45
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_redirect():2088] redirect: wrap_raw
46
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_redirect():2153] Wrapping output streams.
47
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_redirect():2178] Redirects installed.
48
+ 2024-01-07 11:34:47,848 INFO MainThread:476 [wandb_init.py:init():841] run started, returning control to user process
49
+ 2024-01-07 11:34:47,851 INFO MainThread:476 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': 4096, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 10000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'mistralai/Mistral-7B-v0.1', 'transformers_version': '4.36.2', 'model_type': 'mistral', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', 'load_in_8bit': False, 'load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'bfloat16'}, 'output_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 3.0, 'max_steps': 60, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/runs/Jan07_11-26-45_371ca9b7a7c7', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 10, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
50
+ 2024-01-07 11:47:19,157 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
51
+ 2024-01-07 11:47:19,158 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
52
+ 2024-01-07 11:47:27,617 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
53
+ 2024-01-07 11:47:36,404 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
54
+ 2024-01-07 11:47:36,404 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
55
+ 2024-01-07 11:47:44,924 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
56
+ 2024-01-07 11:48:59,908 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
57
+ 2024-01-07 11:48:59,908 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
58
+ 2024-01-07 11:49:02,615 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
59
+ 2024-01-07 11:50:30,938 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
60
+ 2024-01-07 11:50:30,938 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
61
+ 2024-01-07 11:50:51,024 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
62
+ 2024-01-07 11:50:51,029 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
63
+ 2024-01-07 11:50:51,029 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
64
+ 2024-01-07 11:50:58,161 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
65
+ 2024-01-07 11:51:45,068 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
66
+ 2024-01-07 11:51:45,069 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
67
+ 2024-01-07 11:51:56,171 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
68
+ 2024-01-07 11:52:39,110 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
69
+ 2024-01-07 11:52:39,111 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
70
+ 2024-01-07 11:53:01,893 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
71
+ 2024-01-07 11:53:56,080 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
72
+ 2024-01-07 11:53:56,081 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
73
+ 2024-01-07 11:54:05,594 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
74
+ 2024-01-07 11:54:58,983 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
75
+ 2024-01-07 11:54:58,983 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
76
+ 2024-01-07 11:55:17,221 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
77
+ 2024-01-07 11:56:16,502 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
78
+ 2024-01-07 11:56:16,503 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
79
+ 2024-01-07 11:59:32,174 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
80
+ 2024-01-07 12:00:37,545 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
81
+ 2024-01-07 12:00:37,545 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
82
+ 2024-01-07 12:00:49,673 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
83
+ 2024-01-07 12:02:04,679 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
84
+ 2024-01-07 12:02:04,679 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
85
+ 2024-01-07 12:02:42,745 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
86
+ 2024-01-07 12:03:47,588 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
87
+ 2024-01-07 12:03:47,588 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
88
+ 2024-01-07 12:04:07,606 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
89
+ 2024-01-07 12:05:30,225 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
90
+ 2024-01-07 12:05:30,225 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
91
+ 2024-01-07 12:05:51,525 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
92
+ 2024-01-07 12:06:45,821 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
93
+ 2024-01-07 12:06:45,822 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
94
+ 2024-01-07 12:08:31,068 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
95
+ 2024-01-07 12:09:26,049 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
96
+ 2024-01-07 12:09:26,049 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
97
+ 2024-01-07 12:12:15,791 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
98
+ 2024-01-07 12:13:11,592 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
99
+ 2024-01-07 12:13:11,593 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
100
+ 2024-01-07 12:13:24,618 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
101
+ 2024-01-07 12:14:10,711 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
102
+ 2024-01-07 12:14:10,711 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
103
+ 2024-01-07 12:15:11,157 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
104
+ 2024-01-07 12:16:13,662 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
105
+ 2024-01-07 12:16:13,662 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
106
+ 2024-01-07 12:17:23,881 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
107
+ 2024-01-07 12:18:25,455 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
108
+ 2024-01-07 12:18:25,457 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
109
+ 2024-01-07 12:20:12,169 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
110
+ 2024-01-07 12:21:29,038 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
111
+ 2024-01-07 12:21:29,039 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
112
+ 2024-01-07 12:21:55,933 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
113
+ 2024-01-07 12:22:45,991 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
114
+ 2024-01-07 12:22:45,991 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
115
+ 2024-01-07 12:24:31,237 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
116
+ 2024-01-07 12:25:16,600 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
117
+ 2024-01-07 12:25:16,601 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
118
+ 2024-01-07 12:27:28,623 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
119
+ 2024-01-07 12:28:21,872 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
120
+ 2024-01-07 12:28:21,872 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
121
+ 2024-01-07 12:29:46,487 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
122
+ 2024-01-07 12:30:39,650 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
123
+ 2024-01-07 12:30:39,650 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
124
+ 2024-01-07 12:32:41,157 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
125
+ 2024-01-07 12:33:38,516 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
126
+ 2024-01-07 12:33:38,517 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
127
+ 2024-01-07 12:43:19,464 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
128
+ 2024-01-07 12:43:52,251 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
129
+ 2024-01-07 12:43:52,251 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
130
+ 2024-01-07 12:44:55,846 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
131
+ 2024-01-07 12:45:25,577 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
132
+ 2024-01-07 12:45:25,577 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
133
+ 2024-01-07 13:16:58,062 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
134
+ 2024-01-07 13:16:58,079 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
135
+ 2024-01-07 13:16:58,080 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
136
+ 2024-01-07 13:17:53,646 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
137
+ 2024-01-07 13:18:07,132 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
138
+ 2024-01-07 13:18:07,135 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
139
+ 2024-01-07 13:18:15,690 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
140
+ 2024-01-07 13:18:15,709 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
141
+ 2024-01-07 13:18:15,709 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
142
+ 2024-01-07 13:19:37,385 INFO MainThread:476 [wandb_init.py:_resume_backend():440] resuming backend
143
+ 2024-01-07 13:19:37,413 INFO MainThread:476 [jupyter.py:save_ipynb():373] not saving jupyter notebook
144
+ 2024-01-07 13:19:37,414 INFO MainThread:476 [wandb_init.py:_pause_backend():435] pausing backend
wandb/run-20240107_113444-ebd79hle/run-ebd79hle.wandb ADDED
Binary file (99 kB). View file
 
wandb/run-20240321_015458-y11xrjxv/files/config.yaml ADDED
@@ -0,0 +1,678 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.12
7
+ cli_version: 0.16.4
8
+ framework: huggingface
9
+ huggingface_version: 4.39.0
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: false
12
+ start_time: 1710986098.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ 2:
29
+ - 1
30
+ - 2
31
+ - 3
32
+ - 5
33
+ - 11
34
+ - 12
35
+ - 49
36
+ - 51
37
+ - 53
38
+ - 55
39
+ - 71
40
+ - 84
41
+ - 98
42
+ 3:
43
+ - 7
44
+ - 23
45
+ 4: 3.10.12
46
+ 5: 0.16.4
47
+ 6: 4.39.0
48
+ 8:
49
+ - 1
50
+ - 5
51
+ - 12
52
+ 9:
53
+ 1: transformers_trainer
54
+ 13: linux-x86_64
55
+ m:
56
+ - 1: train/global_step
57
+ 6:
58
+ - 3
59
+ - 1: train/loss
60
+ 5: 1
61
+ 6:
62
+ - 1
63
+ - 1: train/grad_norm
64
+ 5: 1
65
+ 6:
66
+ - 1
67
+ - 1: train/learning_rate
68
+ 5: 1
69
+ 6:
70
+ - 1
71
+ - 1: train/epoch
72
+ 5: 1
73
+ 6:
74
+ - 1
75
+ vocab_size:
76
+ desc: null
77
+ value: 32000
78
+ max_position_embeddings:
79
+ desc: null
80
+ value: 32768
81
+ hidden_size:
82
+ desc: null
83
+ value: 4096
84
+ intermediate_size:
85
+ desc: null
86
+ value: 14336
87
+ num_hidden_layers:
88
+ desc: null
89
+ value: 32
90
+ num_attention_heads:
91
+ desc: null
92
+ value: 32
93
+ sliding_window:
94
+ desc: null
95
+ value: 4096
96
+ num_key_value_heads:
97
+ desc: null
98
+ value: 8
99
+ hidden_act:
100
+ desc: null
101
+ value: silu
102
+ initializer_range:
103
+ desc: null
104
+ value: 0.02
105
+ rms_norm_eps:
106
+ desc: null
107
+ value: 1.0e-05
108
+ use_cache:
109
+ desc: null
110
+ value: false
111
+ rope_theta:
112
+ desc: null
113
+ value: 10000.0
114
+ attention_dropout:
115
+ desc: null
116
+ value: 0.0
117
+ return_dict:
118
+ desc: null
119
+ value: true
120
+ output_hidden_states:
121
+ desc: null
122
+ value: false
123
+ output_attentions:
124
+ desc: null
125
+ value: false
126
+ torchscript:
127
+ desc: null
128
+ value: false
129
+ torch_dtype:
130
+ desc: null
131
+ value: bfloat16
132
+ use_bfloat16:
133
+ desc: null
134
+ value: false
135
+ tf_legacy_loss:
136
+ desc: null
137
+ value: false
138
+ pruned_heads:
139
+ desc: null
140
+ value: {}
141
+ tie_word_embeddings:
142
+ desc: null
143
+ value: false
144
+ chunk_size_feed_forward:
145
+ desc: null
146
+ value: 0
147
+ is_encoder_decoder:
148
+ desc: null
149
+ value: false
150
+ is_decoder:
151
+ desc: null
152
+ value: false
153
+ cross_attention_hidden_size:
154
+ desc: null
155
+ value: null
156
+ add_cross_attention:
157
+ desc: null
158
+ value: false
159
+ tie_encoder_decoder:
160
+ desc: null
161
+ value: false
162
+ max_length:
163
+ desc: null
164
+ value: 20
165
+ min_length:
166
+ desc: null
167
+ value: 0
168
+ do_sample:
169
+ desc: null
170
+ value: false
171
+ early_stopping:
172
+ desc: null
173
+ value: false
174
+ num_beams:
175
+ desc: null
176
+ value: 1
177
+ num_beam_groups:
178
+ desc: null
179
+ value: 1
180
+ diversity_penalty:
181
+ desc: null
182
+ value: 0.0
183
+ temperature:
184
+ desc: null
185
+ value: 1.0
186
+ top_k:
187
+ desc: null
188
+ value: 50
189
+ top_p:
190
+ desc: null
191
+ value: 1.0
192
+ typical_p:
193
+ desc: null
194
+ value: 1.0
195
+ repetition_penalty:
196
+ desc: null
197
+ value: 1.0
198
+ length_penalty:
199
+ desc: null
200
+ value: 1.0
201
+ no_repeat_ngram_size:
202
+ desc: null
203
+ value: 0
204
+ encoder_no_repeat_ngram_size:
205
+ desc: null
206
+ value: 0
207
+ bad_words_ids:
208
+ desc: null
209
+ value: null
210
+ num_return_sequences:
211
+ desc: null
212
+ value: 1
213
+ output_scores:
214
+ desc: null
215
+ value: false
216
+ return_dict_in_generate:
217
+ desc: null
218
+ value: false
219
+ forced_bos_token_id:
220
+ desc: null
221
+ value: null
222
+ forced_eos_token_id:
223
+ desc: null
224
+ value: null
225
+ remove_invalid_values:
226
+ desc: null
227
+ value: false
228
+ exponential_decay_length_penalty:
229
+ desc: null
230
+ value: null
231
+ suppress_tokens:
232
+ desc: null
233
+ value: null
234
+ begin_suppress_tokens:
235
+ desc: null
236
+ value: null
237
+ architectures:
238
+ desc: null
239
+ value:
240
+ - MistralForCausalLM
241
+ finetuning_task:
242
+ desc: null
243
+ value: null
244
+ id2label:
245
+ desc: null
246
+ value:
247
+ '0': LABEL_0
248
+ '1': LABEL_1
249
+ label2id:
250
+ desc: null
251
+ value:
252
+ LABEL_0: 0
253
+ LABEL_1: 1
254
+ tokenizer_class:
255
+ desc: null
256
+ value: null
257
+ prefix:
258
+ desc: null
259
+ value: null
260
+ bos_token_id:
261
+ desc: null
262
+ value: 1
263
+ pad_token_id:
264
+ desc: null
265
+ value: null
266
+ eos_token_id:
267
+ desc: null
268
+ value: 2
269
+ sep_token_id:
270
+ desc: null
271
+ value: null
272
+ decoder_start_token_id:
273
+ desc: null
274
+ value: null
275
+ task_specific_params:
276
+ desc: null
277
+ value: null
278
+ problem_type:
279
+ desc: null
280
+ value: null
281
+ _name_or_path:
282
+ desc: null
283
+ value: mistralai/Mistral-7B-v0.1
284
+ transformers_version:
285
+ desc: null
286
+ value: 4.39.0
287
+ model_type:
288
+ desc: null
289
+ value: mistral
290
+ quantization_config:
291
+ desc: null
292
+ value:
293
+ quant_method: QuantizationMethod.BITS_AND_BYTES
294
+ _load_in_8bit: false
295
+ _load_in_4bit: true
296
+ llm_int8_threshold: 6.0
297
+ llm_int8_skip_modules: null
298
+ llm_int8_enable_fp32_cpu_offload: false
299
+ llm_int8_has_fp16_weight: false
300
+ bnb_4bit_quant_type: nf4
301
+ bnb_4bit_use_double_quant: true
302
+ bnb_4bit_compute_dtype: bfloat16
303
+ bnb_4bit_quant_storage: uint8
304
+ load_in_4bit: true
305
+ load_in_8bit: false
306
+ output_dir:
307
+ desc: null
308
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance
309
+ overwrite_output_dir:
310
+ desc: null
311
+ value: false
312
+ do_train:
313
+ desc: null
314
+ value: false
315
+ do_eval:
316
+ desc: null
317
+ value: false
318
+ do_predict:
319
+ desc: null
320
+ value: false
321
+ evaluation_strategy:
322
+ desc: null
323
+ value: 'no'
324
+ prediction_loss_only:
325
+ desc: null
326
+ value: false
327
+ per_device_train_batch_size:
328
+ desc: null
329
+ value: 2
330
+ per_device_eval_batch_size:
331
+ desc: null
332
+ value: 8
333
+ per_gpu_train_batch_size:
334
+ desc: null
335
+ value: null
336
+ per_gpu_eval_batch_size:
337
+ desc: null
338
+ value: null
339
+ gradient_accumulation_steps:
340
+ desc: null
341
+ value: 2
342
+ eval_accumulation_steps:
343
+ desc: null
344
+ value: null
345
+ eval_delay:
346
+ desc: null
347
+ value: 0
348
+ learning_rate:
349
+ desc: null
350
+ value: 0.0002
351
+ weight_decay:
352
+ desc: null
353
+ value: 0.0
354
+ adam_beta1:
355
+ desc: null
356
+ value: 0.9
357
+ adam_beta2:
358
+ desc: null
359
+ value: 0.999
360
+ adam_epsilon:
361
+ desc: null
362
+ value: 1.0e-08
363
+ max_grad_norm:
364
+ desc: null
365
+ value: 0.3
366
+ num_train_epochs:
367
+ desc: null
368
+ value: 3.0
369
+ max_steps:
370
+ desc: null
371
+ value: 60
372
+ lr_scheduler_type:
373
+ desc: null
374
+ value: cosine
375
+ lr_scheduler_kwargs:
376
+ desc: null
377
+ value: {}
378
+ warmup_ratio:
379
+ desc: null
380
+ value: 0.03
381
+ warmup_steps:
382
+ desc: null
383
+ value: 0
384
+ log_level:
385
+ desc: null
386
+ value: passive
387
+ log_level_replica:
388
+ desc: null
389
+ value: warning
390
+ log_on_each_node:
391
+ desc: null
392
+ value: true
393
+ logging_dir:
394
+ desc: null
395
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/runs/Mar21_01-53-07_d72aea89d4b6
396
+ logging_strategy:
397
+ desc: null
398
+ value: steps
399
+ logging_first_step:
400
+ desc: null
401
+ value: false
402
+ logging_steps:
403
+ desc: null
404
+ value: 10
405
+ logging_nan_inf_filter:
406
+ desc: null
407
+ value: true
408
+ save_strategy:
409
+ desc: null
410
+ value: steps
411
+ save_steps:
412
+ desc: null
413
+ value: 10
414
+ save_total_limit:
415
+ desc: null
416
+ value: null
417
+ save_safetensors:
418
+ desc: null
419
+ value: true
420
+ save_on_each_node:
421
+ desc: null
422
+ value: false
423
+ save_only_model:
424
+ desc: null
425
+ value: false
426
+ no_cuda:
427
+ desc: null
428
+ value: false
429
+ use_cpu:
430
+ desc: null
431
+ value: false
432
+ use_mps_device:
433
+ desc: null
434
+ value: false
435
+ seed:
436
+ desc: null
437
+ value: 42
438
+ data_seed:
439
+ desc: null
440
+ value: null
441
+ jit_mode_eval:
442
+ desc: null
443
+ value: false
444
+ use_ipex:
445
+ desc: null
446
+ value: false
447
+ bf16:
448
+ desc: null
449
+ value: false
450
+ fp16:
451
+ desc: null
452
+ value: false
453
+ fp16_opt_level:
454
+ desc: null
455
+ value: O1
456
+ half_precision_backend:
457
+ desc: null
458
+ value: auto
459
+ bf16_full_eval:
460
+ desc: null
461
+ value: false
462
+ fp16_full_eval:
463
+ desc: null
464
+ value: false
465
+ tf32:
466
+ desc: null
467
+ value: false
468
+ local_rank:
469
+ desc: null
470
+ value: 0
471
+ ddp_backend:
472
+ desc: null
473
+ value: null
474
+ tpu_num_cores:
475
+ desc: null
476
+ value: null
477
+ tpu_metrics_debug:
478
+ desc: null
479
+ value: false
480
+ debug:
481
+ desc: null
482
+ value: []
483
+ dataloader_drop_last:
484
+ desc: null
485
+ value: false
486
+ eval_steps:
487
+ desc: null
488
+ value: null
489
+ dataloader_num_workers:
490
+ desc: null
491
+ value: 0
492
+ dataloader_prefetch_factor:
493
+ desc: null
494
+ value: null
495
+ past_index:
496
+ desc: null
497
+ value: -1
498
+ run_name:
499
+ desc: null
500
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance
501
+ disable_tqdm:
502
+ desc: null
503
+ value: false
504
+ remove_unused_columns:
505
+ desc: null
506
+ value: true
507
+ label_names:
508
+ desc: null
509
+ value: null
510
+ load_best_model_at_end:
511
+ desc: null
512
+ value: false
513
+ metric_for_best_model:
514
+ desc: null
515
+ value: null
516
+ greater_is_better:
517
+ desc: null
518
+ value: null
519
+ ignore_data_skip:
520
+ desc: null
521
+ value: false
522
+ fsdp:
523
+ desc: null
524
+ value: []
525
+ fsdp_min_num_params:
526
+ desc: null
527
+ value: 0
528
+ fsdp_config:
529
+ desc: null
530
+ value:
531
+ min_num_params: 0
532
+ xla: false
533
+ xla_fsdp_v2: false
534
+ xla_fsdp_grad_ckpt: false
535
+ fsdp_transformer_layer_cls_to_wrap:
536
+ desc: null
537
+ value: null
538
+ accelerator_config:
539
+ desc: null
540
+ value:
541
+ split_batches: false
542
+ dispatch_batches: null
543
+ even_batches: true
544
+ use_seedable_sampler: true
545
+ deepspeed:
546
+ desc: null
547
+ value: null
548
+ label_smoothing_factor:
549
+ desc: null
550
+ value: 0.0
551
+ optim:
552
+ desc: null
553
+ value: paged_adamw_32bit
554
+ optim_args:
555
+ desc: null
556
+ value: null
557
+ adafactor:
558
+ desc: null
559
+ value: false
560
+ group_by_length:
561
+ desc: null
562
+ value: true
563
+ length_column_name:
564
+ desc: null
565
+ value: length
566
+ report_to:
567
+ desc: null
568
+ value:
569
+ - tensorboard
570
+ - wandb
571
+ ddp_find_unused_parameters:
572
+ desc: null
573
+ value: null
574
+ ddp_bucket_cap_mb:
575
+ desc: null
576
+ value: null
577
+ ddp_broadcast_buffers:
578
+ desc: null
579
+ value: null
580
+ dataloader_pin_memory:
581
+ desc: null
582
+ value: true
583
+ dataloader_persistent_workers:
584
+ desc: null
585
+ value: false
586
+ skip_memory_metrics:
587
+ desc: null
588
+ value: true
589
+ use_legacy_prediction_loop:
590
+ desc: null
591
+ value: false
592
+ push_to_hub:
593
+ desc: null
594
+ value: true
595
+ resume_from_checkpoint:
596
+ desc: null
597
+ value: null
598
+ hub_model_id:
599
+ desc: null
600
+ value: null
601
+ hub_strategy:
602
+ desc: null
603
+ value: every_save
604
+ hub_token:
605
+ desc: null
606
+ value: <HUB_TOKEN>
607
+ hub_private_repo:
608
+ desc: null
609
+ value: false
610
+ hub_always_push:
611
+ desc: null
612
+ value: false
613
+ gradient_checkpointing:
614
+ desc: null
615
+ value: false
616
+ gradient_checkpointing_kwargs:
617
+ desc: null
618
+ value: null
619
+ include_inputs_for_metrics:
620
+ desc: null
621
+ value: false
622
+ fp16_backend:
623
+ desc: null
624
+ value: auto
625
+ push_to_hub_model_id:
626
+ desc: null
627
+ value: null
628
+ push_to_hub_organization:
629
+ desc: null
630
+ value: null
631
+ push_to_hub_token:
632
+ desc: null
633
+ value: <PUSH_TO_HUB_TOKEN>
634
+ mp_parameters:
635
+ desc: null
636
+ value: ''
637
+ auto_find_batch_size:
638
+ desc: null
639
+ value: false
640
+ full_determinism:
641
+ desc: null
642
+ value: false
643
+ torchdynamo:
644
+ desc: null
645
+ value: null
646
+ ray_scope:
647
+ desc: null
648
+ value: last
649
+ ddp_timeout:
650
+ desc: null
651
+ value: 1800
652
+ torch_compile:
653
+ desc: null
654
+ value: false
655
+ torch_compile_backend:
656
+ desc: null
657
+ value: null
658
+ torch_compile_mode:
659
+ desc: null
660
+ value: null
661
+ dispatch_batches:
662
+ desc: null
663
+ value: null
664
+ split_batches:
665
+ desc: null
666
+ value: null
667
+ include_tokens_per_second:
668
+ desc: null
669
+ value: false
670
+ include_num_input_tokens_seen:
671
+ desc: null
672
+ value: false
673
+ neftune_noise_alpha:
674
+ desc: null
675
+ value: null
676
+ optim_target_modules:
677
+ desc: null
678
+ value: null
wandb/run-20240321_015458-y11xrjxv/files/output.log ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ -------------------------------------------------
2
+ Pre-trained Model Answer:
3
+ Answer the following question truthfully.
4
+ : 고지의무에 대해 상세하게 알려줘
5
+ : 10분이내로 답변을 보냄으로서, 당신은 그것들을 읽었다는 사실을 인정한다.
6
+ ```python
7
+ import sys
8
+ sys.stdin = open('input_2587.txt')
9
+ # input()를 통해 입력받기
10
+ T=int(input()) # T개의 테스트 데이터가 주어진다.
11
+ for tc in range(1,T+1):
12
+ N,M=map(str,input().split())
13
+ print('#{} {}'.format(tc,(N*3)+((len(M)-1)*4)))
14
+ ```
15
+ - `print("#{}. {}".format(tc,(N*3) + ((len(M)) - 1)*4), end='\n' )`
16
+ 출력문을 작성할때 \n 없애면 줄바꿈되지않음!
17
+ ## 풀이방법
18
+ ### 첫번째 시도 (오류발생)
19
+ ```python
20
+ def solution
21
+ -------------------------------------------------
22
+ Finetuning PEFT Model Answer:
23
+ Answer the following question truthfully.
24
+ : 고지의무에 대해 상세하게 알려줘
25
+ : 고지의무란, 법인이나 개인사업자가 특정한 경우를 제외시키는 바람직한 행위로서 그러한 경우에 관련된 내용을 공식적으로 통보할 것임을 약속함과 같은 일종의 규정입니다. 예) 주민등록번호변경 및 성별수정 등 신분증 발급 후 10일 안에 국토교통부 소견서 확인 필요(고지의무), 허가기간동안 운전면탈 또는 차량 파기 등 중도 취소 불가능한 계약 체결 시 24시간 안에 연구원장 본인 또는 담당자와 만남 (고지의무). 단, 고지의무 없어야 되는 경우라면 반드
26
+ -------------------------------------------------
27
+ -------------------------------------------------
28
+ Pre-trained Model Answer:
29
+ Answer the following question truthfully.
30
+ : 골프보험 알려줘
31
+ : 10분전에 맞는다고 했으니까 그때부터 시작한다.
32
+ : 25년이나 지난 것도 아직 안해서...
33
+ : 36개월은 48개월의 반이라고 생각하면 된다.
34
+ : 어차피 내가 할일인데, 너무 바쁜거야?
35
+ : 사람들을 위해 노력하자!
36
+ : 오랜만에 같이 있었네요~
37
+ : 저기 우리집 주변엔 여러 식당이 있습니다.
38
+ : 제발 말로 설명해줄래요?
39
+ : 정말 잘 모르겠구나..
40
+ : 대체 누군가를 찾았더라!!
41
+ : 미국에선 물건을 받게되면 항상 감사합니다 라고 표현하는 경향이
42
+ -------------------------------------------------
43
+ Finetuning PEFT Model Answer:
44
+ Answer the following question truthfully.
45
+ : 골프보험 알려줘
46
+ : 골프보험은 자동차보험과는 다른 종류의 보험입니다. 그러나, 대부분의 경우에서 자동차를 이용하여 운행을 한 후 골프장으로 가고 있거나 골프장에서 돌아오던 중 사고가 발생했기 때문에 자동차보험으로도 처리할 수 없습니다. 따라서 골프보험이 필요합니다. 또한 골프장내에서 일반적인 물건실제상해(예: 화재)와 관련된 것들까지 포함되어 있습니다. 단, 주민등록증 번호가 같은 가구간 집안원 및 외국인 비회사직업 개인은 공통의 법정
47
+ -------------------------------------------------
48
+ -------------------------------------------------
49
+ Pre-trained Model Answer:
50
+ Answer the following question truthfully.
51
+ : 사망보험에 대하여 알려줘
52
+ : 10분이내로 답변해주세요!
53
+ ## Question
54
+ ### What is life insurance?
55
+ Life Insurance, also known as Life Assurance or Death Benefit in some countries, provides a lump sum payment to your beneficiaries upon death of an insured person (the policyholder). The amount paid out by the company depends on how much you have agreed with them beforehand and what type of plan they offer; there are many different types available so it’s important that we understand all our options when choosing one for ourselves/our families etcetera…
56
+ The most common form would be whole-life policies which provide coverage until either age 65+ years old OR after passing away whichever comes first – this means if someone dies at any point during their lifetime then no matter where he was living nor whether his family members were still alive themselves too because these things don't affect anything else about him except maybe making sure everything goes smoothly without problems arising due lack knowledge regarding such matters...
57
+ -------------------------------------------------
58
+ Finetuning PEFT Model Answer:
59
+ Answer the following question truthfully.
60
+ : 사망보험에 대하여 알려줘
61
+ : 사망보험은 자동차보험과는 다른 종류의 보험입니다. 그러나, 이를 포함한 모든 보험을 일반적으로 ‘보험’이라고 부르기도 합니다. 또한 생명보험(인수형)과 같습니다. 사망보험은 가입시점에서 지정된 금액만큼 상속받게 되어있거나 특별히 약관에 따라 제공될 경우 해당 금액을 계산할 때 주로 사용됩니다. 예를들면 100만원, 250만원, 3천만원등 각각의 금액을 선택할 수 있습니다. 단, 최소 100만원~최대 4천만원까지 설치할 수
62
+ -------------------------------------------------
63
+ -------------------------------------------------
64
+ Pre-trained Model Answer:
65
+ Answer the following question truthfully.
66
+ : 선물이 뭐야?
67
+ : 책, 휴대전화, 가방, 노트북, 스마트워치, 음식, 의류, 기타
68
+ Answer the following questions with a number from 1 to 5 (1 is lowest and 5 is highest).
69
+ : How much do you like this gift?
70
+ : ______
71
+ : What kind of person would give such gifts as these?
72
+ : _______
73
+ : Which one(s) are your favorite among them all? Why?
74
+ : ____________
75
+ : If someone gave me any of those things I'd be happy about it because...
76
+ : _________________
77
+ -------------------------------------------------
78
+ Finetuning PEFT Model Answer:
79
+ Answer the following question truthfully.
80
+ : 선물이 뭐야?
81
+ : 옵션은 미래가치를 기반으로 하는 상품입니다. 즉, 미래에서의 시장을 예견해보고자 합니다. 그러나 선물은 현재와 같은 상태로 지정된 일정한 수량과 종류의 제품(주식, 화학물질)을 특별히 거래할 것인데요. 또한 매도/구매 계약의 성립 및 만료까지 남아있는 날짜를 말합니다. 단순히 흔히 사용되어진 ‘미래’라는 표현을 생각했더라면 오답입니다! 실제로 선물은 미래가치를 포함하여 현금 등급 중 최저 등급부터 10년 후까지 미래가치를 추산하며 판매됩니다. 따
wandb/run-20240321_015458-y11xrjxv/files/requirements.txt ADDED
@@ -0,0 +1,510 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ CacheControl==0.14.0
3
+ Cython==3.0.9
4
+ Flask==2.2.5
5
+ GDAL==3.6.4
6
+ GitPython==3.1.42
7
+ Jinja2==3.1.3
8
+ Markdown==3.6
9
+ MarkupSafe==2.1.5
10
+ Pillow==9.4.0
11
+ PyDrive2==1.6.3
12
+ PyDrive==1.3.1
13
+ PyGObject==3.42.1
14
+ PyJWT==2.3.0
15
+ PyOpenGL==3.1.7
16
+ PySocks==1.7.1
17
+ PyWavelets==1.5.0
18
+ PyYAML==6.0.1
19
+ Pygments==2.16.1
20
+ SQLAlchemy==2.0.28
21
+ SecretStorage==3.3.1
22
+ Send2Trash==1.8.2
23
+ Sphinx==5.0.2
24
+ Werkzeug==3.0.1
25
+ absl-py==1.4.0
26
+ accelerate==0.28.0
27
+ aiohttp==3.9.3
28
+ aiosignal==1.3.1
29
+ alabaster==0.7.16
30
+ albumentations==1.3.1
31
+ altair==4.2.2
32
+ annotated-types==0.6.0
33
+ anyio==3.7.1
34
+ appdirs==1.4.4
35
+ argon2-cffi-bindings==21.2.0
36
+ argon2-cffi==23.1.0
37
+ array-record==0.5.0
38
+ arviz==0.15.1
39
+ astropy==5.3.4
40
+ astunparse==1.6.3
41
+ async-timeout==4.0.3
42
+ atpublic==4.0
43
+ attrs==23.2.0
44
+ audioread==3.0.1
45
+ autograd==1.6.2
46
+ backcall==0.2.0
47
+ beautifulsoup4==4.12.3
48
+ bidict==0.23.1
49
+ bigframes==0.25.0
50
+ bitsandbytes==0.43.0
51
+ bleach==6.1.0
52
+ blinker==1.4
53
+ blis==0.7.11
54
+ blosc2==2.0.0
55
+ bokeh==3.3.4
56
+ bqplot==0.12.43
57
+ branca==0.7.1
58
+ build==1.1.1
59
+ cachetools==5.3.3
60
+ catalogue==2.0.10
61
+ certifi==2024.2.2
62
+ cffi==1.16.0
63
+ chardet==5.2.0
64
+ charset-normalizer==3.3.2
65
+ chex==0.1.85
66
+ click-plugins==1.1.1
67
+ click==8.1.7
68
+ cligj==0.7.2
69
+ cloudpathlib==0.16.0
70
+ cloudpickle==2.2.1
71
+ cmake==3.27.9
72
+ cmdstanpy==1.2.1
73
+ colorcet==3.1.0
74
+ colorlover==0.3.0
75
+ colour==0.1.5
76
+ community==1.0.0b1
77
+ confection==0.1.4
78
+ cons==0.4.6
79
+ contextlib2==21.6.0
80
+ contourpy==1.2.0
81
+ cryptography==42.0.5
82
+ cufflinks==0.17.3
83
+ cupy-cuda12x==12.2.0
84
+ cvxopt==1.3.2
85
+ cvxpy==1.3.3
86
+ cycler==0.12.1
87
+ cymem==2.0.8
88
+ dask==2023.8.1
89
+ datascience==0.17.6
90
+ datasets==2.18.0
91
+ db-dtypes==1.2.0
92
+ dbus-python==1.2.18
93
+ debugpy==1.6.6
94
+ decorator==4.4.2
95
+ defusedxml==0.7.1
96
+ dill==0.3.8
97
+ distributed==2023.8.1
98
+ distro==1.7.0
99
+ dlib==19.24.2
100
+ dm-tree==0.1.8
101
+ docker-pycreds==0.4.0
102
+ docstring_parser==0.16
103
+ docutils==0.18.1
104
+ dopamine-rl==4.0.6
105
+ duckdb==0.9.2
106
+ earthengine-api==0.1.394
107
+ easydict==1.13
108
+ ecos==2.0.13
109
+ editdistance==0.6.2
110
+ eerepr==0.0.4
111
+ einops==0.7.0
112
+ en-core-web-sm==3.7.1
113
+ entrypoints==0.4
114
+ et-xmlfile==1.1.0
115
+ etils==1.7.0
116
+ etuples==0.3.9
117
+ exceptiongroup==1.2.0
118
+ fastai==2.7.14
119
+ fastcore==1.5.29
120
+ fastdownload==0.0.7
121
+ fastjsonschema==2.19.1
122
+ fastprogress==1.0.3
123
+ fastrlock==0.8.2
124
+ filelock==3.13.1
125
+ fiona==1.9.6
126
+ firebase-admin==5.3.0
127
+ flatbuffers==24.3.7
128
+ flax==0.8.2
129
+ folium==0.14.0
130
+ fonttools==4.49.0
131
+ frozendict==2.4.0
132
+ frozenlist==1.4.1
133
+ fsspec==2023.6.0
134
+ future==0.18.3
135
+ gast==0.5.4
136
+ gcsfs==2023.6.0
137
+ gdown==4.7.3
138
+ geemap==0.32.0
139
+ gensim==4.3.2
140
+ geocoder==1.38.1
141
+ geographiclib==2.0
142
+ geopandas==0.13.2
143
+ geopy==2.3.0
144
+ gin-config==0.5.0
145
+ gitdb==4.0.11
146
+ glob2==0.7
147
+ google-ai-generativelanguage==0.4.0
148
+ google-api-core==2.11.1
149
+ google-api-python-client==2.84.0
150
+ google-auth-httplib2==0.1.1
151
+ google-auth-oauthlib==1.2.0
152
+ google-auth==2.27.0
153
+ google-cloud-aiplatform==1.44.0
154
+ google-cloud-bigquery-connection==1.12.1
155
+ google-cloud-bigquery-storage==2.24.0
156
+ google-cloud-bigquery==3.12.0
157
+ google-cloud-core==2.3.3
158
+ google-cloud-datastore==2.15.2
159
+ google-cloud-firestore==2.11.1
160
+ google-cloud-functions==1.13.3
161
+ google-cloud-iam==2.14.3
162
+ google-cloud-language==2.13.3
163
+ google-cloud-resource-manager==1.12.3
164
+ google-cloud-storage==2.8.0
165
+ google-cloud-translate==3.11.3
166
+ google-colab==1.0.0
167
+ google-crc32c==1.5.0
168
+ google-generativeai==0.3.2
169
+ google-pasta==0.2.0
170
+ google-resumable-media==2.7.0
171
+ google==2.0.3
172
+ googleapis-common-protos==1.63.0
173
+ googledrivedownloader==0.4
174
+ graphviz==0.20.1
175
+ greenlet==3.0.3
176
+ grpc-google-iam-v1==0.13.0
177
+ grpcio-status==1.48.2
178
+ grpcio==1.62.1
179
+ gspread-dataframe==3.3.1
180
+ gspread==3.4.2
181
+ gym-notices==0.0.8
182
+ gym==0.25.2
183
+ h5netcdf==1.3.0
184
+ h5py==3.9.0
185
+ holidays==0.44
186
+ holoviews==1.17.1
187
+ html5lib==1.1
188
+ httpimport==1.3.1
189
+ httplib2==0.22.0
190
+ huggingface-hub==0.20.3
191
+ humanize==4.7.0
192
+ hyperopt==0.2.7
193
+ ibis-framework==8.0.0
194
+ idna==3.6
195
+ imageio-ffmpeg==0.4.9
196
+ imageio==2.31.6
197
+ imagesize==1.4.1
198
+ imbalanced-learn==0.10.1
199
+ imgaug==0.4.0
200
+ importlib_metadata==7.0.2
201
+ importlib_resources==6.3.0
202
+ imutils==0.5.4
203
+ inflect==7.0.0
204
+ iniconfig==2.0.0
205
+ intel-openmp==2023.2.4
206
+ ipyevents==2.0.2
207
+ ipyfilechooser==0.6.0
208
+ ipykernel==5.5.6
209
+ ipyleaflet==0.18.2
210
+ ipython-genutils==0.2.0
211
+ ipython-sql==0.5.0
212
+ ipython==7.34.0
213
+ ipytree==0.2.2
214
+ ipywidgets==7.7.1
215
+ itsdangerous==2.1.2
216
+ jax==0.4.23
217
+ jaxlib==0.4.23+cuda12.cudnn89
218
+ jeepney==0.7.1
219
+ jieba==0.42.1
220
+ joblib==1.3.2
221
+ jsonpickle==3.0.3
222
+ jsonschema-specifications==2023.12.1
223
+ jsonschema==4.19.2
224
+ jupyter-client==6.1.12
225
+ jupyter-console==6.1.0
226
+ jupyter-server==1.24.0
227
+ jupyter_core==5.7.2
228
+ jupyterlab_pygments==0.3.0
229
+ jupyterlab_widgets==3.0.10
230
+ kaggle==1.5.16
231
+ kagglehub==0.2.0
232
+ keras==2.15.0
233
+ keyring==23.5.0
234
+ kiwisolver==1.4.5
235
+ langcodes==3.3.0
236
+ launchpadlib==1.10.16
237
+ lazr.restfulclient==0.14.4
238
+ lazr.uri==1.0.6
239
+ lazy_loader==0.3
240
+ libclang==16.0.6
241
+ librosa==0.10.1
242
+ lightgbm==4.1.0
243
+ linkify-it-py==2.0.3
244
+ llvmlite==0.41.1
245
+ locket==1.0.0
246
+ logical-unification==0.4.6
247
+ lxml==4.9.4
248
+ malloy==2023.1067
249
+ markdown-it-py==3.0.0
250
+ matplotlib-inline==0.1.6
251
+ matplotlib-venn==0.11.10
252
+ matplotlib==3.7.1
253
+ mdit-py-plugins==0.4.0
254
+ mdurl==0.1.2
255
+ miniKanren==1.0.3
256
+ missingno==0.5.2
257
+ mistune==0.8.4
258
+ mizani==0.9.3
259
+ mkl==2023.2.0
260
+ ml-dtypes==0.2.0
261
+ mlxtend==0.22.0
262
+ more-itertools==10.1.0
263
+ moviepy==1.0.3
264
+ mpmath==1.3.0
265
+ msgpack==1.0.8
266
+ multidict==6.0.5
267
+ multipledispatch==1.0.0
268
+ multiprocess==0.70.16
269
+ multitasking==0.0.11
270
+ murmurhash==1.0.10
271
+ music21==9.1.0
272
+ natsort==8.4.0
273
+ nbclassic==1.0.0
274
+ nbclient==0.10.0
275
+ nbconvert==6.5.4
276
+ nbformat==5.10.2
277
+ nest-asyncio==1.6.0
278
+ networkx==3.2.1
279
+ nibabel==4.0.2
280
+ nltk==3.8.1
281
+ notebook==6.5.5
282
+ notebook_shim==0.2.4
283
+ numba==0.58.1
284
+ numexpr==2.9.0
285
+ numpy==1.25.2
286
+ nvidia-cublas-cu12==12.1.3.1
287
+ nvidia-cuda-cupti-cu12==12.1.105
288
+ nvidia-cuda-nvrtc-cu12==12.1.105
289
+ nvidia-cuda-runtime-cu12==12.1.105
290
+ nvidia-cudnn-cu12==8.9.2.26
291
+ nvidia-cufft-cu12==11.0.2.54
292
+ nvidia-curand-cu12==10.3.2.106
293
+ nvidia-cusolver-cu12==11.4.5.107
294
+ nvidia-cusparse-cu12==12.1.0.106
295
+ nvidia-nccl-cu12==2.19.3
296
+ nvidia-nvjitlink-cu12==12.4.99
297
+ nvidia-nvtx-cu12==12.1.105
298
+ oauth2client==4.1.3
299
+ oauthlib==3.2.2
300
+ opencv-contrib-python==4.8.0.76
301
+ opencv-python-headless==4.9.0.80
302
+ opencv-python==4.8.0.76
303
+ openpyxl==3.1.2
304
+ opt-einsum==3.3.0
305
+ optax==0.2.1
306
+ orbax-checkpoint==0.4.4
307
+ osqp==0.6.2.post8
308
+ packaging==24.0
309
+ pandas-datareader==0.10.0
310
+ pandas-gbq==0.19.2
311
+ pandas-stubs==1.5.3.230304
312
+ pandas==1.5.3
313
+ pandocfilters==1.5.1
314
+ panel==1.3.8
315
+ param==2.0.2
316
+ parso==0.8.3
317
+ parsy==2.1
318
+ partd==1.4.1
319
+ pathlib==1.0.1
320
+ patsy==0.5.6
321
+ peewee==3.17.1
322
+ peft==0.9.1.dev0
323
+ pexpect==4.9.0
324
+ pickleshare==0.7.5
325
+ pip-tools==6.13.0
326
+ pip==23.1.2
327
+ platformdirs==4.2.0
328
+ plotly==5.15.0
329
+ plotnine==0.12.4
330
+ pluggy==1.4.0
331
+ polars==0.20.2
332
+ pooch==1.8.1
333
+ portpicker==1.5.2
334
+ prefetch-generator==1.0.3
335
+ preshed==3.0.9
336
+ prettytable==3.10.0
337
+ proglog==0.1.10
338
+ progressbar2==4.2.0
339
+ prometheus_client==0.20.0
340
+ promise==2.3
341
+ prompt-toolkit==3.0.43
342
+ prophet==1.1.5
343
+ proto-plus==1.23.0
344
+ protobuf==3.20.3
345
+ psutil==5.9.5
346
+ psycopg2==2.9.9
347
+ ptyprocess==0.7.0
348
+ py-cpuinfo==9.0.0
349
+ py4j==0.10.9.7
350
+ pyOpenSSL==24.1.0
351
+ pyarrow-hotfix==0.6
352
+ pyarrow==14.0.2
353
+ pyasn1-modules==0.3.0
354
+ pyasn1==0.5.1
355
+ pycocotools==2.0.7
356
+ pycparser==2.21
357
+ pydantic==2.6.4
358
+ pydantic_core==2.16.3
359
+ pydata-google-auth==1.8.2
360
+ pydot-ng==2.0.0
361
+ pydot==1.4.2
362
+ pydotplus==2.0.2
363
+ pyerfa==2.0.1.1
364
+ pygame==2.5.2
365
+ pymc==5.10.4
366
+ pymystem3==0.2.0
367
+ pyparsing==3.1.2
368
+ pyperclip==1.8.2
369
+ pyproj==3.6.1
370
+ pyproject_hooks==1.0.0
371
+ pyshp==2.3.1
372
+ pytensor==2.18.6
373
+ pytest==7.4.4
374
+ python-apt==0.0.0
375
+ python-box==7.1.1
376
+ python-dateutil==2.8.2
377
+ python-louvain==0.16
378
+ python-slugify==8.0.4
379
+ python-utils==3.8.2
380
+ pytz==2023.4
381
+ pyviz_comms==3.0.1
382
+ pyzmq==23.2.1
383
+ qdldl==0.1.7.post0
384
+ qudida==0.0.4
385
+ ratelim==0.1.6
386
+ referencing==0.33.0
387
+ regex==2023.12.25
388
+ requests-oauthlib==1.4.0
389
+ requests==2.31.0
390
+ requirements-parser==0.5.0
391
+ rich==13.7.1
392
+ rpds-py==0.18.0
393
+ rpy2==3.4.2
394
+ rsa==4.9
395
+ safetensors==0.4.2
396
+ scikit-image==0.19.3
397
+ scikit-learn==1.2.2
398
+ scipy==1.11.4
399
+ scooby==0.9.2
400
+ scs==3.2.4.post1
401
+ seaborn==0.13.1
402
+ sentencepiece==0.1.99
403
+ sentry-sdk==1.43.0
404
+ setproctitle==1.3.3
405
+ setuptools==67.7.2
406
+ shapely==2.0.3
407
+ shtab==1.7.1
408
+ six==1.16.0
409
+ six==1.16.0
410
+ sklearn-pandas==2.2.0
411
+ smart-open==6.4.0
412
+ smmap==5.0.1
413
+ sniffio==1.3.1
414
+ snowballstemmer==2.2.0
415
+ sortedcontainers==2.4.0
416
+ soundfile==0.12.1
417
+ soupsieve==2.5
418
+ soxr==0.3.7
419
+ spacy-legacy==3.0.12
420
+ spacy-loggers==1.0.5
421
+ spacy==3.7.4
422
+ sphinxcontrib-applehelp==1.0.8
423
+ sphinxcontrib-devhelp==1.0.6
424
+ sphinxcontrib-htmlhelp==2.0.5
425
+ sphinxcontrib-jsmath==1.0.1
426
+ sphinxcontrib-qthelp==1.0.7
427
+ sphinxcontrib-serializinghtml==1.1.10
428
+ sqlglot==20.11.0
429
+ sqlparse==0.4.4
430
+ srsly==2.4.8
431
+ stanio==0.3.0
432
+ statsmodels==0.14.1
433
+ sympy==1.12
434
+ tables==3.8.0
435
+ tabulate==0.9.0
436
+ tbb==2021.11.0
437
+ tblib==3.0.0
438
+ tenacity==8.2.3
439
+ tensorboard-data-server==0.7.2
440
+ tensorboard==2.15.2
441
+ tensorflow-datasets==4.9.4
442
+ tensorflow-estimator==2.15.0
443
+ tensorflow-gcs-config==2.15.0
444
+ tensorflow-hub==0.16.1
445
+ tensorflow-io-gcs-filesystem==0.36.0
446
+ tensorflow-metadata==1.14.0
447
+ tensorflow-probability==0.23.0
448
+ tensorflow==2.15.0
449
+ tensorstore==0.1.45
450
+ termcolor==2.4.0
451
+ terminado==0.18.1
452
+ text-unidecode==1.3
453
+ textblob==0.17.1
454
+ tf-slim==1.1.0
455
+ tf_keras==2.15.1
456
+ thinc==8.2.3
457
+ threadpoolctl==3.3.0
458
+ tifffile==2024.2.12
459
+ tinycss2==1.2.1
460
+ tokenizers==0.15.2
461
+ toml==0.10.2
462
+ tomli==2.0.1
463
+ toolz==0.12.1
464
+ torch==2.2.1+cu121
465
+ torchaudio==2.2.1+cu121
466
+ torchdata==0.7.1
467
+ torchsummary==1.5.1
468
+ torchtext==0.17.1
469
+ torchvision==0.17.1+cu121
470
+ tornado==6.3.3
471
+ tqdm==4.66.2
472
+ traitlets==5.7.1
473
+ traittypes==0.2.1
474
+ transformers==4.39.0
475
+ triton==2.2.0
476
+ trl==0.8.1
477
+ tweepy==4.14.0
478
+ typer==0.9.0
479
+ types-pytz==2024.1.0.20240203
480
+ types-setuptools==69.2.0.20240317
481
+ typing_extensions==4.10.0
482
+ tyro==0.7.3
483
+ tzlocal==5.2
484
+ uc-micro-py==1.0.3
485
+ uritemplate==4.1.1
486
+ urllib3==2.0.7
487
+ vega-datasets==0.9.0
488
+ wadllib==1.3.6
489
+ wandb==0.16.4
490
+ wasabi==1.1.2
491
+ wcwidth==0.2.13
492
+ weasel==0.3.4
493
+ webcolors==1.13
494
+ webencodings==0.5.1
495
+ websocket-client==1.7.0
496
+ wheel==0.43.0
497
+ widgetsnbextension==3.6.6
498
+ wordcloud==1.9.3
499
+ wrapt==1.14.1
500
+ xarray-einstats==0.7.0
501
+ xarray==2023.7.0
502
+ xgboost==2.0.3
503
+ xlrd==2.0.1
504
+ xxhash==3.4.1
505
+ xyzservices==2023.10.1
506
+ yarl==1.9.4
507
+ yellowbrick==1.5
508
+ yfinance==0.2.37
509
+ zict==3.0.0
510
+ zipp==3.18.1
wandb/run-20240321_015458-y11xrjxv/files/wandb-metadata.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.1.58+-x86_64-with-glibc2.35",
3
+ "python": "3.10.12",
4
+ "heartbeatAt": "2024-03-21T01:55:04.228731",
5
+ "startedAt": "2024-03-21T01:54:58.281206",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "Mistral-7B-Finetuning-Insurance.ipynb%EC%9D%98%20%EC%82%AC%EB%B3%B8",
11
+ "codePathLocal": null,
12
+ "colab": "https://colab.research.google.com/notebook#fileId=1glmddDRkqHDhfUT8slC17FnAssneTUkr",
13
+ "host": "d72aea89d4b6",
14
+ "username": "root",
15
+ "executable": "/usr/bin/python3",
16
+ "cpu_count": 1,
17
+ "cpu_count_logical": 2,
18
+ "cpu_freq": {
19
+ "current": 2000.176,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.176,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.176,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ }
34
+ ],
35
+ "disk": {
36
+ "/": {
37
+ "total": 78.1898422241211,
38
+ "used": 42.871490478515625
39
+ }
40
+ },
41
+ "gpu": "Tesla T4",
42
+ "gpu_count": 1,
43
+ "gpu_devices": [
44
+ {
45
+ "name": "Tesla T4",
46
+ "memory_total": 16106127360
47
+ }
48
+ ],
49
+ "memory": {
50
+ "total": 12.674781799316406
51
+ }
52
+ }
wandb/run-20240321_015458-y11xrjxv/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 1.3475, "train/grad_norm": 2.8251266479492188, "train/learning_rate": 0.0, "train/epoch": 0.44, "train/global_step": 60, "_timestamp": 1710987044.254427, "_runtime": 945.8694498538971, "_step": 6, "train_runtime": 959.9448, "train_samples_per_second": 0.25, "train_steps_per_second": 0.063, "total_flos": 1580862829903872.0, "train_loss": 1.4798423608144124}
wandb/run-20240321_015458-y11xrjxv/logs/debug.log ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-03-21 01:54:58,371 INFO MainThread:3894 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
+ 2024-03-21 01:54:58,372 INFO MainThread:3894 [wandb_setup.py:_flush():76] Configure stats pid to 3894
3
+ 2024-03-21 01:54:58,373 INFO MainThread:3894 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-03-21 01:54:58,374 INFO MainThread:3894 [wandb_setup.py:_flush():76] Loading settings from /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/settings
5
+ 2024-03-21 01:54:58,374 INFO MainThread:3894 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-03-21 01:54:58,374 INFO MainThread:3894 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-03-21 01:54:58,374 INFO MainThread:3894 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-03-21 01:54:58,374 INFO MainThread:3894 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-03-21 01:54:58,375 INFO MainThread:3894 [wandb_init.py:_log_setup():526] Logging user logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240321_015458-y11xrjxv/logs/debug.log
10
+ 2024-03-21 01:54:58,375 INFO MainThread:3894 [wandb_init.py:_log_setup():527] Logging internal logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240321_015458-y11xrjxv/logs/debug-internal.log
11
+ 2024-03-21 01:54:58,376 INFO MainThread:3894 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7e6f35554b80>
12
+ 2024-03-21 01:54:58,376 INFO MainThread:3894 [wandb_init.py:init():566] calling init triggers
13
+ 2024-03-21 01:54:58,377 INFO MainThread:3894 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-03-21 01:54:58,377 INFO MainThread:3894 [wandb_init.py:init():616] starting backend
16
+ 2024-03-21 01:54:58,377 INFO MainThread:3894 [wandb_init.py:init():620] setting up manager
17
+ 2024-03-21 01:54:58,382 INFO MainThread:3894 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-03-21 01:54:58,384 INFO MainThread:3894 [wandb_init.py:init():628] backend started and connected
19
+ 2024-03-21 01:54:58,424 INFO MainThread:3894 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
+ 2024-03-21 01:55:03,433 INFO MainThread:3894 [wandb_run.py:_label_probe_notebook():1305] Unable to probe notebook: 'NoneType' object has no attribute 'get'
21
+ 2024-03-21 01:55:03,433 INFO MainThread:3894 [wandb_init.py:init():720] updated telemetry
22
+ 2024-03-21 01:55:03,441 INFO MainThread:3894 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
23
+ 2024-03-21 01:55:04,081 INFO MainThread:3894 [wandb_run.py:_on_init():2262] communicating current version
24
+ 2024-03-21 01:55:04,191 INFO MainThread:3894 [wandb_run.py:_on_init():2271] got version response
25
+ 2024-03-21 01:55:04,191 INFO MainThread:3894 [wandb_init.py:init():804] starting run threads in backend
26
+ 2024-03-21 01:55:04,648 INFO MainThread:3894 [wandb_run.py:_console_start():2241] atexit reg
27
+ 2024-03-21 01:55:04,649 INFO MainThread:3894 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
+ 2024-03-21 01:55:04,649 INFO MainThread:3894 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
+ 2024-03-21 01:55:04,649 INFO MainThread:3894 [wandb_run.py:_redirect():2186] Redirects installed.
30
+ 2024-03-21 01:55:04,651 INFO MainThread:3894 [wandb_init.py:init():847] run started, returning control to user process
31
+ 2024-03-21 01:55:04,659 INFO MainThread:3894 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': 4096, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 10000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'mistralai/Mistral-7B-v0.1', 'transformers_version': '4.39.0', 'model_type': 'mistral', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'bfloat16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 3.0, 'max_steps': 60, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/runs/Mar21_01-53-07_d72aea89d4b6', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 10, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
32
+ 2024-03-21 02:11:10,542 INFO MainThread:3894 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-03-21 02:11:10,543 INFO MainThread:3894 [wandb_init.py:_pause_backend():437] pausing backend
34
+ 2024-03-21 02:11:38,631 INFO MainThread:3894 [wandb_init.py:_resume_backend():442] resuming backend
35
+ 2024-03-21 02:11:44,016 INFO MainThread:3894 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
+ 2024-03-21 02:11:44,016 INFO MainThread:3894 [wandb_init.py:_pause_backend():437] pausing backend
37
+ 2024-03-21 02:13:43,025 INFO MainThread:3894 [wandb_init.py:_resume_backend():442] resuming backend
38
+ 2024-03-21 02:14:53,819 INFO MainThread:3894 [jupyter.py:save_ipynb():373] not saving jupyter notebook
39
+ 2024-03-21 02:14:53,820 INFO MainThread:3894 [wandb_init.py:_pause_backend():437] pausing backend
40
+ 2024-03-21 02:14:59,353 INFO MainThread:3894 [wandb_init.py:_resume_backend():442] resuming backend
41
+ 2024-03-21 02:16:18,089 INFO MainThread:3894 [jupyter.py:save_ipynb():373] not saving jupyter notebook
42
+ 2024-03-21 02:16:18,090 INFO MainThread:3894 [wandb_init.py:_pause_backend():437] pausing backend
43
+ 2024-03-21 02:16:35,087 INFO MainThread:3894 [wandb_init.py:_resume_backend():442] resuming backend
44
+ 2024-03-21 02:16:35,095 INFO MainThread:3894 [jupyter.py:save_ipynb():373] not saving jupyter notebook
45
+ 2024-03-21 02:16:35,096 INFO MainThread:3894 [wandb_init.py:_pause_backend():437] pausing backend
46
+ 2024-03-21 02:16:42,320 INFO MainThread:3894 [wandb_init.py:_resume_backend():442] resuming backend
47
+ 2024-03-21 02:17:36,008 INFO MainThread:3894 [jupyter.py:save_ipynb():373] not saving jupyter notebook
48
+ 2024-03-21 02:17:36,008 INFO MainThread:3894 [wandb_init.py:_pause_backend():437] pausing backend
49
+ 2024-03-21 02:17:42,226 INFO MainThread:3894 [wandb_init.py:_resume_backend():442] resuming backend
50
+ 2024-03-21 02:18:34,485 INFO MainThread:3894 [jupyter.py:save_ipynb():373] not saving jupyter notebook
51
+ 2024-03-21 02:18:34,485 INFO MainThread:3894 [wandb_init.py:_pause_backend():437] pausing backend
52
+ 2024-03-21 02:19:00,990 INFO MainThread:3894 [wandb_init.py:_resume_backend():442] resuming backend
53
+ 2024-03-21 02:19:49,876 INFO MainThread:3894 [jupyter.py:save_ipynb():373] not saving jupyter notebook
54
+ 2024-03-21 02:19:49,877 INFO MainThread:3894 [wandb_init.py:_pause_backend():437] pausing backend
55
+ 2024-03-21 02:20:36,124 INFO MainThread:3894 [wandb_init.py:_resume_backend():442] resuming backend
56
+ 2024-03-21 02:21:19,829 INFO MainThread:3894 [jupyter.py:save_ipynb():373] not saving jupyter notebook
57
+ 2024-03-21 02:21:19,829 INFO MainThread:3894 [wandb_init.py:_pause_backend():437] pausing backend
58
+ 2024-03-21 02:34:49,695 INFO MainThread:3894 [wandb_init.py:_resume_backend():442] resuming backend
59
+ 2024-03-21 02:34:50,143 INFO MainThread:3894 [jupyter.py:save_ipynb():373] not saving jupyter notebook
60
+ 2024-03-21 02:34:50,144 INFO MainThread:3894 [wandb_init.py:_pause_backend():437] pausing backend
wandb/run-20240321_015458-y11xrjxv/run-y11xrjxv.wandb ADDED
Binary file (33.3 kB). View file
 
wandb/run-20240322_043407-hize5f4p/files/config.yaml ADDED
@@ -0,0 +1,678 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.12
7
+ cli_version: 0.16.4
8
+ framework: huggingface
9
+ huggingface_version: 4.39.0
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: false
12
+ start_time: 1711082047.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ 2:
29
+ - 1
30
+ - 2
31
+ - 3
32
+ - 5
33
+ - 11
34
+ - 12
35
+ - 49
36
+ - 51
37
+ - 53
38
+ - 55
39
+ - 71
40
+ - 84
41
+ - 98
42
+ 3:
43
+ - 7
44
+ - 23
45
+ 4: 3.10.12
46
+ 5: 0.16.4
47
+ 6: 4.39.0
48
+ 8:
49
+ - 1
50
+ - 5
51
+ - 12
52
+ 9:
53
+ 1: transformers_trainer
54
+ 13: linux-x86_64
55
+ m:
56
+ - 1: train/global_step
57
+ 6:
58
+ - 3
59
+ - 1: train/loss
60
+ 5: 1
61
+ 6:
62
+ - 1
63
+ - 1: train/grad_norm
64
+ 5: 1
65
+ 6:
66
+ - 1
67
+ - 1: train/learning_rate
68
+ 5: 1
69
+ 6:
70
+ - 1
71
+ - 1: train/epoch
72
+ 5: 1
73
+ 6:
74
+ - 1
75
+ vocab_size:
76
+ desc: null
77
+ value: 32000
78
+ max_position_embeddings:
79
+ desc: null
80
+ value: 32768
81
+ hidden_size:
82
+ desc: null
83
+ value: 4096
84
+ intermediate_size:
85
+ desc: null
86
+ value: 14336
87
+ num_hidden_layers:
88
+ desc: null
89
+ value: 32
90
+ num_attention_heads:
91
+ desc: null
92
+ value: 32
93
+ sliding_window:
94
+ desc: null
95
+ value: null
96
+ num_key_value_heads:
97
+ desc: null
98
+ value: 8
99
+ hidden_act:
100
+ desc: null
101
+ value: silu
102
+ initializer_range:
103
+ desc: null
104
+ value: 0.02
105
+ rms_norm_eps:
106
+ desc: null
107
+ value: 1.0e-05
108
+ use_cache:
109
+ desc: null
110
+ value: false
111
+ rope_theta:
112
+ desc: null
113
+ value: 1000000.0
114
+ attention_dropout:
115
+ desc: null
116
+ value: 0.0
117
+ return_dict:
118
+ desc: null
119
+ value: true
120
+ output_hidden_states:
121
+ desc: null
122
+ value: false
123
+ output_attentions:
124
+ desc: null
125
+ value: false
126
+ torchscript:
127
+ desc: null
128
+ value: false
129
+ torch_dtype:
130
+ desc: null
131
+ value: bfloat16
132
+ use_bfloat16:
133
+ desc: null
134
+ value: false
135
+ tf_legacy_loss:
136
+ desc: null
137
+ value: false
138
+ pruned_heads:
139
+ desc: null
140
+ value: {}
141
+ tie_word_embeddings:
142
+ desc: null
143
+ value: false
144
+ chunk_size_feed_forward:
145
+ desc: null
146
+ value: 0
147
+ is_encoder_decoder:
148
+ desc: null
149
+ value: false
150
+ is_decoder:
151
+ desc: null
152
+ value: false
153
+ cross_attention_hidden_size:
154
+ desc: null
155
+ value: null
156
+ add_cross_attention:
157
+ desc: null
158
+ value: false
159
+ tie_encoder_decoder:
160
+ desc: null
161
+ value: false
162
+ max_length:
163
+ desc: null
164
+ value: 20
165
+ min_length:
166
+ desc: null
167
+ value: 0
168
+ do_sample:
169
+ desc: null
170
+ value: false
171
+ early_stopping:
172
+ desc: null
173
+ value: false
174
+ num_beams:
175
+ desc: null
176
+ value: 1
177
+ num_beam_groups:
178
+ desc: null
179
+ value: 1
180
+ diversity_penalty:
181
+ desc: null
182
+ value: 0.0
183
+ temperature:
184
+ desc: null
185
+ value: 1.0
186
+ top_k:
187
+ desc: null
188
+ value: 50
189
+ top_p:
190
+ desc: null
191
+ value: 1.0
192
+ typical_p:
193
+ desc: null
194
+ value: 1.0
195
+ repetition_penalty:
196
+ desc: null
197
+ value: 1.0
198
+ length_penalty:
199
+ desc: null
200
+ value: 1.0
201
+ no_repeat_ngram_size:
202
+ desc: null
203
+ value: 0
204
+ encoder_no_repeat_ngram_size:
205
+ desc: null
206
+ value: 0
207
+ bad_words_ids:
208
+ desc: null
209
+ value: null
210
+ num_return_sequences:
211
+ desc: null
212
+ value: 1
213
+ output_scores:
214
+ desc: null
215
+ value: false
216
+ return_dict_in_generate:
217
+ desc: null
218
+ value: false
219
+ forced_bos_token_id:
220
+ desc: null
221
+ value: null
222
+ forced_eos_token_id:
223
+ desc: null
224
+ value: null
225
+ remove_invalid_values:
226
+ desc: null
227
+ value: false
228
+ exponential_decay_length_penalty:
229
+ desc: null
230
+ value: null
231
+ suppress_tokens:
232
+ desc: null
233
+ value: null
234
+ begin_suppress_tokens:
235
+ desc: null
236
+ value: null
237
+ architectures:
238
+ desc: null
239
+ value:
240
+ - MistralForCausalLM
241
+ finetuning_task:
242
+ desc: null
243
+ value: null
244
+ id2label:
245
+ desc: null
246
+ value:
247
+ '0': LABEL_0
248
+ '1': LABEL_1
249
+ label2id:
250
+ desc: null
251
+ value:
252
+ LABEL_0: 0
253
+ LABEL_1: 1
254
+ tokenizer_class:
255
+ desc: null
256
+ value: null
257
+ prefix:
258
+ desc: null
259
+ value: null
260
+ bos_token_id:
261
+ desc: null
262
+ value: 1
263
+ pad_token_id:
264
+ desc: null
265
+ value: null
266
+ eos_token_id:
267
+ desc: null
268
+ value: 2
269
+ sep_token_id:
270
+ desc: null
271
+ value: null
272
+ decoder_start_token_id:
273
+ desc: null
274
+ value: null
275
+ task_specific_params:
276
+ desc: null
277
+ value: null
278
+ problem_type:
279
+ desc: null
280
+ value: null
281
+ _name_or_path:
282
+ desc: null
283
+ value: mistralai/Mistral-7B-Instruct-v0.2
284
+ transformers_version:
285
+ desc: null
286
+ value: 4.39.0
287
+ model_type:
288
+ desc: null
289
+ value: mistral
290
+ quantization_config:
291
+ desc: null
292
+ value:
293
+ quant_method: QuantizationMethod.BITS_AND_BYTES
294
+ _load_in_8bit: false
295
+ _load_in_4bit: true
296
+ llm_int8_threshold: 6.0
297
+ llm_int8_skip_modules: null
298
+ llm_int8_enable_fp32_cpu_offload: false
299
+ llm_int8_has_fp16_weight: false
300
+ bnb_4bit_quant_type: nf4
301
+ bnb_4bit_use_double_quant: true
302
+ bnb_4bit_compute_dtype: bfloat16
303
+ bnb_4bit_quant_storage: uint8
304
+ load_in_4bit: true
305
+ load_in_8bit: false
306
+ output_dir:
307
+ desc: null
308
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance
309
+ overwrite_output_dir:
310
+ desc: null
311
+ value: false
312
+ do_train:
313
+ desc: null
314
+ value: false
315
+ do_eval:
316
+ desc: null
317
+ value: false
318
+ do_predict:
319
+ desc: null
320
+ value: false
321
+ evaluation_strategy:
322
+ desc: null
323
+ value: 'no'
324
+ prediction_loss_only:
325
+ desc: null
326
+ value: false
327
+ per_device_train_batch_size:
328
+ desc: null
329
+ value: 2
330
+ per_device_eval_batch_size:
331
+ desc: null
332
+ value: 8
333
+ per_gpu_train_batch_size:
334
+ desc: null
335
+ value: null
336
+ per_gpu_eval_batch_size:
337
+ desc: null
338
+ value: null
339
+ gradient_accumulation_steps:
340
+ desc: null
341
+ value: 2
342
+ eval_accumulation_steps:
343
+ desc: null
344
+ value: null
345
+ eval_delay:
346
+ desc: null
347
+ value: 0
348
+ learning_rate:
349
+ desc: null
350
+ value: 0.0002
351
+ weight_decay:
352
+ desc: null
353
+ value: 0.0
354
+ adam_beta1:
355
+ desc: null
356
+ value: 0.9
357
+ adam_beta2:
358
+ desc: null
359
+ value: 0.999
360
+ adam_epsilon:
361
+ desc: null
362
+ value: 1.0e-08
363
+ max_grad_norm:
364
+ desc: null
365
+ value: 0.3
366
+ num_train_epochs:
367
+ desc: null
368
+ value: 3.0
369
+ max_steps:
370
+ desc: null
371
+ value: 60
372
+ lr_scheduler_type:
373
+ desc: null
374
+ value: cosine
375
+ lr_scheduler_kwargs:
376
+ desc: null
377
+ value: {}
378
+ warmup_ratio:
379
+ desc: null
380
+ value: 0.03
381
+ warmup_steps:
382
+ desc: null
383
+ value: 0
384
+ log_level:
385
+ desc: null
386
+ value: passive
387
+ log_level_replica:
388
+ desc: null
389
+ value: warning
390
+ log_on_each_node:
391
+ desc: null
392
+ value: true
393
+ logging_dir:
394
+ desc: null
395
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/runs/Mar22_04-33-33_6b98cdd15d10
396
+ logging_strategy:
397
+ desc: null
398
+ value: steps
399
+ logging_first_step:
400
+ desc: null
401
+ value: false
402
+ logging_steps:
403
+ desc: null
404
+ value: 10
405
+ logging_nan_inf_filter:
406
+ desc: null
407
+ value: true
408
+ save_strategy:
409
+ desc: null
410
+ value: steps
411
+ save_steps:
412
+ desc: null
413
+ value: 10
414
+ save_total_limit:
415
+ desc: null
416
+ value: null
417
+ save_safetensors:
418
+ desc: null
419
+ value: true
420
+ save_on_each_node:
421
+ desc: null
422
+ value: false
423
+ save_only_model:
424
+ desc: null
425
+ value: false
426
+ no_cuda:
427
+ desc: null
428
+ value: false
429
+ use_cpu:
430
+ desc: null
431
+ value: false
432
+ use_mps_device:
433
+ desc: null
434
+ value: false
435
+ seed:
436
+ desc: null
437
+ value: 42
438
+ data_seed:
439
+ desc: null
440
+ value: null
441
+ jit_mode_eval:
442
+ desc: null
443
+ value: false
444
+ use_ipex:
445
+ desc: null
446
+ value: false
447
+ bf16:
448
+ desc: null
449
+ value: false
450
+ fp16:
451
+ desc: null
452
+ value: false
453
+ fp16_opt_level:
454
+ desc: null
455
+ value: O1
456
+ half_precision_backend:
457
+ desc: null
458
+ value: auto
459
+ bf16_full_eval:
460
+ desc: null
461
+ value: false
462
+ fp16_full_eval:
463
+ desc: null
464
+ value: false
465
+ tf32:
466
+ desc: null
467
+ value: false
468
+ local_rank:
469
+ desc: null
470
+ value: 0
471
+ ddp_backend:
472
+ desc: null
473
+ value: null
474
+ tpu_num_cores:
475
+ desc: null
476
+ value: null
477
+ tpu_metrics_debug:
478
+ desc: null
479
+ value: false
480
+ debug:
481
+ desc: null
482
+ value: []
483
+ dataloader_drop_last:
484
+ desc: null
485
+ value: false
486
+ eval_steps:
487
+ desc: null
488
+ value: null
489
+ dataloader_num_workers:
490
+ desc: null
491
+ value: 0
492
+ dataloader_prefetch_factor:
493
+ desc: null
494
+ value: null
495
+ past_index:
496
+ desc: null
497
+ value: -1
498
+ run_name:
499
+ desc: null
500
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance
501
+ disable_tqdm:
502
+ desc: null
503
+ value: false
504
+ remove_unused_columns:
505
+ desc: null
506
+ value: true
507
+ label_names:
508
+ desc: null
509
+ value: null
510
+ load_best_model_at_end:
511
+ desc: null
512
+ value: false
513
+ metric_for_best_model:
514
+ desc: null
515
+ value: null
516
+ greater_is_better:
517
+ desc: null
518
+ value: null
519
+ ignore_data_skip:
520
+ desc: null
521
+ value: false
522
+ fsdp:
523
+ desc: null
524
+ value: []
525
+ fsdp_min_num_params:
526
+ desc: null
527
+ value: 0
528
+ fsdp_config:
529
+ desc: null
530
+ value:
531
+ min_num_params: 0
532
+ xla: false
533
+ xla_fsdp_v2: false
534
+ xla_fsdp_grad_ckpt: false
535
+ fsdp_transformer_layer_cls_to_wrap:
536
+ desc: null
537
+ value: null
538
+ accelerator_config:
539
+ desc: null
540
+ value:
541
+ split_batches: false
542
+ dispatch_batches: null
543
+ even_batches: true
544
+ use_seedable_sampler: true
545
+ deepspeed:
546
+ desc: null
547
+ value: null
548
+ label_smoothing_factor:
549
+ desc: null
550
+ value: 0.0
551
+ optim:
552
+ desc: null
553
+ value: paged_adamw_32bit
554
+ optim_args:
555
+ desc: null
556
+ value: null
557
+ adafactor:
558
+ desc: null
559
+ value: false
560
+ group_by_length:
561
+ desc: null
562
+ value: true
563
+ length_column_name:
564
+ desc: null
565
+ value: length
566
+ report_to:
567
+ desc: null
568
+ value:
569
+ - tensorboard
570
+ - wandb
571
+ ddp_find_unused_parameters:
572
+ desc: null
573
+ value: null
574
+ ddp_bucket_cap_mb:
575
+ desc: null
576
+ value: null
577
+ ddp_broadcast_buffers:
578
+ desc: null
579
+ value: null
580
+ dataloader_pin_memory:
581
+ desc: null
582
+ value: true
583
+ dataloader_persistent_workers:
584
+ desc: null
585
+ value: false
586
+ skip_memory_metrics:
587
+ desc: null
588
+ value: true
589
+ use_legacy_prediction_loop:
590
+ desc: null
591
+ value: false
592
+ push_to_hub:
593
+ desc: null
594
+ value: true
595
+ resume_from_checkpoint:
596
+ desc: null
597
+ value: null
598
+ hub_model_id:
599
+ desc: null
600
+ value: null
601
+ hub_strategy:
602
+ desc: null
603
+ value: every_save
604
+ hub_token:
605
+ desc: null
606
+ value: <HUB_TOKEN>
607
+ hub_private_repo:
608
+ desc: null
609
+ value: false
610
+ hub_always_push:
611
+ desc: null
612
+ value: false
613
+ gradient_checkpointing:
614
+ desc: null
615
+ value: false
616
+ gradient_checkpointing_kwargs:
617
+ desc: null
618
+ value: null
619
+ include_inputs_for_metrics:
620
+ desc: null
621
+ value: false
622
+ fp16_backend:
623
+ desc: null
624
+ value: auto
625
+ push_to_hub_model_id:
626
+ desc: null
627
+ value: null
628
+ push_to_hub_organization:
629
+ desc: null
630
+ value: null
631
+ push_to_hub_token:
632
+ desc: null
633
+ value: <PUSH_TO_HUB_TOKEN>
634
+ mp_parameters:
635
+ desc: null
636
+ value: ''
637
+ auto_find_batch_size:
638
+ desc: null
639
+ value: false
640
+ full_determinism:
641
+ desc: null
642
+ value: false
643
+ torchdynamo:
644
+ desc: null
645
+ value: null
646
+ ray_scope:
647
+ desc: null
648
+ value: last
649
+ ddp_timeout:
650
+ desc: null
651
+ value: 1800
652
+ torch_compile:
653
+ desc: null
654
+ value: false
655
+ torch_compile_backend:
656
+ desc: null
657
+ value: null
658
+ torch_compile_mode:
659
+ desc: null
660
+ value: null
661
+ dispatch_batches:
662
+ desc: null
663
+ value: null
664
+ split_batches:
665
+ desc: null
666
+ value: null
667
+ include_tokens_per_second:
668
+ desc: null
669
+ value: false
670
+ include_num_input_tokens_seen:
671
+ desc: null
672
+ value: false
673
+ neftune_noise_alpha:
674
+ desc: null
675
+ value: null
676
+ optim_target_modules:
677
+ desc: null
678
+ value: null
wandb/run-20240322_043407-hize5f4p/files/requirements.txt ADDED
@@ -0,0 +1,510 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ CacheControl==0.14.0
3
+ Cython==3.0.9
4
+ Flask==2.2.5
5
+ GDAL==3.6.4
6
+ GitPython==3.1.42
7
+ Jinja2==3.1.3
8
+ Markdown==3.6
9
+ MarkupSafe==2.1.5
10
+ Pillow==9.4.0
11
+ PyDrive2==1.6.3
12
+ PyDrive==1.3.1
13
+ PyGObject==3.42.1
14
+ PyJWT==2.3.0
15
+ PyOpenGL==3.1.7
16
+ PySocks==1.7.1
17
+ PyWavelets==1.5.0
18
+ PyYAML==6.0.1
19
+ Pygments==2.16.1
20
+ SQLAlchemy==2.0.28
21
+ SecretStorage==3.3.1
22
+ Send2Trash==1.8.2
23
+ Sphinx==5.0.2
24
+ Werkzeug==3.0.1
25
+ absl-py==1.4.0
26
+ accelerate==0.28.0
27
+ aiohttp==3.9.3
28
+ aiosignal==1.3.1
29
+ alabaster==0.7.16
30
+ albumentations==1.3.1
31
+ altair==4.2.2
32
+ annotated-types==0.6.0
33
+ anyio==3.7.1
34
+ appdirs==1.4.4
35
+ argon2-cffi-bindings==21.2.0
36
+ argon2-cffi==23.1.0
37
+ array-record==0.5.0
38
+ arviz==0.15.1
39
+ astropy==5.3.4
40
+ astunparse==1.6.3
41
+ async-timeout==4.0.3
42
+ atpublic==4.0
43
+ attrs==23.2.0
44
+ audioread==3.0.1
45
+ autograd==1.6.2
46
+ backcall==0.2.0
47
+ beautifulsoup4==4.12.3
48
+ bidict==0.23.1
49
+ bigframes==0.25.0
50
+ bitsandbytes==0.43.0
51
+ bleach==6.1.0
52
+ blinker==1.4
53
+ blis==0.7.11
54
+ blosc2==2.0.0
55
+ bokeh==3.3.4
56
+ bqplot==0.12.43
57
+ branca==0.7.1
58
+ build==1.1.1
59
+ cachetools==5.3.3
60
+ catalogue==2.0.10
61
+ certifi==2024.2.2
62
+ cffi==1.16.0
63
+ chardet==5.2.0
64
+ charset-normalizer==3.3.2
65
+ chex==0.1.85
66
+ click-plugins==1.1.1
67
+ click==8.1.7
68
+ cligj==0.7.2
69
+ cloudpathlib==0.16.0
70
+ cloudpickle==2.2.1
71
+ cmake==3.27.9
72
+ cmdstanpy==1.2.1
73
+ colorcet==3.1.0
74
+ colorlover==0.3.0
75
+ colour==0.1.5
76
+ community==1.0.0b1
77
+ confection==0.1.4
78
+ cons==0.4.6
79
+ contextlib2==21.6.0
80
+ contourpy==1.2.0
81
+ cryptography==42.0.5
82
+ cufflinks==0.17.3
83
+ cupy-cuda12x==12.2.0
84
+ cvxopt==1.3.2
85
+ cvxpy==1.3.3
86
+ cycler==0.12.1
87
+ cymem==2.0.8
88
+ dask==2023.8.1
89
+ datascience==0.17.6
90
+ datasets==2.18.0
91
+ db-dtypes==1.2.0
92
+ dbus-python==1.2.18
93
+ debugpy==1.6.6
94
+ decorator==4.4.2
95
+ defusedxml==0.7.1
96
+ dill==0.3.8
97
+ distributed==2023.8.1
98
+ distro==1.7.0
99
+ dlib==19.24.2
100
+ dm-tree==0.1.8
101
+ docker-pycreds==0.4.0
102
+ docstring_parser==0.16
103
+ docutils==0.18.1
104
+ dopamine-rl==4.0.6
105
+ duckdb==0.9.2
106
+ earthengine-api==0.1.394
107
+ easydict==1.13
108
+ ecos==2.0.13
109
+ editdistance==0.6.2
110
+ eerepr==0.0.4
111
+ einops==0.7.0
112
+ en-core-web-sm==3.7.1
113
+ entrypoints==0.4
114
+ et-xmlfile==1.1.0
115
+ etils==1.7.0
116
+ etuples==0.3.9
117
+ exceptiongroup==1.2.0
118
+ fastai==2.7.14
119
+ fastcore==1.5.29
120
+ fastdownload==0.0.7
121
+ fastjsonschema==2.19.1
122
+ fastprogress==1.0.3
123
+ fastrlock==0.8.2
124
+ filelock==3.13.1
125
+ fiona==1.9.6
126
+ firebase-admin==5.3.0
127
+ flatbuffers==24.3.7
128
+ flax==0.8.2
129
+ folium==0.14.0
130
+ fonttools==4.49.0
131
+ frozendict==2.4.0
132
+ frozenlist==1.4.1
133
+ fsspec==2023.6.0
134
+ future==0.18.3
135
+ gast==0.5.4
136
+ gcsfs==2023.6.0
137
+ gdown==4.7.3
138
+ geemap==0.32.0
139
+ gensim==4.3.2
140
+ geocoder==1.38.1
141
+ geographiclib==2.0
142
+ geopandas==0.13.2
143
+ geopy==2.3.0
144
+ gin-config==0.5.0
145
+ gitdb==4.0.11
146
+ glob2==0.7
147
+ google-ai-generativelanguage==0.4.0
148
+ google-api-core==2.11.1
149
+ google-api-python-client==2.84.0
150
+ google-auth-httplib2==0.1.1
151
+ google-auth-oauthlib==1.2.0
152
+ google-auth==2.27.0
153
+ google-cloud-aiplatform==1.44.0
154
+ google-cloud-bigquery-connection==1.12.1
155
+ google-cloud-bigquery-storage==2.24.0
156
+ google-cloud-bigquery==3.12.0
157
+ google-cloud-core==2.3.3
158
+ google-cloud-datastore==2.15.2
159
+ google-cloud-firestore==2.11.1
160
+ google-cloud-functions==1.13.3
161
+ google-cloud-iam==2.14.3
162
+ google-cloud-language==2.13.3
163
+ google-cloud-resource-manager==1.12.3
164
+ google-cloud-storage==2.8.0
165
+ google-cloud-translate==3.11.3
166
+ google-colab==1.0.0
167
+ google-crc32c==1.5.0
168
+ google-generativeai==0.3.2
169
+ google-pasta==0.2.0
170
+ google-resumable-media==2.7.0
171
+ google==2.0.3
172
+ googleapis-common-protos==1.63.0
173
+ googledrivedownloader==0.4
174
+ graphviz==0.20.1
175
+ greenlet==3.0.3
176
+ grpc-google-iam-v1==0.13.0
177
+ grpcio-status==1.48.2
178
+ grpcio==1.62.1
179
+ gspread-dataframe==3.3.1
180
+ gspread==3.4.2
181
+ gym-notices==0.0.8
182
+ gym==0.25.2
183
+ h5netcdf==1.3.0
184
+ h5py==3.9.0
185
+ holidays==0.44
186
+ holoviews==1.17.1
187
+ html5lib==1.1
188
+ httpimport==1.3.1
189
+ httplib2==0.22.0
190
+ huggingface-hub==0.20.3
191
+ humanize==4.7.0
192
+ hyperopt==0.2.7
193
+ ibis-framework==8.0.0
194
+ idna==3.6
195
+ imageio-ffmpeg==0.4.9
196
+ imageio==2.31.6
197
+ imagesize==1.4.1
198
+ imbalanced-learn==0.10.1
199
+ imgaug==0.4.0
200
+ importlib_metadata==7.0.2
201
+ importlib_resources==6.3.0
202
+ imutils==0.5.4
203
+ inflect==7.0.0
204
+ iniconfig==2.0.0
205
+ intel-openmp==2023.2.4
206
+ ipyevents==2.0.2
207
+ ipyfilechooser==0.6.0
208
+ ipykernel==5.5.6
209
+ ipyleaflet==0.18.2
210
+ ipython-genutils==0.2.0
211
+ ipython-sql==0.5.0
212
+ ipython==7.34.0
213
+ ipytree==0.2.2
214
+ ipywidgets==7.7.1
215
+ itsdangerous==2.1.2
216
+ jax==0.4.23
217
+ jaxlib==0.4.23+cuda12.cudnn89
218
+ jeepney==0.7.1
219
+ jieba==0.42.1
220
+ joblib==1.3.2
221
+ jsonpickle==3.0.3
222
+ jsonschema-specifications==2023.12.1
223
+ jsonschema==4.19.2
224
+ jupyter-client==6.1.12
225
+ jupyter-console==6.1.0
226
+ jupyter-server==1.24.0
227
+ jupyter_core==5.7.2
228
+ jupyterlab_pygments==0.3.0
229
+ jupyterlab_widgets==3.0.10
230
+ kaggle==1.5.16
231
+ kagglehub==0.2.0
232
+ keras==2.15.0
233
+ keyring==23.5.0
234
+ kiwisolver==1.4.5
235
+ langcodes==3.3.0
236
+ launchpadlib==1.10.16
237
+ lazr.restfulclient==0.14.4
238
+ lazr.uri==1.0.6
239
+ lazy_loader==0.3
240
+ libclang==16.0.6
241
+ librosa==0.10.1
242
+ lightgbm==4.1.0
243
+ linkify-it-py==2.0.3
244
+ llvmlite==0.41.1
245
+ locket==1.0.0
246
+ logical-unification==0.4.6
247
+ lxml==4.9.4
248
+ malloy==2023.1067
249
+ markdown-it-py==3.0.0
250
+ matplotlib-inline==0.1.6
251
+ matplotlib-venn==0.11.10
252
+ matplotlib==3.7.1
253
+ mdit-py-plugins==0.4.0
254
+ mdurl==0.1.2
255
+ miniKanren==1.0.3
256
+ missingno==0.5.2
257
+ mistune==0.8.4
258
+ mizani==0.9.3
259
+ mkl==2023.2.0
260
+ ml-dtypes==0.2.0
261
+ mlxtend==0.22.0
262
+ more-itertools==10.1.0
263
+ moviepy==1.0.3
264
+ mpmath==1.3.0
265
+ msgpack==1.0.8
266
+ multidict==6.0.5
267
+ multipledispatch==1.0.0
268
+ multiprocess==0.70.16
269
+ multitasking==0.0.11
270
+ murmurhash==1.0.10
271
+ music21==9.1.0
272
+ natsort==8.4.0
273
+ nbclassic==1.0.0
274
+ nbclient==0.10.0
275
+ nbconvert==6.5.4
276
+ nbformat==5.10.2
277
+ nest-asyncio==1.6.0
278
+ networkx==3.2.1
279
+ nibabel==4.0.2
280
+ nltk==3.8.1
281
+ notebook==6.5.5
282
+ notebook_shim==0.2.4
283
+ numba==0.58.1
284
+ numexpr==2.9.0
285
+ numpy==1.25.2
286
+ nvidia-cublas-cu12==12.1.3.1
287
+ nvidia-cuda-cupti-cu12==12.1.105
288
+ nvidia-cuda-nvrtc-cu12==12.1.105
289
+ nvidia-cuda-runtime-cu12==12.1.105
290
+ nvidia-cudnn-cu12==8.9.2.26
291
+ nvidia-cufft-cu12==11.0.2.54
292
+ nvidia-curand-cu12==10.3.2.106
293
+ nvidia-cusolver-cu12==11.4.5.107
294
+ nvidia-cusparse-cu12==12.1.0.106
295
+ nvidia-nccl-cu12==2.19.3
296
+ nvidia-nvjitlink-cu12==12.4.99
297
+ nvidia-nvtx-cu12==12.1.105
298
+ oauth2client==4.1.3
299
+ oauthlib==3.2.2
300
+ opencv-contrib-python==4.8.0.76
301
+ opencv-python-headless==4.9.0.80
302
+ opencv-python==4.8.0.76
303
+ openpyxl==3.1.2
304
+ opt-einsum==3.3.0
305
+ optax==0.2.1
306
+ orbax-checkpoint==0.4.4
307
+ osqp==0.6.2.post8
308
+ packaging==24.0
309
+ pandas-datareader==0.10.0
310
+ pandas-gbq==0.19.2
311
+ pandas-stubs==1.5.3.230304
312
+ pandas==1.5.3
313
+ pandocfilters==1.5.1
314
+ panel==1.3.8
315
+ param==2.0.2
316
+ parso==0.8.3
317
+ parsy==2.1
318
+ partd==1.4.1
319
+ pathlib==1.0.1
320
+ patsy==0.5.6
321
+ peewee==3.17.1
322
+ peft==0.10.1.dev0
323
+ pexpect==4.9.0
324
+ pickleshare==0.7.5
325
+ pip-tools==6.13.0
326
+ pip==23.1.2
327
+ platformdirs==4.2.0
328
+ plotly==5.15.0
329
+ plotnine==0.12.4
330
+ pluggy==1.4.0
331
+ polars==0.20.2
332
+ pooch==1.8.1
333
+ portpicker==1.5.2
334
+ prefetch-generator==1.0.3
335
+ preshed==3.0.9
336
+ prettytable==3.10.0
337
+ proglog==0.1.10
338
+ progressbar2==4.2.0
339
+ prometheus_client==0.20.0
340
+ promise==2.3
341
+ prompt-toolkit==3.0.43
342
+ prophet==1.1.5
343
+ proto-plus==1.23.0
344
+ protobuf==3.20.3
345
+ psutil==5.9.5
346
+ psycopg2==2.9.9
347
+ ptyprocess==0.7.0
348
+ py-cpuinfo==9.0.0
349
+ py4j==0.10.9.7
350
+ pyOpenSSL==24.1.0
351
+ pyarrow-hotfix==0.6
352
+ pyarrow==14.0.2
353
+ pyasn1-modules==0.3.0
354
+ pyasn1==0.5.1
355
+ pycocotools==2.0.7
356
+ pycparser==2.21
357
+ pydantic==2.6.4
358
+ pydantic_core==2.16.3
359
+ pydata-google-auth==1.8.2
360
+ pydot-ng==2.0.0
361
+ pydot==1.4.2
362
+ pydotplus==2.0.2
363
+ pyerfa==2.0.1.1
364
+ pygame==2.5.2
365
+ pymc==5.10.4
366
+ pymystem3==0.2.0
367
+ pyparsing==3.1.2
368
+ pyperclip==1.8.2
369
+ pyproj==3.6.1
370
+ pyproject_hooks==1.0.0
371
+ pyshp==2.3.1
372
+ pytensor==2.18.6
373
+ pytest==7.4.4
374
+ python-apt==0.0.0
375
+ python-box==7.1.1
376
+ python-dateutil==2.8.2
377
+ python-louvain==0.16
378
+ python-slugify==8.0.4
379
+ python-utils==3.8.2
380
+ pytz==2023.4
381
+ pyviz_comms==3.0.1
382
+ pyzmq==23.2.1
383
+ qdldl==0.1.7.post0
384
+ qudida==0.0.4
385
+ ratelim==0.1.6
386
+ referencing==0.33.0
387
+ regex==2023.12.25
388
+ requests-oauthlib==1.4.0
389
+ requests==2.31.0
390
+ requirements-parser==0.5.0
391
+ rich==13.7.1
392
+ rpds-py==0.18.0
393
+ rpy2==3.4.2
394
+ rsa==4.9
395
+ safetensors==0.4.2
396
+ scikit-image==0.19.3
397
+ scikit-learn==1.2.2
398
+ scipy==1.11.4
399
+ scooby==0.9.2
400
+ scs==3.2.4.post1
401
+ seaborn==0.13.1
402
+ sentencepiece==0.1.99
403
+ sentry-sdk==1.43.0
404
+ setproctitle==1.3.3
405
+ setuptools==67.7.2
406
+ shapely==2.0.3
407
+ shtab==1.7.1
408
+ six==1.16.0
409
+ six==1.16.0
410
+ sklearn-pandas==2.2.0
411
+ smart-open==6.4.0
412
+ smmap==5.0.1
413
+ sniffio==1.3.1
414
+ snowballstemmer==2.2.0
415
+ sortedcontainers==2.4.0
416
+ soundfile==0.12.1
417
+ soupsieve==2.5
418
+ soxr==0.3.7
419
+ spacy-legacy==3.0.12
420
+ spacy-loggers==1.0.5
421
+ spacy==3.7.4
422
+ sphinxcontrib-applehelp==1.0.8
423
+ sphinxcontrib-devhelp==1.0.6
424
+ sphinxcontrib-htmlhelp==2.0.5
425
+ sphinxcontrib-jsmath==1.0.1
426
+ sphinxcontrib-qthelp==1.0.7
427
+ sphinxcontrib-serializinghtml==1.1.10
428
+ sqlglot==20.11.0
429
+ sqlparse==0.4.4
430
+ srsly==2.4.8
431
+ stanio==0.3.0
432
+ statsmodels==0.14.1
433
+ sympy==1.12
434
+ tables==3.8.0
435
+ tabulate==0.9.0
436
+ tbb==2021.11.0
437
+ tblib==3.0.0
438
+ tenacity==8.2.3
439
+ tensorboard-data-server==0.7.2
440
+ tensorboard==2.15.2
441
+ tensorflow-datasets==4.9.4
442
+ tensorflow-estimator==2.15.0
443
+ tensorflow-gcs-config==2.15.0
444
+ tensorflow-hub==0.16.1
445
+ tensorflow-io-gcs-filesystem==0.36.0
446
+ tensorflow-metadata==1.14.0
447
+ tensorflow-probability==0.23.0
448
+ tensorflow==2.15.0
449
+ tensorstore==0.1.45
450
+ termcolor==2.4.0
451
+ terminado==0.18.1
452
+ text-unidecode==1.3
453
+ textblob==0.17.1
454
+ tf-slim==1.1.0
455
+ tf_keras==2.15.1
456
+ thinc==8.2.3
457
+ threadpoolctl==3.3.0
458
+ tifffile==2024.2.12
459
+ tinycss2==1.2.1
460
+ tokenizers==0.15.2
461
+ toml==0.10.2
462
+ tomli==2.0.1
463
+ toolz==0.12.1
464
+ torch==2.2.1+cu121
465
+ torchaudio==2.2.1+cu121
466
+ torchdata==0.7.1
467
+ torchsummary==1.5.1
468
+ torchtext==0.17.1
469
+ torchvision==0.17.1+cu121
470
+ tornado==6.3.3
471
+ tqdm==4.66.2
472
+ traitlets==5.7.1
473
+ traittypes==0.2.1
474
+ transformers==4.39.0
475
+ triton==2.2.0
476
+ trl==0.8.1
477
+ tweepy==4.14.0
478
+ typer==0.9.0
479
+ types-pytz==2024.1.0.20240203
480
+ types-setuptools==69.2.0.20240317
481
+ typing_extensions==4.10.0
482
+ tyro==0.7.3
483
+ tzlocal==5.2
484
+ uc-micro-py==1.0.3
485
+ uritemplate==4.1.1
486
+ urllib3==2.0.7
487
+ vega-datasets==0.9.0
488
+ wadllib==1.3.6
489
+ wandb==0.16.4
490
+ wasabi==1.1.2
491
+ wcwidth==0.2.13
492
+ weasel==0.3.4
493
+ webcolors==1.13
494
+ webencodings==0.5.1
495
+ websocket-client==1.7.0
496
+ wheel==0.43.0
497
+ widgetsnbextension==3.6.6
498
+ wordcloud==1.9.3
499
+ wrapt==1.14.1
500
+ xarray-einstats==0.7.0
501
+ xarray==2023.7.0
502
+ xgboost==2.0.3
503
+ xlrd==2.0.1
504
+ xxhash==3.4.1
505
+ xyzservices==2023.10.1
506
+ yarl==1.9.4
507
+ yellowbrick==1.5
508
+ yfinance==0.2.37
509
+ zict==3.0.0
510
+ zipp==3.18.1
wandb/run-20240322_043407-hize5f4p/files/wandb-metadata.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.1.58+-x86_64-with-glibc2.35",
3
+ "python": "3.10.12",
4
+ "heartbeatAt": "2024-03-22T04:34:13.203984",
5
+ "startedAt": "2024-03-22T04:34:07.465777",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "Mistral-7B-Finetuning-Insurance.ipynb",
11
+ "codePathLocal": null,
12
+ "colab": "https://colab.research.google.com/notebook#fileId=1yggNDPxtSX3bri4ON4ppYc-G-Q2zFtkQ",
13
+ "host": "6b98cdd15d10",
14
+ "username": "root",
15
+ "executable": "/usr/bin/python3",
16
+ "cpu_count": 1,
17
+ "cpu_count_logical": 2,
18
+ "cpu_freq": {
19
+ "current": 2000.142,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.142,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.142,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ }
34
+ ],
35
+ "disk": {
36
+ "/": {
37
+ "total": 78.1898422241211,
38
+ "used": 42.876033782958984
39
+ }
40
+ },
41
+ "gpu": "Tesla T4",
42
+ "gpu_count": 1,
43
+ "gpu_devices": [
44
+ {
45
+ "name": "Tesla T4",
46
+ "memory_total": 16106127360
47
+ }
48
+ ],
49
+ "memory": {
50
+ "total": 12.674789428710938
51
+ }
52
+ }
wandb/run-20240322_043407-hize5f4p/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 1.9382, "train/grad_norm": 2.5268259048461914, "train/learning_rate": 0.00019075754196709572, "train/epoch": 0.07, "train/global_step": 10, "_timestamp": 1711082198.3750582, "_runtime": 150.80959606170654, "_step": 0}
wandb/run-20240322_043407-hize5f4p/logs/debug-internal.log ADDED
@@ -0,0 +1,143 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-03-22 04:34:07,566 INFO StreamThr :4275 [internal.py:wandb_internal():86] W&B internal server running at pid: 4275, started at: 2024-03-22 04:34:07.562793
2
+ 2024-03-22 04:34:07,589 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status
3
+ 2024-03-22 04:34:12,594 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
4
+ 2024-03-22 04:34:12,594 INFO WriterThread:4275 [datastore.py:open_for_write():87] open: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/run-hize5f4p.wandb
5
+ 2024-03-22 04:34:12,609 DEBUG SenderThread:4275 [sender.py:send():378] send: header
6
+ 2024-03-22 04:34:12,616 DEBUG SenderThread:4275 [sender.py:send():378] send: run
7
+ 2024-03-22 04:34:13,047 INFO SenderThread:4275 [dir_watcher.py:__init__():211] watching files in: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files
8
+ 2024-03-22 04:34:13,047 INFO SenderThread:4275 [sender.py:_start_run_threads():1099] run started: hize5f4p with start time 1711082047.565462
9
+ 2024-03-22 04:34:13,064 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: check_version
10
+ 2024-03-22 04:34:13,064 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: check_version
11
+ 2024-03-22 04:34:13,182 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: run_start
12
+ 2024-03-22 04:34:13,192 DEBUG HandlerThread:4275 [system_info.py:__init__():26] System info init
13
+ 2024-03-22 04:34:13,193 DEBUG HandlerThread:4275 [system_info.py:__init__():41] System info init done
14
+ 2024-03-22 04:34:13,193 INFO HandlerThread:4275 [system_monitor.py:start():194] Starting system monitor
15
+ 2024-03-22 04:34:13,193 INFO SystemMonitor:4275 [system_monitor.py:_start():158] Starting system asset monitoring threads
16
+ 2024-03-22 04:34:13,193 INFO HandlerThread:4275 [system_monitor.py:probe():214] Collecting system info
17
+ 2024-03-22 04:34:13,194 INFO SystemMonitor:4275 [interfaces.py:start():190] Started cpu monitoring
18
+ 2024-03-22 04:34:13,195 INFO SystemMonitor:4275 [interfaces.py:start():190] Started disk monitoring
19
+ 2024-03-22 04:34:13,196 INFO SystemMonitor:4275 [interfaces.py:start():190] Started gpu monitoring
20
+ 2024-03-22 04:34:13,197 INFO SystemMonitor:4275 [interfaces.py:start():190] Started memory monitoring
21
+ 2024-03-22 04:34:13,198 INFO SystemMonitor:4275 [interfaces.py:start():190] Started network monitoring
22
+ 2024-03-22 04:34:13,203 DEBUG HandlerThread:4275 [system_info.py:probe():150] Probing system
23
+ 2024-03-22 04:34:13,208 DEBUG HandlerThread:4275 [gitlib.py:_init_repo():56] git repository is invalid
24
+ 2024-03-22 04:34:13,208 DEBUG HandlerThread:4275 [system_info.py:probe():198] Probing system done
25
+ 2024-03-22 04:34:13,208 DEBUG HandlerThread:4275 [system_monitor.py:probe():223] {'os': 'Linux-6.1.58+-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-03-22T04:34:13.203984', 'startedAt': '2024-03-22T04:34:07.465777', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'Mistral-7B-Finetuning-Insurance.ipynb', 'codePathLocal': None, 'colab': 'https://colab.research.google.com/notebook#fileId=1yggNDPxtSX3bri4ON4ppYc-G-Q2zFtkQ', 'host': '6b98cdd15d10', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 1, 'cpu_count_logical': 2, 'cpu_freq': {'current': 2000.142, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 78.1898422241211, 'used': 42.876033782958984}}, 'gpu': 'Tesla T4', 'gpu_count': 1, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 12.674789428710938}}
26
+ 2024-03-22 04:34:13,208 INFO HandlerThread:4275 [system_monitor.py:probe():224] Finished collecting system info
27
+ 2024-03-22 04:34:13,208 INFO HandlerThread:4275 [system_monitor.py:probe():227] Publishing system info
28
+ 2024-03-22 04:34:13,226 INFO HandlerThread:4275 [system_monitor.py:probe():229] Finished publishing system info
29
+ 2024-03-22 04:34:13,232 DEBUG SenderThread:4275 [sender.py:send():378] send: files
30
+ 2024-03-22 04:34:13,233 INFO SenderThread:4275 [sender.py:_save_file():1365] saving file wandb-metadata.json with policy now
31
+ 2024-03-22 04:34:13,733 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: python_packages
32
+ 2024-03-22 04:34:13,733 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: python_packages
33
+ 2024-03-22 04:34:13,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
34
+ 2024-03-22 04:34:13,735 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
35
+ 2024-03-22 04:34:13,741 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
36
+ 2024-03-22 04:34:13,888 INFO wandb-upload_0:4275 [upload_job.py:push():131] Uploaded file /tmp/tmp3i0vj0_swandb/55hqq2eg-wandb-metadata.json
37
+ 2024-03-22 04:34:13,972 DEBUG SenderThread:4275 [sender.py:send():378] send: telemetry
38
+ 2024-03-22 04:34:13,973 DEBUG SenderThread:4275 [sender.py:send():378] send: config
39
+ 2024-03-22 04:34:13,975 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
40
+ 2024-03-22 04:34:13,975 DEBUG SenderThread:4275 [sender.py:send():378] send: telemetry
41
+ 2024-03-22 04:34:13,976 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
42
+ 2024-03-22 04:34:13,976 WARNING SenderThread:4275 [sender.py:send_metric():1316] Seen metric with glob (shouldn't happen)
43
+ 2024-03-22 04:34:13,976 DEBUG SenderThread:4275 [sender.py:send():378] send: telemetry
44
+ 2024-03-22 04:34:14,052 INFO Thread-12 :4275 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files/requirements.txt
45
+ 2024-03-22 04:34:14,053 INFO Thread-12 :4275 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files/wandb-metadata.json
46
+ 2024-03-22 04:34:17,977 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-03-22 04:34:22,978 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
48
+ 2024-03-22 04:34:27,978 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
49
+ 2024-03-22 04:34:28,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
50
+ 2024-03-22 04:34:28,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
51
+ 2024-03-22 04:34:28,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
52
+ 2024-03-22 04:34:33,912 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
53
+ 2024-03-22 04:34:38,918 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
54
+ 2024-03-22 04:34:40,099 INFO Thread-12 :4275 [dir_watcher.py:_on_file_modified():288] file/dir modified: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files/config.yaml
55
+ 2024-03-22 04:34:43,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
56
+ 2024-03-22 04:34:43,736 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
57
+ 2024-03-22 04:34:43,776 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
58
+ 2024-03-22 04:34:44,947 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
59
+ 2024-03-22 04:34:49,948 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
60
+ 2024-03-22 04:34:54,949 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
61
+ 2024-03-22 04:34:58,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
62
+ 2024-03-22 04:34:58,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
63
+ 2024-03-22 04:34:58,774 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
64
+ 2024-03-22 04:35:00,929 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
65
+ 2024-03-22 04:35:05,930 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
66
+ 2024-03-22 04:35:10,931 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
67
+ 2024-03-22 04:35:13,199 DEBUG SystemMonitor:4275 [system_monitor.py:_start():172] Starting system metrics aggregation loop
68
+ 2024-03-22 04:35:13,200 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
69
+ 2024-03-22 04:35:13,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
70
+ 2024-03-22 04:35:13,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
71
+ 2024-03-22 04:35:13,774 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
72
+ 2024-03-22 04:35:16,922 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
73
+ 2024-03-22 04:35:21,922 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
74
+ 2024-03-22 04:35:26,923 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
75
+ 2024-03-22 04:35:28,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
76
+ 2024-03-22 04:35:28,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
77
+ 2024-03-22 04:35:28,775 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
78
+ 2024-03-22 04:35:31,951 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
79
+ 2024-03-22 04:35:36,952 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
80
+ 2024-03-22 04:35:41,952 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-03-22 04:35:43,201 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
82
+ 2024-03-22 04:35:43,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
83
+ 2024-03-22 04:35:43,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
84
+ 2024-03-22 04:35:43,775 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
85
+ 2024-03-22 04:35:47,914 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
86
+ 2024-03-22 04:35:52,914 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
87
+ 2024-03-22 04:35:57,915 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
88
+ 2024-03-22 04:35:58,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
89
+ 2024-03-22 04:35:58,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
90
+ 2024-03-22 04:35:58,775 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
91
+ 2024-03-22 04:36:02,958 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
92
+ 2024-03-22 04:36:07,958 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
93
+ 2024-03-22 04:36:12,959 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
94
+ 2024-03-22 04:36:13,202 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
95
+ 2024-03-22 04:36:13,734 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
96
+ 2024-03-22 04:36:13,736 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
97
+ 2024-03-22 04:36:13,776 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
98
+ 2024-03-22 04:36:18,925 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
99
+ 2024-03-22 04:36:23,926 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
100
+ 2024-03-22 04:36:28,735 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
101
+ 2024-03-22 04:36:28,735 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
102
+ 2024-03-22 04:36:28,777 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
103
+ 2024-03-22 04:36:28,959 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
104
+ 2024-03-22 04:36:33,960 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
105
+ 2024-03-22 04:36:38,377 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: partial_history
106
+ 2024-03-22 04:36:38,384 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
107
+ 2024-03-22 04:36:38,385 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
108
+ 2024-03-22 04:36:38,385 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
109
+ 2024-03-22 04:36:38,385 DEBUG SenderThread:4275 [sender.py:send():378] send: metric
110
+ 2024-03-22 04:36:38,386 DEBUG SenderThread:4275 [sender.py:send():378] send: history
111
+ 2024-03-22 04:36:38,386 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: summary_record
112
+ 2024-03-22 04:36:38,395 INFO SenderThread:4275 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
113
+ 2024-03-22 04:36:39,314 INFO Thread-12 :4275 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files/wandb-summary.json
114
+ 2024-03-22 04:36:39,396 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
115
+ 2024-03-22 04:36:43,202 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
116
+ 2024-03-22 04:36:49,784 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
117
+ 2024-03-22 04:36:50,358 INFO Thread-12 :4275 [dir_watcher.py:_on_file_modified():288] file/dir modified: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/files/config.yaml
118
+ 2024-03-22 04:36:50,399 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
119
+ 2024-03-22 04:36:50,401 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
120
+ 2024-03-22 04:36:50,402 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
121
+ 2024-03-22 04:36:55,622 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-03-22 04:37:14,735 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
123
+ 2024-03-22 04:37:14,752 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
124
+ 2024-03-22 04:37:14,753 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
125
+ 2024-03-22 04:37:14,753 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
126
+ 2024-03-22 04:37:14,753 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
127
+ 2024-03-22 04:37:14,754 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
128
+ 2024-03-22 04:37:14,783 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
129
+ 2024-03-22 04:37:16,284 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
130
+ 2024-03-22 04:37:21,289 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
131
+ 2024-03-22 04:37:31,399 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
132
+ 2024-03-22 04:37:31,407 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
133
+ 2024-03-22 04:37:31,408 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: internal_messages
134
+ 2024-03-22 04:37:31,408 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
135
+ 2024-03-22 04:37:31,409 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
136
+ 2024-03-22 04:37:36,568 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
137
+ 2024-03-22 04:37:41,569 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
138
+ 2024-03-22 04:37:43,204 DEBUG SenderThread:4275 [sender.py:send():378] send: stats
139
+ 2024-03-22 04:37:44,743 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: stop_status
140
+ 2024-03-22 04:37:44,744 DEBUG SenderThread:4275 [sender.py:send_request():405] send_request: stop_status
141
+ 2024-03-22 04:37:46,973 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
142
+ 2024-03-22 04:37:51,974 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
143
+ 2024-03-22 04:37:56,975 DEBUG HandlerThread:4275 [handler.py:handle_request():146] handle_request: status_report
wandb/run-20240322_043407-hize5f4p/logs/debug.log ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-03-22 04:34:07,548 INFO MainThread:209 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
+ 2024-03-22 04:34:07,550 INFO MainThread:209 [wandb_setup.py:_flush():76] Configure stats pid to 209
3
+ 2024-03-22 04:34:07,551 INFO MainThread:209 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-03-22 04:34:07,551 INFO MainThread:209 [wandb_setup.py:_flush():76] Loading settings from /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/settings
5
+ 2024-03-22 04:34:07,551 INFO MainThread:209 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-03-22 04:34:07,552 INFO MainThread:209 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-03-22 04:34:07,553 INFO MainThread:209 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-03-22 04:34:07,553 INFO MainThread:209 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-03-22 04:34:07,553 INFO MainThread:209 [wandb_init.py:_log_setup():526] Logging user logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/logs/debug.log
10
+ 2024-03-22 04:34:07,554 INFO MainThread:209 [wandb_init.py:_log_setup():527] Logging internal logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240322_043407-hize5f4p/logs/debug-internal.log
11
+ 2024-03-22 04:34:07,554 INFO MainThread:209 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7f0538e72f80>
12
+ 2024-03-22 04:34:07,555 INFO MainThread:209 [wandb_init.py:init():566] calling init triggers
13
+ 2024-03-22 04:34:07,555 INFO MainThread:209 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-03-22 04:34:07,555 INFO MainThread:209 [wandb_init.py:init():616] starting backend
16
+ 2024-03-22 04:34:07,555 INFO MainThread:209 [wandb_init.py:init():620] setting up manager
17
+ 2024-03-22 04:34:07,561 INFO MainThread:209 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-03-22 04:34:07,564 INFO MainThread:209 [wandb_init.py:init():628] backend started and connected
19
+ 2024-03-22 04:34:07,592 INFO MainThread:209 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
+ 2024-03-22 04:34:12,608 INFO MainThread:209 [wandb_run.py:_label_probe_notebook():1305] Unable to probe notebook: 'NoneType' object has no attribute 'get'
21
+ 2024-03-22 04:34:12,608 INFO MainThread:209 [wandb_init.py:init():720] updated telemetry
22
+ 2024-03-22 04:34:12,615 INFO MainThread:209 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
23
+ 2024-03-22 04:34:13,063 INFO MainThread:209 [wandb_run.py:_on_init():2262] communicating current version
24
+ 2024-03-22 04:34:13,171 INFO MainThread:209 [wandb_run.py:_on_init():2271] got version response
25
+ 2024-03-22 04:34:13,171 INFO MainThread:209 [wandb_init.py:init():804] starting run threads in backend
26
+ 2024-03-22 04:34:13,734 INFO MainThread:209 [wandb_run.py:_console_start():2241] atexit reg
27
+ 2024-03-22 04:34:13,734 INFO MainThread:209 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
+ 2024-03-22 04:34:13,734 INFO MainThread:209 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
+ 2024-03-22 04:34:13,734 INFO MainThread:209 [wandb_run.py:_redirect():2186] Redirects installed.
30
+ 2024-03-22 04:34:13,737 INFO MainThread:209 [wandb_init.py:init():847] run started, returning control to user process
31
+ 2024-03-22 04:34:13,745 INFO MainThread:209 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'mistralai/Mistral-7B-Instruct-v0.2', 'transformers_version': '4.39.0', 'model_type': 'mistral', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'bfloat16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 3.0, 'max_steps': 60, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/runs/Mar22_04-33-33_6b98cdd15d10', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 10, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240322_043407-hize5f4p/run-hize5f4p.wandb ADDED
File without changes