csujeong commited on
Commit
1eb0f11
1 Parent(s): 687f89f

Training in progress, step 10

Browse files
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "revision": null,
21
  "target_modules": [
22
  "gate_proj",
23
- "o_proj",
24
  "q_proj",
 
25
  "v_proj",
26
- "k_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_rslora": false
 
20
  "revision": null,
21
  "target_modules": [
22
  "gate_proj",
 
23
  "q_proj",
24
+ "k_proj",
25
  "v_proj",
26
+ "o_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b751455bac18c260c793188b2414a3cabb61d6d296185708d65553c0970461d
3
  size 184592616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2104965ca16cb4213524612b30d6fa1833cf429a9897695d51b61432dfcc478c
3
  size 184592616
runs/Jan07_11-26-45_371ca9b7a7c7/events.out.tfevents.1704626819.371ca9b7a7c7.476.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9cd26860faf87ca16849fe529b47490601bb308c0982c3b295c6dc1dfb532ed
3
+ size 4883
runs/Jan07_11-26-45_371ca9b7a7c7/events.out.tfevents.1704627272.371ca9b7a7c7.476.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38e4dd55bc2f52ef008bf8942c173163e86faf0b749df4ab917b41a8019c0f1d
3
+ size 5037
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d970283b2db9e26741e20f533939320bbbc53f311abd5aa39921484c70f4190
3
  size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:297df666ae15800cff8c767710d9fd3f9096050bb43177b8c76e73080e339e3d
3
  size 4792
wandb/debug-internal.log ADDED
@@ -0,0 +1,134 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-01-07 11:34:44,819 INFO StreamThr :4664 [internal.py:wandb_internal():86] W&B internal server running at pid: 4664, started at: 2024-01-07 11:34:44.816615
2
+ 2024-01-07 11:34:44,840 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status
3
+ 2024-01-07 11:34:47,222 INFO WriterThread:4664 [datastore.py:open_for_write():85] open: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/run-ebd79hle.wandb
4
+ 2024-01-07 11:34:47,232 DEBUG SenderThread:4664 [sender.py:send():382] send: header
5
+ 2024-01-07 11:34:47,237 DEBUG SenderThread:4664 [sender.py:send():382] send: run
6
+ 2024-01-07 11:34:47,603 INFO SenderThread:4664 [dir_watcher.py:__init__():211] watching files in: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files
7
+ 2024-01-07 11:34:47,603 INFO SenderThread:4664 [sender.py:_start_run_threads():1136] run started: ebd79hle with start time 1704627284.817819
8
+ 2024-01-07 11:34:47,619 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-01-07 11:34:47,620 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: check_version
10
+ 2024-01-07 11:34:47,769 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-01-07 11:34:47,778 DEBUG HandlerThread:4664 [system_info.py:__init__():32] System info init
12
+ 2024-01-07 11:34:47,779 DEBUG HandlerThread:4664 [system_info.py:__init__():47] System info init done
13
+ 2024-01-07 11:34:47,779 INFO HandlerThread:4664 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-01-07 11:34:47,779 INFO SystemMonitor:4664 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-01-07 11:34:47,779 INFO HandlerThread:4664 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-01-07 11:34:47,780 INFO SystemMonitor:4664 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-01-07 11:34:47,781 INFO SystemMonitor:4664 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-01-07 11:34:47,782 INFO SystemMonitor:4664 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-01-07 11:34:47,783 INFO SystemMonitor:4664 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-01-07 11:34:47,784 INFO SystemMonitor:4664 [interfaces.py:start():190] Started network monitoring
21
+ 2024-01-07 11:34:47,788 DEBUG HandlerThread:4664 [system_info.py:probe():196] Probing system
22
+ 2024-01-07 11:34:47,793 DEBUG HandlerThread:4664 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-01-07 11:34:47,793 DEBUG HandlerThread:4664 [system_info.py:probe():244] Probing system done
24
+ 2024-01-07 11:34:47,793 DEBUG HandlerThread:4664 [system_monitor.py:probe():223] {'os': 'Linux-6.1.58+-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-01-07T11:34:47.788548', 'startedAt': '2024-01-07T11:34:44.770262', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'Mistral-7B-Finetuning-Insurance.ipynb', 'codePathLocal': None, 'colab': 'https://colab.research.google.com/notebook#fileId=1yggNDPxtSX3bri4ON4ppYc-G-Q2zFtkQ', 'host': '371ca9b7a7c7', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 1, 'cpu_count_logical': 2, 'cpu_freq': {'current': 2199.998, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2199.998, 'min': 0.0, 'max': 0.0}, {'current': 2199.998, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 78.1898422241211, 'used': 40.238895416259766}}, 'gpu': 'Tesla T4', 'gpu_count': 1, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 12.674789428710938}}
25
+ 2024-01-07 11:34:47,794 INFO HandlerThread:4664 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-01-07 11:34:47,794 INFO HandlerThread:4664 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-01-07 11:34:47,794 DEBUG HandlerThread:4664 [system_info.py:_save_pip():52] Saving list of pip packages installed into the current environment
28
+ 2024-01-07 11:34:47,798 DEBUG HandlerThread:4664 [system_info.py:_save_pip():68] Saving pip packages done
29
+ 2024-01-07 11:34:47,813 INFO HandlerThread:4664 [system_monitor.py:probe():229] Finished publishing system info
30
+ 2024-01-07 11:34:47,822 DEBUG SenderThread:4664 [sender.py:send():382] send: files
31
+ 2024-01-07 11:34:47,823 INFO SenderThread:4664 [sender.py:_save_file():1392] saving file wandb-metadata.json with policy now
32
+ 2024-01-07 11:34:47,846 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
33
+ 2024-01-07 11:34:47,847 DEBUG SenderThread:4664 [sender.py:send():382] send: telemetry
34
+ 2024-01-07 11:34:47,888 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
35
+ 2024-01-07 11:34:47,889 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
36
+ 2024-01-07 11:34:48,400 INFO wandb-upload_0:4664 [upload_job.py:push():131] Uploaded file /tmp/tmpn9igr4cqwandb/gszisqfu-wandb-metadata.json
37
+ 2024-01-07 11:34:48,418 DEBUG SenderThread:4664 [sender.py:send():382] send: config
38
+ 2024-01-07 11:34:48,421 DEBUG SenderThread:4664 [sender.py:send():382] send: metric
39
+ 2024-01-07 11:34:48,422 DEBUG SenderThread:4664 [sender.py:send():382] send: telemetry
40
+ 2024-01-07 11:34:48,422 DEBUG SenderThread:4664 [sender.py:send():382] send: metric
41
+ 2024-01-07 11:34:48,423 WARNING SenderThread:4664 [sender.py:send_metric():1343] Seen metric with glob (shouldn't happen)
42
+ 2024-01-07 11:34:48,423 DEBUG SenderThread:4664 [sender.py:send():382] send: telemetry
43
+ 2024-01-07 11:34:48,608 INFO Thread-12 :4664 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/output.log
44
+ 2024-01-07 11:34:48,609 INFO Thread-12 :4664 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/requirements.txt
45
+ 2024-01-07 11:34:48,609 INFO Thread-12 :4664 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/wandb-metadata.json
46
+ 2024-01-07 11:34:50,454 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-01-07 11:34:55,452 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
48
+ 2024-01-07 11:35:00,453 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
49
+ 2024-01-07 11:35:02,844 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
50
+ 2024-01-07 11:35:02,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
51
+ 2024-01-07 11:35:02,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
52
+ 2024-01-07 11:35:05,942 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
53
+ 2024-01-07 11:35:10,943 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
54
+ 2024-01-07 11:35:15,950 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
55
+ 2024-01-07 11:35:16,664 INFO Thread-12 :4664 [dir_watcher.py:_on_file_modified():288] file/dir modified: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/config.yaml
56
+ 2024-01-07 11:35:17,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
57
+ 2024-01-07 11:35:17,845 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
58
+ 2024-01-07 11:35:17,885 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
59
+ 2024-01-07 11:35:21,975 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
60
+ 2024-01-07 11:35:26,976 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
61
+ 2024-01-07 11:35:31,979 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
62
+ 2024-01-07 11:35:32,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
63
+ 2024-01-07 11:35:32,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
64
+ 2024-01-07 11:35:32,885 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
65
+ 2024-01-07 11:35:37,925 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
66
+ 2024-01-07 11:35:42,925 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
67
+ 2024-01-07 11:35:47,784 DEBUG SystemMonitor:4664 [system_monitor.py:_start():172] Starting system metrics aggregation loop
68
+ 2024-01-07 11:35:47,786 DEBUG SenderThread:4664 [sender.py:send():382] send: stats
69
+ 2024-01-07 11:35:47,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
70
+ 2024-01-07 11:35:47,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
71
+ 2024-01-07 11:35:47,885 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
72
+ 2024-01-07 11:35:47,986 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
73
+ 2024-01-07 11:35:52,987 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
74
+ 2024-01-07 11:35:57,988 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
75
+ 2024-01-07 11:36:02,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
76
+ 2024-01-07 11:36:02,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
77
+ 2024-01-07 11:36:02,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
78
+ 2024-01-07 11:36:03,966 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
79
+ 2024-01-07 11:36:08,966 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
80
+ 2024-01-07 11:36:13,967 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-01-07 11:36:17,786 DEBUG SenderThread:4664 [sender.py:send():382] send: stats
82
+ 2024-01-07 11:36:17,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
83
+ 2024-01-07 11:36:17,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
84
+ 2024-01-07 11:36:17,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
85
+ 2024-01-07 11:36:18,977 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
86
+ 2024-01-07 11:36:23,978 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
87
+ 2024-01-07 11:36:28,978 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
88
+ 2024-01-07 11:36:32,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
89
+ 2024-01-07 11:36:32,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
90
+ 2024-01-07 11:36:32,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
91
+ 2024-01-07 11:36:34,003 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
92
+ 2024-01-07 11:36:39,004 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
93
+ 2024-01-07 11:36:44,004 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
94
+ 2024-01-07 11:36:47,787 DEBUG SenderThread:4664 [sender.py:send():382] send: stats
95
+ 2024-01-07 11:36:47,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
96
+ 2024-01-07 11:36:47,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
97
+ 2024-01-07 11:36:47,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
98
+ 2024-01-07 11:36:49,009 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
99
+ 2024-01-07 11:36:54,010 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
100
+ 2024-01-07 11:36:59,011 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
101
+ 2024-01-07 11:37:02,846 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
102
+ 2024-01-07 11:37:02,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
103
+ 2024-01-07 11:37:02,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
104
+ 2024-01-07 11:37:05,005 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
105
+ 2024-01-07 11:37:10,005 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
106
+ 2024-01-07 11:37:15,006 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
107
+ 2024-01-07 11:37:17,788 DEBUG SenderThread:4664 [sender.py:send():382] send: stats
108
+ 2024-01-07 11:37:17,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
109
+ 2024-01-07 11:37:17,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
110
+ 2024-01-07 11:37:17,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
111
+ 2024-01-07 11:37:20,943 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
112
+ 2024-01-07 11:37:25,944 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
113
+ 2024-01-07 11:37:30,944 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
114
+ 2024-01-07 11:37:32,846 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
115
+ 2024-01-07 11:37:32,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
116
+ 2024-01-07 11:37:32,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
117
+ 2024-01-07 11:37:32,993 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: partial_history
118
+ 2024-01-07 11:37:32,996 DEBUG SenderThread:4664 [sender.py:send():382] send: metric
119
+ 2024-01-07 11:37:32,999 DEBUG SenderThread:4664 [sender.py:send():382] send: metric
120
+ 2024-01-07 11:37:32,999 DEBUG SenderThread:4664 [sender.py:send():382] send: metric
121
+ 2024-01-07 11:37:33,000 DEBUG SenderThread:4664 [sender.py:send():382] send: history
122
+ 2024-01-07 11:37:33,000 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: summary_record
123
+ 2024-01-07 11:37:33,018 INFO SenderThread:4664 [sender.py:_save_file():1392] saving file wandb-summary.json with policy end
124
+ 2024-01-07 11:37:33,944 INFO Thread-12 :4664 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/wandb-summary.json
125
+ 2024-01-07 11:37:36,020 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
126
+ 2024-01-07 11:37:41,026 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
127
+ 2024-01-07 11:37:45,019 INFO Thread-12 :4664 [dir_watcher.py:_on_file_modified():288] file/dir modified: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/output.log
128
+ 2024-01-07 11:37:46,027 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
129
+ 2024-01-07 11:37:47,789 DEBUG SenderThread:4664 [sender.py:send():382] send: stats
130
+ 2024-01-07 11:37:47,846 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
131
+ 2024-01-07 11:37:47,847 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
132
+ 2024-01-07 11:37:47,847 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
133
+ 2024-01-07 11:37:51,936 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
134
+ 2024-01-07 11:37:53,072 INFO Thread-12 :4664 [dir_watcher.py:_on_file_modified():288] file/dir modified: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/config.yaml
wandb/debug.log CHANGED
@@ -1,131 +1,49 @@
1
- 2024-01-07 08:32:15,522 INFO MainThread:684 [wandb_setup.py:_flush():76] Current SDK version is 0.16.1
2
- 2024-01-07 08:32:15,523 INFO MainThread:684 [wandb_setup.py:_flush():76] Configure stats pid to 684
3
- 2024-01-07 08:32:15,523 INFO MainThread:684 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-01-07 08:32:15,524 INFO MainThread:684 [wandb_setup.py:_flush():76] Loading settings from /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/settings
5
- 2024-01-07 08:32:15,524 INFO MainThread:684 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-01-07 08:32:15,524 INFO MainThread:684 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-01-07 08:32:15,524 INFO MainThread:684 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-01-07 08:32:15,524 INFO MainThread:684 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
- 2024-01-07 08:32:15,525 INFO MainThread:684 [wandb_init.py:_log_setup():524] Logging user logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_083215-enryt6zo/logs/debug.log
10
- 2024-01-07 08:32:15,525 INFO MainThread:684 [wandb_init.py:_log_setup():525] Logging internal logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_083215-enryt6zo/logs/debug-internal.log
11
- 2024-01-07 08:32:15,525 INFO MainThread:684 [wandb_init.py:_jupyter_setup():470] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7bfea43750f0>
12
- 2024-01-07 08:32:15,526 INFO MainThread:684 [wandb_init.py:init():564] calling init triggers
13
- 2024-01-07 08:32:15,526 INFO MainThread:684 [wandb_init.py:init():571] wandb.init called with sweep_config: {}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  config: {}
15
- 2024-01-07 08:32:15,526 INFO MainThread:684 [wandb_init.py:init():614] starting backend
16
- 2024-01-07 08:32:15,526 INFO MainThread:684 [wandb_init.py:init():618] setting up manager
17
- 2024-01-07 08:32:15,531 INFO MainThread:684 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
- 2024-01-07 08:32:15,534 INFO MainThread:684 [wandb_init.py:init():624] backend started and connected
19
- 2024-01-07 08:32:15,570 INFO MainThread:684 [wandb_run.py:_label_probe_notebook():1294] probe notebook
20
- 2024-01-07 08:32:17,418 INFO MainThread:684 [wandb_init.py:init():716] updated telemetry
21
- 2024-01-07 08:32:17,453 INFO MainThread:684 [wandb_init.py:init():749] communicating run to backend with 90.0 second timeout
22
- 2024-01-07 08:32:17,964 INFO MainThread:684 [wandb_run.py:_on_init():2254] communicating current version
23
- 2024-01-07 08:32:18,119 INFO MainThread:684 [wandb_run.py:_on_init():2263] got version response
24
- 2024-01-07 08:32:18,120 INFO MainThread:684 [wandb_init.py:init():800] starting run threads in backend
25
- 2024-01-07 08:32:18,210 INFO MainThread:684 [wandb_run.py:_console_start():2233] atexit reg
26
- 2024-01-07 08:32:18,211 INFO MainThread:684 [wandb_run.py:_redirect():2088] redirect: wrap_raw
27
- 2024-01-07 08:32:18,211 INFO MainThread:684 [wandb_run.py:_redirect():2153] Wrapping output streams.
28
- 2024-01-07 08:32:18,211 INFO MainThread:684 [wandb_run.py:_redirect():2178] Redirects installed.
29
- 2024-01-07 08:32:18,213 INFO MainThread:684 [wandb_init.py:init():841] run started, returning control to user process
30
- 2024-01-07 08:32:18,219 INFO MainThread:684 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': 4096, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 10000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'mistralai/Mistral-7B-v0.1', 'transformers_version': '4.36.2', 'model_type': 'mistral', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', 'load_in_8bit': False, 'load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'bfloat16'}, 'output_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 3.0, 'max_steps': 60, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/runs/Jan07_08-30-52_096ae31a5012', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 10, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
31
- 2024-01-07 08:44:03,889 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
32
- 2024-01-07 08:44:03,890 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
33
- 2024-01-07 08:44:26,570 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
34
- 2024-01-07 08:44:34,326 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
35
- 2024-01-07 08:44:34,327 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
36
- 2024-01-07 08:44:46,475 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
37
- 2024-01-07 08:46:05,058 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
38
- 2024-01-07 08:46:05,058 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
39
- 2024-01-07 08:46:13,038 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
40
- 2024-01-07 08:46:18,516 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
41
- 2024-01-07 08:46:18,516 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
42
- 2024-01-07 08:50:09,111 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
43
- 2024-01-07 08:50:13,508 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
44
- 2024-01-07 08:50:13,513 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
45
- 2024-01-07 08:51:38,094 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
46
- 2024-01-07 08:51:38,098 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
47
- 2024-01-07 08:51:38,098 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
48
- 2024-01-07 08:51:41,383 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
49
- 2024-01-07 08:52:12,662 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
50
- 2024-01-07 08:52:12,662 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
51
- 2024-01-07 08:52:45,454 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
52
- 2024-01-07 08:53:09,095 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
53
- 2024-01-07 08:53:09,096 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
54
- 2024-01-07 08:55:59,367 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
55
- 2024-01-07 08:56:00,251 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
56
- 2024-01-07 08:56:00,252 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
57
- 2024-01-07 08:56:06,562 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
58
- 2024-01-07 08:56:10,699 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
59
- 2024-01-07 08:56:10,700 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
60
- 2024-01-07 08:57:55,151 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
61
- 2024-01-07 08:57:58,981 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
62
- 2024-01-07 08:57:58,981 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
63
- 2024-01-07 09:00:08,883 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
64
- 2024-01-07 09:00:18,822 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
65
- 2024-01-07 09:00:18,822 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
66
- 2024-01-07 09:08:01,727 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
67
- 2024-01-07 09:08:06,231 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
68
- 2024-01-07 09:08:06,232 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
69
- 2024-01-07 09:08:25,862 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
70
- 2024-01-07 09:08:25,898 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
71
- 2024-01-07 09:08:25,898 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
72
- 2024-01-07 09:08:35,845 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
73
- 2024-01-07 09:08:37,716 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
74
- 2024-01-07 09:08:37,717 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
75
- 2024-01-07 09:08:40,487 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
76
- 2024-01-07 09:08:40,495 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
77
- 2024-01-07 09:08:40,501 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
78
- 2024-01-07 09:08:45,788 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
79
- 2024-01-07 09:08:45,793 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
80
- 2024-01-07 09:08:45,794 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
81
- 2024-01-07 09:08:49,111 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
82
- 2024-01-07 09:08:49,155 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
83
- 2024-01-07 09:08:49,155 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
84
- 2024-01-07 09:09:44,376 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
85
- 2024-01-07 09:09:44,379 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
86
- 2024-01-07 09:09:44,380 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
87
- 2024-01-07 09:10:16,380 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
88
- 2024-01-07 09:10:16,383 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
89
- 2024-01-07 09:10:16,383 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
90
- 2024-01-07 09:10:25,980 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
91
- 2024-01-07 09:10:26,068 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
92
- 2024-01-07 09:10:26,076 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
93
- 2024-01-07 09:10:52,944 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
94
- 2024-01-07 09:10:52,950 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
95
- 2024-01-07 09:10:52,950 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
96
- 2024-01-07 09:10:54,782 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
97
- 2024-01-07 09:10:54,813 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
98
- 2024-01-07 09:10:54,813 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
99
- 2024-01-07 09:12:03,682 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
100
- 2024-01-07 09:12:03,692 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
101
- 2024-01-07 09:12:03,692 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
102
- 2024-01-07 09:12:06,232 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
103
- 2024-01-07 09:12:06,325 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
104
- 2024-01-07 09:12:06,326 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
105
- 2024-01-07 09:12:33,934 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
106
- 2024-01-07 09:12:34,001 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
107
- 2024-01-07 09:12:34,004 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
108
- 2024-01-07 09:13:00,605 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
109
- 2024-01-07 09:13:00,639 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
110
- 2024-01-07 09:13:00,639 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
111
- 2024-01-07 09:13:06,384 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
112
- 2024-01-07 09:13:06,462 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
113
- 2024-01-07 09:13:06,462 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
114
- 2024-01-07 09:13:24,618 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
115
- 2024-01-07 09:13:24,664 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
116
- 2024-01-07 09:13:24,665 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
117
- 2024-01-07 09:13:52,451 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
118
- 2024-01-07 09:13:52,481 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
119
- 2024-01-07 09:13:52,481 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
120
- 2024-01-07 09:14:01,980 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
121
- 2024-01-07 09:14:02,039 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
122
- 2024-01-07 09:14:02,040 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
123
- 2024-01-07 09:14:37,393 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
124
- 2024-01-07 09:14:42,402 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
125
- 2024-01-07 09:14:42,403 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
126
- 2024-01-07 09:16:27,093 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
127
- 2024-01-07 09:16:27,127 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
128
- 2024-01-07 09:16:27,127 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
129
- 2024-01-07 09:19:48,148 INFO MainThread:684 [wandb_init.py:_resume_backend():440] resuming backend
130
- 2024-01-07 09:19:48,193 INFO MainThread:684 [jupyter.py:save_ipynb():373] not saving jupyter notebook
131
- 2024-01-07 09:19:48,193 INFO MainThread:684 [wandb_init.py:_pause_backend():435] pausing backend
 
1
+ 2024-01-07 11:34:44,805 INFO MainThread:476 [wandb_setup.py:_flush():76] Current SDK version is 0.16.1
2
+ 2024-01-07 11:34:44,806 INFO MainThread:476 [wandb_setup.py:_flush():76] Configure stats pid to 476
3
+ 2024-01-07 11:34:44,806 INFO MainThread:476 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Loading settings from /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/settings
5
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-01-07 11:34:44,807 ERROR MainThread:476 [wandb_setup.py:_flush():76] error
9
+ Traceback (most recent call last):
10
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_init.py", line 1166, in init
11
+ wi.setup(kwargs)
12
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_init.py", line 306, in setup
13
+ wandb_login._login(
14
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_login.py", line 298, in _login
15
+ wlogin.prompt_api_key()
16
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_login.py", line 221, in prompt_api_key
17
+ key, status = self._prompt_api_key()
18
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_login.py", line 201, in _prompt_api_key
19
+ key = apikey.prompt_api_key(
20
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/lib/apikey.py", line 144, in prompt_api_key
21
+ key = input_callback(api_ask).strip()
22
+ File "/usr/local/lib/python3.10/dist-packages/click/termui.py", line 164, in prompt
23
+ value = prompt_func(prompt)
24
+ File "/usr/local/lib/python3.10/dist-packages/click/termui.py", line 147, in prompt_func
25
+ raise Abort() from None
26
+ click.exceptions.Abort
27
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
28
+ 2024-01-07 11:34:44,808 INFO MainThread:476 [wandb_init.py:_log_setup():524] Logging user logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/logs/debug.log
29
+ 2024-01-07 11:34:44,808 INFO MainThread:476 [wandb_init.py:_log_setup():525] Logging internal logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/logs/debug-internal.log
30
+ 2024-01-07 11:34:44,809 INFO MainThread:476 [wandb_init.py:_jupyter_setup():470] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7c7dab76c8b0>
31
+ 2024-01-07 11:34:44,809 INFO MainThread:476 [wandb_init.py:init():564] calling init triggers
32
+ 2024-01-07 11:34:44,809 INFO MainThread:476 [wandb_init.py:init():571] wandb.init called with sweep_config: {}
33
  config: {}
34
+ 2024-01-07 11:34:44,810 INFO MainThread:476 [wandb_init.py:init():614] starting backend
35
+ 2024-01-07 11:34:44,810 INFO MainThread:476 [wandb_init.py:init():618] setting up manager
36
+ 2024-01-07 11:34:44,815 INFO MainThread:476 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
37
+ 2024-01-07 11:34:44,817 INFO MainThread:476 [wandb_init.py:init():624] backend started and connected
38
+ 2024-01-07 11:34:44,842 INFO MainThread:476 [wandb_run.py:_label_probe_notebook():1294] probe notebook
39
+ 2024-01-07 11:34:47,220 INFO MainThread:476 [wandb_init.py:init():716] updated telemetry
40
+ 2024-01-07 11:34:47,236 INFO MainThread:476 [wandb_init.py:init():749] communicating run to backend with 90.0 second timeout
41
+ 2024-01-07 11:34:47,619 INFO MainThread:476 [wandb_run.py:_on_init():2254] communicating current version
42
+ 2024-01-07 11:34:47,757 INFO MainThread:476 [wandb_run.py:_on_init():2263] got version response
43
+ 2024-01-07 11:34:47,758 INFO MainThread:476 [wandb_init.py:init():800] starting run threads in backend
44
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_console_start():2233] atexit reg
45
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_redirect():2088] redirect: wrap_raw
46
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_redirect():2153] Wrapping output streams.
47
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_redirect():2178] Redirects installed.
48
+ 2024-01-07 11:34:47,848 INFO MainThread:476 [wandb_init.py:init():841] run started, returning control to user process
49
+ 2024-01-07 11:34:47,851 INFO MainThread:476 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': 4096, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 10000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'mistralai/Mistral-7B-v0.1', 'transformers_version': '4.36.2', 'model_type': 'mistral', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', 'load_in_8bit': False, 'load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'bfloat16'}, 'output_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 3.0, 'max_steps': 60, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/runs/Jan07_11-26-45_371ca9b7a7c7', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 10, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
wandb/run-20240107_113444-ebd79hle/files/config.yaml ADDED
@@ -0,0 +1,657 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.12
7
+ cli_version: 0.16.1
8
+ framework: huggingface
9
+ huggingface_version: 4.36.2
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: false
12
+ start_time: 1704627284.817819
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ 2:
29
+ - 1
30
+ - 2
31
+ - 3
32
+ - 5
33
+ - 11
34
+ - 12
35
+ - 49
36
+ - 51
37
+ - 53
38
+ - 55
39
+ - 71
40
+ - 84
41
+ - 98
42
+ 3:
43
+ - 7
44
+ - 23
45
+ 4: 3.10.12
46
+ 5: 0.16.1
47
+ 6: 4.36.2
48
+ 8:
49
+ - 1
50
+ - 5
51
+ - 12
52
+ 9:
53
+ 1: transformers_trainer
54
+ 13: linux-x86_64
55
+ m:
56
+ - 1: train/global_step
57
+ 6:
58
+ - 3
59
+ - 1: train/loss
60
+ 5: 1
61
+ 6:
62
+ - 1
63
+ - 1: train/learning_rate
64
+ 5: 1
65
+ 6:
66
+ - 1
67
+ - 1: train/epoch
68
+ 5: 1
69
+ 6:
70
+ - 1
71
+ vocab_size:
72
+ desc: null
73
+ value: 32000
74
+ max_position_embeddings:
75
+ desc: null
76
+ value: 32768
77
+ hidden_size:
78
+ desc: null
79
+ value: 4096
80
+ intermediate_size:
81
+ desc: null
82
+ value: 14336
83
+ num_hidden_layers:
84
+ desc: null
85
+ value: 32
86
+ num_attention_heads:
87
+ desc: null
88
+ value: 32
89
+ sliding_window:
90
+ desc: null
91
+ value: 4096
92
+ num_key_value_heads:
93
+ desc: null
94
+ value: 8
95
+ hidden_act:
96
+ desc: null
97
+ value: silu
98
+ initializer_range:
99
+ desc: null
100
+ value: 0.02
101
+ rms_norm_eps:
102
+ desc: null
103
+ value: 1.0e-05
104
+ use_cache:
105
+ desc: null
106
+ value: false
107
+ rope_theta:
108
+ desc: null
109
+ value: 10000.0
110
+ attention_dropout:
111
+ desc: null
112
+ value: 0.0
113
+ return_dict:
114
+ desc: null
115
+ value: true
116
+ output_hidden_states:
117
+ desc: null
118
+ value: false
119
+ output_attentions:
120
+ desc: null
121
+ value: false
122
+ torchscript:
123
+ desc: null
124
+ value: false
125
+ torch_dtype:
126
+ desc: null
127
+ value: bfloat16
128
+ use_bfloat16:
129
+ desc: null
130
+ value: false
131
+ tf_legacy_loss:
132
+ desc: null
133
+ value: false
134
+ pruned_heads:
135
+ desc: null
136
+ value: {}
137
+ tie_word_embeddings:
138
+ desc: null
139
+ value: false
140
+ is_encoder_decoder:
141
+ desc: null
142
+ value: false
143
+ is_decoder:
144
+ desc: null
145
+ value: false
146
+ cross_attention_hidden_size:
147
+ desc: null
148
+ value: null
149
+ add_cross_attention:
150
+ desc: null
151
+ value: false
152
+ tie_encoder_decoder:
153
+ desc: null
154
+ value: false
155
+ max_length:
156
+ desc: null
157
+ value: 20
158
+ min_length:
159
+ desc: null
160
+ value: 0
161
+ do_sample:
162
+ desc: null
163
+ value: false
164
+ early_stopping:
165
+ desc: null
166
+ value: false
167
+ num_beams:
168
+ desc: null
169
+ value: 1
170
+ num_beam_groups:
171
+ desc: null
172
+ value: 1
173
+ diversity_penalty:
174
+ desc: null
175
+ value: 0.0
176
+ temperature:
177
+ desc: null
178
+ value: 1.0
179
+ top_k:
180
+ desc: null
181
+ value: 50
182
+ top_p:
183
+ desc: null
184
+ value: 1.0
185
+ typical_p:
186
+ desc: null
187
+ value: 1.0
188
+ repetition_penalty:
189
+ desc: null
190
+ value: 1.0
191
+ length_penalty:
192
+ desc: null
193
+ value: 1.0
194
+ no_repeat_ngram_size:
195
+ desc: null
196
+ value: 0
197
+ encoder_no_repeat_ngram_size:
198
+ desc: null
199
+ value: 0
200
+ bad_words_ids:
201
+ desc: null
202
+ value: null
203
+ num_return_sequences:
204
+ desc: null
205
+ value: 1
206
+ chunk_size_feed_forward:
207
+ desc: null
208
+ value: 0
209
+ output_scores:
210
+ desc: null
211
+ value: false
212
+ return_dict_in_generate:
213
+ desc: null
214
+ value: false
215
+ forced_bos_token_id:
216
+ desc: null
217
+ value: null
218
+ forced_eos_token_id:
219
+ desc: null
220
+ value: null
221
+ remove_invalid_values:
222
+ desc: null
223
+ value: false
224
+ exponential_decay_length_penalty:
225
+ desc: null
226
+ value: null
227
+ suppress_tokens:
228
+ desc: null
229
+ value: null
230
+ begin_suppress_tokens:
231
+ desc: null
232
+ value: null
233
+ architectures:
234
+ desc: null
235
+ value:
236
+ - MistralForCausalLM
237
+ finetuning_task:
238
+ desc: null
239
+ value: null
240
+ id2label:
241
+ desc: null
242
+ value:
243
+ '0': LABEL_0
244
+ '1': LABEL_1
245
+ label2id:
246
+ desc: null
247
+ value:
248
+ LABEL_0: 0
249
+ LABEL_1: 1
250
+ tokenizer_class:
251
+ desc: null
252
+ value: null
253
+ prefix:
254
+ desc: null
255
+ value: null
256
+ bos_token_id:
257
+ desc: null
258
+ value: 1
259
+ pad_token_id:
260
+ desc: null
261
+ value: null
262
+ eos_token_id:
263
+ desc: null
264
+ value: 2
265
+ sep_token_id:
266
+ desc: null
267
+ value: null
268
+ decoder_start_token_id:
269
+ desc: null
270
+ value: null
271
+ task_specific_params:
272
+ desc: null
273
+ value: null
274
+ problem_type:
275
+ desc: null
276
+ value: null
277
+ _name_or_path:
278
+ desc: null
279
+ value: mistralai/Mistral-7B-v0.1
280
+ transformers_version:
281
+ desc: null
282
+ value: 4.36.2
283
+ model_type:
284
+ desc: null
285
+ value: mistral
286
+ quantization_config:
287
+ desc: null
288
+ value:
289
+ quant_method: QuantizationMethod.BITS_AND_BYTES
290
+ load_in_8bit: false
291
+ load_in_4bit: true
292
+ llm_int8_threshold: 6.0
293
+ llm_int8_skip_modules: null
294
+ llm_int8_enable_fp32_cpu_offload: false
295
+ llm_int8_has_fp16_weight: false
296
+ bnb_4bit_quant_type: nf4
297
+ bnb_4bit_use_double_quant: true
298
+ bnb_4bit_compute_dtype: bfloat16
299
+ output_dir:
300
+ desc: null
301
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance
302
+ overwrite_output_dir:
303
+ desc: null
304
+ value: false
305
+ do_train:
306
+ desc: null
307
+ value: false
308
+ do_eval:
309
+ desc: null
310
+ value: false
311
+ do_predict:
312
+ desc: null
313
+ value: false
314
+ evaluation_strategy:
315
+ desc: null
316
+ value: 'no'
317
+ prediction_loss_only:
318
+ desc: null
319
+ value: false
320
+ per_device_train_batch_size:
321
+ desc: null
322
+ value: 2
323
+ per_device_eval_batch_size:
324
+ desc: null
325
+ value: 8
326
+ per_gpu_train_batch_size:
327
+ desc: null
328
+ value: null
329
+ per_gpu_eval_batch_size:
330
+ desc: null
331
+ value: null
332
+ gradient_accumulation_steps:
333
+ desc: null
334
+ value: 2
335
+ eval_accumulation_steps:
336
+ desc: null
337
+ value: null
338
+ eval_delay:
339
+ desc: null
340
+ value: 0
341
+ learning_rate:
342
+ desc: null
343
+ value: 0.0002
344
+ weight_decay:
345
+ desc: null
346
+ value: 0.0
347
+ adam_beta1:
348
+ desc: null
349
+ value: 0.9
350
+ adam_beta2:
351
+ desc: null
352
+ value: 0.999
353
+ adam_epsilon:
354
+ desc: null
355
+ value: 1.0e-08
356
+ max_grad_norm:
357
+ desc: null
358
+ value: 0.3
359
+ num_train_epochs:
360
+ desc: null
361
+ value: 3.0
362
+ max_steps:
363
+ desc: null
364
+ value: 60
365
+ lr_scheduler_type:
366
+ desc: null
367
+ value: cosine
368
+ lr_scheduler_kwargs:
369
+ desc: null
370
+ value: {}
371
+ warmup_ratio:
372
+ desc: null
373
+ value: 0.03
374
+ warmup_steps:
375
+ desc: null
376
+ value: 0
377
+ log_level:
378
+ desc: null
379
+ value: passive
380
+ log_level_replica:
381
+ desc: null
382
+ value: warning
383
+ log_on_each_node:
384
+ desc: null
385
+ value: true
386
+ logging_dir:
387
+ desc: null
388
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/runs/Jan07_11-26-45_371ca9b7a7c7
389
+ logging_strategy:
390
+ desc: null
391
+ value: steps
392
+ logging_first_step:
393
+ desc: null
394
+ value: false
395
+ logging_steps:
396
+ desc: null
397
+ value: 10
398
+ logging_nan_inf_filter:
399
+ desc: null
400
+ value: true
401
+ save_strategy:
402
+ desc: null
403
+ value: steps
404
+ save_steps:
405
+ desc: null
406
+ value: 10
407
+ save_total_limit:
408
+ desc: null
409
+ value: null
410
+ save_safetensors:
411
+ desc: null
412
+ value: true
413
+ save_on_each_node:
414
+ desc: null
415
+ value: false
416
+ save_only_model:
417
+ desc: null
418
+ value: false
419
+ no_cuda:
420
+ desc: null
421
+ value: false
422
+ use_cpu:
423
+ desc: null
424
+ value: false
425
+ use_mps_device:
426
+ desc: null
427
+ value: false
428
+ seed:
429
+ desc: null
430
+ value: 42
431
+ data_seed:
432
+ desc: null
433
+ value: null
434
+ jit_mode_eval:
435
+ desc: null
436
+ value: false
437
+ use_ipex:
438
+ desc: null
439
+ value: false
440
+ bf16:
441
+ desc: null
442
+ value: false
443
+ fp16:
444
+ desc: null
445
+ value: false
446
+ fp16_opt_level:
447
+ desc: null
448
+ value: O1
449
+ half_precision_backend:
450
+ desc: null
451
+ value: auto
452
+ bf16_full_eval:
453
+ desc: null
454
+ value: false
455
+ fp16_full_eval:
456
+ desc: null
457
+ value: false
458
+ tf32:
459
+ desc: null
460
+ value: false
461
+ local_rank:
462
+ desc: null
463
+ value: 0
464
+ ddp_backend:
465
+ desc: null
466
+ value: null
467
+ tpu_num_cores:
468
+ desc: null
469
+ value: null
470
+ tpu_metrics_debug:
471
+ desc: null
472
+ value: false
473
+ debug:
474
+ desc: null
475
+ value: []
476
+ dataloader_drop_last:
477
+ desc: null
478
+ value: false
479
+ eval_steps:
480
+ desc: null
481
+ value: null
482
+ dataloader_num_workers:
483
+ desc: null
484
+ value: 0
485
+ past_index:
486
+ desc: null
487
+ value: -1
488
+ run_name:
489
+ desc: null
490
+ value: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance
491
+ disable_tqdm:
492
+ desc: null
493
+ value: false
494
+ remove_unused_columns:
495
+ desc: null
496
+ value: true
497
+ label_names:
498
+ desc: null
499
+ value: null
500
+ load_best_model_at_end:
501
+ desc: null
502
+ value: false
503
+ metric_for_best_model:
504
+ desc: null
505
+ value: null
506
+ greater_is_better:
507
+ desc: null
508
+ value: null
509
+ ignore_data_skip:
510
+ desc: null
511
+ value: false
512
+ fsdp:
513
+ desc: null
514
+ value: []
515
+ fsdp_min_num_params:
516
+ desc: null
517
+ value: 0
518
+ fsdp_config:
519
+ desc: null
520
+ value:
521
+ min_num_params: 0
522
+ xla: false
523
+ xla_fsdp_grad_ckpt: false
524
+ fsdp_transformer_layer_cls_to_wrap:
525
+ desc: null
526
+ value: null
527
+ deepspeed:
528
+ desc: null
529
+ value: null
530
+ label_smoothing_factor:
531
+ desc: null
532
+ value: 0.0
533
+ optim:
534
+ desc: null
535
+ value: paged_adamw_32bit
536
+ optim_args:
537
+ desc: null
538
+ value: null
539
+ adafactor:
540
+ desc: null
541
+ value: false
542
+ group_by_length:
543
+ desc: null
544
+ value: true
545
+ length_column_name:
546
+ desc: null
547
+ value: length
548
+ report_to:
549
+ desc: null
550
+ value:
551
+ - tensorboard
552
+ - wandb
553
+ ddp_find_unused_parameters:
554
+ desc: null
555
+ value: null
556
+ ddp_bucket_cap_mb:
557
+ desc: null
558
+ value: null
559
+ ddp_broadcast_buffers:
560
+ desc: null
561
+ value: null
562
+ dataloader_pin_memory:
563
+ desc: null
564
+ value: true
565
+ dataloader_persistent_workers:
566
+ desc: null
567
+ value: false
568
+ skip_memory_metrics:
569
+ desc: null
570
+ value: true
571
+ use_legacy_prediction_loop:
572
+ desc: null
573
+ value: false
574
+ push_to_hub:
575
+ desc: null
576
+ value: true
577
+ resume_from_checkpoint:
578
+ desc: null
579
+ value: null
580
+ hub_model_id:
581
+ desc: null
582
+ value: null
583
+ hub_strategy:
584
+ desc: null
585
+ value: every_save
586
+ hub_token:
587
+ desc: null
588
+ value: <HUB_TOKEN>
589
+ hub_private_repo:
590
+ desc: null
591
+ value: false
592
+ hub_always_push:
593
+ desc: null
594
+ value: false
595
+ gradient_checkpointing:
596
+ desc: null
597
+ value: false
598
+ gradient_checkpointing_kwargs:
599
+ desc: null
600
+ value: null
601
+ include_inputs_for_metrics:
602
+ desc: null
603
+ value: false
604
+ fp16_backend:
605
+ desc: null
606
+ value: auto
607
+ push_to_hub_model_id:
608
+ desc: null
609
+ value: null
610
+ push_to_hub_organization:
611
+ desc: null
612
+ value: null
613
+ push_to_hub_token:
614
+ desc: null
615
+ value: <PUSH_TO_HUB_TOKEN>
616
+ mp_parameters:
617
+ desc: null
618
+ value: ''
619
+ auto_find_batch_size:
620
+ desc: null
621
+ value: false
622
+ full_determinism:
623
+ desc: null
624
+ value: false
625
+ torchdynamo:
626
+ desc: null
627
+ value: null
628
+ ray_scope:
629
+ desc: null
630
+ value: last
631
+ ddp_timeout:
632
+ desc: null
633
+ value: 1800
634
+ torch_compile:
635
+ desc: null
636
+ value: false
637
+ torch_compile_backend:
638
+ desc: null
639
+ value: null
640
+ torch_compile_mode:
641
+ desc: null
642
+ value: null
643
+ dispatch_batches:
644
+ desc: null
645
+ value: null
646
+ split_batches:
647
+ desc: null
648
+ value: false
649
+ include_tokens_per_second:
650
+ desc: null
651
+ value: false
652
+ include_num_input_tokens_seen:
653
+ desc: null
654
+ value: false
655
+ neftune_noise_alpha:
656
+ desc: null
657
+ value: null
wandb/run-20240107_113444-ebd79hle/files/output.log ADDED
@@ -0,0 +1 @@
 
 
1
+
wandb/run-20240107_113444-ebd79hle/files/requirements.txt ADDED
@@ -0,0 +1,497 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.4.0
2
+ accelerate==0.25.0
3
+ aiohttp==3.9.1
4
+ aiosignal==1.3.1
5
+ alabaster==0.7.13
6
+ albumentations==1.3.1
7
+ altair==4.2.2
8
+ anyio==3.7.1
9
+ appdirs==1.4.4
10
+ argon2-cffi-bindings==21.2.0
11
+ argon2-cffi==23.1.0
12
+ array-record==0.5.0
13
+ arviz==0.15.1
14
+ astropy==5.3.4
15
+ astunparse==1.6.3
16
+ async-timeout==4.0.3
17
+ atpublic==4.0
18
+ attrs==23.1.0
19
+ audioread==3.0.1
20
+ autograd==1.6.2
21
+ babel==2.14.0
22
+ backcall==0.2.0
23
+ beautifulsoup4==4.11.2
24
+ bidict==0.22.1
25
+ bigframes==0.17.0
26
+ bitsandbytes==0.41.3.post2
27
+ bleach==6.1.0
28
+ blinker==1.4
29
+ blis==0.7.11
30
+ blosc2==2.0.0
31
+ bokeh==3.3.2
32
+ bqplot==0.12.42
33
+ branca==0.7.0
34
+ build==1.0.3
35
+ cachecontrol==0.13.1
36
+ cachetools==5.3.2
37
+ catalogue==2.0.10
38
+ certifi==2023.11.17
39
+ cffi==1.16.0
40
+ chardet==5.2.0
41
+ charset-normalizer==3.3.2
42
+ chex==0.1.7
43
+ click-plugins==1.1.1
44
+ click==8.1.7
45
+ cligj==0.7.2
46
+ cloudpickle==2.2.1
47
+ cmake==3.27.9
48
+ cmdstanpy==1.2.0
49
+ colorcet==3.0.1
50
+ colorlover==0.3.0
51
+ colour==0.1.5
52
+ community==1.0.0b1
53
+ confection==0.1.4
54
+ cons==0.4.6
55
+ contextlib2==21.6.0
56
+ contourpy==1.2.0
57
+ cryptography==41.0.7
58
+ cufflinks==0.17.3
59
+ cupy-cuda12x==12.2.0
60
+ cvxopt==1.3.2
61
+ cvxpy==1.3.2
62
+ cycler==0.12.1
63
+ cymem==2.0.8
64
+ cython==3.0.7
65
+ dask==2023.8.1
66
+ datascience==0.17.6
67
+ datasets==2.16.1
68
+ db-dtypes==1.2.0
69
+ dbus-python==1.2.18
70
+ debugpy==1.6.6
71
+ decorator==4.4.2
72
+ defusedxml==0.7.1
73
+ dill==0.3.7
74
+ diskcache==5.6.3
75
+ distributed==2023.8.1
76
+ distro==1.7.0
77
+ dlib==19.24.2
78
+ dm-tree==0.1.8
79
+ docker-pycreds==0.4.0
80
+ docstring-parser==0.15
81
+ docutils==0.18.1
82
+ dopamine-rl==4.0.6
83
+ duckdb==0.9.2
84
+ earthengine-api==0.1.384
85
+ easydict==1.11
86
+ ecos==2.0.12
87
+ editdistance==0.6.2
88
+ eerepr==0.0.4
89
+ einops==0.7.0
90
+ en-core-web-sm==3.6.0
91
+ entrypoints==0.4
92
+ et-xmlfile==1.1.0
93
+ etils==1.6.0
94
+ etuples==0.3.9
95
+ exceptiongroup==1.2.0
96
+ fastai==2.7.13
97
+ fastcore==1.5.29
98
+ fastdownload==0.0.7
99
+ fastjsonschema==2.19.0
100
+ fastprogress==1.0.3
101
+ fastrlock==0.8.2
102
+ filelock==3.13.1
103
+ fiona==1.9.5
104
+ firebase-admin==5.3.0
105
+ flask==2.2.5
106
+ flatbuffers==23.5.26
107
+ flax==0.7.5
108
+ folium==0.14.0
109
+ fonttools==4.47.0
110
+ frozendict==2.3.10
111
+ frozenlist==1.4.1
112
+ fsspec==2023.6.0
113
+ future==0.18.3
114
+ gast==0.5.4
115
+ gcsfs==2023.6.0
116
+ gdal==3.4.3
117
+ gdown==4.6.6
118
+ geemap==0.29.6
119
+ gensim==4.3.2
120
+ geocoder==1.38.1
121
+ geographiclib==2.0
122
+ geopandas==0.13.2
123
+ geopy==2.3.0
124
+ gin-config==0.5.0
125
+ gitdb==4.0.11
126
+ gitpython==3.1.40
127
+ glob2==0.7
128
+ google-ai-generativelanguage==0.4.0
129
+ google-api-core==2.11.1
130
+ google-api-python-client==2.84.0
131
+ google-auth-httplib2==0.1.1
132
+ google-auth-oauthlib==1.2.0
133
+ google-auth==2.17.3
134
+ google-cloud-aiplatform==1.38.1
135
+ google-cloud-bigquery-connection==1.12.1
136
+ google-cloud-bigquery-storage==2.24.0
137
+ google-cloud-bigquery==3.12.0
138
+ google-cloud-core==2.3.3
139
+ google-cloud-datastore==2.15.2
140
+ google-cloud-firestore==2.11.1
141
+ google-cloud-functions==1.13.3
142
+ google-cloud-iam==2.13.0
143
+ google-cloud-language==2.9.1
144
+ google-cloud-resource-manager==1.11.0
145
+ google-cloud-storage==2.8.0
146
+ google-cloud-translate==3.11.3
147
+ google-colab==1.0.0
148
+ google-crc32c==1.5.0
149
+ google-generativeai==0.3.2
150
+ google-pasta==0.2.0
151
+ google-resumable-media==2.7.0
152
+ google==2.0.3
153
+ googleapis-common-protos==1.62.0
154
+ googledrivedownloader==0.4
155
+ graphviz==0.20.1
156
+ greenlet==3.0.2
157
+ grpc-google-iam-v1==0.13.0
158
+ grpcio-status==1.48.2
159
+ grpcio==1.60.0
160
+ gspread-dataframe==3.3.1
161
+ gspread==3.4.2
162
+ gym-notices==0.0.8
163
+ gym==0.25.2
164
+ h5netcdf==1.3.0
165
+ h5py==3.9.0
166
+ holidays==0.39
167
+ holoviews==1.17.1
168
+ html5lib==1.1
169
+ httpimport==1.3.1
170
+ httplib2==0.22.0
171
+ huggingface-hub==0.20.1
172
+ humanize==4.7.0
173
+ hyperopt==0.2.7
174
+ ibis-framework==6.2.0
175
+ idna==3.6
176
+ imageio-ffmpeg==0.4.9
177
+ imageio==2.31.6
178
+ imagesize==1.4.1
179
+ imbalanced-learn==0.10.1
180
+ imgaug==0.4.0
181
+ importlib-metadata==7.0.0
182
+ importlib-resources==6.1.1
183
+ imutils==0.5.4
184
+ inflect==7.0.0
185
+ iniconfig==2.0.0
186
+ install==1.3.5
187
+ intel-openmp==2023.2.3
188
+ ipyevents==2.0.2
189
+ ipyfilechooser==0.6.0
190
+ ipykernel==5.5.6
191
+ ipyleaflet==0.18.1
192
+ ipython-genutils==0.2.0
193
+ ipython-sql==0.5.0
194
+ ipython==7.34.0
195
+ ipytree==0.2.2
196
+ ipywidgets==7.7.1
197
+ itsdangerous==2.1.2
198
+ jax==0.4.23
199
+ jaxlib==0.4.23+cuda12.cudnn89
200
+ jeepney==0.7.1
201
+ jieba==0.42.1
202
+ jinja2==3.1.2
203
+ joblib==1.3.2
204
+ jsonpickle==3.0.2
205
+ jsonschema-specifications==2023.11.2
206
+ jsonschema==4.19.2
207
+ jupyter-client==6.1.12
208
+ jupyter-console==6.1.0
209
+ jupyter-core==5.5.1
210
+ jupyter-server==1.24.0
211
+ jupyterlab-pygments==0.3.0
212
+ jupyterlab-widgets==3.0.9
213
+ kaggle==1.5.16
214
+ kagglehub==0.1.4
215
+ keras==2.15.0
216
+ keyring==23.5.0
217
+ kiwisolver==1.4.5
218
+ langcodes==3.3.0
219
+ launchpadlib==1.10.16
220
+ lazr.restfulclient==0.14.4
221
+ lazr.uri==1.0.6
222
+ lazy-loader==0.3
223
+ libclang==16.0.6
224
+ librosa==0.10.1
225
+ lida==0.0.10
226
+ lightgbm==4.1.0
227
+ linkify-it-py==2.0.2
228
+ llmx==0.0.15a0
229
+ llvmlite==0.41.1
230
+ locket==1.0.0
231
+ logical-unification==0.4.6
232
+ lxml==4.9.4
233
+ malloy==2023.1067
234
+ markdown-it-py==3.0.0
235
+ markdown==3.5.1
236
+ markupsafe==2.1.3
237
+ matplotlib-inline==0.1.6
238
+ matplotlib-venn==0.11.9
239
+ matplotlib==3.7.1
240
+ mdit-py-plugins==0.4.0
241
+ mdurl==0.1.2
242
+ minikanren==1.0.3
243
+ missingno==0.5.2
244
+ mistune==0.8.4
245
+ mizani==0.9.3
246
+ mkl==2023.2.0
247
+ ml-dtypes==0.2.0
248
+ mlxtend==0.22.0
249
+ more-itertools==10.1.0
250
+ moviepy==1.0.3
251
+ mpmath==1.3.0
252
+ msgpack==1.0.7
253
+ multidict==6.0.4
254
+ multipledispatch==1.0.0
255
+ multiprocess==0.70.15
256
+ multitasking==0.0.11
257
+ murmurhash==1.0.10
258
+ music21==9.1.0
259
+ natsort==8.4.0
260
+ nbclassic==1.0.0
261
+ nbclient==0.9.0
262
+ nbconvert==6.5.4
263
+ nbformat==5.9.2
264
+ nest-asyncio==1.5.8
265
+ networkx==3.2.1
266
+ nibabel==4.0.2
267
+ nltk==3.8.1
268
+ notebook-shim==0.2.3
269
+ notebook==6.5.5
270
+ numba==0.58.1
271
+ numexpr==2.8.8
272
+ numpy==1.23.5
273
+ oauth2client==4.1.3
274
+ oauthlib==3.2.2
275
+ opencv-contrib-python==4.8.0.76
276
+ opencv-python-headless==4.8.1.78
277
+ opencv-python==4.8.0.76
278
+ openpyxl==3.1.2
279
+ opt-einsum==3.3.0
280
+ optax==0.1.7
281
+ orbax-checkpoint==0.4.4
282
+ osqp==0.6.2.post8
283
+ packaging==23.2
284
+ pandas-datareader==0.10.0
285
+ pandas-gbq==0.19.2
286
+ pandas-stubs==1.5.3.230304
287
+ pandas==1.5.3
288
+ pandocfilters==1.5.0
289
+ panel==1.3.6
290
+ param==2.0.1
291
+ parso==0.8.3
292
+ parsy==2.1
293
+ partd==1.4.1
294
+ pathlib==1.0.1
295
+ pathy==0.10.3
296
+ patsy==0.5.4
297
+ peewee==3.17.0
298
+ peft==0.7.2.dev0
299
+ pexpect==4.9.0
300
+ pickleshare==0.7.5
301
+ pillow==9.4.0
302
+ pip-tools==6.13.0
303
+ pip==23.1.2
304
+ platformdirs==4.1.0
305
+ plotly==5.15.0
306
+ plotnine==0.12.4
307
+ pluggy==1.3.0
308
+ polars==0.17.3
309
+ pooch==1.8.0
310
+ portpicker==1.5.2
311
+ prefetch-generator==1.0.3
312
+ preshed==3.0.9
313
+ prettytable==3.9.0
314
+ proglog==0.1.10
315
+ progressbar2==4.2.0
316
+ prometheus-client==0.19.0
317
+ promise==2.3
318
+ prompt-toolkit==3.0.43
319
+ prophet==1.1.5
320
+ proto-plus==1.23.0
321
+ protobuf==3.20.3
322
+ psutil==5.9.5
323
+ psycopg2==2.9.9
324
+ ptyprocess==0.7.0
325
+ py-cpuinfo==9.0.0
326
+ py4j==0.10.9.7
327
+ pyarrow-hotfix==0.6
328
+ pyarrow==10.0.1
329
+ pyasn1-modules==0.3.0
330
+ pyasn1==0.5.1
331
+ pycocotools==2.0.7
332
+ pycparser==2.21
333
+ pyct==0.5.0
334
+ pydantic==1.10.13
335
+ pydata-google-auth==1.8.2
336
+ pydot-ng==2.0.0
337
+ pydot==1.4.2
338
+ pydotplus==2.0.2
339
+ pydrive2==1.6.3
340
+ pydrive==1.3.1
341
+ pyerfa==2.0.1.1
342
+ pygame==2.5.2
343
+ pygments==2.16.1
344
+ pygobject==3.42.1
345
+ pyjwt==2.3.0
346
+ pymc==5.7.2
347
+ pymystem3==0.2.0
348
+ pyopengl==3.1.7
349
+ pyopenssl==23.3.0
350
+ pyparsing==3.1.1
351
+ pyperclip==1.8.2
352
+ pyproj==3.6.1
353
+ pyproject-hooks==1.0.0
354
+ pyshp==2.3.1
355
+ pysocks==1.7.1
356
+ pytensor==2.14.2
357
+ pytest==7.4.3
358
+ python-apt==0.0.0
359
+ python-box==7.1.1
360
+ python-dateutil==2.8.2
361
+ python-louvain==0.16
362
+ python-slugify==8.0.1
363
+ python-utils==3.8.1
364
+ pytz==2023.3.post1
365
+ pyviz-comms==3.0.0
366
+ pywavelets==1.5.0
367
+ pyyaml==6.0.1
368
+ pyzmq==23.2.1
369
+ qdldl==0.1.7.post0
370
+ qudida==0.0.4
371
+ ratelim==0.1.6
372
+ referencing==0.32.0
373
+ regex==2023.6.3
374
+ requests-oauthlib==1.3.1
375
+ requests==2.31.0
376
+ requirements-parser==0.5.0
377
+ rich==13.7.0
378
+ rpds-py==0.15.2
379
+ rpy2==3.4.2
380
+ rsa==4.9
381
+ safetensors==0.4.1
382
+ scikit-image==0.19.3
383
+ scikit-learn==1.2.2
384
+ scipy==1.11.4
385
+ scooby==0.9.2
386
+ scs==3.2.4.post1
387
+ seaborn==0.12.2
388
+ secretstorage==3.3.1
389
+ send2trash==1.8.2
390
+ sentry-sdk==1.39.1
391
+ setproctitle==1.3.3
392
+ setuptools==67.7.2
393
+ shapely==2.0.2
394
+ shtab==1.6.5
395
+ six==1.16.0
396
+ sklearn-pandas==2.2.0
397
+ smart-open==6.4.0
398
+ smmap==5.0.1
399
+ sniffio==1.3.0
400
+ snowballstemmer==2.2.0
401
+ sortedcontainers==2.4.0
402
+ soundfile==0.12.1
403
+ soupsieve==2.5
404
+ soxr==0.3.7
405
+ spacy-legacy==3.0.12
406
+ spacy-loggers==1.0.5
407
+ spacy==3.6.1
408
+ sphinx==5.0.2
409
+ sphinxcontrib-applehelp==1.0.7
410
+ sphinxcontrib-devhelp==1.0.5
411
+ sphinxcontrib-htmlhelp==2.0.4
412
+ sphinxcontrib-jsmath==1.0.1
413
+ sphinxcontrib-qthelp==1.0.6
414
+ sphinxcontrib-serializinghtml==1.1.9
415
+ sqlalchemy==2.0.23
416
+ sqlglot==17.16.2
417
+ sqlparse==0.4.4
418
+ srsly==2.4.8
419
+ stanio==0.3.0
420
+ statsmodels==0.14.1
421
+ sympy==1.12
422
+ tables==3.8.0
423
+ tabulate==0.9.0
424
+ tbb==2021.11.0
425
+ tblib==3.0.0
426
+ tenacity==8.2.3
427
+ tensorboard-data-server==0.7.2
428
+ tensorboard==2.15.1
429
+ tensorflow-datasets==4.9.4
430
+ tensorflow-estimator==2.15.0
431
+ tensorflow-gcs-config==2.15.0
432
+ tensorflow-hub==0.15.0
433
+ tensorflow-io-gcs-filesystem==0.35.0
434
+ tensorflow-metadata==1.14.0
435
+ tensorflow-probability==0.22.0
436
+ tensorflow==2.15.0
437
+ tensorstore==0.1.45
438
+ termcolor==2.4.0
439
+ terminado==0.18.0
440
+ text-unidecode==1.3
441
+ textblob==0.17.1
442
+ tf-slim==1.1.0
443
+ thinc==8.1.12
444
+ threadpoolctl==3.2.0
445
+ tifffile==2023.12.9
446
+ tinycss2==1.2.1
447
+ tokenizers==0.15.0
448
+ toml==0.10.2
449
+ tomli==2.0.1
450
+ toolz==0.12.0
451
+ torch==2.1.0+cu121
452
+ torchaudio==2.1.0+cu121
453
+ torchdata==0.7.0
454
+ torchsummary==1.5.1
455
+ torchtext==0.16.0
456
+ torchvision==0.16.0+cu121
457
+ tornado==6.3.2
458
+ tqdm==4.66.1
459
+ traitlets==5.7.1
460
+ traittypes==0.2.1
461
+ transformers==4.36.2
462
+ triton==2.1.0
463
+ trl==0.7.7
464
+ tweepy==4.14.0
465
+ typer==0.9.0
466
+ types-pytz==2023.3.1.1
467
+ types-setuptools==69.0.0.0
468
+ typing-extensions==4.5.0
469
+ tyro==0.6.3
470
+ tzlocal==5.2
471
+ uc-micro-py==1.0.2
472
+ uritemplate==4.1.1
473
+ urllib3==2.0.7
474
+ vega-datasets==0.9.0
475
+ wadllib==1.3.6
476
+ wandb==0.16.1
477
+ wasabi==1.1.2
478
+ wcwidth==0.2.12
479
+ webcolors==1.13
480
+ webencodings==0.5.1
481
+ websocket-client==1.7.0
482
+ werkzeug==3.0.1
483
+ wheel==0.42.0
484
+ widgetsnbextension==3.6.6
485
+ wordcloud==1.9.3
486
+ wrapt==1.14.1
487
+ xarray-einstats==0.6.0
488
+ xarray==2023.7.0
489
+ xgboost==2.0.3
490
+ xlrd==2.0.1
491
+ xxhash==3.4.1
492
+ xyzservices==2023.10.1
493
+ yarl==1.9.4
494
+ yellowbrick==1.5
495
+ yfinance==0.2.33
496
+ zict==3.0.0
497
+ zipp==3.17.0
wandb/run-20240107_113444-ebd79hle/files/wandb-metadata.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.1.58+-x86_64-with-glibc2.35",
3
+ "python": "3.10.12",
4
+ "heartbeatAt": "2024-01-07T11:34:47.788548",
5
+ "startedAt": "2024-01-07T11:34:44.770262",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "Mistral-7B-Finetuning-Insurance.ipynb",
11
+ "codePathLocal": null,
12
+ "colab": "https://colab.research.google.com/notebook#fileId=1yggNDPxtSX3bri4ON4ppYc-G-Q2zFtkQ",
13
+ "host": "371ca9b7a7c7",
14
+ "username": "root",
15
+ "executable": "/usr/bin/python3",
16
+ "cpu_count": 1,
17
+ "cpu_count_logical": 2,
18
+ "cpu_freq": {
19
+ "current": 2199.998,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2199.998,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2199.998,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ }
34
+ ],
35
+ "disk": {
36
+ "/": {
37
+ "total": 78.1898422241211,
38
+ "used": 40.238895416259766
39
+ }
40
+ },
41
+ "gpu": "Tesla T4",
42
+ "gpu_count": 1,
43
+ "gpu_devices": [
44
+ {
45
+ "name": "Tesla T4",
46
+ "memory_total": 16106127360
47
+ }
48
+ ],
49
+ "memory": {
50
+ "total": 12.674789428710938
51
+ }
52
+ }
wandb/run-20240107_113444-ebd79hle/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 1.7048, "train/learning_rate": 0.00019075754196709572, "train/epoch": 0.07, "train/global_step": 10, "_timestamp": 1704627452.9922113, "_runtime": 168.17439222335815, "_step": 0}
wandb/run-20240107_113444-ebd79hle/logs/debug-internal.log ADDED
@@ -0,0 +1,134 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-01-07 11:34:44,819 INFO StreamThr :4664 [internal.py:wandb_internal():86] W&B internal server running at pid: 4664, started at: 2024-01-07 11:34:44.816615
2
+ 2024-01-07 11:34:44,840 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status
3
+ 2024-01-07 11:34:47,222 INFO WriterThread:4664 [datastore.py:open_for_write():85] open: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/run-ebd79hle.wandb
4
+ 2024-01-07 11:34:47,232 DEBUG SenderThread:4664 [sender.py:send():382] send: header
5
+ 2024-01-07 11:34:47,237 DEBUG SenderThread:4664 [sender.py:send():382] send: run
6
+ 2024-01-07 11:34:47,603 INFO SenderThread:4664 [dir_watcher.py:__init__():211] watching files in: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files
7
+ 2024-01-07 11:34:47,603 INFO SenderThread:4664 [sender.py:_start_run_threads():1136] run started: ebd79hle with start time 1704627284.817819
8
+ 2024-01-07 11:34:47,619 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-01-07 11:34:47,620 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: check_version
10
+ 2024-01-07 11:34:47,769 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-01-07 11:34:47,778 DEBUG HandlerThread:4664 [system_info.py:__init__():32] System info init
12
+ 2024-01-07 11:34:47,779 DEBUG HandlerThread:4664 [system_info.py:__init__():47] System info init done
13
+ 2024-01-07 11:34:47,779 INFO HandlerThread:4664 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-01-07 11:34:47,779 INFO SystemMonitor:4664 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-01-07 11:34:47,779 INFO HandlerThread:4664 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-01-07 11:34:47,780 INFO SystemMonitor:4664 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-01-07 11:34:47,781 INFO SystemMonitor:4664 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-01-07 11:34:47,782 INFO SystemMonitor:4664 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-01-07 11:34:47,783 INFO SystemMonitor:4664 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-01-07 11:34:47,784 INFO SystemMonitor:4664 [interfaces.py:start():190] Started network monitoring
21
+ 2024-01-07 11:34:47,788 DEBUG HandlerThread:4664 [system_info.py:probe():196] Probing system
22
+ 2024-01-07 11:34:47,793 DEBUG HandlerThread:4664 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-01-07 11:34:47,793 DEBUG HandlerThread:4664 [system_info.py:probe():244] Probing system done
24
+ 2024-01-07 11:34:47,793 DEBUG HandlerThread:4664 [system_monitor.py:probe():223] {'os': 'Linux-6.1.58+-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-01-07T11:34:47.788548', 'startedAt': '2024-01-07T11:34:44.770262', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'Mistral-7B-Finetuning-Insurance.ipynb', 'codePathLocal': None, 'colab': 'https://colab.research.google.com/notebook#fileId=1yggNDPxtSX3bri4ON4ppYc-G-Q2zFtkQ', 'host': '371ca9b7a7c7', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 1, 'cpu_count_logical': 2, 'cpu_freq': {'current': 2199.998, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2199.998, 'min': 0.0, 'max': 0.0}, {'current': 2199.998, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 78.1898422241211, 'used': 40.238895416259766}}, 'gpu': 'Tesla T4', 'gpu_count': 1, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 12.674789428710938}}
25
+ 2024-01-07 11:34:47,794 INFO HandlerThread:4664 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-01-07 11:34:47,794 INFO HandlerThread:4664 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-01-07 11:34:47,794 DEBUG HandlerThread:4664 [system_info.py:_save_pip():52] Saving list of pip packages installed into the current environment
28
+ 2024-01-07 11:34:47,798 DEBUG HandlerThread:4664 [system_info.py:_save_pip():68] Saving pip packages done
29
+ 2024-01-07 11:34:47,813 INFO HandlerThread:4664 [system_monitor.py:probe():229] Finished publishing system info
30
+ 2024-01-07 11:34:47,822 DEBUG SenderThread:4664 [sender.py:send():382] send: files
31
+ 2024-01-07 11:34:47,823 INFO SenderThread:4664 [sender.py:_save_file():1392] saving file wandb-metadata.json with policy now
32
+ 2024-01-07 11:34:47,846 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
33
+ 2024-01-07 11:34:47,847 DEBUG SenderThread:4664 [sender.py:send():382] send: telemetry
34
+ 2024-01-07 11:34:47,888 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
35
+ 2024-01-07 11:34:47,889 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
36
+ 2024-01-07 11:34:48,400 INFO wandb-upload_0:4664 [upload_job.py:push():131] Uploaded file /tmp/tmpn9igr4cqwandb/gszisqfu-wandb-metadata.json
37
+ 2024-01-07 11:34:48,418 DEBUG SenderThread:4664 [sender.py:send():382] send: config
38
+ 2024-01-07 11:34:48,421 DEBUG SenderThread:4664 [sender.py:send():382] send: metric
39
+ 2024-01-07 11:34:48,422 DEBUG SenderThread:4664 [sender.py:send():382] send: telemetry
40
+ 2024-01-07 11:34:48,422 DEBUG SenderThread:4664 [sender.py:send():382] send: metric
41
+ 2024-01-07 11:34:48,423 WARNING SenderThread:4664 [sender.py:send_metric():1343] Seen metric with glob (shouldn't happen)
42
+ 2024-01-07 11:34:48,423 DEBUG SenderThread:4664 [sender.py:send():382] send: telemetry
43
+ 2024-01-07 11:34:48,608 INFO Thread-12 :4664 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/output.log
44
+ 2024-01-07 11:34:48,609 INFO Thread-12 :4664 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/requirements.txt
45
+ 2024-01-07 11:34:48,609 INFO Thread-12 :4664 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/wandb-metadata.json
46
+ 2024-01-07 11:34:50,454 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-01-07 11:34:55,452 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
48
+ 2024-01-07 11:35:00,453 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
49
+ 2024-01-07 11:35:02,844 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
50
+ 2024-01-07 11:35:02,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
51
+ 2024-01-07 11:35:02,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
52
+ 2024-01-07 11:35:05,942 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
53
+ 2024-01-07 11:35:10,943 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
54
+ 2024-01-07 11:35:15,950 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
55
+ 2024-01-07 11:35:16,664 INFO Thread-12 :4664 [dir_watcher.py:_on_file_modified():288] file/dir modified: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/config.yaml
56
+ 2024-01-07 11:35:17,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
57
+ 2024-01-07 11:35:17,845 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
58
+ 2024-01-07 11:35:17,885 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
59
+ 2024-01-07 11:35:21,975 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
60
+ 2024-01-07 11:35:26,976 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
61
+ 2024-01-07 11:35:31,979 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
62
+ 2024-01-07 11:35:32,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
63
+ 2024-01-07 11:35:32,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
64
+ 2024-01-07 11:35:32,885 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
65
+ 2024-01-07 11:35:37,925 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
66
+ 2024-01-07 11:35:42,925 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
67
+ 2024-01-07 11:35:47,784 DEBUG SystemMonitor:4664 [system_monitor.py:_start():172] Starting system metrics aggregation loop
68
+ 2024-01-07 11:35:47,786 DEBUG SenderThread:4664 [sender.py:send():382] send: stats
69
+ 2024-01-07 11:35:47,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
70
+ 2024-01-07 11:35:47,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
71
+ 2024-01-07 11:35:47,885 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
72
+ 2024-01-07 11:35:47,986 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
73
+ 2024-01-07 11:35:52,987 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
74
+ 2024-01-07 11:35:57,988 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
75
+ 2024-01-07 11:36:02,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
76
+ 2024-01-07 11:36:02,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
77
+ 2024-01-07 11:36:02,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
78
+ 2024-01-07 11:36:03,966 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
79
+ 2024-01-07 11:36:08,966 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
80
+ 2024-01-07 11:36:13,967 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-01-07 11:36:17,786 DEBUG SenderThread:4664 [sender.py:send():382] send: stats
82
+ 2024-01-07 11:36:17,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
83
+ 2024-01-07 11:36:17,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
84
+ 2024-01-07 11:36:17,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
85
+ 2024-01-07 11:36:18,977 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
86
+ 2024-01-07 11:36:23,978 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
87
+ 2024-01-07 11:36:28,978 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
88
+ 2024-01-07 11:36:32,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
89
+ 2024-01-07 11:36:32,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
90
+ 2024-01-07 11:36:32,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
91
+ 2024-01-07 11:36:34,003 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
92
+ 2024-01-07 11:36:39,004 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
93
+ 2024-01-07 11:36:44,004 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
94
+ 2024-01-07 11:36:47,787 DEBUG SenderThread:4664 [sender.py:send():382] send: stats
95
+ 2024-01-07 11:36:47,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
96
+ 2024-01-07 11:36:47,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
97
+ 2024-01-07 11:36:47,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
98
+ 2024-01-07 11:36:49,009 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
99
+ 2024-01-07 11:36:54,010 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
100
+ 2024-01-07 11:36:59,011 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
101
+ 2024-01-07 11:37:02,846 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
102
+ 2024-01-07 11:37:02,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
103
+ 2024-01-07 11:37:02,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
104
+ 2024-01-07 11:37:05,005 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
105
+ 2024-01-07 11:37:10,005 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
106
+ 2024-01-07 11:37:15,006 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
107
+ 2024-01-07 11:37:17,788 DEBUG SenderThread:4664 [sender.py:send():382] send: stats
108
+ 2024-01-07 11:37:17,845 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
109
+ 2024-01-07 11:37:17,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
110
+ 2024-01-07 11:37:17,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
111
+ 2024-01-07 11:37:20,943 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
112
+ 2024-01-07 11:37:25,944 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
113
+ 2024-01-07 11:37:30,944 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
114
+ 2024-01-07 11:37:32,846 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
115
+ 2024-01-07 11:37:32,846 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
116
+ 2024-01-07 11:37:32,886 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
117
+ 2024-01-07 11:37:32,993 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: partial_history
118
+ 2024-01-07 11:37:32,996 DEBUG SenderThread:4664 [sender.py:send():382] send: metric
119
+ 2024-01-07 11:37:32,999 DEBUG SenderThread:4664 [sender.py:send():382] send: metric
120
+ 2024-01-07 11:37:32,999 DEBUG SenderThread:4664 [sender.py:send():382] send: metric
121
+ 2024-01-07 11:37:33,000 DEBUG SenderThread:4664 [sender.py:send():382] send: history
122
+ 2024-01-07 11:37:33,000 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: summary_record
123
+ 2024-01-07 11:37:33,018 INFO SenderThread:4664 [sender.py:_save_file():1392] saving file wandb-summary.json with policy end
124
+ 2024-01-07 11:37:33,944 INFO Thread-12 :4664 [dir_watcher.py:_on_file_created():271] file/dir created: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/wandb-summary.json
125
+ 2024-01-07 11:37:36,020 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
126
+ 2024-01-07 11:37:41,026 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
127
+ 2024-01-07 11:37:45,019 INFO Thread-12 :4664 [dir_watcher.py:_on_file_modified():288] file/dir modified: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/output.log
128
+ 2024-01-07 11:37:46,027 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
129
+ 2024-01-07 11:37:47,789 DEBUG SenderThread:4664 [sender.py:send():382] send: stats
130
+ 2024-01-07 11:37:47,846 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: stop_status
131
+ 2024-01-07 11:37:47,847 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: internal_messages
132
+ 2024-01-07 11:37:47,847 DEBUG SenderThread:4664 [sender.py:send_request():409] send_request: stop_status
133
+ 2024-01-07 11:37:51,936 DEBUG HandlerThread:4664 [handler.py:handle_request():146] handle_request: status_report
134
+ 2024-01-07 11:37:53,072 INFO Thread-12 :4664 [dir_watcher.py:_on_file_modified():288] file/dir modified: /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/files/config.yaml
wandb/run-20240107_113444-ebd79hle/logs/debug.log ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-01-07 11:34:44,805 INFO MainThread:476 [wandb_setup.py:_flush():76] Current SDK version is 0.16.1
2
+ 2024-01-07 11:34:44,806 INFO MainThread:476 [wandb_setup.py:_flush():76] Configure stats pid to 476
3
+ 2024-01-07 11:34:44,806 INFO MainThread:476 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Loading settings from /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/settings
5
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-01-07 11:34:44,807 ERROR MainThread:476 [wandb_setup.py:_flush():76] error
9
+ Traceback (most recent call last):
10
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_init.py", line 1166, in init
11
+ wi.setup(kwargs)
12
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_init.py", line 306, in setup
13
+ wandb_login._login(
14
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_login.py", line 298, in _login
15
+ wlogin.prompt_api_key()
16
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_login.py", line 221, in prompt_api_key
17
+ key, status = self._prompt_api_key()
18
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/wandb_login.py", line 201, in _prompt_api_key
19
+ key = apikey.prompt_api_key(
20
+ File "/usr/local/lib/python3.10/dist-packages/wandb/sdk/lib/apikey.py", line 144, in prompt_api_key
21
+ key = input_callback(api_ask).strip()
22
+ File "/usr/local/lib/python3.10/dist-packages/click/termui.py", line 164, in prompt
23
+ value = prompt_func(prompt)
24
+ File "/usr/local/lib/python3.10/dist-packages/click/termui.py", line 147, in prompt_func
25
+ raise Abort() from None
26
+ click.exceptions.Abort
27
+ 2024-01-07 11:34:44,807 INFO MainThread:476 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
28
+ 2024-01-07 11:34:44,808 INFO MainThread:476 [wandb_init.py:_log_setup():524] Logging user logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/logs/debug.log
29
+ 2024-01-07 11:34:44,808 INFO MainThread:476 [wandb_init.py:_log_setup():525] Logging internal logs to /content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/wandb/run-20240107_113444-ebd79hle/logs/debug-internal.log
30
+ 2024-01-07 11:34:44,809 INFO MainThread:476 [wandb_init.py:_jupyter_setup():470] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7c7dab76c8b0>
31
+ 2024-01-07 11:34:44,809 INFO MainThread:476 [wandb_init.py:init():564] calling init triggers
32
+ 2024-01-07 11:34:44,809 INFO MainThread:476 [wandb_init.py:init():571] wandb.init called with sweep_config: {}
33
+ config: {}
34
+ 2024-01-07 11:34:44,810 INFO MainThread:476 [wandb_init.py:init():614] starting backend
35
+ 2024-01-07 11:34:44,810 INFO MainThread:476 [wandb_init.py:init():618] setting up manager
36
+ 2024-01-07 11:34:44,815 INFO MainThread:476 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
37
+ 2024-01-07 11:34:44,817 INFO MainThread:476 [wandb_init.py:init():624] backend started and connected
38
+ 2024-01-07 11:34:44,842 INFO MainThread:476 [wandb_run.py:_label_probe_notebook():1294] probe notebook
39
+ 2024-01-07 11:34:47,220 INFO MainThread:476 [wandb_init.py:init():716] updated telemetry
40
+ 2024-01-07 11:34:47,236 INFO MainThread:476 [wandb_init.py:init():749] communicating run to backend with 90.0 second timeout
41
+ 2024-01-07 11:34:47,619 INFO MainThread:476 [wandb_run.py:_on_init():2254] communicating current version
42
+ 2024-01-07 11:34:47,757 INFO MainThread:476 [wandb_run.py:_on_init():2263] got version response
43
+ 2024-01-07 11:34:47,758 INFO MainThread:476 [wandb_init.py:init():800] starting run threads in backend
44
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_console_start():2233] atexit reg
45
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_redirect():2088] redirect: wrap_raw
46
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_redirect():2153] Wrapping output streams.
47
+ 2024-01-07 11:34:47,846 INFO MainThread:476 [wandb_run.py:_redirect():2178] Redirects installed.
48
+ 2024-01-07 11:34:47,848 INFO MainThread:476 [wandb_init.py:init():841] run started, returning control to user process
49
+ 2024-01-07 11:34:47,851 INFO MainThread:476 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': 4096, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 10000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'mistralai/Mistral-7B-v0.1', 'transformers_version': '4.36.2', 'model_type': 'mistral', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', 'load_in_8bit': False, 'load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'bfloat16'}, 'output_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 3.0, 'max_steps': 60, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance/runs/Jan07_11-26-45_371ca9b7a7c7', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 10, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/content/gdrive/MyDrive/LLM/Mistral-7B-Finetuning-Insurance', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
wandb/run-20240107_113444-ebd79hle/run-ebd79hle.wandb ADDED
File without changes