Upload face004/logs/GPU-使用率-温度检测20240529141058/wandb/run-20240529_141120-82eq2cmy/files/output.log with huggingface_hub
Browse files
face004/logs/GPU-使用率-温度检测20240529141058/wandb/run-20240529_141120-82eq2cmy/files/output.log
ADDED
@@ -0,0 +1,100 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
epoch 1/40
|
3 |
+
epoch 2/40
|
4 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m Step only supports monotonically increasing values, use define_metric to set a custom x axis. For details see: https://wandb.me/define-metric
|
5 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 1 is less than current step: 25. Dropping entry: {'loss/epoch': 0.1278427829110653, '_timestamp': 1716992086.0651565}).
|
6 |
+
saving checkpoint: /kaggle/working/face4-000002.safetensors
|
7 |
+
epoch 3/40
|
8 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 2 is less than current step: 50. Dropping entry: {'loss/epoch': 0.09913232154688056, '_timestamp': 1716992289.1800776}).
|
9 |
+
epoch 4/40
|
10 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 3 is less than current step: 75. Dropping entry: {'loss/epoch': 0.11722434113486385, '_timestamp': 1716992493.2378407}).
|
11 |
+
saving checkpoint: /kaggle/working/face4-000004.safetensors
|
12 |
+
epoch 5/40
|
13 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 4 is less than current step: 100. Dropping entry: {'loss/epoch': 0.10072355275992684, '_timestamp': 1716992696.256858}).
|
14 |
+
epoch 6/40
|
15 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 5 is less than current step: 125. Dropping entry: {'loss/epoch': 0.11311971934094113, '_timestamp': 1716992900.0773351}).
|
16 |
+
saving checkpoint: /kaggle/working/face4-000006.safetensors
|
17 |
+
epoch 7/40
|
18 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 6 is less than current step: 150. Dropping entry: {'loss/epoch': 0.13957346594721384, '_timestamp': 1716993103.513536}).
|
19 |
+
epoch 8/40
|
20 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 7 is less than current step: 175. Dropping entry: {'loss/epoch': 0.12509408833611071, '_timestamp': 1716993307.2909555}).
|
21 |
+
saving checkpoint: /kaggle/working/face4-000008.safetensors
|
22 |
+
epoch 9/40
|
23 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 8 is less than current step: 200. Dropping entry: {'loss/epoch': 0.11486740752446409, '_timestamp': 1716993510.4338708}).
|
24 |
+
epoch 10/40
|
25 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 9 is less than current step: 225. Dropping entry: {'loss/epoch': 0.13417981069877136, '_timestamp': 1716993713.7542496}).
|
26 |
+
saving checkpoint: /kaggle/working/face4-000010.safetensors
|
27 |
+
epoch 11/40
|
28 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 10 is less than current step: 250. Dropping entry: {'loss/epoch': 0.1213072134438446, '_timestamp': 1716993916.3510857}).
|
29 |
+
epoch 12/40
|
30 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 11 is less than current step: 275. Dropping entry: {'loss/epoch': 0.0936942304608089, '_timestamp': 1716994119.6631804}).
|
31 |
+
saving checkpoint: /kaggle/working/face4-000012.safetensors
|
32 |
+
epoch 13/40
|
33 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 12 is less than current step: 300. Dropping entry: {'loss/epoch': 0.07895276304429436, '_timestamp': 1716994322.5038333}).
|
34 |
+
epoch 14/40
|
35 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 13 is less than current step: 325. Dropping entry: {'loss/epoch': 0.10570306812260984, '_timestamp': 1716994525.8621683}).
|
36 |
+
saving checkpoint: /kaggle/working/face4-000014.safetensors
|
37 |
+
epoch 15/40
|
38 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 14 is less than current step: 350. Dropping entry: {'loss/epoch': 0.1009276533138235, '_timestamp': 1716994728.5323217}).
|
39 |
+
epoch 16/40
|
40 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 15 is less than current step: 375. Dropping entry: {'loss/epoch': 0.09803004021642311, '_timestamp': 1716994931.918454}).
|
41 |
+
saving checkpoint: /kaggle/working/face4-000016.safetensors
|
42 |
+
epoch 17/40
|
43 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 16 is less than current step: 400. Dropping entry: {'loss/epoch': 0.10391505943535238, '_timestamp': 1716995134.346352}).
|
44 |
+
epoch 18/40
|
45 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 17 is less than current step: 425. Dropping entry: {'loss/epoch': 0.11970815372330193, '_timestamp': 1716995338.028467}).
|
46 |
+
saving checkpoint: /kaggle/working/face4-000018.safetensors
|
47 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 18 is less than current step: 450. Dropping entry: {'loss/epoch': 0.09248148356279244, '_timestamp': 1716995540.6438024}).
|
48 |
+
epoch 19/40
|
49 |
+
epoch 20/40
|
50 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 19 is less than current step: 475. Dropping entry: {'loss/epoch': 0.1322992009737966, '_timestamp': 1716995743.8193698}).
|
51 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 20 is less than current step: 500. Dropping entry: {'loss/epoch': 0.12299238996846336, '_timestamp': 1716995946.3782701}).
|
52 |
+
saving checkpoint: /kaggle/working/face4-000020.safetensors
|
53 |
+
epoch 21/40
|
54 |
+
epoch 22/40
|
55 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 21 is less than current step: 525. Dropping entry: {'loss/epoch': 0.08653344316598104, '_timestamp': 1716996149.6951635}).
|
56 |
+
saving checkpoint: /kaggle/working/face4-000022.safetensors
|
57 |
+
epoch 23/40
|
58 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 22 is less than current step: 550. Dropping entry: {'loss/epoch': 0.11906309035720722, '_timestamp': 1716996352.5354953}).
|
59 |
+
epoch 24/40
|
60 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 23 is less than current step: 575. Dropping entry: {'loss/epoch': 0.1258385900255977, '_timestamp': 1716996555.8893914}).
|
61 |
+
saving checkpoint: /kaggle/working/face4-000024.safetensors
|
62 |
+
epoch 25/40
|
63 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 24 is less than current step: 600. Dropping entry: {'loss/epoch': 0.08512429271976711, '_timestamp': 1716996758.919449}).
|
64 |
+
epoch 26/40
|
65 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 25 is less than current step: 625. Dropping entry: {'loss/epoch': 0.13288933675432082, '_timestamp': 1716996962.1542478}).
|
66 |
+
saving checkpoint: /kaggle/working/face4-000026.safetensors
|
67 |
+
epoch 27/40
|
68 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 26 is less than current step: 650. Dropping entry: {'loss/epoch': 0.1215609906868515, '_timestamp': 1716997164.8041437}).
|
69 |
+
epoch 28/40
|
70 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 27 is less than current step: 675. Dropping entry: {'loss/epoch': 0.09894117455435347, '_timestamp': 1716997367.8791773}).
|
71 |
+
saving checkpoint: /kaggle/working/face4-000028.safetensors
|
72 |
+
epoch 29/40
|
73 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 28 is less than current step: 700. Dropping entry: {'loss/epoch': 0.11361698019413316, '_timestamp': 1716997570.4917023}).
|
74 |
+
epoch 30/40
|
75 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 29 is less than current step: 725. Dropping entry: {'loss/epoch': 0.1302316881412146, '_timestamp': 1716997773.7550204}).
|
76 |
+
saving checkpoint: /kaggle/working/face4-000030.safetensors
|
77 |
+
epoch 31/40
|
78 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 30 is less than current step: 750. Dropping entry: {'loss/epoch': 0.11116170267127834, '_timestamp': 1716997976.2618535}).
|
79 |
+
epoch 32/40
|
80 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 31 is less than current step: 775. Dropping entry: {'loss/epoch': 0.11247560868457872, '_timestamp': 1716998179.6457098}).
|
81 |
+
saving checkpoint: /kaggle/working/face4-000032.safetensors
|
82 |
+
epoch 33/40
|
83 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 32 is less than current step: 800. Dropping entry: {'loss/epoch': 0.10462229371032848, '_timestamp': 1716998381.9376934}).
|
84 |
+
epoch 34/40
|
85 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 33 is less than current step: 825. Dropping entry: {'loss/epoch': 0.11725937242486648, '_timestamp': 1716998584.9216814}).
|
86 |
+
saving checkpoint: /kaggle/working/face4-000034.safetensors
|
87 |
+
epoch 35/40
|
88 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 34 is less than current step: 850. Dropping entry: {'loss/epoch': 0.12117312730251982, '_timestamp': 1716998787.7442768}).
|
89 |
+
epoch 36/40
|
90 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 35 is less than current step: 875. Dropping entry: {'loss/epoch': 0.0777922967135222, '_timestamp': 1716998990.8795753}).
|
91 |
+
saving checkpoint: /kaggle/working/face4-000036.safetensors
|
92 |
+
epoch 37/40
|
93 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 36 is less than current step: 900. Dropping entry: {'loss/epoch': 0.09929781914593613, '_timestamp': 1716999193.6439993}).
|
94 |
+
epoch 38/40
|
95 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 37 is less than current step: 925. Dropping entry: {'loss/epoch': 0.1320431025565735, '_timestamp': 1716999396.871643}).
|
96 |
+
saving checkpoint: /kaggle/working/face4-000038.safetensors
|
97 |
+
epoch 39/40
|
98 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 38 is less than current step: 950. Dropping entry: {'loss/epoch': 0.10733248616511724, '_timestamp': 1716999599.3448024}).
|
99 |
+
epoch 40/40
|
100 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m (User provided step: 39 is less than current step: 975. Dropping entry: {'loss/epoch': 0.1096145034368549, '_timestamp': 1716999802.5580628}).
|