Mahadih534
commited on
Commit
•
5c37aa4
1
Parent(s):
f51c629
Upload 22 files
Browse files- train-v1/F1_curve.png +0 -0
- train-v1/PR_curve.png +0 -0
- train-v1/P_curve.png +0 -0
- train-v1/R_curve.png +0 -0
- train-v1/args.yaml +109 -0
- train-v1/confusion_matrix.png +0 -0
- train-v1/confusion_matrix_normalized.png +0 -0
- train-v1/events.out.tfevents.1727939194.0c30dad15f51.93.0 +3 -0
- train-v1/labels.jpg +0 -0
- train-v1/labels_correlogram.jpg +0 -0
- train-v1/results.csv +101 -0
- train-v1/results.png +0 -0
- train-v1/train_batch0.jpg +0 -0
- train-v1/train_batch1.jpg +0 -0
- train-v1/train_batch2.jpg +0 -0
- train-v1/train_batch90.jpg +0 -0
- train-v1/train_batch91.jpg +0 -0
- train-v1/train_batch92.jpg +0 -0
- train-v1/val_batch0_labels.jpg +0 -0
- train-v1/val_batch0_pred.jpg +0 -0
- train-v1/weights/best.pt +3 -0
- train-v1/weights/last.pt +3 -0
train-v1/F1_curve.png
ADDED
train-v1/PR_curve.png
ADDED
train-v1/P_curve.png
ADDED
train-v1/R_curve.png
ADDED
train-v1/args.yaml
ADDED
@@ -0,0 +1,109 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
task: obb
|
2 |
+
mode: train
|
3 |
+
model: yolov8s-obb.pt
|
4 |
+
data: /opt/conda/lib/python3.10/site-packages/ultralytics/cfg/datasets/dota8.yaml
|
5 |
+
epochs: 100
|
6 |
+
time: null
|
7 |
+
patience: 100
|
8 |
+
batch: 32
|
9 |
+
imgsz: 640
|
10 |
+
save: true
|
11 |
+
save_period: -1
|
12 |
+
cache: false
|
13 |
+
device:
|
14 |
+
- 0
|
15 |
+
- 1
|
16 |
+
workers: 16
|
17 |
+
project: null
|
18 |
+
name: train
|
19 |
+
exist_ok: false
|
20 |
+
pretrained: true
|
21 |
+
optimizer: auto
|
22 |
+
verbose: true
|
23 |
+
seed: 42
|
24 |
+
deterministic: true
|
25 |
+
single_cls: false
|
26 |
+
rect: false
|
27 |
+
cos_lr: false
|
28 |
+
close_mosaic: 10
|
29 |
+
resume: false
|
30 |
+
amp: true
|
31 |
+
fraction: 1.0
|
32 |
+
profile: false
|
33 |
+
freeze: null
|
34 |
+
multi_scale: false
|
35 |
+
overlap_mask: true
|
36 |
+
mask_ratio: 4
|
37 |
+
dropout: 0.0
|
38 |
+
val: true
|
39 |
+
split: val
|
40 |
+
save_json: false
|
41 |
+
save_hybrid: false
|
42 |
+
conf: null
|
43 |
+
iou: 0.7
|
44 |
+
max_det: 300
|
45 |
+
half: false
|
46 |
+
dnn: false
|
47 |
+
plots: true
|
48 |
+
source: null
|
49 |
+
vid_stride: 1
|
50 |
+
stream_buffer: false
|
51 |
+
visualize: false
|
52 |
+
augment: false
|
53 |
+
agnostic_nms: false
|
54 |
+
classes: null
|
55 |
+
retina_masks: false
|
56 |
+
embed: null
|
57 |
+
show: false
|
58 |
+
save_frames: false
|
59 |
+
save_txt: false
|
60 |
+
save_conf: false
|
61 |
+
save_crop: false
|
62 |
+
show_labels: true
|
63 |
+
show_conf: true
|
64 |
+
show_boxes: true
|
65 |
+
line_width: null
|
66 |
+
format: torchscript
|
67 |
+
keras: false
|
68 |
+
optimize: false
|
69 |
+
int8: false
|
70 |
+
dynamic: false
|
71 |
+
simplify: true
|
72 |
+
opset: null
|
73 |
+
workspace: 4
|
74 |
+
nms: false
|
75 |
+
lr0: 0.01
|
76 |
+
lrf: 0.01
|
77 |
+
momentum: 0.937
|
78 |
+
weight_decay: 0.0005
|
79 |
+
warmup_epochs: 3.0
|
80 |
+
warmup_momentum: 0.8
|
81 |
+
warmup_bias_lr: 0.1
|
82 |
+
box: 7.5
|
83 |
+
cls: 0.5
|
84 |
+
dfl: 1.5
|
85 |
+
pose: 12.0
|
86 |
+
kobj: 1.0
|
87 |
+
label_smoothing: 0.0
|
88 |
+
nbs: 64
|
89 |
+
hsv_h: 0.015
|
90 |
+
hsv_s: 0.7
|
91 |
+
hsv_v: 0.4
|
92 |
+
degrees: 0.0
|
93 |
+
translate: 0.1
|
94 |
+
scale: 0.5
|
95 |
+
shear: 0.0
|
96 |
+
perspective: 0.0
|
97 |
+
flipud: 0.0
|
98 |
+
fliplr: 0.5
|
99 |
+
bgr: 0.0
|
100 |
+
mosaic: 1.0
|
101 |
+
mixup: 0.0
|
102 |
+
copy_paste: 0.0
|
103 |
+
copy_paste_mode: flip
|
104 |
+
auto_augment: randaugment
|
105 |
+
erasing: 0.4
|
106 |
+
crop_fraction: 1.0
|
107 |
+
cfg: null
|
108 |
+
tracker: botsort.yaml
|
109 |
+
save_dir: runs/obb/train
|
train-v1/confusion_matrix.png
ADDED
train-v1/confusion_matrix_normalized.png
ADDED
train-v1/events.out.tfevents.1727939194.0c30dad15f51.93.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a8d576830361dd0d0b699fd1bc9a0946068a695422fd557de5c4ee87ebfb4b55
|
3 |
+
size 275342
|
train-v1/labels.jpg
ADDED
train-v1/labels_correlogram.jpg
ADDED
train-v1/results.csv
ADDED
@@ -0,0 +1,101 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch, train/box_loss, train/cls_loss, train/dfl_loss, metrics/precision(B), metrics/recall(B), metrics/mAP50(B), metrics/mAP50-95(B), val/box_loss, val/cls_loss, val/dfl_loss, lr/pg0, lr/pg1, lr/pg2
|
2 |
+
1, 1.9371, 5.4822, 1.5856, 0, 0, 0, 0, 0.6096, 4.4185, 1.3397, 0, 0, 0
|
3 |
+
2, 1.0918, 5.5125, 1.2627, 0, 0, 0, 0, 0.61123, 4.4188, 1.3336, 7.0693e-06, 7.0693e-06, 7.0693e-06
|
4 |
+
3, 1.28, 5.9668, 1.213, 0, 0, 0, 0, 0.60045, 4.4309, 1.3342, 1.3997e-05, 1.3997e-05, 1.3997e-05
|
5 |
+
4, 1.4446, 5.3258, 1.5162, 0, 0, 0, 0, 0.59769, 4.4314, 1.3274, 2.0784e-05, 2.0784e-05, 2.0784e-05
|
6 |
+
5, 1.323, 5.7224, 1.525, 0, 0, 0, 0, 0.59318, 4.4255, 1.3412, 2.7429e-05, 2.7429e-05, 2.7429e-05
|
7 |
+
6, 0.84641, 5.9268, 1.1927, 0, 0, 0, 0, 0.59303, 4.43, 1.3273, 3.3933e-05, 3.3933e-05, 3.3933e-05
|
8 |
+
7, 1.6462, 5.396, 1.2974, 0, 0, 0, 0, 0.58735, 4.4306, 1.3301, 4.0295e-05, 4.0295e-05, 4.0295e-05
|
9 |
+
8, 1.0818, 5.4421, 1.8967, 0, 0, 0, 0, 0.59388, 4.4285, 1.3246, 4.6516e-05, 4.6516e-05, 4.6516e-05
|
10 |
+
9, 1.1875, 5.8695, 1.4828, 0, 0, 0, 0, 0.58322, 4.4289, 1.3317, 5.2596e-05, 5.2596e-05, 5.2596e-05
|
11 |
+
10, 1.1339, 5.1997, 2.017, 0, 0, 0, 0, 0.5773, 4.4305, 1.3312, 5.8534e-05, 5.8534e-05, 5.8534e-05
|
12 |
+
11, 1.1019, 5.2016, 1.9295, 0, 0, 0, 0, 0.58576, 4.4291, 1.3345, 6.4331e-05, 6.4331e-05, 6.4331e-05
|
13 |
+
12, 1.2039, 5.112, 1.4187, 0, 0, 0, 0, 0.59884, 4.426, 1.3661, 6.9987e-05, 6.9987e-05, 6.9987e-05
|
14 |
+
13, 0.93501, 5.3407, 1.2059, 0, 0, 0, 0, 0.60102, 4.4242, 1.376, 7.5501e-05, 7.5501e-05, 7.5501e-05
|
15 |
+
14, 0.99984, 4.9424, 1.7481, 0, 0, 0, 0, 0.604, 4.4204, 1.4011, 8.0874e-05, 8.0874e-05, 8.0874e-05
|
16 |
+
15, 0.56343, 4.7601, 1.2252, 0, 0, 0, 0, 0.61481, 4.4256, 1.3984, 8.6106e-05, 8.6106e-05, 8.6106e-05
|
17 |
+
16, 1.2092, 5.3076, 1.4878, 0, 0, 0, 0, 0.61558, 4.4326, 1.3984, 9.1196e-05, 9.1196e-05, 9.1196e-05
|
18 |
+
17, 0.74554, 5.3813, 1.4725, 0, 0, 0, 0, 0.62448, 4.4302, 1.3996, 9.6144e-05, 9.6144e-05, 9.6144e-05
|
19 |
+
18, 0.62091, 4.9868, 1.2827, 0, 0, 0, 0, 0.62541, 4.4317, 1.4162, 0.00010095, 0.00010095, 0.00010095
|
20 |
+
19, 1.1029, 4.9585, 1.634, 0, 0, 0, 0, 0.6262, 4.4501, 1.4187, 0.00010562, 0.00010562, 0.00010562
|
21 |
+
20, 1.0567, 4.4954, 2.1528, 0, 0, 0, 0, 0.63449, 4.4441, 1.4176, 0.00011014, 0.00011014, 0.00011014
|
22 |
+
21, 0.99971, 4.3917, 1.2582, 0, 0, 0, 0, 0.62248, 4.4509, 1.4156, 0.00011453, 0.00011453, 0.00011453
|
23 |
+
22, 1.4189, 4.2809, 1.2863, 0, 0, 0, 0, 0.62239, 4.4556, 1.4196, 0.00011877, 0.00011877, 0.00011877
|
24 |
+
23, 1.1773, 4.1219, 1.6193, 0, 0, 0, 0, 0.62444, 4.459, 1.4022, 0.00012287, 0.00012287, 0.00012287
|
25 |
+
24, 1.1835, 4.0819, 1.6085, 0, 0, 0, 0, 0.62788, 4.4558, 1.4171, 0.00012683, 0.00012683, 0.00012683
|
26 |
+
25, 1.0057, 4.0691, 1.442, 0, 0, 0, 0, 0.62713, 4.4635, 1.4246, 0.00013064, 0.00013064, 0.00013064
|
27 |
+
26, 0.91728, 4.1129, 1.389, 0, 0, 0, 0, 0.63463, 4.4582, 1.4277, 0.00013432, 0.00013432, 0.00013432
|
28 |
+
27, 1.1708, 3.8488, 1.5143, 0, 0, 0, 0, 0.62967, 4.4597, 1.4316, 0.00013786, 0.00013786, 0.00013786
|
29 |
+
28, 0.70989, 4.2745, 1.4559, 0, 0, 0, 0, 0.63995, 4.4532, 1.4515, 0.00014125, 0.00014125, 0.00014125
|
30 |
+
29, 1.3619, 4.1827, 1.4965, 0, 0, 0, 0, 0.63776, 4.4554, 1.4489, 0.0001445, 0.0001445, 0.0001445
|
31 |
+
30, 1.1156, 3.6929, 1.7537, 0, 0, 0, 0, 0.66107, 4.4524, 1.4635, 0.00014761, 0.00014761, 0.00014761
|
32 |
+
31, 0.64132, 4.2313, 1.4215, 0, 0, 0, 0, 0.72044, 4.4477, 1.4741, 0.00015058, 0.00015058, 0.00015058
|
33 |
+
32, 0.93578, 3.5977, 1.8066, 0, 0, 0, 0, 0.73391, 4.4346, 1.4825, 0.00015341, 0.00015341, 0.00015341
|
34 |
+
33, 1.1019, 3.8091, 1.2542, 0, 0, 0, 0, 0.74412, 4.4364, 1.5038, 0.0001561, 0.0001561, 0.0001561
|
35 |
+
34, 0.84539, 3.4436, 1.0689, 0, 0, 0, 0, 0.74807, 4.4378, 1.497, 0.00015864, 0.00015864, 0.00015864
|
36 |
+
35, 0.91252, 3.2498, 1.6328, 0, 0, 0, 0, 0.75948, 4.4437, 1.4959, 0.00016105, 0.00016105, 0.00016105
|
37 |
+
36, 1.8134, 3.8593, 2.6437, 0, 0, 0, 0, 0.77561, 4.437, 1.4952, 0.00016331, 0.00016331, 0.00016331
|
38 |
+
37, 0.84347, 3.5063, 2.0216, 0, 0, 0, 0, 0.78959, 4.4256, 1.4934, 0.00016543, 0.00016543, 0.00016543
|
39 |
+
38, 1.2059, 3.2259, 1.4299, 0, 0, 0, 0, 0.77798, 4.4162, 1.4938, 0.00016741, 0.00016741, 0.00016741
|
40 |
+
39, 0.98883, 3.2292, 2.0503, 0, 0, 0, 0, 0.76268, 4.4108, 1.5055, 0.00016925, 0.00016925, 0.00016925
|
41 |
+
40, 0.85596, 3.5144, 1.4032, 0, 0, 0, 0, 0.76868, 4.4041, 1.5219, 0.00017095, 0.00017095, 0.00017095
|
42 |
+
41, 1.0008, 2.7508, 1.609, 0, 0, 0, 0, 0.75267, 4.39, 1.5238, 0.0001725, 0.0001725, 0.0001725
|
43 |
+
42, 0.93227, 2.8545, 1.7097, 0, 0, 0, 0, 0.75255, 4.3755, 1.5329, 0.00017392, 0.00017392, 0.00017392
|
44 |
+
43, 0.87062, 2.9192, 1.3736, 0, 0, 0, 0, 0.74059, 4.382, 1.5275, 0.00017519, 0.00017519, 0.00017519
|
45 |
+
44, 0.62348, 3.0319, 1.6643, 0, 0, 0, 0, 0.75048, 4.3778, 1.5212, 0.00017632, 0.00017632, 0.00017632
|
46 |
+
45, 1.0943, 3.714, 1.6488, 0, 0, 0, 0, 0.76142, 4.3807, 1.5111, 0.00017731, 0.00017731, 0.00017731
|
47 |
+
46, 1.0184, 2.4912, 1.1645, 0, 0, 0, 0, 0.77397, 4.3942, 1.4981, 0.00017816, 0.00017816, 0.00017816
|
48 |
+
47, 0.86511, 3.1191, 1.5055, 0, 0, 0, 0, 0.78181, 4.4017, 1.4931, 0.00017887, 0.00017887, 0.00017887
|
49 |
+
48, 0.99275, 2.1952, 1.6829, 0, 0, 0, 0, 0.77224, 4.4213, 1.4693, 0.00017943, 0.00017943, 0.00017943
|
50 |
+
49, 0.99244, 2.2928, 1.2764, 0, 0, 0, 0, 0.81467, 4.3883, 1.527, 0.00017986, 0.00017986, 0.00017986
|
51 |
+
50, 0.95475, 1.6466, 1.6346, 0, 0, 0, 0, 0.86053, 4.3532, 1.5764, 0.00018014, 0.00018014, 0.00018014
|
52 |
+
51, 1.1038, 1.964, 1.5556, 0, 0, 0, 0, 0.86053, 4.3532, 1.5764, 0.00018028, 0.00018028, 0.00018028
|
53 |
+
52, 0.93106, 1.9387, 1.4589, 0, 0, 0, 0, 0.84523, 4.3657, 1.5789, 0.00018029, 0.00018029, 0.00018029
|
54 |
+
53, 0.856, 1.5944, 1.2103, 0, 0, 0, 0, 0.84523, 4.3657, 1.5789, 0.00018015, 0.00018015, 0.00018015
|
55 |
+
54, 0.89147, 2.3165, 1.6776, 0, 0, 0, 0, 0.85971, 4.3414, 1.603, 0.00017986, 0.00017986, 0.00017986
|
56 |
+
55, 0.74464, 2.8799, 1.2972, 0, 0, 0, 0, 0.85971, 4.3414, 1.603, 0.00017944, 0.00017944, 0.00017944
|
57 |
+
56, 1.0111, 1.6111, 1.4149, 0, 0, 0, 0, 0.87764, 4.3188, 1.6072, 0.00017887, 0.00017887, 0.00017887
|
58 |
+
57, 0.89807, 2.1541, 1.4189, 0, 0, 0, 0, 0.87764, 4.3188, 1.6072, 0.00017817, 0.00017817, 0.00017817
|
59 |
+
58, 0.91826, 1.7801, 1.4295, 0, 0, 0, 0, 0.92343, 4.2986, 1.5956, 0.00017732, 0.00017732, 0.00017732
|
60 |
+
59, 1.1303, 1.9554, 1.6349, 0, 0, 0, 0, 0.92343, 4.2986, 1.5956, 0.00017633, 0.00017633, 0.00017633
|
61 |
+
60, 0.79234, 2.0986, 1.1406, 0, 0, 0, 0, 0.94124, 4.2925, 1.5863, 0.0001752, 0.0001752, 0.0001752
|
62 |
+
61, 0.6392, 2.7829, 1.4924, 0, 0, 0, 0, 0.94124, 4.2925, 1.5863, 0.00017393, 0.00017393, 0.00017393
|
63 |
+
62, 0.99579, 1.6707, 1.5865, 0.33333, 0.11111, 0.22222, 0.17777, 0.97114, 4.2894, 1.5881, 0.00017252, 0.00017252, 0.00017252
|
64 |
+
63, 0.78802, 2.9204, 1.5541, 0.33333, 0.11111, 0.22222, 0.17777, 0.97114, 4.2894, 1.5881, 0.00017096, 0.00017096, 0.00017096
|
65 |
+
64, 0.93819, 2.3589, 1.4424, 0.33333, 0.33333, 0.33167, 0.2746, 0.99228, 4.2713, 1.6014, 0.00016927, 0.00016927, 0.00016927
|
66 |
+
65, 0.85676, 2.7978, 1.5046, 0.33333, 0.33333, 0.33167, 0.2746, 0.99228, 4.2713, 1.6014, 0.00016743, 0.00016743, 0.00016743
|
67 |
+
66, 0.7237, 2.843, 1.4457, 0.33333, 0.33333, 0.33167, 0.28385, 1.0134, 4.2616, 1.6142, 0.00016545, 0.00016545, 0.00016545
|
68 |
+
67, 0.6095, 2.4592, 1.1793, 0.33333, 0.33333, 0.33167, 0.28385, 1.0134, 4.2616, 1.6142, 0.00016333, 0.00016333, 0.00016333
|
69 |
+
68, 0.96159, 1.9286, 1.6123, 0.33333, 0.33333, 0.33167, 0.2985, 0.99968, 4.2549, 1.6091, 0.00016107, 0.00016107, 0.00016107
|
70 |
+
69, 0.50476, 2.4994, 1.0056, 0.33333, 0.33333, 0.33167, 0.2985, 0.99968, 4.2549, 1.6091, 0.00015867, 0.00015867, 0.00015867
|
71 |
+
70, 0.98407, 1.4497, 1.367, 0.33333, 0.33333, 0.33167, 0.2985, 0.97197, 4.2422, 1.5888, 0.00015612, 0.00015612, 0.00015612
|
72 |
+
71, 0.88926, 1.4476, 1.4651, 0.33333, 0.33333, 0.33167, 0.2985, 0.97197, 4.2422, 1.5888, 0.00015344, 0.00015344, 0.00015344
|
73 |
+
72, 0.63365, 2.5137, 1.7397, 0.33333, 0.33333, 0.33167, 0.2985, 0.96227, 4.2225, 1.574, 0.00015061, 0.00015061, 0.00015061
|
74 |
+
73, 1.2456, 1.6977, 1.3765, 0.33333, 0.33333, 0.33167, 0.2985, 0.96227, 4.2225, 1.574, 0.00014764, 0.00014764, 0.00014764
|
75 |
+
74, 0.6147, 2.4727, 1.1627, 0.33333, 0.33333, 0.33167, 0.30591, 0.93656, 4.2307, 1.5664, 0.00014453, 0.00014453, 0.00014453
|
76 |
+
75, 1.0792, 1.8578, 1.2547, 0.33333, 0.33333, 0.33167, 0.30591, 0.93656, 4.2307, 1.5664, 0.00014128, 0.00014128, 0.00014128
|
77 |
+
76, 0.82627, 1.0721, 1.5261, 0.33333, 0.33333, 0.33167, 0.28757, 0.94145, 4.2119, 1.5616, 0.00013789, 0.00013789, 0.00013789
|
78 |
+
77, 0.77034, 1.2569, 1.5233, 0.33333, 0.33333, 0.33167, 0.28757, 0.94145, 4.2119, 1.5616, 0.00013436, 0.00013436, 0.00013436
|
79 |
+
78, 0.73682, 2.4203, 1.4725, 0.33333, 0.33333, 0.33167, 0.29124, 0.96029, 4.2203, 1.5463, 0.00013068, 0.00013068, 0.00013068
|
80 |
+
79, 0.72919, 2.4714, 1.1651, 0.33333, 0.33333, 0.33167, 0.29124, 0.96029, 4.2203, 1.5463, 0.00012687, 0.00012687, 0.00012687
|
81 |
+
80, 0.92404, 2.0944, 1.1016, 0.33333, 0.33333, 0.33167, 0.29124, 0.96435, 4.2195, 1.5434, 0.00012291, 0.00012291, 0.00012291
|
82 |
+
81, 0.6737, 2.4201, 1.0469, 0.33333, 0.33333, 0.33167, 0.29124, 0.96435, 4.2195, 1.5434, 0.00011881, 0.00011881, 0.00011881
|
83 |
+
82, 0.81295, 2.4266, 1.2519, 0.33333, 0.33333, 0.33167, 0.28757, 0.95992, 4.198, 1.5409, 0.00011457, 0.00011457, 0.00011457
|
84 |
+
83, 0.93353, 1.4744, 1.4787, 0.33333, 0.33333, 0.33167, 0.28757, 0.95992, 4.198, 1.5409, 0.00011019, 0.00011019, 0.00011019
|
85 |
+
84, 0.77733, 1.2069, 1.359, 0.33333, 0.33333, 0.33167, 0.29124, 0.9614, 4.1868, 1.5303, 0.00010566, 0.00010566, 0.00010566
|
86 |
+
85, 0.537, 2.1854, 1.1132, 0.33333, 0.33333, 0.33167, 0.29124, 0.9614, 4.1868, 1.5303, 0.000101, 0.000101, 0.000101
|
87 |
+
86, 0.89207, 1.306, 1.4278, 0.33333, 0.33333, 0.33167, 0.29124, 0.9609, 4.1791, 1.5319, 9.6194e-05, 9.6194e-05, 9.6194e-05
|
88 |
+
87, 1.1278, 1.5631, 1.6268, 0.33333, 0.33333, 0.33167, 0.29124, 0.9609, 4.1791, 1.5319, 9.1246e-05, 9.1246e-05, 9.1246e-05
|
89 |
+
88, 0.89423, 1.1934, 1.2682, 0.33333, 0.33333, 0.33167, 0.27274, 0.96047, 4.1719, 1.5429, 8.6158e-05, 8.6158e-05, 8.6158e-05
|
90 |
+
89, 0.79279, 1.1858, 1.3171, 0.33333, 0.33333, 0.33167, 0.27274, 0.96047, 4.1719, 1.5429, 8.0928e-05, 8.0928e-05, 8.0928e-05
|
91 |
+
90, 0.88971, 1.2936, 1.3713, 0.25, 0.33333, 0.33167, 0.27182, 0.9443, 4.1712, 1.5199, 7.5556e-05, 7.5556e-05, 7.5556e-05
|
92 |
+
91, 0.38342, 2.0782, 0.71686, 0.25, 0.33333, 0.33167, 0.27182, 0.9443, 4.1712, 1.5199, 7.0043e-05, 7.0043e-05, 7.0043e-05
|
93 |
+
92, 0.90025, 3.371, 1.3138, 0.25, 0.33333, 0.33167, 0.27182, 0.93702, 4.1551, 1.5193, 6.4389e-05, 6.4389e-05, 6.4389e-05
|
94 |
+
93, 0.40505, 2.4194, 0.92424, 0.25, 0.33333, 0.33167, 0.27182, 0.93702, 4.1551, 1.5193, 5.8594e-05, 5.8594e-05, 5.8594e-05
|
95 |
+
94, 0.38864, 2.1736, 0.76463, 0.25, 0.33333, 0.33167, 0.27182, 0.93705, 4.1393, 1.521, 5.2657e-05, 5.2657e-05, 5.2657e-05
|
96 |
+
95, 0.56513, 2.521, 1.137, 0.25, 0.33333, 0.33167, 0.27182, 0.93705, 4.1393, 1.521, 4.6579e-05, 4.6579e-05, 4.6579e-05
|
97 |
+
96, 0.94875, 1.2061, 1.2379, 0.2, 0.33333, 0.33167, 0.28236, 0.94382, 4.1368, 1.5246, 4.0359e-05, 4.0359e-05, 4.0359e-05
|
98 |
+
97, 0.90396, 2.6939, 1.788, 0.2, 0.33333, 0.33167, 0.28236, 0.94382, 4.1368, 1.5246, 3.3998e-05, 3.3998e-05, 3.3998e-05
|
99 |
+
98, 0.51437, 1.8355, 1.3005, 0.2, 0.33333, 0.33167, 0.28236, 0.95617, 4.1498, 1.5246, 2.7495e-05, 2.7495e-05, 2.7495e-05
|
100 |
+
99, 0.97043, 2.9099, 2.2593, 0.2, 0.33333, 0.33167, 0.28236, 0.95617, 4.1498, 1.5246, 2.0852e-05, 2.0852e-05, 2.0852e-05
|
101 |
+
100, 1.4045, 3.2643, 1.6, 0.2, 0.33333, 0.33167, 0.28236, 0.96427, 4.1468, 1.5347, 1.4067e-05, 1.4067e-05, 1.4067e-05
|
train-v1/results.png
ADDED
train-v1/train_batch0.jpg
ADDED
train-v1/train_batch1.jpg
ADDED
train-v1/train_batch2.jpg
ADDED
train-v1/train_batch90.jpg
ADDED
train-v1/train_batch91.jpg
ADDED
train-v1/train_batch92.jpg
ADDED
train-v1/val_batch0_labels.jpg
ADDED
train-v1/val_batch0_pred.jpg
ADDED
train-v1/weights/best.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e29260d58a0297b4143e69214d78ad2592309c57fafb8e07a532eb425c0893ef
|
3 |
+
size 20624792
|
train-v1/weights/last.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c193207e61d6871a6ad78b31648f37f3c92970f8e3f6cccc581bc0396c77fa0
|
3 |
+
size 20624792
|