micheller7171
commited on
Commit
•
8989ead
1
Parent(s):
ff23053
added model card info
Browse files
README.md
CHANGED
@@ -1,3 +1,68 @@
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
4 |
+
|
5 |
+
## Model Card for butterfly_segmentation_yolo_v8
|
6 |
+
|
7 |
+
This model takes in an image of a butterfly (with or without body attached to wings) and segments out any existing hindwings and forewings, in addition to pictured equipment described below.
|
8 |
+
|
9 |
+
## Model Details
|
10 |
+
|
11 |
+
yolov8m_shear_10.0_scale_0.5_translate_0.1_fliplr_0.0_best.pt is the butterfly segmentation model.
|
12 |
+
|
13 |
+
The butterfly segmentation model was trained on a dataset of 800 total images from the Jiggins, OM_STRI, and Monteiro datasets. The model architecture is based on YOLO v8 (yolov8m-seg.pt), which we fine-tune further on our dataset of 800 images.
|
14 |
+
|
15 |
+
## Model Description
|
16 |
+
|
17 |
+
The model is responsible for taking an input image (RGB) and generating segmentation masks for all classes below that are found in the image. Data augmentations applied during training include shear (10.0), scale (0.5), and translate (0.1). The model was trained for 50 epochs with an image size of 256. Note that despite defining an image size of 256, the normalized masks predicted by yolo can be rescaled to the original image size.
|
18 |
+
|
19 |
+
### Segmentation Classes
|
20 |
+
[`pixel class`] corresponding category
|
21 |
+
|
22 |
+
- [0] background
|
23 |
+
- [1] right_forewing
|
24 |
+
- [2] left_forewing
|
25 |
+
- [3] right_hindwing
|
26 |
+
- [4] left_hindwing
|
27 |
+
- [5] ruler
|
28 |
+
- [6] white_balance
|
29 |
+
- [7] label
|
30 |
+
- [8] color_card
|
31 |
+
- [9] body
|
32 |
+
|
33 |
+
### Details
|
34 |
+
|
35 |
+
model.train(data=YAML,
|
36 |
+
imgsz=256,
|
37 |
+
epochs=50,
|
38 |
+
batch=16,
|
39 |
+
device=DEVICE,
|
40 |
+
optimizer='auto',
|
41 |
+
verbose=True,
|
42 |
+
val=True,
|
43 |
+
shear=10.0,
|
44 |
+
scale=0.5,
|
45 |
+
translate=0.1,
|
46 |
+
fliplr = 0.0
|
47 |
+
)
|
48 |
+
|
49 |
+
## Metrics
|
50 |
+
|
51 |
+
Class Images Instances mAP50-95
|
52 |
+
all 64 358
|
53 |
+
background 64 3 0.20946
|
54 |
+
right_forewing 64 58 0.9845
|
55 |
+
left_forewing 64 51 0.9682
|
56 |
+
right_hindwing 64 59 0.95296
|
57 |
+
left_hindwing 64 50 0.93961
|
58 |
+
ruler 64 31 0.73608
|
59 |
+
white_balance 64 18 0.90686
|
60 |
+
label 64 50 0.80865
|
61 |
+
color_card 64 24 0.92653
|
62 |
+
body 64 14 0.78283
|
63 |
+
|
64 |
+
**Developed by:** Michelle Ramirez
|
65 |
+
|
66 |
+
## How to Get Started with the Model
|
67 |
+
|
68 |
+
To view applications of how to load in the model file and predict masks on images, please refer to [this github repository](https://github.com/Imageomics/wing-segmentation)
|