wiktorlazarski
commited on
Commit
•
366cb20
1
Parent(s):
1b7c7c6
Update README.md
Browse files
README.md
CHANGED
@@ -14,16 +14,26 @@ The decoder is enhanced through the incorporation of Spatial and Channel "Squeez
|
|
14 |
|
15 |
## Dataset
|
16 |
|
17 |
-
Importantly, no Worldcoin user data was used to train or fine-tune the IRIS pipeline. Rather, a research
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
-
|
25 |
-
|
26 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
|
28 |
## Performance
|
29 |
|
@@ -33,17 +43,33 @@ The experimental dataset contained a total of 9 957 manually annotated IR images
|
|
33 |
|
34 |
| **Metric** | **Value** |
|
35 |
|-----------------------------|-----------|
|
36 |
-
| _eyeball_IoU_ | 0.
|
37 |
-
| _iris_IoU_ | 0.
|
38 |
-
| _pupil_IoU_ | 0.
|
39 |
-
| _eyelashes_IoU_ | 0.
|
40 |
-
| _mIoU_ | 0.
|
41 |
-
| _eyelashes_inside_iris_IoU_ | 0.
|
42 |
-
| _boundary_confidence_pupil_ | 0.
|
43 |
-
| _boundary_confidence_iris_ | 0.
|
44 |
-
| _chamfer_eyeball_ | 1.
|
45 |
-
| _chamfer_iris_ | 1.
|
46 |
-
| _chamfer_pupil_ | 0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
47 |
|
48 |
### Time Performance
|
49 |
|
@@ -76,36 +102,25 @@ The model yields a tensor characterized by dimensions of Nx4x640x480, denoting b
|
|
76 |
|
77 |
Within the ambit of each class, the model formulates probability estimates pertaining to the likelihood of a given pixel being attributed to a particular class.
|
78 |
|
79 |
-
###
|
80 |
|
81 |
**Note**: The provided input image has been subjected to the processing methodology described earlier, prior to its introduction into the model. Moreover, for the intent of visualization, the IR image presented has been anonymized to safeguard the identity of the user. It is also worth to note that the inference process was conducted on the original, non-anonymized version of the image.
|
82 |
|
83 |
#### Input image
|
84 |
|
85 |
-
<img src="https://github.com/worldcoin/open-iris/blob/main/docs/model_card/anonymized.png
|
86 |
|
87 |
#### Output masks
|
88 |
|
89 |
**Note**: A threshold of 0.5 was applied on every resultant probability matrix to generate a definitive binary mask for each individual class. Also, every outputted segmentation map was resized to the input image resolution.
|
90 |
|
91 |
-
<img src="https://github.com/worldcoin/open-iris/blob/main/docs/model_card/overlayed_segmaps.png
|
92 |
|
93 |
## Limitations
|
94 |
|
95 |
Thorough examination of the results enabled us to pinpoint situations where the segmentation model experiences declines in performance. These instances are as follows:
|
96 |
-
- Segmenting images
|
97 |
-
-
|
98 |
-
- Data based on which the model was trained were captured in the constrained environment with cooperative users. Therefore, in practise model is expected to produce poor segmentation maps for cases like: offgazes, misaligned eyes, blurry images etc.
|
99 |
-
|
100 |
-
## License
|
101 |
-
Unless otherwise specified, the contents of this repository are dual-licensed under either:
|
102 |
-
|
103 |
-
- MIT License (LICENSE-MIT)
|
104 |
-
- Apache License, Version 2.0 (LICENSE-APACHE)
|
105 |
-
|
106 |
-
at your option. This means you may select the license you prefer to use.
|
107 |
-
|
108 |
-
Any contribution intentionally submitted for inclusion in the work by you, as defined in the Apache-2.0 license, shall be dual licensed as above, without any additional terms or conditions.
|
109 |
|
110 |
## Further reading
|
111 |
|
@@ -113,3 +128,9 @@ Any contribution intentionally submitted for inclusion in the work by you, as de
|
|
113 |
2. Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., & Chen, L. C. (2018). MobileNetV2: Inverted Residuals and Linear Bottlenecks. (https://arxiv.org/abs/1801.04381)
|
114 |
3. Roy, A. G., Navab, N., & Wachinger, C. (2018). Recalibrating Fully Convolutional Networks with Spatial and Channel “Squeeze and Excitation” Blocks. (https://arxiv.org/abs/1808.08127v1)
|
115 |
4. Bowyer, K. , Flynn, P. (2016), The ND-IRIS-0405 Iris Image Dataset (https://arxiv.org/abs/1606.04853)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
|
15 |
## Dataset
|
16 |
|
17 |
+
Importantly, no Worldcoin user data was used to train or fine-tune the IRIS pipeline. Rather, a research datasets were used to train model. Datasets used during training comes from:
|
18 |
+
- ND-IRIS-0405 [4]
|
19 |
+
- CASIA-Iris-Interval [5]
|
20 |
+
- CASIA-Iris-Lamp [5]
|
21 |
+
- CASIA-Iris-Thousand [5]
|
22 |
+
- CASIA-Iris-Twins [5]
|
23 |
+
- IIT Delhi Iris Database (IITD) [6]
|
24 |
+
- LivDet2013-Iris [7]
|
25 |
+
- LivDet2015-Iris [8]
|
26 |
+
- LivDet2017-Iris [9]
|
27 |
+
- Multimedia University Iris Database (MMU) [10]
|
28 |
+
|
29 |
+
The experimental dataset contained a total of 18 431 annotated IR images. Table below presents dataset split used during training semantic segmentation model.
|
30 |
+
|
31 |
+
|
32 |
+
| **Dataset type**| **Number of images** |
|
33 |
+
|-----------------------------|-----------|
|
34 |
+
| train | 14685 |
|
35 |
+
| validation | 1880 |
|
36 |
+
| test | 1866 |
|
37 |
|
38 |
## Performance
|
39 |
|
|
|
43 |
|
44 |
| **Metric** | **Value** |
|
45 |
|-----------------------------|-----------|
|
46 |
+
| _eyeball_IoU_ | 0.986 |
|
47 |
+
| _iris_IoU_ | 0.978 |
|
48 |
+
| _pupil_IoU_ | 0.978 |
|
49 |
+
| _eyelashes_IoU_ | 0.798 |
|
50 |
+
| _mIoU_ | 0.943 |
|
51 |
+
| _eyelashes_inside_iris_IoU_ | 0.791 |
|
52 |
+
| _boundary_confidence_pupil_ | 0.965 |
|
53 |
+
| _boundary_confidence_iris_ | 0.907 |
|
54 |
+
| _chamfer_eyeball_ | 1.689 |
|
55 |
+
| _chamfer_iris_ | 1.868 |
|
56 |
+
| _chamfer_pupil_ | 0.680 |
|
57 |
+
|
58 |
+
#### Measured on particular dataset
|
59 |
+
|
60 |
+
|Name |test_IoU_eyeball|test_IoU_eyelashes|test_IoU_iris|test_IoU_pupil|test_boundary_confidence_iris|test_boundary_confidence_pupil|test_chamfer_dist_eyeball|test_chamfer_dist_iris|test_chamfer_dist_pupil|test_mIoU |test_mIoU_eyelashes_inside_iris|
|
61 |
+
|---------------------------------------------------------------|------------------------|--------------------------|---------------------|----------------------|-----------------------------|------------------------------|---------------------------------|------------------------------|-------------------------------|------------------|---------------------------------------|
|
62 |
+
|CASIA-Iris-Interval |0.992 |0.885 |0.977 |0.983 |0.868 |0.959 |1.230 |3.132 |0.673 |0.970 |0.876 |
|
63 |
+
|CASIA-Iris-Lamp |0.987 |0.844 |0.988 |0.983 |0.979 |0.988 |0.683 |0.482 |0.295 |0.953 |0.836 |
|
64 |
+
|CASIA-Iris-Thousand |0.987 |0.823 |0.989 |0.981 |0.981 |0.988 |0.637 |0.430 |0.252 |0.956 |0.767 |
|
65 |
+
|CASIA-Iris-Twins |0.986 |0.837 |0.986 |0.981 |0.966 |0.984 |0.998 |0.865 |0.710 |0.949 |0.840 |
|
66 |
+
|IITD |0.990 |0.868 |0.975 |0.979 |0.843 |0.944 |3.424 |3.564 |0.939 |0.966 |0.861 |
|
67 |
+
|LivDet2013-Iris |0.980 |0.853 |0.985 |0.977 |0.973 |0.977 |1.242 |0.518 |0.393 |0.945 |0.851 |
|
68 |
+
|LivDet2015-Iris |0.978 |0.844 |0.984 |0.974 |0.955 |0.958 |1.524 |0.794 |0.608 |0.938 |0.870 |
|
69 |
+
|LivDet2017-Iris |0.978 |0.791 |0.981 |0.956 |0.956 |0.973 |1.920 |1.274 |1.204 |0.921 |0.738 |
|
70 |
+
|MMU |0.983 |0.845 |0.985 |0.980 |0.960 |0.980 |1.405 |0.821 |0.500 |0.946 |0.765 |
|
71 |
+
|ND-IRIS-0405 |0.981 |0.750 |0.977 |0.975 |0.918 |0.969 |1.757 |1.653 |0.739 |0.917 |0.676 |
|
72 |
+
|
73 |
|
74 |
### Time Performance
|
75 |
|
|
|
102 |
|
103 |
Within the ambit of each class, the model formulates probability estimates pertaining to the likelihood of a given pixel being attributed to a particular class.
|
104 |
|
105 |
+
### Example inference results
|
106 |
|
107 |
**Note**: The provided input image has been subjected to the processing methodology described earlier, prior to its introduction into the model. Moreover, for the intent of visualization, the IR image presented has been anonymized to safeguard the identity of the user. It is also worth to note that the inference process was conducted on the original, non-anonymized version of the image.
|
108 |
|
109 |
#### Input image
|
110 |
|
111 |
+
<img src="https://github.com/worldcoin/open-iris/blob/main/docs/model_card/anonymized.png" alt="anonymized input image" width=640 height=480>
|
112 |
|
113 |
#### Output masks
|
114 |
|
115 |
**Note**: A threshold of 0.5 was applied on every resultant probability matrix to generate a definitive binary mask for each individual class. Also, every outputted segmentation map was resized to the input image resolution.
|
116 |
|
117 |
+
<img src="https://github.com/worldcoin/open-iris/blob/main/docs/model_card/overlayed_segmaps.png" alt="segmaps">
|
118 |
|
119 |
## Limitations
|
120 |
|
121 |
Thorough examination of the results enabled us to pinpoint situations where the segmentation model experiences declines in performance. These instances are as follows:
|
122 |
+
- Segmenting images with high specular reflection coming usually from glasses may lead to bad segmentation map predictions.
|
123 |
+
- Data based on which the model was trained were captured in the constrained environment with cooperative users. Therefore, in practice model is expected to produce poor segmentation maps for cases like: offgaze, misaligned eyes, blurry images etc.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
124 |
|
125 |
## Further reading
|
126 |
|
|
|
128 |
2. Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., & Chen, L. C. (2018). MobileNetV2: Inverted Residuals and Linear Bottlenecks. (https://arxiv.org/abs/1801.04381)
|
129 |
3. Roy, A. G., Navab, N., & Wachinger, C. (2018). Recalibrating Fully Convolutional Networks with Spatial and Channel “Squeeze and Excitation” Blocks. (https://arxiv.org/abs/1808.08127v1)
|
130 |
4. Bowyer, K. , Flynn, P. (2016), The ND-IRIS-0405 Iris Image Dataset (https://arxiv.org/abs/1606.04853)
|
131 |
+
5. http://biometrics.idealtest.org/
|
132 |
+
6. http://www4.comp.polyu.edu.hk/csajaykr/IITD/Database_Iris.htm
|
133 |
+
7. D. Yambay, J.S. Doyle, K.W. Bowyer, A. Czajka, S. Schuckers Livdet-iris 2013 - iris liveness detection competition 2013
|
134 |
+
8. D. Yambay, B. Walczak, S. Schuckers, A. Czajka Livdet-iris 2015 - iris liveness detection competition 2015
|
135 |
+
9. D. Yambay, B. Becker, N. Kohli, D. Yadav, A. Czajka, K.W. Bowyer, S. Schuckers, R. Singh, M.Vatsa, A. Noore, D. Gragnaniello, C. Sansone, L. Verdoliva, L. He, Y. Ru, H. Li, N. Liu, Z. Sun, T. Tan Livdet iris 2017 - iris liveness detection competition 2017
|
136 |
+
10. https://mmuexpert.mmu.edu.my/ccteo
|