2D Image Based Digital Anthropometry Using Deep Learning Approach
Keywords:
Digital Anthropometry, Semantic Segmentation, DeepLabV3 , Pose Estimation, BlazePoseAbstract
Anthropometry is a tool which is widely used for human body parts measurement across diverse field of science. There are several conventional tools available for measurement like measure tape, clippers etc. These conventional anthropometric devices are quick being changed via way of means of modern AI based systems. Digital anthropometry (DA) is a relatively new technique for measuring the dimensions of human body parts. Estimating the pose of a human with the assist of a photograph or a video has these days acquired extensive interest from the medical community. An aim of the research work is to introduce Deep learning concept in digital anthropometry and to develop a novel 2D image based digital measurement system which is more efficient to deal with various limitations of existing techniques. Here for body parts measurement, advanced models of the segmentation and pose estimation is employed to get better results. Also, existing models for anthropometry is implemented. The analysis and comparison of the results with the other methods is presented for better understanding.
Downloads
References
L. C. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, “Encoder-decoder with atrous separable convolution for semantic image segmentation,” Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 11211 LNCS, pp. 833–851, 2018, doi: 10.1007/978-3-030-01234-2_49.
X. Li, Z. Liu, P. Luo, C. C. Loy, and X. Tang, “Not All Pixels Are Equal: Difficulty-Aware Semantic Segmentation via Deep Layer Cascade,” IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3193-3202, 2017.
. Wang et al., “Understanding Convolution for Semantic Segmentation,” Proceedings - 2018 IEEE Winter Conference on Applications of Computer Vision, WACV 2018, vol. 2018-Janua, pp. 1451–1460, 2018, doi: 10.1109/WACV.2018.00163.
C. Peng, X. Zhang, G. Yu, G. Luo and J. Sun, "Large Kernel Matters — Improve Semantic Segmentation by Global Convolutional Network," 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 2017, pp. 1743-1751, doi: 10.1109/CVPR.2017.189.
G. Lin, A. Milan, C. Shen and I. Reid, "RefineNet: Multi-path Refinement Networks for High-Resolution Semantic Segmentation," 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 2017, pp. 5168-5177, doi: 10.1109/CVPR.2017.549.
Wu, Zifeng, C. Shen, and A. Van Den Hengel. "Wider or deeper: Revisiting the resnet model for visual recognition." Pattern recognition 90 (2019): 119-133.FLEX Chip Signal Processor (MC68175/D), Motorola, vol. 15, no. 3, pp. 250-275, 1996.
Zhao, Hengshuang, J. Shi, Xiaojuan Qi, X. Wang, and J. Jia. "Pyramid scene parsing network." In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2881-2890. 2017.
Wang, Guangrun, P. Luo, L. Lin, and X. Wang. "Learning object interactions and descriptions for semantic image segmentation." In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5859-5867. 2017.
Fu, Jun, J. Liu, Y. Wang, J. Zhou, C. Wang, and H. Lu. "Stacked deconvolutional network for semantic segmentation." IEEE Transactions on Image Processing, pp. 01-12. 2019.
G. Wang, P. Luo, L. Lin and X. Wang, "Learning Object Interactions and Descriptions for Semantic Image Segmentation," 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 2017, pp. 5235-5243, doi: 10.1109/CVPR.2017.556.
Chen, L. Chieh, G. Papandreou, F. Schroff, and H. Adam. "Rethinking atrous convolution for semantic image segmentation." arXiv preprint arXiv:1706.05587 (2017).
Gong, Wenjuan, X. Zhang, J. Gonzàlez, A. Sobral, T. Bouwmans, C. Tu, and E. Zahzah. "Human pose estimation from monocular images: A comprehensive survey." Sensors 16, no. 12 (2016): 1966.
P. Wang, W. Li, P. Ogunbona, J. Wan, and S. Escalera, “RGB-D-based human motion recognition with deep learning: A survey,” Computer Vision and Image Understanding, vol. 171, no. Wanqing Li, pp. 118–139, 2018, doi: 10.1016/j.cviu.2018.04.007.
Z. Cao, G. Hidalgo, T. Simon, S. E. Wei, and Y. Sheikh, “OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields,” IEEE Trans Pattern Anal Mach Intel, vol. 43, no. 1, pp. 172–186, 2021, doi: 10.1109/TPAMI.2019.2929257.
Bazarevsky, Valentin, Ivan Grishchenko, Karthik Raveendran, Tyler Zhu, Fan Zhang, and Matthias Grundmann. "Blazepose: On-device real-time body pose tracking." arXiv preprint arXiv:2006.10204 (2020).
A. Toshev and C. Szegedy, “DeepPose: Human pose estimation via deep neural networks,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 1653–1660, 2014, doi: 10.1109/CVPR.2014.214.
A. Newell, K. Yang, and J. Deng, “Stacked hourglass networks for human pose estimation,” Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 9912 LNCS, pp. 483–499, 2016, doi: 10.1007/978-3-319-46484-8_29.
Bazarevsky, Valentin, Y. Kartynnik, A. Vakunov, K. Raveendran, and M. Grundmann. "Blazeface: Sub-millisecond neural face detection on mobile GPUs." arXiv preprint arXiv:1907.05047 (2019).
Downloads
Published
How to Cite
Issue
Section
License

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
All papers should be submitted electronically. All submitted manuscripts must be original work that is not under submission at another journal or under consideration for publication in another form, such as a monograph or chapter of a book. Authors of submitted papers are obligated not to submit their paper for publication elsewhere until an editorial decision is rendered on their submission. Further, authors of accepted papers are prohibited from publishing the results in other publications that appear before the paper is published in the Journal unless they receive approval for doing so from the Editor-In-Chief.
IJISAE open access articles are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. This license lets the audience to give appropriate credit, provide a link to the license, and indicate if changes were made and if they remix, transform, or build upon the material, they must distribute contributions under the same license as the original.