Identification of Facial Features in Face Templates Using Deep Neural Network Models

Authors

  • Matas Malickas Neurotechnology, Vilnius, Lithuania
  • Pavel Stefanovič Department of Information Systems, Vilnius Gediminas Technical University, Lithuania
  • Simona Ramanauskaitė Department of Information Technology, Vilnius Gediminas Technical University, Lithuania

DOI:

https://doi.org/10.15837/ijccc.2026.2.7294

Keywords:

face templates, face recognition, deep learning, machine learning, hyperparameter optimization

Abstract

In this research, facial features are identified from face templates using deep neural network models. Facial templates are widely used in modern biometric systems, enabling efficient and reliable individual recognition. Such templates are compact, easy to process, and are often used in real-time systems. During the research, a dataset with 115,023 items (face templates), used by the Neurotechnology company to identify persons, was prepared as a new dataset. The dataset has been prepared to study three facial features from face templates: gender, race, and age. The facial templates are lighter-weight than real photos and could be reused to estimate the facial features. However, it is not yet known how effective facial feature identification will be, as some data are lost when converting a photo into a face template. Several methods have been proposed for classifying original image attributes solely from face templates. Both deep neural networks and classical machine learning algorithms were used in the experiments. The experiments revealed that gender is the most accurately predictable attribute, with the best model achieving 93 % accuracy. Classification of people’s race and age from face templates is more challenging, likely because face template generation models are designed to eliminate information unrelated to identity. A comparison of machine learning methods showed that deep neural networks are better suited to this task than classical classification algorithms.

References

Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. (2019). Arcface: Additive angular margin loss for deep face recognition, In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, IEEE, 2019, pp. 4690-4699, 2019. https://doi.org/10.1109/CVPR.2019.00482

Fu, X.; Ma, Q.; Yang, F.; Zhang, C.; Zhao, X.; Chang, F.; Han, L. (2024). Crop pest image recognition based on the improved ViT method, Information Processing in Agriculture, 11(2), pp. 249-259, 2024. https://doi.org/10.1016/j.inpa.2023.02.007

Kotwal, K.; Marcel, S. (2025). Review of demographic fairness in face recognition, IEEE Transactions on Biometrics, Behavior, and Identity Science, 2025.

Kozłowski, M.; Racewicz, S.; Wierzbicki, S. (2024). Image Analysis in Autonomous Vehicles: A Review of the Latest AI Solutions and Their Comparison, Applied Sciences, 14(18), p. 8150, 2024. https://doi.org/10.3390/app14188150

Kumar, Y.; Shrivastav, S.; Garg, K.; Modi, N.; Wiltos, K.; Woźniak, M.; Ijaz, M. F. (2024). Automating cancer diagnosis using advanced deep learning techniques for multi-cancer image classification, Scientific Reports, 14(1), p. 25006, 2024. https://doi.org/10.1038/s41598-024-75876-2

LeCun, Y.; Bengio, Y.; Hinton, G. (2015). Deep learning, nature, 521(7553), pp. 436-444, 2015. https://doi.org/10.1038/nature14539

Li, M. (2025). Research and Analysis of Facial Recognition Based on FaceNet, DeepFace, and OpenFace, In ITM Web of Conferences. EDP Sciences, 70, p. 03009, 2025. https://doi.org/10.1051/itmconf/20257003009

Liu, W.; Wen, Y.; Yu, Z.; Li, M.; Raj, B.; Song, L. (2017). Sphereface: Deep hypersphere embedding for face recognition, In Proceedings of the IEEE conference on computer vision and pattern recognition, IEEE, pp. 212-220, 2017. https://doi.org/10.1109/CVPR.2017.713

Meftah, I.; Hu, J.; Asham, M. A.; Meftah, A.; Zhen, L.; Wu, R. (2024). Visual detection of road cracks for autonomous vehicles based on deep learning, Sensors, 24(5), p. 1647, 2024. https://doi.org/10.3390/s24051647

Neurotechnology (2025). MegaMatcher SDK (Version 2025.1) [Software], https://neurotechnology.com/megamatcher.html

Praneeth, D.; Sree, T. R.; Viswanatha, V.; Sai, M. A. (2024). AI Enabled Home Security System Using Object Detection and Face Recognition with Haar Cascade Algorithm, In 2024 International Conference on Recent Advances in Science and Engineering Technology (ICRASET), IEEE, pp. 1-6, 2024. https://doi.org/10.1109/ICRASET63057.2024.10895751

Ranjan, R.; Sankaranarayanan, S.; Castillo, C. D.; Chellappa, R. (2017). AI Enabled Home Security System Using Object Detection and Face Recognition with Haar Cascade Algorithm, An all-in-one convolutional neural network for face analysis. In 2017 12th IEEE International Conference on automatic face & gesture recognition (FG 2017), IEEE, pp. 17-24, 2017. https://doi.org/10.1109/FG.2017.137

Rashid, A. B.; Kausik, A. K. (2024). AI revolutionizing industries worldwide: A comprehensive overview of its diverse applications, Hybrid Advances, p. 100277, 2024. https://doi.org/10.1016/j.hybadv.2024.100277

Schroff, F.; Kalenichenko, D.; Philbin, J. (2015). Facenet: A unified embedding for face recognition and clustering, In Proceedings of the IEEE conference on computer vision and pattern recognition, IEEE, pp. 815-823, 2015. https://doi.org/10.1109/CVPR.2015.7298682

Stefanovič, P.; Ramanauskait˙e, S. (2023). Travel direction recommendation˙ model based on photos of user social network profile, IEEE Access, 11, pp. 28252-28262, 2023. https://doi.org/10.1109/ACCESS.2023.3260103

SV, S. (2021). Security Monitoring System Using FaceNet For Wireless Sensor Network, arXiv preprint, arXiv:2112.01305., 2021.

Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Rabinovich, A. (2015). Going deeper with convolutions, In Proceedings of the IEEE conference on computer vision and pattern recognition, IEEE, pp. 1-9, 2015. https://doi.org/10.1109/CVPR.2015.7298594

Taigman, Y.; Yang, M.; Ranzato, M. A.; Wolf, L. (2014). Deepface: Closing the gap to humanlevel performance in face verification, In Proceedings of the IEEE conference on computer vision and pattern recognition, IEEE, pp. 1701-1708, 2014. https://doi.org/10.1109/CVPR.2014.220

Virbukait˙e, S.; Bernatavičien˙e, J.; Imbrasien˙e, D. (2024). Glaucoma Identification Using Convolutional Neural Networks Ensemble for Optic Disc and Cup Segmentation, IEEE Access, 12, pp. 82720-82729. 2024. https://doi.org/10.1109/ACCESS.2024.3412185

Wang, H.; Wang, Y.; Zhou, Z.; Ji, X.; Gong, D.; Zhou, J.; Liu, W. (2018). Cosface: Large margin cosine loss for deep face recognition, In Proceedings of the IEEE conference on computer vision and pattern recognition, IEEE, pp. 5265-5274, 2018. https://doi.org/10.1109/CVPR.2018.00552

Wang, M.; Deng, W.; Hu, J.; Su, S. (2025). Marginal debiased network for fair visual recognition, Pattern Recognition, 158, p. 111046, 2025. https://doi.org/10.1016/j.patcog.2024.111046

Wu, S.; Lin, X.; Yao, T. (2022). Research on Face Recognition Technology of Subway Automatic Ticketing System based on Neural Network and Deep Learning, International Journal of Advanced Computer Science and Applications, 13(12), 2022. https://doi.org/10.14569/IJACSA.2022.01312123

Xin, Y.; Zhou, Y.; Jiang, J. (2024). Self-Attentioned Arcface Towards Improved Efficiency for Deep Face Recognition, SSRN, Available at SSRN 4898101, 2024. https://doi.org/10.2139/ssrn.4898101

Xu, Q. A.; Jayne, C.; Chang, V. (2024). An emoji feature-incorporated multi-view deep learning for explainable sentiment classification of social media reviews, Technological Forecasting and Social Change, 202, p. 123326, 2024. https://doi.org/10.1016/j.techfore.2024.123326

Yangxiaoxiao, Z. (2021). mage modal analysis in art design and image recognition using AI techniques, Journal of Intelligent & Fuzzy Systems, 40(4), pp. 6961-6971, 2021. https://doi.org/10.3233/JIFS-189526

Additional Files

Published

2026-03-12

Most read articles by the same author(s)

Obs.: This plugin requires at least one statistics/report plugin to be enabled. If your statistics plugins provide more than one metric then please also select a main metric on the admin's site settings page and/or on the journal manager's settings pages.