CNN based facial aesthetics analysis through dynamic robust losses and ensemble regression
dc.contributor.author | Bougourzi, Fares | |
dc.contributor.author | Dornaika, Fadi | |
dc.contributor.author | Barrena Orueechebarria, Nagore | |
dc.contributor.author | Distante, Cosimo | |
dc.contributor.author | Taleb-Ahmed, Abdelmalik | |
dc.date.accessioned | 2023-06-29T15:40:24Z | |
dc.date.available | 2023-06-29T15:40:24Z | |
dc.date.issued | 2023-05 | |
dc.identifier.citation | Applied Intelligence 53(9) : 10825-10842 (2023) | es_ES |
dc.identifier.issn | 0924-669X | |
dc.identifier.issn | 1573-7497 | |
dc.identifier.uri | http://hdl.handle.net/10810/61795 | |
dc.description.abstract | In recent years, estimating beauty of faces has attracted growing interest in the fields of computer vision and machine learning. This is due to the emergence of face beauty datasets (such as SCUT-FBP, SCUT-FBP5500 and KDEF-PT) and the prevalence of deep learning methods in many tasks. The goal of this work is to leverage the advances in Deep Learning architectures to provide stable and accurate face beauty estimation from static face images. To this end, our proposed approach has three main contributions. To deal with the complicated high-level features associated with the FBP problem by using more than one pre-trained Convolutional Neural Network (CNN) model, we propose an architecture with two backbones (2B-IncRex). In addition to 2B-IncRex, we introduce a parabolic dynamic law to control the behavior of the robust loss parameters during training. These robust losses are ParamSmoothL1, Huber, and Tukey. As a third contribution, we propose an ensemble regression based on five regressors, namely Resnext-50, Inception-v3 and three regressors based on our proposed 2B-IncRex architecture. These models are trained with the following dynamic loss functions: Dynamic ParamSmoothL1, Dynamic Tukey, Dynamic ParamSmoothL1, Dynamic Huber, and Dynamic Tukey, respectively. To evaluate the performance of our approach, we used two datasets: SCUT-FBP5500 and KDEF-PT. The dataset SCUT-FBP5500 contains two evaluation scenarios provided by the database developers: 60-40% split and five- fold cross-validation. Our approach outperforms state-of-the-art methods on several metrics in both evaluation scenarios of SCUT-FBP5500. Moreover, experiments on the KDEF-PT dataset demonstrate the efficiency of our approach for estimating facial beauty using transfer learning, despite the presence of facial expressions and limited data. These comparisons highlight the effectiveness of the proposed solutions for FBP. They also show that the proposed Dynamic robust losses lead to more flexible and accurate estimators. | es_ES |
dc.description.sponsorship | Open Access funding provided thanks to the CRUE-CSIC agreement with Springer Nature. | es_ES |
dc.language.iso | eng | es_ES |
dc.publisher | Springer | es_ES |
dc.rights | info:eu-repo/semantics/openAccess | es_ES |
dc.rights.uri | http://creativecommons.org/licenses/by/3.0/es/ | * |
dc.subject | facial beauty prediction | es_ES |
dc.subject | convolutional neural network | es_ES |
dc.subject | deep learning | es_ES |
dc.subject | ensemble regression | es_ES |
dc.subject | robust loss functions | es_ES |
dc.title | CNN based facial aesthetics analysis through dynamic robust losses and ensemble regression | es_ES |
dc.type | info:eu-repo/semantics/article | es_ES |
dc.rights.holder | © The Author(s) 2022. This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons. org/licenses/by/4.0/. | es_ES |
dc.rights.holder | Atribución 3.0 España | * |
dc.relation.publisherversion | https://link.springer.com/article/10.1007/s10489-022-03943-0 | es_ES |
dc.identifier.doi | 10.1007/s10489-022-03943-0 | |
dc.departamentoes | Lenguajes y sistemas informáticos | es_ES |
dc.departamentoeu | Hizkuntza eta sistema informatikoak | es_ES |
Files in this item
This item appears in the following Collection(s)
Except where otherwise noted, this item's license is described as © The Author(s) 2022. This article is licensed under a Creative Commons
Attribution 4.0 International License, which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as
long as you give appropriate credit to the original author(s) and the
source, provide a link to the Creative Commons licence, and indicate
if changes were made. The images or other third party material in this
article are included in the article’s Creative Commons licence, unless
indicated otherwise in a credit line to the material. If material is not
included in the article’s Creative Commons licence and your intended
use is not permitted by statutory regulation or exceeds the permitted
use, you will need to obtain permission directly from the copyright
holder. To view a copy of this licence, visit http://creativecommons.
org/licenses/by/4.0/.