IEEE Access (Jan 2024)

Anchor-Net: Distance-Based Self-Supervised Learning Model for Facial Beauty Prediction

  • Jiho Bae,
  • Seok-Jun Buu,
  • Suwon Lee

DOI
https://doi.org/10.1109/ACCESS.2024.3394870
Journal volume & issue
Vol. 12
pp. 61375 – 61387

Abstract

Read online

In today’s society, beauty is more than just aesthetics, it has a profound impact on many aspects of life, including social interactions, self-confidence, and job opportunities. To quantify this beauty, the field of Facial Beauty Prediction (FBP) is gaining traction. In the field of FBP, traditional methods often fall short due to their reliance on absolute beauty scores, which do not fully capture the subjective nature of human aesthetic perception. This study presents a novel approach to address this gap through the development of Anchor-Net, a self-supervised learning model that predicts differences in relative beauty scores by comparing images. The objective of this research is to offer a more nuanced understanding of facial beauty by employing a reference image (anchor) alongside a prediction image, thereby aligning closer with how humans perceive aesthetic differences. To construct Anchor-Net, we first developed a Base model that predicts beauty scores using a model pre-trained with VGGFace2. This Base model was then adapted into Anchor-Net, which is designed to train on the difference in beauty scores between a reference image and a prediction image. Our methodology involved two transfer learning steps to leverage the strengths of pre existing models while tailoring them to our specific research problem. The experimental validation of Anchor-Net was conducted on the SCUT-FBP5500 benchmark dataset, utilizing a 6:4 training-testing split and 5-fold cross-validation to ensure robust testing of the model’s predictive capabilities. The results demonstrate that Anchor-Net outperforms other state-of-the-art deep learning algorithms on all metrics: Pearson Correlation (PC), Mean Absolute Error (MAE), and Root Mean Square Error (RMSE). Anchor-Net outperformed other models with a PC of 0.0021, MAE of 0.0055, and RMSE of 0.0065 on a 6:4 training-test split. On average, it achieved a PC of 0.0034, MAE of 0.0155, and RMSE of 0.0135 on 5-fold cross-validation. This research proposes a novel approach to FBP and suggests a broader application of relative comparison methodologies in fields where absolute measurements fall short.

Keywords