IET Computer Vision (Mar 2024)

Representation constraint‐based dual‐channel network for face antispoofing

  • Zuhe Li,
  • Yuhao Cui,
  • Fengqin Wang,
  • Weihua Liu,
  • Yongshuang Yang,
  • Zeqi Yu,
  • Bin Jiang,
  • Hui Chen

DOI
https://doi.org/10.1049/cvi2.12245
Journal volume & issue
Vol. 18, no. 2
pp. 289 – 303

Abstract

Read online

Abstract Although multimodal face data have obvious advantages in describing live and spoofed features, single‐modality face antispoofing technologies are still widely used when it is difficult to obtain multimodal face images or inconvenient to integrate and deploy multimodal sensors. Since the live/spoofed representations in visible light facial images include considerable face identity information interference, existing deep learning‐based face antispoofing models achieve poor performance when only the visible light modality is used. To address the above problems, the authors design a dual‐channel network structure and a constrained representation learning method for face antispoofing. First, they design a dual‐channel attention mechanism‐based grouped convolutional neural network (CNN) to learn important deceptive cues in live and spoofed faces. Second, they design inner contrastive estimation‐based representation constraints for both live and spoofed samples to minimise the sample similarity loss to prevent the CNN from learning more facial appearance information. This increases the distance between live and spoofed faces and enhances the network's ability to identify deceptive cues. The evaluation results indicate that the framework we designed achieves an average classification error rate (ACER) of 2.37% on the visible light modality subset of the CASIA‐SURF dataset and an ACER of 2.4% on the CASIA‐SURF CeFA dataset, outperforming existing methods. The proposed method achieves low ACER scores in cross‐dataset testing, demonstrating its advantage in domain generalisation.

Keywords