欢迎访问《图学学报》 分享到:

图学学报

• 图像处理与计算机视觉 • 上一篇    下一篇

基于多模态融合的人脸反欺骗技术

  

  1. (安徽大学电气工程与自动化学院,安徽 合肥 230601)
  • 出版日期:2020-10-31 发布日期:2020-11-05
  • 通讯作者: 李 腾(1980),男,安徽凤台人,教授,博士,硕士生导师。主要研究方向为计算机视觉、图像处理。E-mail:liteng@ahu.edu.cn
  • 作者简介:穆大强(1993?),男,安徽亳州人,硕士研究生。主要研究方向为计算机视觉、图像处理等。E-mail:1203986876@qq.com
  • 基金资助:
    国家自然科学基金项目(61572029);安徽省杰出青年基金项目(1908085J25)

Face anti-spoofing technology based on multi-modal fusion

  1. (School of Electrical Engineering and Automation, Anhui University, Hefei Anhui 230601, China)
  • Online:2020-10-31 Published:2020-11-05
  • Contact: LI Teng (1980–), male, professor, Ph.D. His main research interests cover computer vision and image processing. E-mail:liteng@ahu.edu.cn
  • About author:MU Da-qiang (1993–), male, master student. His main research interests cover computer vision and image processing. E-mail:1203986876@qq.com
  • Supported by:
    National Natural Science Foundation Project (61572029); Anhui Provincial Outstanding Youth Fund (1908085J25)

摘要: 在先前的人脸反欺骗方法中大多使用手工提取的特征或者仅使用单一模态上的人 脸特征,并且很少注意到多通道色度的差异,因此得到的人脸反欺骗模型的鲁棒性较差以至于 无法有效地区分真假面孔。鉴于此,卷积神经网络(CNN)被用作特征提取器来代替手工特征的 提取,并且一种有效的多输入 CNN 模型被提出,以融合多种模态上的人脸特征以实现更具有 鲁棒性的人脸反欺骗。通过对人脸图像上的 2 个不同颜色特征(即 HSV 和 YCbCr)以及时间特征 进行联合建模,探索了人脸反欺骗的最佳鲁棒表示。在 REPLAY_ATTACK 和 CASIA-FASD 2 个基准数据集上进行的大量实验表明,该方法可实现最先进的性能。且在 REPLAY_ATTACK 上获得 0.23%的错误率(ERR)与 0.49%的半错误率(HTER)和在 CASIA-FASD 数据库上获得 1.76%的错误率与 3.05%的半错误率。

关键词: 人脸反欺骗, 多模态特征, 多输入卷积神经网络, 模型鲁棒性, 融合

Abstract: Most of the previous face anti-spoofing methods employed the manually extracted features or only face features on a single modality, and rarely noticed the differences in multi-channel chroma. Therefore, the face anti-spoofing model was so low in robustness that it could not effectively distinguish between real and fake faces. In view of this, a convolutional neural network (CNN), which substituted the hand-crafted features extraction, was utilized as a feature extractor, and an effective multi-input CNN model was proposed to fuse face features on multiple modalities to achieve more robust face anti-spoofing. Through the joint modeling of features concerning two different color features (i.e. HSV and YCbCr) of the local face patches, as well as the temporal feature, the optimal robust representation of face anti-spoofing was explored. A large number of experiments on two benchmarks of REPLAY_ATTACK and CASIA-FASD show that this method can attain the most advanced performance. Specifically, 0.23% error rate (ERR) and 0.49% half total error rate (HTER) were obtained on REPLAY_ATTACK, and 1.76% error rate and 3.05% half total error rate were yielded on the CASIA-FASD database.

Key words: face anti-spoofing, multimodal feature, multi-input convolutional neural network, model robustness, fusion