Skip to main navigation Skip to search Skip to main content

Face Anti-Spoofing with Deep Neural Network Distillation

  • Haoliang Li
  • , Shiqi Wang
  • , Peisong He*
  • , Anderson Rocha
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

One challenging aspect in face anti-spoofing (or presentation attack detection, PAD) refers to the difficulty of collecting enough and representative attack samples for an application-specific environment. In view of this, we tackle the problem of training a robust PAD model with limited data in an application-specific domain. We propose to leverage data from a richer and related domain to learn meaningful features through the concept of neural network distilling. We first train a deep neural network based on reasonably sufficient labeled data in an attempt to 'teach' a neural network for the application-specific domain for which training samples are scarce. Subsequently, we form training sample pairs from both domains and formulate a novel optimization function by considering the cross-entropy loss, as well as maximum mean discrepancy of features and paired sample similarity embedding for network distillation. Thus, we expect to capture spoofing-specific information and train a discriminative deep neural network on the application-specific domain. Extensive experiments validate the effectiveness of the proposed scheme in face anti-spoofing setups.

Original languageEnglish
Article number9115256
Pages (from-to)933-946
Number of pages14
JournalIEEE Journal on Selected Topics in Signal Processing
Volume14
Issue number5
DOIs
StatePublished - Aug 2020
Externally publishedYes

Keywords

  • deep learning
  • Face anti-spoofing
  • neural network distilling

Fingerprint

Dive into the research topics of 'Face Anti-Spoofing with Deep Neural Network Distillation'. Together they form a unique fingerprint.

Cite this