Backdoor Poisoning Attack Using De-Identified Face Images
In this section, we describe a backdoor poisoning attack method against face spoofing attack detection. The proposed method aims to induce the target model to misclassify specific spoofing attacks by injecting poisoned data into the training dataset.
Poisoned Data Generation
The poisoned data is generated by embedding features extracted from a specific spoofed face image (trigger face image) into a live face image (cover image) without any perceptible visual alterations. To achieve this, we leverage a face de-identification method that encourages the face features extracted from the generated image to be closer to the features of the trigger image, while ensuring the visual quality remains similar to the cover image. This characteristic makes it extremely difficult for an administrator to detect the poisoned data.
Procedure of Backdoor Poisoning Attack
The attack consists of three phases: 1) Poisoned data generation, 2) Model training on the dataset where a portion of live images is replaced by the poisoned data, and 3) Model evaluation, where the poisoned model exhibits the property of incorrectly classifying the trigger image as "Live" with minimal degradation in overall detection accuracy.
Figure 1: Overview of the proposed backdoor poisoning attack, which consists of Poisoned data generation, Model training, and Model evaluation.