The Custom Silicone Mask Attack Dataset (CSMAD) contains presentation attacks made of six custom-made silicone masks. Each mask cost about USD 4000. The dataset is designed for face presentation attack detection experiments.
The Custom Silicone Mask Attack Dataset (CSMAD) has been collected at the Idiap Research Institute. It is intended for face presentation attack detection experiments, where the presentation attacks have been mounted using a custom-made silicone mask of the person (or identity) being attacked.
The dataset contains videos of face-presentations, as a set of files specifying the experimental protocol corresponding the experiments presented in the corresponding publication.
Reference
If you publish results using this dataset, please cite the following publication.
Sushil Bhattacharjee, Amir Mohammadi and Sebastien Marcel: "Spoofing Deep Face Recognition With Custom Silicone Masks." in Proceedings of International Conference on Biometrics: Theory, Applications, and Systems (BTAS), 2018. 10.1109/BTAS.2018.8698550 http://publications.idiap.ch/index.php/publications/show/3887
Data Collection
Face-biometric data has been collected from 14 subjects to create this dataset. Subjects participating in this data-collection have played three roles: targets, attackers, and bona-fide clients. The subjects represented in the dataset are referred to here with letter-codes: A .. N. The subjects A..F have also been targets. That is, face-data for these six subjects has been used to construct their corresponding flexible masks (made of silicone). These masks have been made by Nimba Creations Ltd., a special effects company.
Bona fide presentations have been recorded for all subjects A..N. Attack presentations (presentations where the subject wears one of 6 masks) have been recorded for all six targets, made by different subjects. That is, each target has been attacked several times, each time by a different attacker wearing the mask in question. This is one way of increasing the variability in the dataset. Another way we have augmented the variability of the dataset is by capturing presentations under different illumination conditions. Presentations have been captured in four different lighting conditions:
flourescent ceiling light only
halogen lamp illuminating from the left of the subject only
halogen lamp illuminating from the right only
both halogen lamps illuminating from both sides simultaneously
All presentations have been captured with a green uniform background. See the paper mentioned above for more details of the data-collection process.
Dataset Structure
The dataset is organized in three subdirectories: ‘attack’, ‘bonafide’, ‘protocols’. The two directories: ‘attack’ and ‘bonafide’ contain presentation-videos and still images for attacks and bona fide presentations, respectively. The folder ‘protocols’ contains text files specifying the experimental protocol for vulnerability analysis of face-recognition (FR) systems.
The number of data-files per category are as follows:
‘bonafide’: 87 videos, and 17 still images (in .JPG format). The still images are frontal face images captured using a Nikon Coolpix digital camera.
‘attack’: 159, organized in two sub-folders – ‘WEAR’ (108 videos), and ‘STAND’ (51 videos)
The folder ‘attack/WEAR’ contains videos where the attack has been made by a person (attacker) wearing the mask of the target being attacked. The ‘attack/STAND’ folder contains videos where the attack has been made using a the target’s mask mounted on an appropriate stand.
Video File Format
The video files for the face-presentations are in ‘hdf5’ format (with file-extensions ‘.h5’. The folder structure of the hdf5 file is shown in Figure 1. Each file contains data collected using two cameras:
RealSense SR300 (from Intel): collects images/videos in visible-light (RGB color) , near infrared (NIR) @ 860nm wavelength, and depth maps
Compact Pro (from Seek Thermal): collects thermal (long-wave infrared (LWIR)) images.
As shown in Figure 1, frames from the different channels (color, infrared, depth, thermal) from he two cameras are stored in separate directory-hierarchies in the hdf5 file. Each file respresents a video of approximately 10 seconds, or roughly, 300 frames.
In the hdf5 file, the directory for SR300 also contains a subdirectory named ‘aligned_color_to_depth’. This folder contains post-processed data, where the frames of depth channel have been aligned with those of the color channel based on the time-stamps of the frames.
Experimental Protocol
The ‘protocols’ folder contains text files that specify the protocols for vulnerability analysis experiments reported in the paper mentioned above. Please see the README file in the protocols folder for details.
Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
Silicone Mask Attack dataset
The dataset contains 6,500+ videos of attacks from 50 different people, filmed using 5 devices, providing a valuable resource for researching presentation attacks in facial recognition technologies. By focusing on this area, the dataset facilitates experiments designed to improve biometric security and anti-spoofing measures, ultimately aiding in the creation of more robust and reliable authentication systems. By utilizing this dataset, researchers can… See the full description on the dataset page: https://huggingface.co/datasets/UniDataPro/silicone-mask-attack.
CSMAD-Mobile is a dataset for mobile face recognition and presentation attack detection (anti-spoofing). The dataset contains face and silicon masks images captured with different smartphones. This dataset consists of images captured from 8 different Bona Fide subjects using three different smartphones (iPhone X, Samsung S7 and Samsung S8). For each subject within the database, varying number of samples are collected using all the three phones. Similarly, the silicone masks of each of the subject is collected using three phones.
Description
The eXtended Custom Silicone Mask Attack Dataset (XCSMAD) consists of 535 short video recordings of both bona fide and presentation attacks (PA) from 72 subjects. The attacks have been created from custom silicone masks. Videos have been recorded in RGB (visual spectra), near infrared (NIR), and thermal (LWIR) channels.
A complete preprocessed data for the aforementioned videos and bona fide images (as a part of experiments related to vulnerability assessment) have been provided to facilitate reproducing experiments from the reference publication, as well as to conduct new experiments. The details of preprocessing can be found in the reference publication.
The implementation of all experiments described in the reference publication is available at https://gitlab.idiap.ch/bob/bob.paper.xcsmad_facepad
Experimental protocols
The reference publication considers two experimental protocols: grandtest and cross-validation (cv). For a frame-level evaluation, 50 frames from each video have been used in both protocols. For the grandtest protocol, videos were divided into train, dev, and eval groups. Each group consists of unique subset of clients. (The videos corresponding to any specific subjects in one group are a part of single group).
For cross-validation (cv) experiments, a 5-fold protocol has been devised. Videos from XCSMAD have been split into 5 folds with non-overlapping clients. Using these five partitions, 5 testprotocols (cv0, · · · , cv4) have been created such that in each protocol, four of the partitions are used for training, and the remaining one is used for evaluation.
Reference
If you use this dataset, please cite the following publication:
@article{Kotwal_TBIOM_2019, author = {Kotwal, Ketan and Bhattacharjee, Sushil and Marcel, S\'{e}bastien}, title = {Multispectral Deep Embeddings As a Countermeasure To Custom Silicone Mask Presentation Attacks}, journal = {IEEE Transactions on Biometrics, Behavior, and Identity Science}, publisher = {{IEEE}}, year = {2019}, }
Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
Face mask dataset for facial recognition
This dataset contains over 11,100+ video recordings of people wearing latex masks, captured using 5 different devices.It is designed for liveness detection algorithms, specifically aimed at enhancing anti-spoofing capabilities in biometric security systems. By utilizing this dataset, researchers can develop more accurate facial recognition technologies, which is crucial for achieving the iBeta Level 2 certification, a benchmark for robust and… See the full description on the dataset page: https://huggingface.co/datasets/UniDataPro/latex-mask-attack.
Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
Wrapped 3D Attacks Dataset is designed to enhance Liveness Detection models by simulating Wrapped 3D Attacks — a more advanced version of 3D Print Attacks, where facial prints include 3D elements and additional attributes. It is particularly useful for iBeta Level 2 certification and anti-spoofing model training. It is more advanced version of 3D paper mask attack Dataset
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F20109613%2Fc3ecd3043c4139ef185a461e47ba257c%2F2025-03-26%20%2022.44.44.png?generation=1743018625916049&alt=media" alt="">
The videos capture realistic spoofing conditions using different recording devices and variety of environments. Additionally, each attack video employs a zoom-in effect, adding to its effectiveness in active liveness detection. The videos were shot using a back-facing camera
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F20109613%2F6eda9058fdaf9ca0118e3fdceee02f30%2FFrame%2065.png?generation=1743018709824907&alt=media" alt="">
This is a more advanced version of 3D paper mask attack dataset for Liveness
Volume: In Wrapped 3D masks are really three-dimensional (nose, chin and forehead protrude), while in 3D Paper the image is flat, the volume is created only by the angle and surroundings
Body: In Wrapped 3D the mask is put on a mannequin, you can see a part of the body with clothes. In 3D Paper, only the printed head and shoulders are visible
Detail: Wrapped 3D uses wigs, glasses and other realistic details. 3D Paper is just a paper photo, with no extras
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Silicone Mask Biometric Attack Dataset for Liveness Detection
10,000+ videos of attacks with Silicone 3D Masks for iBeta 2. The Dataset is designed to address security challenges in liveness detection systems through 3D silicone mask attacks. These presentation attacks are key for testing Passive Liveness Detection systems crucial for iBeta Level 2 certification. This dataset significantly enhances the capabilities of liveness detection models
Full version of dataset is… See the full description on the dataset page: https://huggingface.co/datasets/AxonData/iBeta_level_2_Silicone_masks.
The MLFP dataset consists of face presentation attacks captured with seven 3D latex masks and three 2D print attacks. The dataset contains videos captured from color, thermal and infrared channels.
Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
Web Camera Face Liveness Detection
The dataset consists of videos featuring individuals wearing various types of masks. Videos are recorded under different lighting conditions and with different attributes (glasses, masks, hats, hoods, wigs, and mustaches for men). In the dataset, there are 7 types of videos filmed on a web camera:
Silicone Mask - demonstration of a silicone mask attack (silicone) 2D mask with holes for eyes - demonstration of an attack with a paper/cardboard mask… See the full description on the dataset page: https://huggingface.co/datasets/TrainingDataPro/web-camera-face-liveness-detection.
Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
Textile 3D Face Mask Attack Dataset
This Dataset is specifically designed to enhance Face Anti-Spoofing and Liveness Detection models by simulating Nylon Mask Attacks — an accessible alternative to expensive silicone and latex mask datasets. These attacks utilize thin elastic fabric masks worn like a balaclava, featuring printed facial images that conform to the wearer's head shape through textile elasticity. The dataset is particularly valuable for PAD model training and iBeta… See the full description on the dataset page: https://huggingface.co/datasets/AxonData/3d_cloth_face_mask_spoofing_dataset.
Description
This dataset contains images of 5 subjects. The images have been captured using the Intel Realsense SR300 camera, and the Xenics Gobi thermal camera. The SR300 returns 3 kinds of data: color (RGB) images, near-infrared (NIR) images, and depth information.
For four subjects (subject1 – subject4), images have been captured with both cameras under two conditions:
1with the face visible, and
with the subject wearing a rigid (resin-coated) mask
Each subject has used 3 sets of rigid masks (corresponding to three identities (‘id0’, ‘id1’, ‘id2’, not necessarily corresponding to the subjects in this dataset), with two masks (‘mask0’, ‘mask1’) per identity.
For subject5, data has been captured using both cameras under two conditions:
with the face visible
with the subject wearing a flexible (silicone) mask resembling subject5.
For each combination (subject, camera, condition), several seconds of video have been captured, and the video-frames have been stored in uncompressed form (in .png) format.
All images in .png format have been captured at a resolution of 640x480 pixels.
Reference
If you use this database, please cite the following publication:
Sushil Bhattacharjee and Sébastien Marcel, "What you can't see can help you -- extended-range imaging for 3Dmask presentation attack detection", BIOSIG2017. 10.23919/BIOSIG.2017.8053524 https://publications.idiap.ch/index.php/publications/show/3710
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Silicone Mask Biometric Attack Dataset
Wrapped 3D Attacks Dataset
Full version of dataset is availible for commercial usage - leave a request on our website Axon Labs to purchase the dataset 💰
Introduction
This dataset is designed to enhance Liveness Detection models by simulating Wrapped 3D Attacks — a more advanced version of 3D Print Attacks, where facial prints include 3D elements and additional attributes. It is particularly useful for iBeta Level 2… See the full description on the dataset page: https://huggingface.co/datasets/AxonData/Wrapped_3D_Attacks.
Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
Latex Mask Dataset for Face Anti-Spoofing and Liveness Detection
Anti spoofing dataset with Latex 3D mask attacks (4 000 videos) for iBeta 2. The Biometric Attack Dataset offers a robust solution for enhancing security in liveness detection systems by simulating 3D latex mask attacks. This dataset is invaluable for assessing and fine-tuning Passive Liveness Detection models, an essential step toward achieving iBeta Level 2 certification. By integrating diverse realistic presentation… See the full description on the dataset page: https://huggingface.co/datasets/AxonData/Latex_Mask_dataset.
Description
The in-Vehicle Face Presentation Attack Detection (VFPAD) dataset consists of 4046 bona-fide recordings from 40 subjects, and 1790 attack presentation videos from a total of 89 PAIs (presentation attack instruments). These presentations have been captured using an NIR camera (940 nm) placed on the steering wheel of the car, while NIR illuminators have been fixed on both front pillars (adjacent to the wind-shield) of the car. The bona-fide videos represent 24 male and 16 female subjects of various ethnicities. The PAI species used to construct this dataset include photo-prints, digital displays (for replay attacks), rigid 3D masks, and flexible 3D masks made of silicone.
Data Collection
The videos comprising this dataset represent bona-fide and attack presentations under a range of variations:
Citation
If you use the dataset, please cite the following publication:
@article{IEEE_TBIOM_2021,
author = {Kotwal, Ketan and Bhattacharjee, Sushil and Abbet, Philip and Mostaani, Zohreh and Wei, Huang and Wenkang, Xu and Yaxi, Zhao and Marcel, S\'{e}bastien},
title = {Domain-Specific Adaptation of CNN for Detecting Face Presentation Attacks in NIR},
journal = {IEEE Transactions on Biometrics, Behavior, and Identity Science},
publisher = {{IEEE}},
year={2022},
volume={4},
number={1},
pages={135--147},
doi={10.1109/TBIOM.2022.3143569}
}
Not seeing a result you expected?
Learn how you can add new datasets to our index.
The Custom Silicone Mask Attack Dataset (CSMAD) contains presentation attacks made of six custom-made silicone masks. Each mask cost about USD 4000. The dataset is designed for face presentation attack detection experiments.
The Custom Silicone Mask Attack Dataset (CSMAD) has been collected at the Idiap Research Institute. It is intended for face presentation attack detection experiments, where the presentation attacks have been mounted using a custom-made silicone mask of the person (or identity) being attacked.
The dataset contains videos of face-presentations, as a set of files specifying the experimental protocol corresponding the experiments presented in the corresponding publication.
Reference
If you publish results using this dataset, please cite the following publication.
Sushil Bhattacharjee, Amir Mohammadi and Sebastien Marcel: "Spoofing Deep Face Recognition With Custom Silicone Masks." in Proceedings of International Conference on Biometrics: Theory, Applications, and Systems (BTAS), 2018. 10.1109/BTAS.2018.8698550 http://publications.idiap.ch/index.php/publications/show/3887
Data Collection
Face-biometric data has been collected from 14 subjects to create this dataset. Subjects participating in this data-collection have played three roles: targets, attackers, and bona-fide clients. The subjects represented in the dataset are referred to here with letter-codes: A .. N. The subjects A..F have also been targets. That is, face-data for these six subjects has been used to construct their corresponding flexible masks (made of silicone). These masks have been made by Nimba Creations Ltd., a special effects company.
Bona fide presentations have been recorded for all subjects A..N. Attack presentations (presentations where the subject wears one of 6 masks) have been recorded for all six targets, made by different subjects. That is, each target has been attacked several times, each time by a different attacker wearing the mask in question. This is one way of increasing the variability in the dataset. Another way we have augmented the variability of the dataset is by capturing presentations under different illumination conditions. Presentations have been captured in four different lighting conditions:
flourescent ceiling light only
halogen lamp illuminating from the left of the subject only
halogen lamp illuminating from the right only
both halogen lamps illuminating from both sides simultaneously
All presentations have been captured with a green uniform background. See the paper mentioned above for more details of the data-collection process.
Dataset Structure
The dataset is organized in three subdirectories: ‘attack’, ‘bonafide’, ‘protocols’. The two directories: ‘attack’ and ‘bonafide’ contain presentation-videos and still images for attacks and bona fide presentations, respectively. The folder ‘protocols’ contains text files specifying the experimental protocol for vulnerability analysis of face-recognition (FR) systems.
The number of data-files per category are as follows:
‘bonafide’: 87 videos, and 17 still images (in .JPG format). The still images are frontal face images captured using a Nikon Coolpix digital camera.
‘attack’: 159, organized in two sub-folders – ‘WEAR’ (108 videos), and ‘STAND’ (51 videos)
The folder ‘attack/WEAR’ contains videos where the attack has been made by a person (attacker) wearing the mask of the target being attacked. The ‘attack/STAND’ folder contains videos where the attack has been made using a the target’s mask mounted on an appropriate stand.
Video File Format
The video files for the face-presentations are in ‘hdf5’ format (with file-extensions ‘.h5’. The folder structure of the hdf5 file is shown in Figure 1. Each file contains data collected using two cameras:
RealSense SR300 (from Intel): collects images/videos in visible-light (RGB color) , near infrared (NIR) @ 860nm wavelength, and depth maps
Compact Pro (from Seek Thermal): collects thermal (long-wave infrared (LWIR)) images.
As shown in Figure 1, frames from the different channels (color, infrared, depth, thermal) from he two cameras are stored in separate directory-hierarchies in the hdf5 file. Each file respresents a video of approximately 10 seconds, or roughly, 300 frames.
In the hdf5 file, the directory for SR300 also contains a subdirectory named ‘aligned_color_to_depth’. This folder contains post-processed data, where the frames of depth channel have been aligned with those of the color channel based on the time-stamps of the frames.
Experimental Protocol
The ‘protocols’ folder contains text files that specify the protocols for vulnerability analysis experiments reported in the paper mentioned above. Please see the README file in the protocols folder for details.